source
stringlengths
31
203
text
stringlengths
28
2k
https://en.wikipedia.org/wiki/Dahlander%20pole%20changing%20motor
A Dahlander motor (also known as a pole changing motor, dual- or two speed-motor) is a type of multispeed three-phase induction motor, in which the speed of the motor is varied by altering the number of poles; this is achieved by altering the wiring connections inside the motor. The motor may have fixed or variable torque depending on the stator winding. It is named after its inventor Robert Dahlander (1870–1935). Invention Robert Dahlander, a Swedish engineer working for ASEA, discovered that switching the poles in a motor led to a reduction in the speed of the motor. In 1897 he invented an electrical configuration to switch between poles in a motor for which he was granted a patent along with his co-worker Karl Arvid Lindström. The new connection was named the "Dahlander connection" and a motor having such a configuration is commonly referred to as a "pole changing motor" or a "Dahlander motor". Operation The Dahlander motor is based on a 'consequent pole' connection. The primary factor in determining the speed of an induction motor is the number of poles, given by the formula (RPM) where ns = Synchronous speed, in revolutions per minute f = AC power frequency p = Number of poles per phase winding A regular induction motor has an equal number of opposite poles; that is, at any instant, there are an equal number of North and South magnetic poles. Some smaller induction motors are connected so that all the poles are identical, causing the motor to act as though there is an equal number of opposite poles in between. A Dahlander motor achieves different speeds by switching the configuration of the electrical windings, indirectly adding or removing poles and thus varying the rotor speed. The poles can be varied at a ratio of 1:2 and thus the speed can be varied at 2:1. Normally, the electrical configuration of windings is varied from a delta connection (Δ) to a double star connection (YY) configuration in order to change the speed of the motor for constant tor
https://en.wikipedia.org/wiki/Phrygionis%20argentata
Phrygionis argentata is a species of moth in the family Geometridae. It was first described by Dru Drury in 1773 from Jamaica. Description Upper Side. Antennae filiform. Body grey. Wings pale yellowish grey. A narrow yellow bar rises near the middle of the anterior wings, which, crossing them and the posterior, ends a little below the body on the abdominal edges; another small bar crosses the anterior wings near the shoulders, both of them being verged with silver. A small dark spot, surrounded with silver, is also placed close to the external edges of the posterior wings; and above it is a yellowish patch reaching to the upper corners. Under Side. Wings pale light-coloured, almost white, immaculate. Margins of the wings entire. Wing-span 1¾ inches (44 mm). References Baptini Moths described in 1773 Taxa named by Dru Drury Descriptions from Illustrations of Exotic Entomology
https://en.wikipedia.org/wiki/Petit%20Computer
Petit Computer is a software development application for the Nintendo DSi and later systems, developed by SmileBoom in Sapporo, Japan. The application is built around a custom dialect of BASIC known as SmileBASIC (not to be confused with the 3DS sequel with the same name). Users can write games and other software using the onscreen keyboard and run the applications from within Petit Computer. The platform supports text-based console applications, visual applications, and any combination of the two. Input is available via hardware buttons, touchscreen input, or the onscreen keyboard. In addition to the code editor and interpreter, Petit Computer includes a simple shell for file management, as well as file sharing functionality. Files can be shared by a direct wireless connection between two DS systems, or by the use of QR codes. The usage of QR codes enabled some users to develop desktop software that can be used to write SmileBASIC and generate a QR code for easy transfer to the DS. Petit Computer comes with several simple sample applications, 5 sample games, and several graphics-editing applications, all written in SmileBASIC with viewable source code. The latter applications can be used to create sprites, backgrounds, and other resources that can then be used within user-created software. Hundreds of premade sprites and tiles are included with Petit Computer. An extensive manual is available from within Petit that describes the basic features and limitations of SmileBASIC, as well as brief descriptions of most of the commands and their syntax. SmileBASIC language Petit Computer uses a customized dialect of BASIC known as SmileBASIC designed specifically for the DSi. Applications written in SmileBASIC can read input from all of the DS's hardware buttons except the Select button (which is always used to terminate the current application) as well as the touch screen, draw graphics and sprites to both screens, and play music written in Music Macro Language. Standa
https://en.wikipedia.org/wiki/Grothendieck%27s%20T%C3%B4hoku%20paper
The article "Sur quelques points d'algèbre homologique" by Alexander Grothendieck, now often referred to as the Tôhoku paper, was published in 1957 in the Tôhoku Mathematical Journal. It revolutionized the subject of homological algebra, a purely algebraic aspect of algebraic topology. It removed the need to distinguish the cases of modules over a ring and sheaves of abelian groups over a topological space. Background Material in the paper dates from Grothendieck's year at the University of Kansas in 1955–6. Research there allowed him to put homological algebra on an axiomatic basis, by introducing the abelian category concept. A textbook treatment of homological algebra, "Cartan–Eilenberg" after the authors Henri Cartan and Samuel Eilenberg, appeared in 1956. Grothendieck's work was largely independent of it. His abelian category concept had at least partially been anticipated by others. David Buchsbaum in his doctoral thesis written under Eilenberg had introduced a notion of "exact category" close to the abelian category concept (needing only direct sums to be identical); and had formulated the idea of "enough injectives". The Tôhoku paper contains an argument to prove that a Grothendieck category (a particular type of abelian category, the name coming later) has enough injectives; the author indicated that the proof was of a standard type. In showing by this means that categories of sheaves of abelian groups admitted injective resolutions, Grothendieck went beyond the theory available in Cartan–Eilenberg, to prove the existence of a cohomology theory in generality. Later developments After the Gabriel–Popescu theorem of 1964, it was known that every Grothendieck category is a quotient category of a module category. The Tôhoku paper also introduced the Grothendieck spectral sequence associated to the composition of derived functors. In further reconsideration of the foundations of homological algebra, Grothendieck introduced and developed with Jean-Louis Verdi
https://en.wikipedia.org/wiki/Occam%20learning
In computational learning theory, Occam learning is a model of algorithmic learning where the objective of the learner is to output a succinct representation of received training data. This is closely related to probably approximately correct (PAC) learning, where the learner is evaluated on its predictive power of a test set. Occam learnability implies PAC learning, and for a wide variety of concept classes, the converse is also true: PAC learnability implies Occam learnability. Introduction Occam Learning is named after Occam's razor, which is a principle stating that, given all other things being equal, a shorter explanation for observed data should be favored over a lengthier explanation. The theory of Occam learning is a formal and mathematical justification for this principle. It was first shown by Blumer, et al. that Occam learning implies PAC learning, which is the standard model of learning in computational learning theory. In other words, parsimony (of the output hypothesis) implies predictive power. Definition of Occam learning The succinctness of a concept in concept class can be expressed by the length of the shortest bit string that can represent in . Occam learning connects the succinctness of a learning algorithm's output to its predictive power on unseen data. Let and be concept classes containing target concepts and hypotheses respectively. Then, for constants and , a learning algorithm is an -Occam algorithm for using iff, given a set of samples labeled according to a concept , outputs a hypothesis such that is consistent with on (that is, ), and where is the maximum length of any sample . An Occam algorithm is called efficient if it runs in time polynomial in , , and We say a concept class is Occam learnable with respect to a hypothesis class if there exists an efficient Occam algorithm for using The relation between Occam and PAC learning Occam learnability implies PAC learnability, as the following theorem of Blu
https://en.wikipedia.org/wiki/3D%20Content%20Retrieval
A 3D Content Retrieval system is a computer system for browsing, searching and retrieving three dimensional digital contents (e.g.: Computer-aided design, molecular biology models, and cultural heritage 3D scenes, etc.) from a large database of digital images. The most original way of doing 3D content retrieval uses methods to add description text to 3D content files such as the content file name, link text, and the web page title so that related 3D content can be found through text retrieval. Because of the inefficiency of manually annotating 3D files, researchers have investigated ways to automate the annotation process and provide a unified standard to create text descriptions for 3D contents. Moreover, the increase in 3D content has demanded and inspired more advanced ways to retrieve 3D information. Thus, shape matching methods for 3D content retrieval have become popular. Shape matching retrieval is based on techniques that compare and contrast similarities between 3D models. 3D retrieval methods Derive a high level description (e.g.: a skeleton) and then find matching results This method describes 3D models by using a skeleton. The skeleton encodes the geometric and topological information in the form of a skeletal graph and uses graph matching techniques to match the skeletons and compare them. However, this method requires a 2-manifold input model, and it is very sensitive to noise and details. Many of the existing 3D models are created for visualization purposes, while missing the input quality standard for the skeleton method. The skeleton 3D retrieval method needs more time and effort before it can be used widely. Compute a feature vector based on statistics Unlike Skeleton modeling, which requires a high quality standard for the input source, statistical methods do not put restriction on the validity of an input source. Shape histograms, feature vectors composed of global geo-metic properties such as circularity and eccentricity, and feature vector
https://en.wikipedia.org/wiki/B%C3%B6hm%27s%20language
Böhm's language refers to the language, machine and a translation method developed by Corrado Böhm during the latter part of 1950. Böhm used this work as part of his dissertation, submitted in 1951 (amended after submission), published in 1954. The compiler Böhm's work described the first complete meta-circular compiler. The code for the compiler was remarkably precise, and consisted of only 114 lines of code. Since the language accepted only two kinds of expressions: fully parenthesized or without parenthesis, but with operator precedence, therefore the code of the compiler split into two parts. 59 lines were used to handle formulas with parenthesis, 51 to handle operator precedence expressions and 4 to decide between those two cases. Böhm's parsing technique for expressions had only linear complexity. It generated instructions to a structure similar to a binary tree. The language Böhm's language consisted of only assignment operations. It had no special constructs like user defined functions, control structures. Variables represented only non-negative integers. To perform a jump one had to write to a special π variable. To perform I/O ? symbol was used. An example program which loads 11-element array from an input would look as follows. A. Set i = 0 (plus the π → G base address 100 for 100 → i the input array a). B → π B. Let a new input a[i] be π' → B given. Increase i by unity, ? → ↓i and stop if i > 10, i+1 → i otherwise repeat B. [(1∩(i∸110))∙Ω]+[(1∸(i∸110))∙B] → π ∩ represents a minimum operator and ∸ logical difference. References Sources Knuth, Donald E.; Pardo, Luis Trabb (1976). "Early development of programming languages". Stanford University, Computer Science Department. Translation Computer languages Computer data Encodings
https://en.wikipedia.org/wiki/4G-LTE%20filter
4G- filter is a low-pass filter or notch filter, to be used in OTA-reception installations (both collective and individual) if one is using an older/unfiltered TV antenna, without the risk of interference caused by the cellular transmitters on the higher channels. (See following section.) These filters are usually used for existing facilities as for new installations in urban or rural areas, antennas and amplifiers sold from the application of the new standard may be already be configured to receive, with good signal gain, only TV channels from 14 to 51 of the UHF band, the other higher channels (former TV channels 52 to 83) being attenuated. Interference between 4G LTE and DTT 4G LTE is the fourth generation mobile phone standard. In urban areas, the 4G uses a frequency band located between 1800 MHz and 2600 MHz, and therefore is quite far from the TV band for causing any type of interference problem. In rural areas, however, the major operators asked to use part of the UHF band. Since the UHF frequency band is not expandable, it was agreed that television broadcasting should limit its number of channels. Thus, the frequency band dedicated to TV became between 470 MHz and 700 MHz (channels 14-52), whilst 4G LTE uses the frequency bands between 700 and 900 MHz (former TV channels 52 to 83), resulting in an interval separating the two bands (DTT and 4G) of about 1 MHz, so that there is a risk of interference in the areas close to the 4G-LTE transmitting towers. In practice, these bands used 698 MHz to 960 MHz (depending on the carrier). See previous section on Filters. This re-allocation of TV bandwidth to 4G is called Digital dividend. Digital dividend (Europe) Digital dividend to frequency 698 to 806 MHz (TV Channels 61 to 69) assigned by the plan for the New UHF Frequency Band distribution agreed in The World Radio Congress (WRC-07) which identified 108 MHz of Digital Dividend Spectrum from 698 to 806 MHz for ITU-R Regions 2-1 and nine countries in Re
https://en.wikipedia.org/wiki/Lyssa%20patroclus
Lyssa patroclus is a species of moth in the family Uraniidae. The species was described by Carl Linnaeus in his 1758 10th edition of Systema Naturae from the Moluccas. Description Upperside: The antennae are about an inch (25 mm) long, slender, setaceous, and gradually diminishing from the base to the extremities. The head is small. The thorax is clothed with long soft hair, and, with the abdomen, is of a darkish brown. A remarkable straight narrow line, or bar, of a cream colour, arises from the middle of the anterior edge of each of the forewings, and, crossing both anterior and posterior wings, ends at the abdominal edges, about half an inch below the abdomen; so that, when the wings are extended, as in the figure, these lines, with the anterior edges, form an equilateral triangle. The space within the triangle is dark brown; but the parts, near the shoulders, are lighter, having a greyish cast or hue, and contain many small transverse curved streaks, extending to the anterior edges, where they are large, black, and very conspicuous, like stripes. Some transverse markings of this kind, are dispersed on the posterior edges of the superior wings, and also on the abdominal edges of the posterior. On the outside the triangle, both on the anterior and posterior wings, is a fascia, of light brown, about half an inch broad, which deepens into a dark brown. On the posterior wings, after becoming dark, it softens again into the same light colour, continuing to the external edges. Each of these wings is ornamented with two tails, the inner ones the longest and near an inch in length, the tips of which incline towards each other; the lesser or outer tails, are about half an inch long, strengthened by the tendons of the wings passing through the middle of them; all of them being bordered with a soft ray of dark brown. Underside: The bars or lines, which form the triangle on the upperside, are not visible on this; but the inclosed triangular field appears of a light greyish
https://en.wikipedia.org/wiki/Wouzee
wouzee is an interactive broadcast platform for streaming video that allows users to visualize and share play and stream video using different devices video camera or a computer through Internet. The user produces, shares and consumes information actively. In addition to the free support, Wouzee offers a premium service through livestreaming event production. Wouzee became known nationally when one of their users decided to upload a video in which Infanta Cristina stated on the Noos Case. afterwards, was published by the Spanish newspaper El Mundo. References External links Official site Video
https://en.wikipedia.org/wiki/CROC
JSC CROC incorporated () is an technology company operating in the Russian IT market. CROC was established by Boris Bobrovnikov in 1992. The company provides systems integration and managed B2B services, off-the-shelf products and promising end-to-end technologies, such as Big data, blockchain, artificial intelligence, Internet of things, RPA, and machine learning, while also offering cloud and data center services based on its own data centers. It is headquartered in Moscow. History Founded in 1992 in Moscow, CROC first supplied computers, peripherals, and network equipment and provided relevant maintenance services. In 1995, CROC focused on systems integration as its core business and then, in 1996–1998, expanded its portfolio with computing and structured cabling systems, networks, enterprise telephony, and storage systems. In 2000–2001, the company increased the number of its regional Compaq authorized service centers. Also, in 2001, the CROC North-West branch was opened in St. Petersburg, and later, in 2004, another CROC Ural branch was launched in Yekaterinburg. Thus, CROC has been one of the largest national IT service providers since the early 2000s. In 2005, CROC proceeded with the creation and support of data storage and processing systems. Addressing the economic recession of 2008—2009 the company completed more virtualization projects, container data center implementations and relocations, as well as outsourcing engagements, while CROC was also developing expertise in the deployment of video conferencing and AV systems. In early 2009, CROC's Moscow offices moved to a single headquarters located at Volochaevskaya Street. In 2010, the company launched Russia's first domestic cloud platform. The first renewable energy projects were completed in 2011. The next year, CROC innovative force started to implement virtual reality (VR) projects. In 2016, CROC Disk was launched, a secure and handy corporate file exchange service. CROC installed security, TV
https://en.wikipedia.org/wiki/SIM.JS
SIM.JS is an event-based discrete-event simulation library based on standard JavaScript. The library has been written in order to enable simulation within standard browsers by utilizing web technology. SIM.JS supports entities, resources (Facility, Buffers and Stores), communication (via Timers, Events and Messages) and statistics (with Data Series, Time Series and Population statistics). The SIM.JS distribution contains tutorials, in-depth documentation, and a large number of examples. SIM.JS is released as open source software under the LGPL license. The first version was released in January 2011. Example There are several examples bundled with the library download. Trafficlight simulation is a standard simulation problem, which may be simulated as within this example: External links Analysis of the potential role of open source discrete event simulation software in the manufacturing sector issued by Proceedings of the Operational Research Society Simulation Workshop 2012 comparing SIM.JS among other open source simulation solutions Bachelor thesis on Web-Based Single-Player Project Simulation Game reviewing SIM.JS Interactive Model-Centric Systems Engineering report by Systems Engineering Research Center managed by Stevens Institute of Technology mentioning SIM.JS "Just-In-Time parallel simulation" compares simulation programmed in JIT languages versus AOT languages referencing SIM.JS Modeling and Simulation of Tape Libraries for Hierarchical Storage Management Systems, Master Thesis reviewing DES frameworks, including SIM.JS Cross-platform software Free science software Simulation programming languages
https://en.wikipedia.org/wiki/Deanna%20M.%20Church
Deanna Church is a scientist working in the areas of bioinformatics and genomics. She is known for her work on the human genome, "making the genome a friendlier place". Life Church graduated with a bachelor's degree from the University of Virginia in 1990. She received a doctorate in Genomics from University of California, Irvine in 1997. Church describes her passion for bioinformatics as connected to her enjoyment of problem solving and being in a team that has direct impacts on people's medical care. Work Church worked for the National Center for Biotechnology Information (NCBI) from 1999 until 2013. While there, she headed NCBI's team in the Genome Reference Consortium, an international group focusing on refining data on the human genome. She was part of the group involved in releasing GRCh38, a build of the human genome that included centromere sequences for the first time. In 2013, she joined Personalis as Senior Director of Genomics and Content where she worked towards improving bioinformatics for better analysis of the human genome. In 2016, she joined 10x Genomics as Senior Director of Applications. Church has had over 35 publications in her career. See also Human genome project Mouse Genome Informatics References 21st-century American biologists Human Genome Project scientists Living people American biotechnologists American bioinformaticians University of California, Irvine alumni University of Virginia alumni 21st-century American economists Year of birth missing (living people)
https://en.wikipedia.org/wiki/Weber%27s%20glands
The Weber's glands are muciparous glands on the side of the tongue. They are a minor salivary gland in the peritonsillar space. The glands are named after German anatomist Moritz Ignaz Weber. They clear the peritonsillar space of debris. References Glands Tongue Anatomy named for one who described it
https://en.wikipedia.org/wiki/Band-gap%20engineering
Band-gap engineering is the process of controlling or altering the band gap of a material. This is typically done to semiconductors by controlling the composition of alloys, constructing layered materials with alternating compositions, or by inducing strain either epitaxially or topologically. A band gap is the range in a solid where no electron state can exist. The band gap of insulators is much larger than in semiconductors. Conductors or metals have a much smaller or nonexistent band gap than semiconductors since the valence and conduction bands overlap. Controlling the band gap allows for the creation of desirable electrical properties. Molecular-beam epitaxy (MBE) Molecular-beam epitaxy is a technique used to construct thin epitaxial films of materials ranging from oxides to semiconductors to metals. Different beams of atoms and molecules in an ultra-high vacuum environment are shot onto a nearly atomically clean crystal, creating a layering effect. This is a type of thin-film deposition. Semiconductors are the most commonly used material due to their use in electronics. Technologies such as quantum well devices, super-lattices, and lasers are possible with MBE. Epitaxial films are useful due to their ability to be produced with electrical properties different from those of the substrate, either higher purity, or fewer defects or with a different concentration of electrically active impurities as desired. Varying the composition of the material alters the band gap due to bonding of different atoms with differing energy level gaps. Strain-induced band-gap engineering Semiconducting materials are able to be altered with strain-inducing from tunable sizes and shapes due to quantum confinement effects. A larger tunable bandgap range is possible due to the high elastic limit of semiconducting nanostructures (Guerra, and Guerra and Vezenov). Strain is the ratio of extension to original length, and can be used on the nanoscale. Thulin and Guerra (2008) theoreti
https://en.wikipedia.org/wiki/Approach%20and%20departure%20angles
Approach angle is the maximum angle of a ramp onto which a vehicle can climb from a horizontal plane without interference. It is defined as the angle between the ground and the line drawn between the front tire and the lowest-hanging part of the vehicle at the front overhang. Departure angle is its counterpart at the rear of the vehicle – the maximum ramp angle from which the car can descend without damage. Approach and departure angles are also referred to as ramp angles. Approach and departure angles are indicators of off-road ability of the vehicle: they indicate how steep obstacles, such as rocks or logs, the vehicle can negotiate according to its body shape alone. See also Breakover angle Overhang (automotive) Ride height References External links Approach and Departure Angles at Why High End? Automotive engineering
https://en.wikipedia.org/wiki/Edimax
Edimax Technology Co., Ltd. () is a Taiwanese manufacturer of data networking products. It also operates in the telephone and telegraph apparatus sector. The company was founded in 1986 and has been listed at the Taiwan Stock Exchange since 2001. See also List of companies of Taiwan References External links Open source software support Manufacturing companies based in Taipei Telecommunications companies of Taiwan Networking hardware companies Taiwanese companies established in 1986 Electronics companies of Taiwan Companies listed on the Taiwan Stock Exchange Taiwanese brands
https://en.wikipedia.org/wiki/Rate%20making
Rate making, or insurance pricing, is the determination of rates charged by insurance companies. The benefit of rate making is to ensure insurance companies are setting fair and adequate premiums given the competitive nature. Fundamental rate-making definitions The following are fundamental terms that are commonly used in rate making. A rate "is the price per unit of insurance for each exposure unit, which is the unit of measurement used in insurance pricing". The exposure unit is used to establish insurance premiums by examining parallel groups. The pure premium "refers to that portion of that rate needed to pay losses and loss-adjustment expenses". The loading "refers to the amount of the premium necessary to cover other expenses, particularly sales expenses, and to allow for a profit". The gross rate "is the pure premium and the loading per exposure unit". Finally, the gross premium is the premium paid by the insured consisting of the gross rate multiplied by the number of exposure units. Objectives in rate making Rate making has several objectives under regulatory requirements regulated by the states and business objectives due to the goal of profitability: The goal of insurance regulation is to protect the public and three regulatory objectives are placed to meet certain standards: The first regulatory requirement is that rates must be adequate; meaning the rates the insurers charge should be able to cover expenses. The second regulatory requirement is that rates must not be excessive; meaning rates should not be so high that policyholders are paying more than the actual value of their protection. The third regulatory objective is the rates must not be unfairly discriminatory; meaning exposures that are similar with respect to losses and expenses should not be charged significantly different rates. The business objectives are set as a guide for insurers while designing the rating system. The rating system should meet each of the four objectives: For
https://en.wikipedia.org/wiki/List%20of%20Nigerian%20engineers
This is a list of notable engineers in Nigeria arranged in alphabetical order. A Akinsola Olusegun Faluyi (Born, 13 November 1934) is a Nigerian Mechanical engineer and former President of the Council for the Regulation of Engineering (COREN) in Nigeria. Abubakar Sani Sambo, Officer of the Order of the Niger, OON (Born 31 July 1955) is a Nigerian Mechanical engineer, incumbent Director-General of the Energy Commission of Nigeria (ECN) Mayen Adetiba (born 1954) is a Nigerian actor who went on to be a leading civil engineer. Ahmad Salihijo Ahmad is a Nigerian Engineer Godman Akinlabi is a Nigerian pastor, author, public speaker and engineer. Deji Akinwande Professor of Electrical and Computer Engineering Demola Aladekomo is a Computer Engineer Bolaji Aluko a Professor of Chemical Engineering Edet Amana (born December 11, 1938) is a Nigerian engineer. Idiat Amusu female agricultural engineer in Nigeria Aderemi Aaron-Anthony Atayero Ifedayo Akintunde, Nigerian engineer, first African to win the WFEO Medal of Engineering Excellence. Distinguished Fellow of the Nigerian Society of Engineers (NSE), and former President of the society. ALHAO rahman babatunde, male born august 25, 1974. is a civil engineer in Nigeria. leading civil engineer B Babalola Borishade an electrical engineer, Maikanti Baru Olufemi Bamiro Professor of Mechanical Engineering Rahmon Ade Bello s a Nigerian professor of Chemical engineering, educational administrator and former vice chancellor of the University of Lagos, Nigeria D Danladi Slim Matawal, (Born 30 October 1955) is a Nigerian professor of Civil engineering and the Director-General of the Nigerian Building and Road Research Institute (NBRRI) John Dabiri a Nigerian-American aeronautics engineer E Ebele Ofunneamaka Okeke (14 June 1948) is a Nigerian civil engineer and former Head of Nigerian Civil Service Edet James Amana (11 December 1938) is a Nigerian Civil engineer and former President of the Nigerian Academy of Engineering
https://en.wikipedia.org/wiki/Marcelo%20Viana
Marcelo Miranda Viana da Silva (born 4 March 1962) is a Brazilian mathematician working in dynamical systems theory. He was a Guggenheim Fellow in 1993. He received the TWAS Prize in 1998 and in 2005 he was awarded the inaugural ICTP Ramanujan Prize for his research achievements. Viana was vice-president of the International Mathematical Union in 2011–2014, and president of the Brazilian Mathematical Society (2013–2015). In 1998, he was a plenary speaker at the International Congress of Mathematicians, in Berlin. Viana is director elected of the IMPA (for the period 2016–2019). Viana is a columnist for Folha de S.Paulo. He is the chair of the executive committee for the 2018 International Congress of Mathematicians, Rio de Janeiro. Biography Viana was born in Rio de Janeiro, Brazil, his parents being Portuguese. He grew up in Portugal, and his undergraduate studies were at the University of Porto. He received his Ph.D. degree from the IMPA in Rio de Janeiro, with Jacob Palis as advisor. He is now director at IMPA. Work Viana's work concerns chaotic dynamical systems and strange attractors. Selected publications jointly with AVILA, A., "Simplicity of Lyapunov spectra: proof of the Zorich–Kontsevich conjecture". Acta Mathematica. vol. 198 (2007), no. 1, pp. 1–56. jointly with PALIS, J., "High dimension diffeomorphisms displaying infinitely many periodic attractors". Annals of Mathematics. vol. 140 (1994), no. 1, pp. 207–250. jointly with MORA, L., "Abundance of strange attractors". Acta Mathematica. vol. 171 (1993), no. 1, pp. 1–71. References External links Home page of Marcelo Viana Unión Matemática de América Latina y el Caribe – UMALCA Award in Mathematics 2000: Marcelo Viana Viana teaching a course on ordinary differential equations (in Portuguese) Interview with Marcelo Viana, conducted by Maria Manuel Clementino and Jorge Picado 1962 births Living people University of Porto alumni Members of the Brazilian Academy of Sciences Dynamical
https://en.wikipedia.org/wiki/Intermediate%20band%20photovoltaics
Intermediate band photovoltaics in solar cell research provides methods for exceeding the Shockley–Queisser limit on the efficiency of a cell. It introduces an intermediate band (IB) energy level in between the valence and conduction bands. Theoretically, introducing an IB allows two photons with energy less than the bandgap to excite an electron from the valence band to the conduction band. This increases the induced photocurrent and thereby efficiency. Limiting efficiencies One band Luque and Marti first derived a theoretical limit for an IB device with one midgap energy level using detailed balance. They assumed no carriers were collected at the IB and that the device was under full concentration. They found the maximum efficiency to be 63.2%, for a bandgap of 1.95eV with the IB 0.71eV from either the valence or conduction band. Under one sun illumination the limiting efficiency is 47%. Infinite bands Green and Brown expanded upon these results by deriving the theoretical efficiency limit for a device with infinite IBs. By introducing more IB's, even more of the incident spectrum can be utilized. After performing the detailed balance, they found the maximum efficiency to be 77.2%. This efficiency is less than that of a multijunction cell with infinite junctions. This is because in multijunction cells, electrons are captured exactly after being excited to a higher energy state, while in an IB device, the electrons still need another energy transition to reach the conduction band and be collected. Current technology IBs have theoretical potential to become high efficiency devices, but they are hard to make. Introducing an IB greatly increases non-radiative recombination mechanisms. Additionally, IBs need to be partially filled to allow for carrier movement to and from the IB. This often requires donor carriers. The three current methods of fabricating IB devices are described below. Quantum dots The first method is to introduce small, homogeneous quantum dot
https://en.wikipedia.org/wiki/Aquifer%20properties
The aquifer properties of the aquifer essentially depend upon the composition of the aquifer. The most important properties of the aquifer are porosity and specific yield which in turn give its capacity to release the water in the pores and its ability to transmit the flow with ease. Porosity In soil mechanics and hydrology, porosity is defined as the ratio of volume of voids to the total volume of porous medium. Mathematically it can be represented by the equation expressed as where is void ratio, is porosity, VV is the volume of void-space (air and water), VS is the volume of solids, and VT is the total or bulk volume of medium. The significance of the porosity is that it gives the idea of water storage capacity of the aquifer. Qualitatively, porosity less than 5% is considered to be small, between 5 and 20% as medium and the percentage exceeding 20% is considered as large. Specific yield Porosity gives a measure of the water storage capability of soil but not all the water present in the soil pores is available for extraction by pumping for the use of humans or draining by gravity. The pores in the soil hold back sufficient quantity of water on account of forces like surface tension and molecular attraction. Hence the actual amount of water that can be extracted from the unit volume of aquifer by pumping or under the action of gravity is called as specific yield. The fraction of water held back in the aquifer is known as specific retention. Thus it can be said that porosity is the sum of specific yield and specific retention. Specific yield of soils differ from each other in the sense that some soil types have strong molecular attraction with the water held in their pores while others have less. It is found experimentally that cohesionless soils have high specific yield than cohesive soils because the former has significantly less molecular attraction then the latter. Coarse-grained soils or rocks such as coarse sandstone can have specific yields that are cl
https://en.wikipedia.org/wiki/Verification-based%20message-passing%20algorithms%20in%20compressed%20sensing
Verification-based message-passing algorithms (VB-MPAs) in compressed sensing (CS), a branch of digital signal processing that deals with measuring sparse signals, are some methods to efficiently solve the recovery problem in compressed sensing. One of the main goal in compressed sensing is the recovery process. Generally speaking, recovery process in compressed sensing is a method by which the original signal is estimated using the knowledge of the compressed signal and the measurement matrix. Mathematically, the recovery process in Compressed Sensing is finding the sparsest possible solution of an under-determined system of linear equations. Based on the nature of the measurement matrix one can employ different reconstruction methods. If the measurement matrix is also sparse, one efficient way is to use Message Passing Algorithms for signal recovery. Although there are message passing approaches that deals with dense matrices, the nature of those algorithms are to some extent different from the algorithms working on sparse matrices. Overview The main problem in recovery process in CS is to find the sparsest possible solution to the following under-determined system of linear equations where is the measurement matrix, is the original signal to be recovered and is the compresses known signal. When the matrix is sparse, one can represent this matrix by a bipartite graph for better understanding. is the set of variable nodes in which represents the set of elements of and also is the set of check nodes corresponding to the set of elements of . Besides, there is an edge between and if the corresponding elements in is non-zero, i.e. . Moreover, the weight of the edge . Here is an example of a binary sparse measurement matrix where the weights of the edges are either zero or one. The basic idea behind message passing algorithms in CS is to transmit appropriate messages between variable nodes and check nodes in an iterative manner in order to efficiently f
https://en.wikipedia.org/wiki/Bismuth%E2%80%93indium
The elements bismuth and indium have relatively low melting points when compared to other metals, and their alloy bismuth–indium (Bi–In) is classified as a fusible alloy. It has a melting point lower than the eutectic point of the tin–lead alloy. The most common application of the Bi-In alloy is as a low temperature solder, which can also contain, besides bismuth and indium, lead, cadmium, and tin. Metals Bismuth Bismuth has many unique characteristics. When solidifying, bismuth's volume expands by roughly 2.32%. Its electrical resistance is twice as high in its solid state than in its liquid form. Bismuth has one of the lowest thermal conductivities of pure elemental metals. It is fragile, highly diamagnetic and it has a magnetic susceptibility of −1.68×10−5 mks. Bismuth is used as catalyst in the production of plastics and cosmetics, as an additive in steel alloys, and in electronics. It has a rhombohedral (Biα) structure, with an atomic radius of 1.54 Å, electronegativity of 1.83, and valence of +3 and +5. Indium Indium is a metal softer than lead (hardness of 0.9 HB), permitting it to be scratched by a nail. It is also malleable, ductile and has a thermal conductivity value of 0.78 W/m°C (85 °C). It also has the capacity of wetting glass, quartz and other ceramic materials. It maintains the plasticity and ductility when exposed to cryogenic environments and has a big gap between the melting point and the boiling point (156.6 °C and 2080 °C respectively). Under compression, it has high plasticity that allows almost unlimited deformation (2.14 MPa of compression resistance) and under tensile it has low elongation (4 MPa of tensile resistance). Indium is used in dental alloys, semiconductor components, nuclear reactor panels, sodium lamps, strengthening factor in lead-based welds and low melting temperature welds. The metal has a body centered tetragonal structure, atomic radius of 1.63 Å, electronegativity of 1.81 and valence of +3 or +5, being the trivalent
https://en.wikipedia.org/wiki/Zuhair%20Fayez%20Partnership
Zuhair Fayez Partnership (ZFP) (Arabic: شركة زهير فايز ومشاركوه), formerly Zuhair Fayez and Associates (ZFA), is an Architectural, Engineering, Project Management, Construction Management, and Engineering Information System consultancy firm in the Kingdom of Saudi Arabia. History Zuhair Fayez Partnership was established in 1975 as Zuhair Fayez and Associates by Zuhair H. Fayez and started with a small number of professional architects, engineers, planners, and support staff. The growing number of commissions and projects eventually prompted the organization to open an office in London to meet the demand. Eight departments were initially formed: Design, Architectural Production, Structural Engineering, Civil Engineering, HVAC, Specifications, and Quantity Surveying departments. However, the demand for professional management and supervision of large scale projects led to the creation of the Project Management and Construction Management departments to effectively monitor and control all contractors and subcontractors' work on any construction project. In 1991, the organization launched its Information Technology division to provide clients with software and hardware. The division handled the software integration of Maintenance Management Systems (MMS), Geographical Information Systems (GIS), and Facilities Management Systems (FMS) into a single Graphical User Interface (GUI). In the same year, the Engineering Information Systems division was formed and is responsible for on-site implementation and provision of GIS, MMS, and FMS based on the clients' needs and requirements. ZFP partnered with Avaya in May 2011 to create a virtual computing environment (VCE). The purpose is to enable the organization to provide virtual services in a cloud-based environment that would serve 3000 users in 5 branches and 20 remote sites and maximize the efficiency of its existing IT resources. In July 2012, SNC-Lavalin acquired ZFP's industrial division in Khobar to form SNC-Lavalin
https://en.wikipedia.org/wiki/Verband%20Deutscher%20Tonmeister
The Verband Deutscher Tonmeister e.V. (VDT) is a registered association for audio industry professionals. The VDT has evolved from the Deutsche Filmtonmeister-Vereinigung (which focused on film sound professionals) that was founded in Munich in 1950. There are currently more than 1900 members in the VDT that are either freelancers or employed in various institutions; e.g. in the film business, radio broadcasting, television, recording studios, record labels, theaters and performance venues, the audio equipment producing industry, research and development, multimedia education and other audio related areas of occupation. Students preparing for a job in the audio industry are also members of the VDT, making up roughly 10%. Even though the designation of Tonmeister is a fixed part of the name of the association its members carry many other job titles like audio engineer, sound director, music director, sound designer, producer, record producer and music supervisor. The occupational title Tonmeister, although created in the 1930s, is not protected in Germany and does not describe a clearly defined occupation. Thematically the goals and activities of the VDT are related to those of the Audio Engineering Society (AES), even if there is no formal connection. The biennial Tonmeistertagung (VDT International convention) combines an international scientific conference and workshop program with a trade fair, focusing on applied research and technical excellence. It has been in existence since 1949. Similarly the International Tonmeister Symposium, dedicated to one thematic complex, is analogous to the AES-conferences. The Tonmeister Academy as initiated by the VDT offers educational classes and advanced training sessions. The VDT award Golden Bobby is awarded in seven categories for outstanding performance in sound recording and mixing. Since 2002 the VDT also bestows a medal of honour on members who have rendered outstanding services to the profession of sound engineers
https://en.wikipedia.org/wiki/Cloud%20robotics
Cloud robotics is a field of robotics that attempts to invoke cloud technologies such as cloud computing, cloud storage, and other Internet technologies centered on the benefits of converged infrastructure and shared services for robotics. When connected to the cloud, robots can benefit from the powerful computation, storage, and communication resources of modern data center in the cloud, which can process and share information from various robots or agent (other machines, smart objects, humans, etc.). Humans can also delegate tasks to robots remotely through networks. Cloud computing technologies enable robot systems to be endowed with powerful capability whilst reducing costs through cloud technologies. Thus, it is possible to build lightweight, low-cost, smarter robots with an intelligent "brain" in the cloud. The "brain" consists of data center, knowledge base, task planners, deep learning, information processing, environment models, communication support, etc. Components A cloud for robots potentially has at least six significant components: Building a "cloud brain" for robots. It is the main object of cloud robotics. Offering a global library of images, maps, and object data, often with geometry and mechanical properties, expert system, knowledge base (i.e. semantic web, data centres); Massively-parallel computation on demand for sample-based statistical modelling and motion planning, task planning, multi-robot collaboration, scheduling and coordination of system; Robot sharing of outcomes, trajectories, and dynamic control policies and robot learning support; Human sharing of "open-source" code, data, and designs for programming, experimentation, and hardware construction; On-demand human guidance and assistance for evaluation, learning, and error recovery; Augmented human–robot interaction through various way (Semantics knowledge base, Apple SIRI like service etc.). Applications Autonomous mobile robots Google's self-driving cars are cloud robots.
https://en.wikipedia.org/wiki/Matrix%20regularization
In the field of statistical learning theory, matrix regularization generalizes notions of vector regularization to cases where the object to be learned is a matrix. The purpose of regularization is to enforce conditions, for example sparsity or smoothness, that can produce stable predictive functions. For example, in the more common vector framework, Tikhonov regularization optimizes over to find a vector that is a stable solution to the regression problem. When the system is described by a matrix rather than a vector, this problem can be written as where the vector norm enforcing a regularization penalty on has been extended to a matrix norm on . Matrix regularization has applications in matrix completion, multivariate regression, and multi-task learning. Ideas of feature and group selection can also be extended to matrices, and these can be generalized to the nonparametric case of multiple kernel learning. Basic definition Consider a matrix to be learned from a set of examples, , where goes from to , and goes from to . Let each input matrix be , and let be of size . A general model for the output can be posed as where the inner product is the Frobenius inner product. For different applications the matrices will have different forms, but for each of these the optimization problem to infer can be written as where defines the empirical error for a given , and is a matrix regularization penalty. The function is typically chosen to be convex and is often selected to enforce sparsity (using -norms) and/or smoothness (using -norms). Finally, is in the space of matrices with Frobenius inner product . General applications Matrix completion In the problem of matrix completion, the matrix takes the form where and are the canonical basis in and . In this case the role of the Frobenius inner product is to select individual elements from the matrix . Thus, the output is a sampling of entries from the matrix . The problem of r
https://en.wikipedia.org/wiki/Genetics%20of%20post-traumatic%20stress%20disorder
The genetic influences of post-traumatic stress disorder (PTSD) are not understood well due to the limitations of any genetic study of mental illness; in that, it cannot be ethically induced in selected groups. Because of this, all studies must use naturally occurring groups with genetic similarities and differences, thus the amount of data is limited. Still, genetics play some role in the development of PTSD. Research and potential influences Approximately 30% of the variance in PTSD is caused by genetics alone. For twins exposed to combat in the Vietnam War, a monozygotic (identical) twin with PTSD was associated with an increased risk of the co-twin having PTSD, as compared to dizygotic (non-identical) twins; additionally, assaultive trauma (compared to non-assaultive trauma) was more likely to exacerbate these effects. There is also evidence that those with a genetically smaller hippocampus are more likely to develop PTSD following a traumatic event. Research has also found that PTSD shares many genetic influences common to other psychiatric disorders. Panic and generalized anxiety disorders and PTSD share 60% of the same genetic variance. Alcohol, nicotine, and drug dependence share greater than 40% genetic similarities. Additional disorders—such as depression, schizophrenia, and bipolar disorder—share the same fundamental genetic phenotypes as PTSD. Nature vs. nurture An individual's potential for onset of many psychological disorders is heavily affected by genetic phenotypes, yet this is not the only contributing factor. Environment plays an important role as well, especially for trauma-based disorders such as PTSD, considering that certain life experiences can trigger the activation of an underlying genetic phenotype which might have been previously dormant. This can be further understood by examining the diathesis-stress model for the onset of psychological disorders, which explains that certain individuals, due to their genetic phenotypes, are more s
https://en.wikipedia.org/wiki/M-theory%20%28learning%20framework%29
In machine learning and computer vision, M-theory is a learning framework inspired by feed-forward processing in the ventral stream of visual cortex and originally developed for recognition and classification of objects in visual scenes. M-theory was later applied to other areas, such as speech recognition. On certain image recognition tasks, algorithms based on a specific instantiation of M-theory, HMAX, achieved human-level performance. The core principle of M-theory is extracting representations invariant under various transformations of images (translation, scale, 2D and 3D rotation and others). In contrast with other approaches using invariant representations, in M-theory they are not hardcoded into the algorithms, but learned. M-theory also shares some principles with compressed sensing. The theory proposes multilayered hierarchical learning architecture, similar to that of visual cortex. Intuition Invariant representations A great challenge in visual recognition tasks is that the same object can be seen in a variety of conditions. It can be seen from different distances, different viewpoints, under different lighting, partially occluded, etc. In addition, for particular classes objects, such as faces, highly complex specific transformations may be relevant, such as changing facial expressions. For learning to recognize images, it is greatly beneficial to factor out these variations. It results in much simpler classification problem and, consequently, in great reduction of sample complexity of the model. A simple computational experiment illustrates this idea. Two instances of a classifier were trained to distinguish images of planes from those of cars. For training and testing of the first instance, images with arbitrary viewpoints were used. Another instance received only images seen from a particular viewpoint, which was equivalent to training and testing the system on invariant representation of the images. One can see that the second classifier perfo
https://en.wikipedia.org/wiki/Conservative%20replacement
A conservative replacement (also called a conservative mutation or a conservative substitution) is an amino acid replacement in a protein that changes a given amino acid to a different amino acid with similar biochemical properties (e.g. charge, hydrophobicity and size). Conversely, a radical replacement, or radical substitution, is an amino acid replacement that exchanges an initial amino acid by a final amino acid with different physicochemical properties. Description There are 20 naturally occurring amino acids, however some of these share similar characteristics. For example, leucine and isoleucine are both aliphatic, branched hydrophobes. Similarly, aspartic acid and glutamic acid are both small, negatively charged residues. Although there are many ways to classify amino acids, they are often sorted into six main classes on the basis of their structure and the general chemical characteristics of their side chains (R groups). Physicochemical distances aim at quantifying the intra-class and inter-class dissimilarity between amino acids based on their measurable properties, and many such measures have been proposed in the literature. Owing to their simplicity, two of the most commonly used measures are the ones of Grantham (1974) and Miyata et al (1979). A conservative replacement is therefore an exchange between two amino acids separated by a small physicochemical distance. Conversely, a radical replacement is an exchange between two amino acids separated by a large physicochemical distance. Impact on function Conservative replacements in proteins often have a better effect on function than non-conservative replacements. The reduced effect of conservative replacements on function can also be seen in the occurrence of different replacements in nature. Non-conservative replacements between proteins are far more likely to be removed by natural selection due to their deleterious effects. See also Segregating site Ultra-conserved element Sequence alignm
https://en.wikipedia.org/wiki/Splice%20%28platform%29
Splice is a cloud-based music creation platform founded by Matt Aimonetti and Steve Martocci which includes a sample library, audio plug-ins on a subscription basis, and integration with several digital audio workstations (DAWs). The program is available for MacOS, Windows, iOS and Android. History The site and the MacOS version of Studio Audiosed was launched in private beta in October 2013. A Windows version was released a few months later and the offer became available to the public in September 2014. Splice Studio allows musicians to remotely collaborate through the cloud. The technology is compatible with several popular digital audio workstations (DAWs) programs including Ableton Live, Logic Pro X, FL Studio, Garageband, and Studio One. Its subscription-based sample marketplace, Splice Sounds, was launched in 2015. In 2016, Splice introduced Rent-To-Own, which allowed users to pay a monthly fee to access premium products like synths. The model enabled users to put their rental balance toward the ownership cost and own the product outright or move on to a new program without paying the full market price. The Rent-To-Own program expanded in 2017 to products like Ozone 8 and Neutron 2, prompting The Verge to proclaim: "Splice is making it much easier to buy expensive music production plugins". In 2019, Splice launched its proprietary AI-driven feature Similar Sounds, giving users machine-learning based similarity search across the entire marketplace of audio samples. In November, the company introduced a mobile version of its app on iOS and Android. In 2020 Splice launched the inaugural Splice Awards, recognizing artists in multiple categories, including "Splice Producer of the Year", "Splice Breakout Artist of the Year", "Splice New Standard", "Best Use of Splice Sounds" and "Best Sound Design". The Awards also recognized Splice's "Most Downloaded One Shot", "Most Downloaded Loop", and "Most Downloaded Pack & Vocal Pack". In June 2021, Splice launched Sk
https://en.wikipedia.org/wiki/Address%20generation%20unit
The address generation unit (AGU), sometimes also called address computation unit (ACU), is an execution unit inside central processing units (CPUs) that calculates addresses used by the CPU to access main memory. By having address calculations handled by separate circuitry that operates in parallel with the rest of the CPU, the number of CPU cycles required for executing various machine instructions can be reduced, bringing performance improvements. While performing various operations, CPUs need to calculate memory addresses required for fetching data from the memory; for example, in-memory positions of array elements must be calculated before the CPU can fetch the data from actual memory locations. Those address-generation calculations involve different integer arithmetic operations, such as addition, subtraction, modulo operations, or bit shifts. Often, calculating a memory address involves more than one general-purpose machine instruction, which do not necessarily decode and execute quickly. By incorporating an AGU into a CPU design, together with introducing specialized instructions that use the AGU, various address-generation calculations can be offloaded from the rest of the CPU, and can often be executed quickly in a single CPU cycle. Capabilities of an AGU depend on a particular CPU and its architecture. Thus, some AGUs implement and expose more address-calculation operations, while some also include more advanced specialized instructions that can operate on multiple operands at a time. Furthermore, some CPU architectures include multiple AGUs so more than one address-calculation operation can be executed simultaneously, bringing further performance improvements by capitalizing on the superscalar nature of advanced CPU designs. For example, Intel incorporates multiple AGUs into its Sandy Bridge and Haswell microarchitectures, which increase bandwidth of the CPU memory subsystem by allowing multiple memory-access instructions to be executed in parall
https://en.wikipedia.org/wiki/Code%20Project
Code Project (formerly The Code Project) is a community for computer programmers with articles on different topics and programming languages such as web development, software development, C++, Java, and other topics. Once a visitor registers a user account on the site, they can gain reputation which allows users to unlock different privileges such as the ability to store personal files in the user's account area, have live hyperlinks in their profile biography, and more. Members can also write and upload their own articles and code for other visitors to view. Articles can be related to general programming, GUI design, algorithms or collaboration. Most of the articles are uploaded by visitors and do not come from an external source. Nearly every article is accompanied with source code and examples which can be downloaded independently. Most articles and sources are released under the Code Project Open License (CPOL), although the license can be configured by the user. These articles either go through a moderation and editing phase or are immediately posted as unedited reader contributions. Code Project employs a rating and comment system that helps to filter the good articles from the poor. It also has forums, and is a resource for resolving difficult software development issues. Rather than being just a collection of samples, contributors are encouraged to explain concepts and ideas, and discuss design decisions. A separate format, "Tips and Tricks", was introduced in 2010 as a place to post short code snippets that don't fit the requirements for an article. Code Project strives to be a wealth of information and a valuable resource. The site encourages users to share what source code or knowledge they can in order to give back to the community. Code Project also conducts interviews with notable developers. Code Project also awards Code Project Members Choice Awards in various categories. These awards are based on the votes of Code Project members and editors,
https://en.wikipedia.org/wiki/High%20Bandwidth%20Memory
High Bandwidth Memory (HBM) is a computer memory interface for 3D-stacked synchronous dynamic random-access memory (SDRAM) initially from Samsung, AMD and SK Hynix. It is used in conjunction with high-performance graphics accelerators, network devices, high-performance datacenter AI ASICs, as on-package cache in CPUs and on-package RAM in upcoming CPUs, and FPGAs and in some supercomputers (such as the NEC SX-Aurora TSUBASA and Fujitsu A64FX). The first HBM memory chip was produced by SK Hynix in 2013, and the first devices to use HBM were the AMD Fiji GPUs in 2015. High Bandwidth Memory was adopted by JEDEC as an industry standard in October 2013. The second generation, HBM2, was accepted by JEDEC in January 2016. Technology HBM achieves higher bandwidth while using less power in a substantially smaller form factor than DDR4 or GDDR5. This is achieved by stacking up to eight DRAM dies and an optional base die which can include buffer circuitry and test logic. The stack is often connected to the memory controller on a GPU or CPU through a substrate, such as a silicon interposer. Alternatively, the memory die could be stacked directly on the CPU or GPU chip. Within the stack the die are vertically interconnected by through-silicon vias (TSVs) and microbumps. The HBM technology is similar in principle but incompatible with the Hybrid Memory Cube (HMC) interface developed by Micron Technology. HBM memory bus is very wide in comparison to other DRAM memories such as DDR4 or GDDR5. An HBM stack of four DRAM dies (4Hi) has two 128bit channels per die for a total of 8 channels and a width of 1024 bits in total. A graphics card/GPU with four 4Hi HBM stacks would therefore have a memory bus with a width of 4096 bits. In comparison, the bus width of GDDR memories is 32 bits, with 16 channels for a graphics card with a 512bit memory interface. HBM supports up to 4 GB per package. The larger number of connections to the memory, relative to DDR4 or GDDR5, required a new met
https://en.wikipedia.org/wiki/SPARQL%20Query%20Results%20XML%20Format
A SPARQL Query Results XML(also sometimes called SPARQL Results Document) is a file stores data (value, uri and text) in XML. This document is generally the response by default of a RDF database after a SPARQL query. See also SPARQL SPARQL Query Results JSON Format References Data serialization formats SPARQL
https://en.wikipedia.org/wiki/Protein%20mimetic
A protein mimetic is a molecule such as a peptide, a modified peptide or any other molecule that biologically mimics the action or activity of some other protein. Protein mimetics are commonly used in drug design and discovery. Types of mimetics There are a number of different distinct classes of protein mimetics. Antibody mimetic - Molecules that mimic antigen binding activity of antibodies Peptidomimetic - Small protein-like chains designed to mimic larger peptide. Phosphomimetics - An amino acid substitution or modification which mimic the effect of protein phosphorylation. See also Homology (biology) Drug discovery Drugs blocking protein-protein interactions Proteins Molecular biology
https://en.wikipedia.org/wiki/Identity%20type
In type theory, the identity type represents the concept of equality. It is also known as propositional equality to differentiate it from "judgemental equality". Equality in type theory is a complex topic and has been the subject of research, such as the field of homotopy type theory. Comparison with Judgemental Equality The identity type is one of 2 different notions of equality in type theory. The more fundamental notion is "judgemental equality", which is a judgement. Beyond Judgemental Equality The identity type can do more than what judgemental equality can do. It can be used to show "for all ", which is impossible to show with judgemental equality. This is accomplished by using the eliminator (or "recursor") of the natural numbers, known as "R". The "R" function let's us define a new function on the natural numbers. That new function "P" is defined to be "(λ x:nat . x+1 = 1+x)". The other arguments act like the parts of an induction proof. The argument "PZ : P 0" becomes the base case "0+1 = 1+0", which is the term "refl nat 1". The argument "PS : P n → {\displaystyle \to } \to P (S n)" becomes the inductive case. Essentially, this says that when "x+1 = 1+x" has "x" replaced with a canonical value, the expression will be the same as "refl nat (x+1)". Versions of the Identity Type The identity type is complex and is the subject of research in type theory. While every version agrees on the constructor, "refl". Their properties and eliminator functions differ dramatically. For "extensional" versions, any identity type can be converted into a judgemental equality. A computational version is known as "Axiom K" due to Thomas Streicher. These are not very popular lately. Complexity of Identity Type Martin Hoffman and Thomas Streicher refuted that idea type theory required all terms of the identity type to be the same. A popular branches of research into the identity type are homotopy type theory and its Cubical type theory. References
https://en.wikipedia.org/wiki/Center%20for%20Data-Driven%20Discovery
The Center for Data-Driven Discovery is a multi-division research group at the California Institute of Technology, focusing on the methodologies for handling and analysis of large and complex data sets, facilitating the data-to-discovery process. It supports all applications of data-driven computing in various scientific domains, such as biology, physics, astronomy, geophysics, etc. It also functions as a catalyst for new collaborations and projects between different scientific disciplines, and between the campus and JPL, with especial interest in the sharing and transfer of methodologies, where the solutions from one field can be reapplied in another one. The Center for Data-Driven Discovery is a part of a joint initiative with the Center for Data Science and Technology at the Jet Propulsion Laboratory. It became operational in the fall of 2014. Directors Stanislav George Djorgovski (2014 – present) References Computational science Jet Propulsion Laboratory California Institute of Technology 2014 establishments in California
https://en.wikipedia.org/wiki/Scale%20co-occurrence%20matrix
Scale co-occurrence matrix (SCM) is a method for image feature extraction within scale space after wavelet transformation, proposed by Wu Jun and Zhao Zhongming (Institute of Remote Sensing Application, China). In practice, we first do discrete wavelet transformation for one gray image and get sub images with different scales. Then we construct a series of scale based concurrent matrices, every matrix describing the gray level variation between two adjacent scales. Last we use selected functions (such as Harris statistical approach) to calculate measurements with SCM and do feature extraction and classification. One basis of the method is the fact: way texture information changes from one scale to another can represent that texture in some extent thus it can be used as a criterion for feature extraction. The matrix captures the relation of features between different scales rather than the features within a single scale space, which can represent the scale property of texture better. Also, there are several experiments showing that it can get more accurate results for texture classification than the traditional texture classification. Background Texture can be regarded as a similarity grouping in an image. Traditional texture analysis can be divided into four major issues: feature extraction, texture discrimination, texture classification and shape from texture(to reconstruct 3D surface geometry from texture information). For tradition feature extraction, approaches are usually categorized into structural, statistical, model based and transform. Wavelet transformation is a popular method in numerical analysis and functional analysis, which captures both frequency and location information. Gray level co-occurrence matrix provides an important basis for SCM construction. SCM based on discrete wavelet frame transformation make use of both correlations and feature information so that it combines structural and statistical benefits. Discrete wavelet frame (DWF) In
https://en.wikipedia.org/wiki/Gap%20reduction
In computational complexity theory, a gap reduction is a reduction to a particular type of decision problem, known as a c-gap problem. Such reductions provide information about the hardness of approximating solutions to optimization problems. In short, a gap problem refers to one wherein the objective is to distinguish between cases where the best solution is above one threshold from cases where the best solution is below another threshold, such that the two thresholds have a gap in between. Gap reductions can be used to demonstrate inapproximability results, as if a problem may be approximated to a better factor than the size of gap, then the approximation algorithm can be used to solve the corresponding gap problem. -gap problem We define a -gap problem as follows: given an optimization (maximization or minimization) problem , the equivalent -gap problem distinguishes between two cases, for an input and an instance of problem : . Here, the best solution to instance of problem has a cost, or score, below . . Here, the best solution to instance of problem has cost above The gap between the two thresholds is thus . Note that whenever falls between the thresholds, there is no requirement on what the output should be. valid algorithm for the -gap problem may answer anything if is in the middle of the gap. The value does not need to be constant; it can depend on the size of the instance of . Note that -approximating the solution to an instance of is at least as hard as solving the -gap version of . One can define an -gap problem similarly. The difference is that the thresholds do not depend on the input; instead, the lower threshold is and the upper threshold is . Gap-producing reduction A gap-producing reduction is a reduction from an optimization problem to a c-gap problem, so that solving the c-gap problem quickly would enable solving the optimization problem quickly. The term gap-producing arises from the nature of the reduction: the op
https://en.wikipedia.org/wiki/Membrane%20osmometer
A membrane osmometer is a device used to indirectly measure the number average molecular weight () of a polymer sample. One chamber contains pure solvent and the other chamber contains a solution in which the solute is a polymer with an unknown . The osmotic pressure of the solvent across the semipermeable membrane is measured by the membrane osmometer. This osmotic pressure measurement is used to calculate for the sample. Basic operation A low concentration solution is created by adding a small amount of polymer to a solvent. This solution is separated from pure solvent by a semipermeable membrane. Solute cannot cross the semipermeable membrane but the solvent is able to cross the membrane. Solvent flows across the membrane to dilute the solution. The pressure required to stop the flow across the membrane is called the osmotic pressure. The osmotic pressure is measured and used to calculate . In an ideally dilute solution, van ‘t Hoff's law of osmotic pressure can be used to calculate from osmotic pressure. , number average molecular weight, mass/mole , gas constant , absolute temperature, typically Kelvin , concentration of polymer, mass/volume , osmotic pressure Virial equations In practice, the osmotic pressure produced by an ideally dilute solution would be too small to be accurately measured. For accurate measurements, solutions are not ideally dilute and a virial equation is used to account for deviations from ideal behavior and allow the calculation of . The virial equation takes a form similar to van ‘t Hoff's law of osmotic pressure, but contains additional constants to account for non-ideal behavior: where are constants and is still the concentration of polymer 1. This virial equation may be represented in different additional forms: where and are constants and . Different membrane osmometry devices Static membrane osmometry Capillary tubes are attached to both the solvent and the solution compartments. In this case the osmotic pr
https://en.wikipedia.org/wiki/Hamilton%27s%20optico-mechanical%20analogy
Hamilton's optico-mechanical analogy is a conceptual parallel between trajectories in classical mechanics and wavefronts in optics, introduced by William Rowan Hamilton around 1831. It may be viewed as linking Huygens' principle of optics with Maupertuis' principle of mechanics. While Hamilton discovered the analogy in 1831, it was not applied practically until Hans Busch used it to explain electron beam focusing in 1925. According to Cornelius Lanczos, the analogy has been important in the development of ideas in quantum physics. Erwin Schrödinger cites the analogy in the very first sentence of his paper introducing his wave mechanics. Later in the body of his paper he says: Quantitative and formal analysis based on the analogy use the Hamilton–Jacobi equation; conversely the analogy provides an alternative and more accessible path for introducing the Hamilton–Jacobi equation approach to mechanics. The orthogonality of mechanical trajectories characteristic of geometrical optics to the optical wavefronts characteristic of a full wave equation, resulting from the variational principle, leads to the corresponding differential equations. Hamilton's analogy The propagation of light can be considered in terms of rays and wavefronts in ordinary physical three-dimensional space. The wavefronts are two-dimensional curved surfaces; the rays are one-dimensional curved lines. Hamilton's analogy amounts to two interpretations of a figure like the one shown here. In the optical interpretation, the green wavefronts are lines of constant phase and the orthogonal red lines are the rays of geometrical optics. In the mechanical interpretation, the green lines denote constant values of action derived by applying Hamilton's principle to mechanical motion and the red lines are the orthogonal object trajectories. The orthogonality of the wavefronts to rays (or equal-action surfaces to trajectories) means we can compute one set from the other set. This explains how Kirchhoff's diff
https://en.wikipedia.org/wiki/Vegetation%20classification
Vegetation classification is the process of classifying and mapping the vegetation over an area of the earth's surface. Vegetation classification is often performed by state based agencies as part of land use, resource and environmental management. Many different methods of vegetation classification have been used. In general, there has been a shift from structural classification used by forestry for the mapping of timber resources, to floristic community mapping for biodiversity management. Whereas older forestry-based schemes considered factors such as height, species and density of the woody canopy, floristic community mapping shifts the emphasis onto ecological factors such as climate, soil type and floristic associations. Classification mapping is usually now done using geographic information systems (GIS) software. Classification schemes Following, some important classification schemes. Köppen (1884) Although this scheme is in fact of a climate classification, it has a deep relationship with vegetation studies: Class A Tropical rainforest (Af) Tropical monsoon (Am) Tropical savanna (Aw, As) Class B Desert (BWh, BWk) Semi-arid (BSh, BSk) Class C Humid subtropical (Cfa, Cwa) Oceanic (Cfb, Cwb, Cfc, Cwc) Mediterranean (Csa, Csb, Csc) Class D Humid continental (Dfa, Dwa, Dfb, Dwb, Dsa, Dsb) Subarctic (Dfc, Dwc, Dfd, Dwd, Dsc, Dsd) Class E Tundra (ET) Ice cap (EF) Alpine (ET, EF) Wagner & von Sydow (1888) Wagner & von Sydow (1888) scheme: Vegetationsgürtel (vegetation belts): Tundren (tundra) Hochgebirgsflora (mountain flora) Vegetationsarme Gebiete (Wüsten) (vegetation poor areas [deserts]) der gemässigten zone (the temperate zone) Grasland (prairie) Vorherrschend Nadelwald (mainly coniferous forest) Wald (Laub und Nadelwald) und Kulturland (forest [deciduous and coniferous forest] and cultivated land) in tropischen und subtropischen Gebieten (in tropical and subtropical areas) Grasland (prairie) Wald und Kulturland (forest and cul
https://en.wikipedia.org/wiki/PRIME%20%28labeling%20technique%29
PRIME (probe incorporation mediated by enzymes) is a molecular biology research tool developed by Alice Y. Ting and the Ting Lab at MIT for site-specific labeling of proteins in living cells with chemical probes. Probes often have useful biophysical properties, such as fluorescence, and allow imaging of proteins. Ultimately, PRIME enables scientists to study functions of specific proteins of interest. Significance Protein labeling with fluorescent molecules allows the visualization of protein dynamics, localization, and protein-protein interactions, and therefore serves as an important technique to understand protein functions and networks in living cells. The protein labeling should have a high selectivity towards the protein of interest, and should not interfere with the natural functions of the protein. Although genetic coding of fluorescent proteins, such as the green fluorescent protein (GFP), is the most popular technique due to its high specificity, fluorescent proteins are likely to interfere with the functions of the protein to which they are fused because of their large sizes. There are multiple tagging tools, such as HaloTag, SNAP tag, and FlAsH, developed in order to overcome the weakness of traditional protein labeling with fluorescent proteins. However, they still have significant shortcomings either due to the large size of a tag or the low specificity of the labeling process. PRIME has been developed in order to achieve a high labeling specificity comparable to fluorescent proteins with small molecules. Principles In PRIME, a mutant enzyme LplA (lipoic acid ligase from Escherichia coli) first catalyzes the conjugation of the "functional group handle" and LplA acceptor peptide (LAP), which is genetically fused to the protein of interest. “Functional group handle” indicates a bridge molecule connecting a LAP tag to a fluorescent probe or fluorophore. Fluorescent probe reacts with the “functional group handle” connected to the tag, and ultimately la
https://en.wikipedia.org/wiki/System%20for%20Cross-domain%20Identity%20Management
System for Cross-domain Identity Management (SCIM) is a standard for automating the exchange of user identity information between identity domains, or IT systems. One example might be that as a company onboards new employees and separates from existing employees, they are added and removed from the company's electronic employee directory. SCIM could be used to automatically add/delete (or, provision/de-provision) accounts for those users in external systems such as Google Workspace, Office 365, or Salesforce.com. Then, a new user account would exist in the external systems for each new employee, and the user accounts for former employees might no longer exist in those systems. In addition to simple user-record management (creating and deleting), SCIM can also be used to share information about user attributes, attribute schema, and group membership. Attributes could range from user contact information to group membership. Group membership or other attribute values are generally used to manage user permissions. Attribute values and group assignments can change, adding to the challenge of maintaining the relevant data across multiple identity domains. The SCIM standard has grown in popularity and importance, as organizations use more SaaS tools. A large organization can have hundreds or thousands of hosted applications (internal and external) and related servers, databases and file shares that require user provisioning. Without a standard connection method, companies must write custom software connectors to join these systems and their Identity Management (IdM) system. SCIM uses a standardised API through REST with data formatted in JSON or XML. History The first version, SCIM 1.0, was released in 2011 by a SCIM standard working group organized under the Open Web Foundation. In 2011, it was transferred to the IETF, and the current standard, SCIM 2.0 was released as IETF RFC in 2015. SCIM 2.0 was completed in September 2015 and is published as IETF RFCs 7643 and
https://en.wikipedia.org/wiki/Hostucan
Hostucan is an online web hosting search and review website. It is affiliated to its parent company Shanghai Racent Information Tech Co. Ltd. Overview The company has a platform where users can search for the web hosting, domain, keyword, SSL solutions and customer reviews. History The company was launched a decade ago and was established by Ray Zheng. The website comes along with its Chinese version website Hostucan.cn as well. In 2012, HostUcan released its Search Engine Ranking tool in beta, with which webmasters could check the keyword ranking of their sites in the major search engines like Google and Bing. In 2013, Hostucan revamped their website and added new features. They improved the domain search tool which enables users to find research and buy domain names. The company also released SSL, Whois and 'Who is hosting this' tool intended for professional webmaster tutorials. On September 30, 2013, Hostucan co-organized WHD.china2013 with WorldHostingDays in Shanghai, the first non-official hosting event in China. On May 22, 2014, Hostucan partnered with Ressellerclub and took part in HostingCon in Shanghai. On November 21, 2014, Hostucan partnered with WorldHostingDays as the local host in Beijing for WHD.china2014. Related websites Other websites owned by Racent include Pinnok, Top10InAction and a Chinese website Shangbangla. References Web hosting
https://en.wikipedia.org/wiki/Confederation%20of%20European%20Data%20Protection%20Organisations
The Confederation of European Data Protection Organisations (CEDPO) is a European umbrella organisation of data protection organisations. Data protection umbrella organisation CEDPO was founded in September 2011 by national data protection organisations as their European umbrella organisation, so far (2014) without any explicit legal form. CEDPO pursues to promote the role of the Data Protection Officer and to provide advice on balanced, practicable, and effective data protection. In addition, CEDPO aims to contribute to a better harmonisation of data protection law and practices in the European Union (EU) / European Economic Area (EEA). Founder members The four founder members of CEDPO are the Association française des correspondants à la protection des données à caractère personnel (AFCDP, France), the Asociación Profesional Española de Privacidad (APEP, Spain), the Gesellschaft für Datenschutz und Datensicherheit e.V. (GDD, Germany) and the Nederlands Genootschap van Functionarissen voor de Gegevensbescherming (NGFG, Netherlands). Members Present legal entities of CEDPO include: Association of Data Protection Officers (ADPO), Ireland, http://www.dpo.ie/ Association Française des Correspondants à la Protection des Données à Caractère Personnel (AFCDP), France, http://www.afcdp.net/ Asociación Profesional Española de Privacidad (APEP), Spanien, http://www.apep.es/ , Austria, http://www.argedaten.at/ Association Data Protection Officer (ASSO DPO), Italy, https://www.assodpo.it/ , Germany, https://www.gdd.de/ Nederlands Genootschap van Functionarissen voor de Gegevensbescherming (NGFG), Netherlands, http://www.ngfg.nl/ Stowarzyszenie Administratorów Bezpieczeństwa Informacji (SABI), Poland, http://www.sabi.org.pl/ Associação de Encarregados de Proteção de Dados (AEPD), Portugal, https://www.aepd.pt/ Asociația Specialiștilor în Confidențialitate și Protecția Datelor (ASCPD), Romania, https://ascpd.ro/ References External links Data protection 20
https://en.wikipedia.org/wiki/Circle%20Hough%20Transform
The circle Hough Transform (CHT) is a basic feature extraction technique used in digital image processing for detecting circles in imperfect images. The circle candidates are produced by “voting” in the Hough parameter space and then selecting local maxima in an accumulator matrix. It is a specialization of the Hough transform. Theory In a two-dimensional space, a circle can be described by: where (a,b) is the center of the circle, and r is the radius. If a 2D point (x,y) is fixed, then the parameters can be found according to (1). The parameter space would be three dimensional, (a, b, r). And all the parameters that satisfy (x, y) would lie on the surface of an inverted right-angled cone whose apex is at (x, y, 0). In the 3D space, the circle parameters can be identified by the intersection of many conic surfaces that are defined by points on the 2D circle. This process can be divided into two stages. The first stage is fixing radius then find the optimal center of circles in a 2D parameter space. The second stage is to find the optimal radius in a one dimensional parameter space. Find parameters with known radius R If the radius is fixed, then the parameter space would be reduced to 2D (the position of the circle center). For each point (x, y) on the original circle, it can define a circle centered at (x, y) with radius R according to (1). The intersection point of all such circles in the parameter space would be corresponding to the center point of the original circle. Consider 4 points on a circle in the original image (left). The circle Hough transform is shown in the right. Note that the radius is assumed to be known. For each (x,y) of the four points (white points) in the original image, it can define a circle in the Hough parameter space centered at (x, y) with radius r. An accumulator matrix is used for tracking the intersection point. In the parameter space, the voting number of points through which the circle passing would be increased by one. Then t
https://en.wikipedia.org/wiki/Sprague%20Electric
Sprague Electric Company was an electronic component maker founded by Robert C. Sprague in 1926. Sprague was best known for making a large line of capacitors used in a wide variety of electrical and electronic in commercial, industrial and military/space applications. Other products include resistive components, magnetic components (transformers and coils), filter assemblies, semiconductors and integrated circuits. History overview Sprague used $25,000 of his savings to open Sprague Specialties Company at his home in Quincy, MA, in 1926. One of his first products was the mini condenser (an old name for capacitor). Mini condensers were commonly used in radio applications for noise filtering, signal coupling and tone controls. Early capacitors were two pieces of metal foil wrapped between wax paper or any other type of suitable insulation material. The type of insulating material determined the capacitor's capacitance and maximum voltage. Capacitors are also useful in high power applications like motors, and soon Sprague turned his attention those areas as well. Sprague found a sustainable product line in capacitors. The increase in the types of radios using AC created demand for many different types of capacitors. By 1929, Sprague Specialties Company needed a bigger facility, and in 1930 Sprague purchased a plant on Beaver Street in North Adams, MA, in Berkshire County. When local residents heard the company was expanding, Sprague received all kinds of incentives from the banks and other businesses to relocate there. Sprague chose the area because he wanted to open a shop where his father Frank had grown up. By the mid-1930s Sprague had become a recommended source for capacitors by radio manufacturers, radio repair and many electrical applications. As the size of the company grew there was a desire from the manufacturing workers to form an organized labor union. The Wagner Act of 1935 prohibited company unions. In 1937 the company agreed with the workers to form
https://en.wikipedia.org/wiki/Supercritical%20liquid%E2%80%93gas%20boundaries
Supercritical liquid–gas boundaries are lines in the pressure-temperature (pT) diagram that delimit more liquid-like and more gas-like states of a supercritical fluid. They comprise the Fisher–Widom line, the Widom line, and the Frenkel line. Overview According to textbook knowledge, it is possible to transform a liquid continuously into a gas, without undergoing a phase transition, by heating and compressing strongly enough to go around the critical point. However, different criteria still allow to distinguish liquid-like and more gas-like states of a supercritical fluid. These criteria result in different boundaries in the pT plane. These lines emanate either from the critical point, or from the liquid–vapor boundary (boiling curve) somewhat below the critical point. They do not correspond to first or second order phase transitions, but to weaker singularities. The Fisher–Widom line is the boundary between monotonic and oscillating asymptotics of the pair correlation function . The Widom line is a generalization thereof, apparently so named by H. Eugene Stanley. However, it was first measured experimentally in 1956 by Jones and Walker, and subsequently named the 'hypercritical line' by Bernal in 1964, who suggested a structural interpretation. The Frenkel line is a boundary between "rigid" and "non-rigid" fluids characterized by the onset of transverse sound modes. One of the above mentioned criteria is based on the velocity autocorrelation function (vacf): below the Frenkel line the vacf demonstrates oscillatory behaviour, while above it the vacf monotonically decays to zero. The second criterion is based on the fact that at moderate temperatures liquids can sustain transverse excitations, which disappear upon heating. One further criterion is based on isochoric heat capacity measurements. The isochoric heat capacity per particle of a monatomic liquid near to the melting line is close to (where is the Boltzmann constant). The contribution to the heat capa
https://en.wikipedia.org/wiki/Embryokine
Embryokines (Greek: embryuon "embryo" + kinōs "movement") are regulatory molecules produced by the oviduct and endometrium in the reproductive tract that modulate embryonic growth and development. Embryokines include growth factors such as insulin-like growth factor-1, and activin a transforming growth factor; cytokines such as colony stimulating factor 2, WNT regulatory proteins including DKK1; Other small molecule amino acids are included that regulate embryonic development through the mTOR signalling pathway. Prostacyclin 1 can activate peroxisome proliferator-activated receptor 6 to increase blastocyst hatching, and cannabinoids that can also act to regulate implantation and development. References Embryology Developmental biology Human pregnancy Fertility
https://en.wikipedia.org/wiki/Quadratic%20integrate%20and%20fire
The quadratic integrate and fire (QIF) model is a biological neuron model that describes action potentials in neurons. In contrast to physiologically accurate but computationally expensive neuron models like the Hodgkin–Huxley model, the QIF model seeks only to produce action potential-like patterns by ignoring the dynamics of transmembrane currents and ion channels. Thus, the QIF model is computationally efficient and has found ubiquitous use in computational neuroscience. An idealized model of neural spiking is given by the autonomous differential equation, where represents the membrane voltage and represents an input current. A solution to this differential equation is the function, where is an arbitrary shift dependent on the initial condition (specifically by the formula ). This solution "blows up" in finite time, namely at for all , which resembles the rhythmic action potentials generated by neurons stimulated by some input current. Thus a "spike" is said to have occurred when the solution reaches positive infinity. Just after this point in time, the solution resets to negative infinity by definition. When implementing this model in a numerical simulation, a threshold crossing value () and a reset value () is assigned, so that when the solution rises above the threshold, , the solution is immediately reset to . The above equation is directly related to an alternative form of the QIF model, , where is the membrane time constant. References Computational neuroscience Mathematical modeling Nonlinear systems
https://en.wikipedia.org/wiki/Setra%20Systems
Setra Systems, Inc. is an American electronics company. It is a manufacturer of electronic instruments that measure pressure, acceleration, and weight. The products designed and developed by Setra Systems, Inc. are used chiefly in the industrial manufacturing industry. Setra Systems Inc. was a fully owned subsidiary of Fortive. History Founded in 1967, Setra Systems, Inc. is a USA-based designer and manufacturer of pressure, humidity and current sensing products for the HVAC industry. In 2001, Setra was acquired by Danaher Corporation. Setra Technology Park is located in Boxborough, MA.[2] Setra's range of products also cater to agricultural, medical, and semi-conductor industries apart from the existing OEM, barometric, HVAC/R, environmental, test and measurement, food and pharmaceutical industries. Management The company was founded by Dr. Y. T. Li and Dr. S. Y. Lee, former professors of engineering at MIT. The founders are also the co–developers of the variable capacitance transduction principle, the force sensing technology which is the basis of Setra's products. Products Setra's range of products include differential pressure transmitters, low differential pressure transducer, wet-to-wet differential pressure transducer, gauge pressure transmitters, current switches and sensors, humidity sensors and vacuum sensors. The engineering and research efforts of Setra have been put into the development of transmitters and transducers designed using the variable capacitive transduction principle. Setra has applications in various industries such as OEM, barometric, HVAC/R, environmental, test and measurement, food and pharmaceutical industries and has now ventured into building automation, agricultural, medical and semi-conductor industries. References External links Setra Systems Official Website Sensors 1967 establishments in the United States Electronics companies established in 1967 Electronics companies of the United States Manufacturing companies based in M
https://en.wikipedia.org/wiki/SubOS
In computing, a SubOS may mean several related concepts: A process-specific protection mechanism allowing potentially dangerous applications to run in a restricted environment. It worked by setting a sub-user id which was user id of the owner of the file rather than the person running the file. A substitute-operating system, which simulated a full operating system. These were mainly developed by the GameMaker community. An interface (graphical or terminal based) that provides additional functions or command for a specific audience target. It can also make processes easier and add details to the main operating system. See also Virtual machine Sandbox (computer security) Computing terminology Operating system security Virtualization software
https://en.wikipedia.org/wiki/Rate-zonal%20centrifugation
Rate-zonal centrifugation is a centrifugation technique employed to effectively separate particles of different sizes. The tube is first filled with different concentrations of sucrose or another solute establishing layers with different densities and viscosities, forming a density gradient, within which the particles to be separated are added. The larger particles will be able to travel to the bottom layer because they are more massive. The greater mass allows the particles to travel through layers with a greater viscosity, while the smaller particles will remain at the top, as they lack the mass to travel through the more viscous layers. Once the centrifugation is over, fractions are collected. References External links Rate-zonal centrifugation at ScienceDirect Molecular biology techniques
https://en.wikipedia.org/wiki/Chomsky%E2%80%93Sch%C3%BCtzenberger%20enumeration%20theorem
In formal language theory, the Chomsky–Schützenberger enumeration theorem is a theorem derived by Noam Chomsky and Marcel-Paul Schützenberger about the number of words of a given length generated by an unambiguous context-free grammar. The theorem provides an unexpected link between the theory of formal languages and abstract algebra. Statement In order to state the theorem, a few notions from algebra and formal language theory are needed. Let denote the set of nonnegative integers. A power series over is an infinite series of the form with coefficients in . The multiplication of two formal power series and is defined in the expected way as the convolution of the sequences and : In particular, we write , , and so on. In analogy to algebraic numbers, a power series is called algebraic over , if there exists a finite set of polynomials each with rational coefficients such that A context-free grammar is said to be unambiguous if every string generated by the grammar admits a unique parse tree or, equivalently, only one leftmost derivation. Having established the necessary notions, the theorem is stated as follows. Chomsky–Schützenberger theorem. If is a context-free language admitting an unambiguous context-free grammar, and is the number of words of length in , then is a power series over that is algebraic over . Proofs of this theorem are given by , and by . Usage Asymptotic estimates The theorem can be used in analytic combinatorics to estimate the number of words of length n generated by a given unambiguous context-free grammar, as n grows large. The following example is given by : the unambiguous context-free grammar G over the alphabet {0,1} has start symbol S and the following rules S → M | U M → 0M1M | ε U → 0S | 0M1U. To obtain an algebraic representation of the power series associated with a given context-free grammar G, one transforms the grammar into a system of equations. This is achieved by replacing each occurrence of a termin
https://en.wikipedia.org/wiki/Chomsky%E2%80%93Sch%C3%BCtzenberger%20representation%20theorem
In formal language theory, the Chomsky–Schützenberger representation theorem is a theorem derived by Noam Chomsky and Marcel-Paul Schützenberger about representing a given context-free language in terms of two simpler languages. These two simpler languages, namely a regular language and a Dyck language, are combined by means of an intersection and a homomorphism. A few notions from formal language theory are in order. A context-free language is regular, if it can be described by a regular expression, or, equivalently, if it is accepted by a finite automaton. A homomorphism is based on a function which maps symbols from an alphabet to words over another alphabet ; If the domain of this function is extended to words over in the natural way, by letting for all words and , this yields a homomorphism . A matched alphabet is an alphabet with two equal-sized sets; it is convenient to think of it as a set of parentheses types, where contains the opening parenthesis symbols, whereas the symbols in contains the closing parenthesis symbols. For a matched alphabet , the Dyck language is given by Chomsky–Schützenberger theorem. A language L over the alphabet is context-free if and only if there exists a matched alphabet a regular language over , and a homomorphism such that . Proofs of this theorem are found in several textbooks, e.g. or . References Noam Chomsky Formal languages Theorems in discrete mathematics
https://en.wikipedia.org/wiki/McIntosh%20MC-2300
The McIntosh MC-2300 is a solid-state power amplifier which was built by the American high-end audio company McIntosh Laboratory between 1971 and 1980. McIntosh produced the MC-2105 (with blue meters) and the MC-2100 (without) between 1969 and 1977. Both 100 watt-per-channel stereo amps (200 watts monophonic) sold. The MC-2300 was succeeded by the more powerful MC-2500 (500 WPC stereo/1000 watts mono), sold from 1980 to 1990; and then the MC-2600 (600 WPC stereo/1200 watts mono), which was available from 1990 to 1995. Design Physically, the MC-2300 is a very large and sturdy amplifier, measuring 10.5 in (26.7 cm) high x 19 in (48.3 cm) wide x 17 in (43.2 cm) deep, and weighing an impressive 128 lbs (58 kg). During its production run, 4545 units were made. Today, the MC-2300 remains a very sought-after amplifier for audiophiles and collectors. The MC-2300 can be utilized either as a 300-watt-per-channel stereo amp, or a 600-watt monoblock, and was rated by its manufacturer as being able to produce this amount of power continuously, with very little (less than 0.25%) distortion. McIntosh's ratings were conservative, however, because like many of their amplifiers, when bench-tested the MC-2300 has frequently been found to produce an even higher level of clean power. As such, it was ideal for use in demanding, professional applications. During the 1970s, the MC-2300 was an expensive piece of audio equipment, with a retail price of $1799 by the time of its discontinuation in 1980. That being said, its outstanding power and sound production quality made it a valued part of many recording studios and although some people prefer the sound of tube amplifiers, the overall greater reliability and freedom from repair of the newer solid-state amps was a major vote in their favor. History The improvisational rock band the Grateful Dead employed 48 McIntosh MC-2300 amps as the main power source for their enormous public-address system, the Wall of Sound. Designed by Owsley
https://en.wikipedia.org/wiki/Hattendorff%27s%20theorem
Hattendorff's Theorem, attributed to K. Hattendorff (1868), is a theorem in actuarial science that describes the allocation of the variance or risk of the loss random variable over the lifetime of an actuarial reserve. In other words, Hattendorff's theorem demonstrates that the variation in the present value of the loss of an issued insurance policy can be allocated to the future years during which the insured is still alive. This, in turn, facilitates the management of risk prevalent in such insurance contracts over short periods of time. Hattendorff's Theorem The main result of the theorem has three equivalent formulations: where: In its above formulation, and in particular the first result, Hattendorff's theorem states that the variance of , the insurer's total loss over the remaining life of the policy at time h, can be calculated by discounting the variances of the yearly net losses (cash losses plus changes in net liabilities) in future years. Background In the most general stochastic setting in which the analysis of reserves is carried out, consider an insurance policy written at time zero, over which the insured pays yearly premiums at the beginning of each year starting today until the year of death of the insured. Furthermore, the insured receives a benefit of , at the end of the year of death, equal to . No other payments are received nor paid over the lifetime of the policy. Suppose an insurance company is interested to know the cash loss from this policy over the year (h, h+1). Of course, if the death of the insured happens prior to time h, or when , then there is no remaining loss and . If the death of the insured occurs exactly at time h, or when , then the loss on the policy is equal to the present value of the benefit paid in the following year, , less the premium paid at time h. Hence in this case Lastly, if the death of the insured occurs after time h, or when , then the cash loss in the year (h, h+1) is just the negative of the premiu
https://en.wikipedia.org/wiki/Anton%20blood%20group%20antigen
The Anton blood group antigen is a cell surface receptor found on some human red blood cells. It has been observed to play a role in Haemophilus influenzae infections. Studies showed that bacterium can adhere to this receptor and cause human red blood cells to agglutinate. References Transmembrane receptors Blood cells Microbiology
https://en.wikipedia.org/wiki/REFInd
rEFInd is a boot manager for UEFI and EFI-based machines. It can be used to boot multiple operating systems that are installed on a single non-volatile device. It also provides a way to launch UEFI applications. It was forked from discontinued rEFIt in 2012, with 0.2.0 as its first release. rEFind supports x86, x86-64, and AArch64 architecture. Features rEFInd has several features: Automatic operating systems detection. Customisable OS launch options. Graphical or text mode. Theme is customisable. Mac-specific features, including spoofing booting process to enable secondary video chipsets on some Mac. Linux-specific features, including autodetecting EFI stub loader to boot Linux kernel directly and using fstab in lieu of rEFInd configuration file for boot order. Support for Secure Boot. Adoption rEFInd is the default Unified Extensible Firmware Interface (UEFI) boot manager for TrueOS. rEFInd is included in official repositories of major Linux distributions. Development GNU-EFI and TianoCore are supported as main development platforms for writing binary UEFI applications in C to launch right from the rEFInd GUI menu. Typical purposes of an EFI application are fixing boot problems and programmatically modifying settings within UEFI environment, which would otherwise be performed from within the BIOS of a personal computer (PC) without UEFI. rEFInd can be built with either GNU-EFI or TianoCore EDK2/UDK. Fork RefindPlus is a fork of rEFInd that add several features and improvements for Mac devices, specifically MacPro3,1 and MacPro5,1, and equivalent Xserve. See also GNU GRUB - Another boot loader for Unix-like systems Comparison of boot loaders References Free boot loaders Free system software Macintosh firmware Software using the BSD license Software forks
https://en.wikipedia.org/wiki/Digimon%20Story%3A%20Cyber%20Sleuth
is a role-playing video game developed by Media.Vision and published by Bandai Namco Entertainment that was released in Japan on March 12, 2015 for PlayStation Vita and PlayStation 4. Part of the Digimon franchise, the game is the fifth installment in the Digimon Story series, following 2011's Super Xros Wars, and the first to be released on home consoles. The game would be released in North America on February 2, 2016, becoming the first installment of the Digimon Story series to be released in North America since 2007's Digimon World Dawn and Dusk, and the first to be released under its original title. A sequel, titled Digimon Story: Cyber Sleuth – Hacker's Memory, was released in Japan in 2017 and in Western territories in 2018. In July 2019, a port of the game and its sequel for Nintendo Switch and Windows, was announced for release on October 18, 2019, as Digimon Story Cyber Sleuth: Complete Edition, although the PC version was released a day early. Gameplay Digimon Story: Cyber Sleuth is a role-playing game, played from a third-person perspective where players control a human character with the ability to command Digimon, digital creatures with their own unique abilities who do battle against other Digimon. Players can choose between either Palmon, Terriermon or Hagurumon as their starting partner at the beginning of the game, with more able to be obtained as they make their way into new areas. A total of 249 unique Digimon are featured, including seven that were available as DLC throughout the life of the game, and two which were exclusive to the Western release. The title features a New Game Plus mode where players retain all of their Digimon, non-key items, money, memory, sleuth rank, scan percentages, and Digifarm progress. The Complete Edition includes the 92 new Digimon from Hacker's Memory, for a total of 341 Digimon. Plot Players assume the role of Takumi Aiba (male) or Ami Aiba (female), a young Japanese student living in Tokyo while their moth
https://en.wikipedia.org/wiki/Language-based%20security
In computer science, language-based security (LBS) is a set of techniques that may be used to strengthen the security of applications on a high level by using the properties of programming languages. LBS is considered to enforce computer security on an application-level, making it possible to prevent vulnerabilities which traditional operating system security is unable to handle. Software applications are typically specified and implemented in certain programming languages, and in order to protect against attacks, flaws and bugs an application's source code might be vulnerable to, there is a need for application-level security; security evaluating the applications behavior with respect to the programming language. This area is generally known as language-based security. Motivation The use of large software systems, such as SCADA, is taking place all around the world and computer systems constitute the core of many infrastructures. The society relies greatly on infrastructure such as water, energy, communication and transportation, which again all rely on fully functionally working computer systems. There are several well known examples of when critical systems fail due to bugs or errors in software, such as when shortage of computer memory caused LAX computers to crash and hundreds of flights to be delayed (April 30, 2014). Traditionally, the mechanisms used to control the correct behavior of software are implemented at the operating system level. The operating system handles several possible security violations such as memory access violations, stack overflow violations, access control violations, and many others. This is a crucial part of security in computer systems, however by securing the behavior of software on a more specific level, even stronger security can be achieved. Since a lot of properties and behavior of the software is lost in compilation, it is significantly more difficult to detect vulnerabilities in machine code. By evaluating the source code
https://en.wikipedia.org/wiki/Prismatic%20compass
A prismatic compass is a navigation and surveying instrument which is extensively used to find out the bearing of the traversing and included angles between them, waypoints (an endpoint of the lcourse) and direction. Compass surveying is a type of surveying in which the directions of surveying lines are determined with a magnetic compass, and the length of the surveying lines are measured with a tape or chain or laser range finder. The compass is generally used to run a traverse line. The compass calculates bearings of lines with respect to magnetic needle. The included angles can then be calculated using suitable formulas in case of clockwise and anti-clockwise traverse respectively. For each survey line in the traverse, surveyors take two bearings that is fore bearing and back bearing which should exactly differ by 180° if local attraction is negligible. The name Prismatic compass is given to it because it essentially consists of a prism which is used for taking observations more accurately. Least count Least count means the minimum value that an instrument can read which is 30 minutes in case of prismatic compass. It means compass can read only those observations which are multiples of 30 minutes, 5° 30', 16° 00', 35° 30', for example. Bearings The compass calculates the bearings in whole circle bearing system which determines the angle which the survey line makes with the magnetic north in the clockwise direction. The whole circle bearing system also known as the azimuthal system varies from 0 degrees to 360 degrees in the clockwise direction. The included angles can be calculated by the formulas F-P ±180 in case of anti-clockwise traverse and P-F ±180 in case of clockwise traverse, where 'F' is the fore bearing of forward line in the direction of survey work and 'P' is the fore bearing of previous line. References Surveying Civil engineering
https://en.wikipedia.org/wiki/Ethoxyresorufin-O-deethylase
Ethoxyresorufin-O-deethylase (EROD) is used as a biomarker in fish bioassays through catalytic measurement of cytochrome p4501A1 induction. References Biomarkers EC 1.14
https://en.wikipedia.org/wiki/Iron%20hand%20%28prosthesis%29
Iron hands are metal prostheses for hands and upper extremities from the middle ages and early modern period. These designs combined cosmetic and functional properties. The most famous example of an iron hand was made around the year 1530, being the second prosthetic hand made for the German knight Götz von Berlichingen. Most iron hands are based on the same constructive principles, although there are considerable differences in complexity. Fingers can be flexed passively (for example using the healthy hand) and are locked in place by a ratchet mechanism, similar to those of contemporary flintlocks. Extension of the fingers works by spring pressure. Bibliography Löffler, Liebhard: Der Ersatz für die obere Extremität: die Entwicklung von den ersten Zeugnissen bis heute. Stuttgart: Enke 1984, . Putti, Vittorio: Historical Prostheses. In: Journal of Hand Surgery. Vol. 30, No. 3, Edinburgh: 2005, pp. 310–325. References Prosthetics History of mechanical engineering
https://en.wikipedia.org/wiki/Bibliography%20of%20encyclopedias%3A%20biology
This is a list of encyclopedias as well as encyclopedic and biographical dictionaries published on the subject of biology in any language. Entries are in the English language unless specifically stated as otherwise. General biology Becher, Anne, Joseph Richey. American environmental leaders: From colonial times to the present. Grey House, 2008. . Butcher, Russell D., Stephen E. Adair, Lynn A. Greenwalt. America's national wildlife refuges: A complete guide. Roberts Rinehart Publishers in cooperation with Ducks Unlimited, 2003. . Ecological Internet, Inc. EcoEarth.info: Environment portal and search engine. Ecological Internet, Inc. . Friday, Adrian & Davis S. Ingram. The Cambridge Encyclopedia of Life Sciences. Cambridge, 1985. Gaither, Carl C., Alma E. Cavazos-Gaither, Andrew Slocombe. Naturally speaking: A dictionary of quotations on biology, botany, nature and zoology. Institute of Physics, 2001. . Gibson, Daniel, National Audubon Society. Audubon guide to the national wildlife refuges. Southwest: Arizona, Nevada, New Mexico, Texas. St. Martin's Griffin, 2000. . Goudie, Andrew, David J. Cuff. Encyclopedia of global change: Environmental change and human society. Oxford University Press, 2002. . Gove, Doris. Audubon guide to the national wildlife refuges. Southeast : Alabama, Florida, Georgia, Kentucky, Mississippi, North Carolina, Puerto Rico, South Carolina, Tennessee, U.S. Virgin Islands. St. Martin's Griffin, 2000. . Grassy, John. Audubon guide to the national wildlife refuges: Northern Midwest: Illinois, Indiana, Iowa, Michigan, Minnesota, Nebraska, North Dakota, Ohio, South Dakota, Wisconsin. St. Martin's Griffin, c2000. . Grassy, John. Audubon guide to the national wildlife refuges: Rocky Mountains: Colorado, Idaho, Montana, Utah, Wyoming. St. Martin's Griffin, 2000. . Gray, Peter. Encyclopedia of the Biological Sciences. Krieger, 1981. Grinstein, Louise S., Carol A. Biermann, Rose K. Rose. Women in the biological sciences: A biobibliographic sourceboo
https://en.wikipedia.org/wiki/Francis%20F.%20Lee
Francis Fan Lee (李凡, born January 28, 1927) is an inventor, entrepreneur, and professor emeritus of Electrical Engineering and Computer Science at the Massachusetts Institute of Technology (MIT). Lee is the founder of Lexicon (company) (originally American Data Sciences). He is best known for three inventions: the Digital Cardiac Monitor (1969), the Digital Audio Signal Processor (1971), and the Digital Time Compression System (1972). In 1984, Lexicon won an Emmy Award for Engineering Excellence for the Model 1200 Audio Time Compressor and Expander, widely used in the television industry. Education Lee was born January 28, 1927, in Nanjing, China. In September 1948, during the Chinese Civil War, Lee left Shanghai aboard the USS General W.H. Gordon to complete his undergraduate education at the Massachusetts Institute of Technology. He studied Electrical Engineering and earned his Bachelor of Science in 1950 and Master of Science in 1951. In Fall 1952, Lee entered the PhD program at MIT. He withdrew in 1954 to pursue his career, becoming a naturalized US citizen on November 15, 1954. He returned to academia in 1964, and completed his PhD in Electrical Engineering in 1965. Early career In 1954, Lee took a full-time position as Research Engineer with the Servomechanism Laboratory. He was part of a team working on the first Digitally Controlled Milling Machine In 1955, Lee joined the Bizmac Computer Division of RCA. He left a year later to join the UNIVAC super-computer division of Remington Rand. Project MAC In 1963, Lee accepted a one-year appointment to work on Project MAC, a time-sharing Multiple Access Computer being developed at MIT under the direction of Robert Fano. Fano launched Project MAC with a 6-week summer session that drew 57 people (including Lee) from universities, industry, and government for brainstorming and collaboration. At the end of the session, Lee described his work on speeding up computer memory in "Lookaside Memory Implementation" (1963).
https://en.wikipedia.org/wiki/Cantilever%20method
The cantilever method is an approximate method for calculating shear forces and moments developed in beams and columns of a frame or structure due to lateral loads. The applied lateral loads typically include wind loads and earthquake loads, which must be taken into consideration while designing buildings. The assumptions used in this method are that the points of contraflexure (or points of inflection of the moment diagram) in both the vertical and horizontal members are located at the midpoint of the member, and that the direct stresses in the columns are proportional to their distances from the centroidal axis of the frame. The frame is analysed in step-wise (iterative) fashion, and the results can then be described by force diagrams drawn up at the end of the process. The method is quite versatile and can be used to analyse frames of any number of storeys or floors. The position of the centroidal axis (the center of gravity line for the frame) is determined by using the areas of the end columns and interior columns. The cantilever method is considered one of the two primary approximate methods (the other being the portal method) for indeterminate structural analysis of frames for lateral loads. Its use is recommended for frames that are taller than they are wide, and therefore behave similar to a beam cantilevered up from the ground. See also Conjugate beam method References Civil engineering Structural analysis
https://en.wikipedia.org/wiki/Jes%C3%BAs%20Ildefonso%20D%C3%ADaz
Jesús Ildefonso Díaz is a Spanish mathematician who works in partial differential equations. He is a professor at Complutense University of Madrid (UCM) and a member of the Spanish Royal Academy of Sciences. Biography Díaz was born in Toledo, Spain on December 11, 1950. He graduated in mathematics from UCM in 1973, and obtained his PhD from the same university in 1976. His Ph.D. thesis advisors were Alberto Dou and Haïm Brezis. Career Díaz joined the faculty at UCM as an Associate Professor in Mathematical Analysis in 1978. He moved briefly to the University of Santander in 1980, before returning to UCM as a full professor in 1983. In 1998, he co-founded the journal Revista Matemática de la UCM and served on its editorial board from 1988 to 1995. He founded the Department of Applied Mathematics at the Facultad de Matemáticas of UCM in the early 1980s and led it for several years. In 2006, he founded the Instituto de Matemática Interdisciplinar (IMI), serving as Director from 2006 to 2008 and again from 2012 to 2016. He is an energetic teacher, organizing six summer courses at UCM, two of them with Jacques Louis Lions. Díaz has worked in many areas of applied mathematics, such as theoretical and applied aspects of nonlinear partial differential equations, fluid mechanics models, geophysical models, reaction-diffusion models, elasticity and homogenization models and control theory models, among others. He has also worked in areas closer to pure mathematics, such as nonlinear analysis, focusing on accretive operators, rearrangement and gradient estimates. Other activities include contributions in science history, science communication and scientific management. His mentors and influential colleagues include Philippe Benilan and Jacques Louis Lions. As of July 2019, his research publications included more than 250 papers in research journals, 141 contributions published in proceedings of meetings, seven books, eight book chapters and 20 edited volumes. His popul
https://en.wikipedia.org/wiki/Underwater%20computer%20vision
Underwater computer vision is a subfield of computer vision. In recent years, with the development of underwater vehicles ( ROV, AUV, gliders), the need to be able to record and process huge amounts of information has become increasingly important. Applications range from inspection of underwater structures for the offshore industry to the identification and counting of fishes for biological research. However, no matter how big the impact of this technology can be to industry and research, it still is in a very early stage of development compared to traditional computer vision. One reason for this is that, the moment the camera goes into the water, a whole new set of challenges appear. On one hand, cameras have to be made waterproof, marine corrosion deteriorates materials quickly and access and modifications to experimental setups are costly, both in time and resources. On the other hand, the physical properties of the water make light behave differently, changing the appearance of a same object with variations of depth, organic material, currents, temperature etc. Applications Seafloor survey Vehicle navigation and positioning Biological monitoring Video mosaics as visual navigation maps Pipeline inspection Wreckage visualization Maintenance of underwater structures Drowning detection, e.g. pool safety Medium differences Illumination In air, light comes from the whole hemisphere on cloudy days, and is dominated by the sun. In water lighting comes from a finite cone above the scene. This phenomenon is called Snell's window. Light attenuation Unlike air, water attenuates light exponentially. This results in hazy images with very low contrast. The main reasons for light attenuation are light absorption (where energy is removed from the light) and light scattering, by which the direction of light is changed. Light scattering can further be divided into forward scattering, which results in an increased blurriness and backward scattering that limit
https://en.wikipedia.org/wiki/RAD5500
The RAD5500 is a radiation-hardened 64-bit processor core design created by BAE Systems Electronics, Intelligence & Support based on the PowerPC e5500 designed by IBM and Freescale Semiconductor. Successor of the RAD750, the RAD5500 processor platform is for use in high radiation environments experienced on board satellites and spacecraft. The RAD5500 core supports VPX high speed connectors, DDR2/DDR3 memory, serialize/deserialize (SerDes), and SpaceWire IO. Processors The RAD5500 core is based on those of the Freescale Semiconductor e5500-based QorIQ system-on-chip. The RAD5510, RAD5515, RAD5545, and RADSPEED-HB (host bridge) are four system on a chip processors implemented with RAD5500 cores produced with 45 nm SOI technology from the IBM Trusted Foundry. RAD5510 and RAD5515 The RAD5510 and RAD5515 processors employ a single RAD5500 core and are intended for medium processing capability in environments that require low power consumption (11.5 and 13.7 watts respectively). This processor provides up to 1.4 giga operations per second (GOPS) and 0.9 GFLOPS of performance. RAD5545 The RAD5545 processor employs four RAD5500 cores, achieving performance characteristics of up to 5.6 giga-operations per second (GOPS) and over 3.7 GFLOPS. Power consumption is 20 watts with all peripherals operating. RADSPEED-HB (host bridge) Based on the RAD5545, the RADSPEED-HB is intended for host processing and data management support for one to four RADSPEED DSPs. The RADSPEED-HB replaces a secondary DDR2/DDR3 memory interface connection found on the RAD5545 with connections for RADSPEED DSPs instead. (Note that RADSPEED DSPs are entirely different processors that are specialized for digital signal processing and are not to be confused with the RADSPEED-HB, which serves as a host bridge.) Single-board computer The RAD5545 SpaceVPX single-board computer is for use in the harsh environmental conditions of outer space; designed for operating in temperatures between −55 °C
https://en.wikipedia.org/wiki/Human%20Factors%20in%20Engineering%20and%20Design
Human Factors in Engineering and Design is an engineering textbook, currently in its seventh edition. The book, first published in 1957, is considered a classic in human factors and ergonomics, and one of the best-established texts in the field. It is frequently taught in upper-level and graduate courses in the U.S., and is relied on by practicing human factors and ergonomics professionals. The text is divided into six sections: Introduction; Information Input; Human Output and Control; Work Space and Arrangement; Environment; and Human Factors: Selected Topics. See also Anthropometry Industrial and organizational psychology References Engineering textbooks McGraw-Hill books Ergonomics Occupational safety and health Industrial engineering Systems psychology
https://en.wikipedia.org/wiki/AudioQuest
AudioQuest is a company that was founded in 1980 by William E. Low that provides audio/video cables, digital-to-analog converters, headphones, power-conditioning products, and various audio/video accessories. The company is based in Irvine, California, has offices in the Netherlands and distributes its products to approximately 65 countries throughout the world. History AudioQuest's founder, William E. Low, has described himself as "an absolute hedonist." In the December 2008 issue of The Absolute Sound, Low explained to TAS's Neil Gader, "Everything I’ve learned about hi-fi or cables is purely the result of being interested in getting high on music." In his early days of selling high-end audio equipment, William E. Low allegedly discovered that the sound of an audio system was easily influenced by the quality of the cables connecting its various components. Hi-fi journalist, Richard Hardesty explained: "With experimentation Bill found that better interconnect and speaker cables could make bigger audible improvements than many costly upgrades to amplifiers and speakers. And he recognized the opportunities afforded by this new category of audio components. He founded AudioQuest to explore and develop new and innovative wire, connectors and accessories." Although AudioQuest remains best known for its analog and digital cables, the company has entered other product categories, such as the DragonFly USB digital-to-analog converter/headphone amplifier, recipient of numerous awards, including: Stereophile 's 2012 "Computer Audio Component of the Year" and 2012 "Budget Component of the Year;". At the 2015 Consumer Electronics Show, AudioQuest entered the headphone market with its NightHawk over-the-ear headphones designed by Skylar Gray. In October 2014, NightHawk was named 2015 CES Innovation Award Honoree (Headphones) and 2015 Best of Innovation Winner (Eco-Design and Sustainable Technologies). The NightHawk was followed by the critically successful NightOwl headpho
https://en.wikipedia.org/wiki/Zen%20%28microarchitecture%29
Zen is the codename for a family of computer processor microarchitectures from AMD, first launched in February 2017 with the first generation of its Ryzen CPUs. It is used in Ryzen (desktop and mobile), Ryzen Threadripper (workstation/high end desktop), and Epyc (server). Comparison History First generation The first generation Zen was launched with the Ryzen 1000 series of CPUs (codenamed Summit Ridge) in February 2017. The first Zen-based preview system was demonstrated at E3 2016, and first substantially detailed at an event hosted a block away from the Intel Developer Forum 2016. The first Zen-based CPUs reached the market in early March 2017, and Zen-derived Epyc server processors (codenamed "Naples") launched in June 2017 and Zen-based APUs (codenamed "Raven Ridge") arrived in November 2017. This first iteration of Zen utilized GlobalFoundries' 14 nm manufacturing process. Modified Zen-based processors for the Chinese market were built. First generation refresh Zen+ was first released in April 2018, powering the second generation of Ryzen processors, known as Ryzen 2000 (codenamed "Pinnacle Ridge") for mainstream desktop systems, and Threadripper 2000 (codenamed "Colfax") for high-end desktop setups. Zen+ used GlobalFoundries' 12 nm process, an enhanced version of their 14 nm node. Second generation The Ryzen 3000 series CPUs were released on July 7, 2019, while the Zen 2-based Epyc server CPUs (codename "Rome") were released on August 7, 2019. Zen 2 Matisse products were the first consumer CPUs to use TSMC's 7 nm process node. Zen 2 introduced the chiplet based architecture, where desktop, workstation, and server CPUs are all produced as multi-chip modules (MCMs); these Zen 2 products utilise the same core chiplets but are attached to different uncore silicon (different IO dies) in a hub and spoke topology. This approach differs from Zen 1 products, where the same die (Zeppelin) is used in a simple monolithic package for Summit Ridge products (Ryzen 1
https://en.wikipedia.org/wiki/HackerNest
HackerNest is a not-for-profit organization and global movement founded on January 11, 2011. The organization unites local technology communities around the world through community events and socially beneficial hackathons to further its mission of economic development through technological proliferation. It is Canada's largest, most prolific technology community with growing international reach. Background HackerNest was founded on the belief that the fastest, most permanent way to improve the world is to build supportive local technology and innovation communities characterized by trust, sharing, and respect. The rationale is that the technology community is the cornerstone of economic development enabling collaboration, innovation, knowledge-sharing, recruiting, and scientific progress. Growing and strengthening the community lets businesses hire better, perform better, and create more jobs, which ultimately increases economic prosperity. The organization's ideology is rooted in the idea that minor tweaks at the start of a process in a dynamic system can have a major impact on the end result. . HackerNest "splinter cells" (chapters) regularly host "Tech Socials" open to anyone interested in technology. The events vary slightly by city, but maintain the same core tenets: all are friendly and down-to-earth. The first Tech Social was held in Toronto on Monday, January 31, 2011. HackerNest Toronto is currently the world's largest Meetup group for programmers and Canada's largest technologist community. As of July 2017, HackerNest splinter cells have run over 550+ events in 34 cities across 14 countries on 5 continents. Activities In 2014, HackerNest produced Construct, Canada's largest hardware hackathon and DementiaHack for the British government, the world's first hackathon dedicated to helping people with dementia and their caregivers. In 2015, the organization produced Deloitte's first internal innovation hackathon as well as DementiaHack Facebook as
https://en.wikipedia.org/wiki/Guard%20theory
Guard theory is a branch of immunology which concerns the innate sensing of stereotypical consequences of a virulence factor or pathogen. This is in contrast to the classical understanding of recognition by the innate immune system, which involves recognition of distinct microbial structures- pathogen-associated molecular patterns (PAMPs)- with pattern recognition receptors (PRRs). Some of these stereotypical consequences of virulence factors and pathogens may include altered endosomal trafficking and changes in the cytoskeleton. These recognition mechanisms would work to complement classical pattern recognition mechanisms. Mechanism In plants In plants, guard theory is also known as indirect recognition. This is because rather than direct recognition of a virulence factor or pathogen, there is instead recognition of the end result of a process mediated by a virulence factor or pathogen. In these cases, the virulence factor appears to target an accessory protein that is either a target or a structural mimic of the target of that virulence factor, allowing for plant defences to respond to a specific strategy of pathogenesis rather structures that may evolve and change over time at a faster rate than the plant can adapt to. The interaction between pathogen and accessory protein results in some modification of the accessory protein, which allows for recognition by plant NBS-LRR proteins, which monitor for infection. This model is best illustrated by RIN4 protein in A. thaliana. RIN4 forms a complex with the NB-LRR proteins RPM1 and RPS2. Protease effector AvrRpt2 is able to degrade RIN4, causing de-repression of RPS2. On the other hand, AvrB or AvrRPM1-mediated phosphorylation of RIN4 results in activation of RPM1. In short, this example elucidates how one NBS-LRR protein is able to recognize the effects of more than one virulence factor or effector. Guard defences in humans and relationship with allergies Little is known concerning guard receptors in humans. One
https://en.wikipedia.org/wiki/WaferCatalyst
WaferCatalyst is a Multi-Project Wafer (MPW) consolidation service by King Abdulaziz City for Science and Technology (KACST), Saudi Arabia. WaferCatalyst is a concept to silicon service and provides a number of tools for community building in the field of integrated circuit (IC) design. These include Multi-project wafer service fabrication, multi-layer mask (MLM), design support, consultancy services and fabrication support. History Microsystems Infrastructure Development initiative (MIDI) was launched by Micro-Sensors Division (MSD) under the National Center of Electronics, Communications and Photonics in early 2012 and was chartered to create and develop integrated circuit ecosystem in the Kingdom of Saudi Arabia. The service developed as a result of this initiative is called 'WaferCatalyst' which has been chartered to serve Saudi Arabia, greater Middle East & North Africa region as well as the global community in semiconductor design. It was formally launched on 28 April 2013 by H.H. Prince Dr. Turki bin Saud bin Mohammad Al Saud, Vice President for Research Institutes of the Kingdom of Saudi Arabia. Programs WaferCatalyst has a number of programs to enhance the ecosystem in IC development area. These include support services through its Support and Design Portal, support on process development kits, support for universities in taping out of ICs, project titles for undergraduate/post-graduate students and partners program. WaferCatalyst has been working to develop close coordination and partnerships among the various institutions, research and commercial organizations to add value for themselves while also contributing to the development of a virtual cluster ecosystem. Related links Multi-project wafer service External links WaferCatalyst Website www.wafercat.com WaferCatalyst Portal http://portal-wafercat.com King Abdulaziz City for Science and Technology www.kacst.edu.sa References 2013 establishments in Saudi Arabia Science and technology in Saud
https://en.wikipedia.org/wiki/Process%20performance%20qualification%20protocol
Process performance qualification protocol is a component of process validation: process qualification. This step is vital in maintaining ongoing production quality by recording and having available for review essential conditions, controls, testing, and expected manufacturing outcome of a production process. The Food and Drug Administration recommends the following criteria be included in a PPQ protocol: Manufacturing conditions: Operating parameters, equipment limits, and component inputs What data should be recorded and analyzed What tests should be performed to ensure quality at each production step A sampling plan to outline sampling methods both during and between production batches Analysis methodology that allows for data scientific and risk oriented decision making based on statistical data. Variability limits should be defined and contingencies in the event of non-conforming data established Approval of PPQ protocol from relevant departments Deviations from the standard operation procedures should be made within the framework of the protocol and at the approval of relevant quality control departments. The FDA further recommends a documentation of the protocol be published internally. The report should include: A summation of relevant data and analysis from the protocol An explanation of unexpected data and any other results not mandated by the protocol and its effects on production quality Identify correlating effects and suggest changes to existing processes Conclude if the process performance is adequately qualified to meet performance standards. Should production standards not be met appropriate changes should be outlined References External links casss.org Process Validation Guidance Formal methods Enterprise modelling Business process management Validity (statistics) Pharmaceutical industry
https://en.wikipedia.org/wiki/Mysms
mysms is a cross-platform text messaging service, based on Short Message Service (SMS), for smartphones, tablet computers, and laptop and desktop personal computers. History mysms is a brand of sms.at, which is a company of Up to Eleven Digital Solutions GmbH, and based in Graz, Austria. Up to Eleven was founded by Martin and Jürgen Pansy. Toto Wolff and Rene Berger belong to its shareholders. The global launch of mysms was in July 2012 and its efforts were initially focused in Germany, Austria and Switzerland. The apps are available for several operating systems, including Android, iOS, Mac OS X, Windows, and web browsers. As of December 2014, mysms for Android phone exhibits 1,000,000–5,000,000 downloads on the Google Play Store. Functions mysms synchronizes the existing text messaging history from the smartphone to desktop computers and tablet computers. In contrast to instant messaging services like WhatsApp, mysms focuses on Short Message Service (SMS). Messages can be sent to any mobile phone number, regardless of device type or operating system. Similar to iMessage, users can also send and receive instant messages to other mysms users over the Internet. Partnerships and integrations mysms cooperates with other cloud-services like Evernote and implemented Dropbox and Google Drive integrations to allow users to archive their SMS and multimedia messages. The app for Windows 8 is pre-installed on Hewlett-Packard (HP) laptops. Also, the service is offered as white-label product. For this purpose, mysms partnered with the French mobile network operator Bouygues Telecom. Open API mysms provides an open application programming interface (API) to other developers that allows them to build their own application based on mysms' core features. Achievements In 2012, mysms achieved second place in the Austrian Fast Forward Award. The German magazine Computer Bild mentioned mysms as winner of their messenger test. As of December 22, 2014, the mysms app for Android
https://en.wikipedia.org/wiki/Chemiresistor
A chemiresistor is a material that changes its electrical resistance in response to changes in the nearby chemical environment. Chemiresistors are a class of chemical sensors that rely on the direct chemical interaction between the sensing material and the analyte. The sensing material and the analyte can interact by covalent bonding, hydrogen bonding, or molecular recognition. Several different materials have chemiresistor properties: metal-oxide semiconductors, some conductive polymers, and nanomaterials like graphene, carbon nanotubes and nanoparticles. Typically these materials are used as partially selective sensors in devices like electronic tongues or electronic noses. A basic chemiresistor consists of a sensing material that bridges the gap between two electrodes or coats a set of interdigitated electrodes. The resistance between the electrodes can be easily measured. The sensing material has an inherent resistance that can be modulated by the presence or absence of the analyte. During exposure, analytes interact with the sensing material. These interactions cause changes in the resistance reading. In some chemiresistors the resistance changes simply indicate the presence of analyte. In others, the resistance changes are proportional to the amount of analyte present; this allows for the amount of analyte present to be measured. History As far back as 1965 there are reports of semiconductor materials exhibiting electrical conductivities that are strongly affected by ambient gases and vapours. However, it was not until 1985 that Wohltjen and Snow coined the term chemiresistor. The chemiresistive material they investigated was copper phthalocyanine, and they demonstrated that its resistivity decreased in the presence of ammonia vapour at room temperature. In recent years chemiresistor technology has been used to develop promising sensors for many applications, including conductive polymer sensors for secondhand smoke, carbon nanotube sensors for gaseous ammo
https://en.wikipedia.org/wiki/Dolby%20Cinema
Dolby Cinema is a premium cinema created by Dolby Laboratories that combines Dolby proprietary technologies such as Dolby Vision and Dolby Atmos, as well as other signature entrance and intrinsic design features. The technology competes with IMAX and other premium large formats such as Cinemark XD and Regal's RPX (Regal Premium Experience). History The first installations featuring Dolby Cinema were JT (now Vue) Bioscopen Cinema in Eindhoven, Netherlands on 18 December 2014; followed by Cinesa La Maquinista in Barcelona, Spain. Dolby Laboratories has partner contracts with Cinesa, Vue Cinemas, AMC Theatres (known as Dolby Cinema at AMC, formerly Dolby Cinema at AMC Prime until early 2017), Cineplexx Cinemas, Wanda Cinemas, Jackie Chan Cinema, Reel Cinemas and Odeon Cinemas to install Dolby Cinemas. On May 26, 2017, Dolby announced they made a deal with Les Cinémas Gaumont Pathé to open 10 new locations in Europe. Seven will be located in France and three will be located in The Netherlands. Technology Dolby Vision Dolby Cinema utilizes a Dolby Vision projection system developed by Dolby Laboratories in conjunction with Christie Digital. The system consists of dual Christie 4K 6P (primary) modular laser projectors featuring a custom design to allow for unique light path. The system is capable of delivering up to 14 foot-lambert (48 nits) on unity-gain matte-white screens for 3-D (and up to 31 foot-lambert (106 nits) for 2-D), a substantial improvement on current generation 3-D systems which deliver in the range of 3 to 4 foot-lambert for 3-D. The result is improved brightness, color and contrast compared to traditional xenon projectors. The first theaters temporarily used off-the-shelf dual Christie 4K laser projectors until the Dolby Vision-capable ones were shipped out in spring 2015. Dolby 3D uses spectrum separation, where the two projectors function in stacked operation with each projector emitting a slightly different wavelength of red, green, and blue pr
https://en.wikipedia.org/wiki/Griffiths%20group
In mathematics, more specifically in algebraic geometry, the Griffiths group of a projective complex manifold X measures the difference between homological equivalence and algebraic equivalence, which are two important equivalence relations of algebraic cycles. More precisely, it is defined as where denotes the group of algebraic cycles of some fixed codimension k and the subscripts indicate the groups that are homologically trivial, respectively algebraically equivalent to zero. This group was introduced by Phillip Griffiths who showed that for a general quintic in (projective 4-space), the group is not a torsion group. Notes References Algebraic geometry
https://en.wikipedia.org/wiki/Tri-level%20sync
Tri-level sync is an analogue video synchronization pulse primarily used for the locking of high-definition video signals (genlock). It is preferred in HD environments over black and burst, as timing jitter is reduced due to the nature of its higher frequency. It also benefits from having no DC content, as the pulses are in both polarities. Synchronization Modern real-time multi-source HD facilities have many pieces of equipment that all output HD-SDI video. If this baseband video is to be mixed, switched or luma keyed with any other sources, then they will need to be synchronous, i.e. the first pixel of the first line must be transmitted at the same time (within a few microseconds). This then allows the switcher to cut, mix or key these sources together with a minimal amount of delay (~1 HD video line 1/(1125×25) seconds for 50i video). This synchronization is done by supplying each piece of equipment with either a tri-level sync, or black-and-burst input. There are video switchers that do not require synchronous sources, but these operate with a much bigger delay Waveform The main pulse definition is as follows: a negative-going pulse of 300 mV lasting 40 sample clocks followed by a positive-going pulse of 300 mV lasting 40 sample clocks. The allowed rise/fall time for each of the transitions is 4 sample clocks. This is with a clock rate of 74.25 MHz. References Synchronization Film and video technology Broadcast engineering Television terminology
https://en.wikipedia.org/wiki/Design%20space%20verification
Design space verification is defined by the European Medicines Agency as the verification that material inputs and processes are able to scale to commercial manufacturing levels while maintaining a standard of quality. Therefore, it is difficult to conduct design space verification while not operating at target levels and should be conducted over the manufacturing lifecycle. Changes in manufacturing output within the design space should not present any risks. Should the manufacturing load exceed the boundaries defined as normal operating ranges unanticipated scale-dependent issues can occur. Design space verification is a part of process validation as defined by the EMA in conjunction with the FDA. Its purpose is to guarantee end product quality within a range of manufacturing boundaries. The effects of scale up activities should be fully understood by the manufacturer. Most initial design space conclusions are based upon laboratory testing or pilot batches with scale up effects being inferred by experimentation or based on statistical evidence, simulations, or studies. Ongoing design space verification should be dependent upon the results of an assessment of risk involved with scale up activities. More specifically, how scaling up production affects scale-dependent variables. Design space verification is much more focused in scope than overall process validation. Design space verification specifically aims to confirm output quality within a given operating range. This allows for changes in operating level flexibility while guaranteeing production quality, and allows for changes in production quantities without necessitating a reevaluation of the production process. References External links NSF Health Sciences BioPharm International Quality management Formal methods Pharmaceutical industry
https://en.wikipedia.org/wiki/Orbital%20Mechanics%20for%20Engineering%20Students
Orbital Mechanics for Engineering Students is an aerospace engineering textbook by Howard D. Curtis, in its fourth edition . The book provides an introduction to orbital mechanics, while assuming an undergraduate-level background in physics, rigid body dynamics, differential equations, and linear algebra. Topics covered by the text include a review of kinematics and Newtonian dynamics, the two-body problem, Kepler's laws of planetary motion, orbit determination, orbital maneuvers, relative motion and rendezvous, and interplanetary trajectories. The text focuses primarily on orbital mechanics, but also includes material on rigid body dynamics, rocket vehicle dynamics, and attitude control. Control theory and spacecraft control systems are less thoroughly covered. The textbook includes exercises at the end of each chapter, and supplemental material is available online, including MATLAB code for orbital mechanics projects. References Engineering textbooks Elsevier books Astrodynamics Orbital maneuvers Spacecraft attitude control Spacecraft propulsion
https://en.wikipedia.org/wiki/ADISA%20certification
Founded in 2010, ADISA Certification Limited (formerly Asset Disposal and Information Security Alliance) is a certification body for companies that provide IT Asset Disposal services. In July 2021, the ADISA ICT Asset Recovery Standard 8.0 was formally approved by the UK Information Commissioner's Office as a UK GDPR Certification Scheme. In 2019, ADISA launched the ADISA Research Centre (ARC). ARC delivers product certification schemes for software and hardware data sanitization tools. The ADISA Product Claims and Product Assurance Schemes are different levels of product testing for data sanitization tools.   The Product Claims Test (PCT) scientifically evaluates the claim behind the data sanitization capabilities of a software or hardware device to determine its validity. The Product Assurance Test offers a higher level of assurance than the PCT; it requires a larger sample size to be forensically analyzed and measures the vendors of the software or hardware device against a range of requirements. ADISA won the 2020 Computer Security Magazine "One to Watch" award to follow up on previous wins including the 2019 "Computer Security Compliance Company of the Year " award and Training Provider of the Year (2015). The ADISA Standard is recognized as an industry standard of merit by the UK Defence Infosec Product Co-Operation Group (DIPCOG) and is listed on the National Cyber Security Centre's guidance for companies when disposing of IT assets. ADISA owns a YouTube channel called "ADISA Media Centre" where they provide information from the group in various formats. See also Computer recycling Data remanence Electronic waste Sanitization (classified information) References Electronic waste Information technology organisations based in the United Kingdom Organisations based in the City of Westminster Waste organizations
https://en.wikipedia.org/wiki/Jes%C3%BAs%20A.%20De%20Loera
Jesús Antonio De Loera (born January 18, 1966) is a Mexican-American mathematician at the University of California, Davis, specializing in discrete mathematics and discrete geometry. Education De Loera did his undergraduate studies at the National Autonomous University of Mexico, earning a B.S. in mathematics in 1989. After earning a master's degree from Western Michigan University a year later, and another master's degree in 1993 from Cornell University, he finished his doctorate from Cornell in 1995 under the supervision of Bernd Sturmfels. Career De Loera joined the UC Davis faculty as an assistant professor in 1998, and was promoted to full professor in 2005. He served on the Scientific Advisory Board for the Institute for Computational and Experimental Research in Mathematics (ICERM) through 2021. Awards and honors In 2014, he became a fellow of the American Mathematical Society "for contributions to discrete geometry and combinatorial optimization as well as for service to the profession, including mentoring and diversity". In 2019 he was named a SIAM Fellow "for contributions to discrete geometry and optimization, polynomial algebra, and mathematical software". In 2020 he received the INFORMS optimization society's Farkas Prize "in recognition of his pioneering work at the intersection of discrete mathematics, optimization and algebraic geometry". He delivered an invited plenary address at the 2021 National Math Festival. In 2021, he was elected Vice President of the AMS. References External links Home page 1966 births Living people 20th-century American mathematicians Mexican mathematicians Combinatorialists Researchers in geometric algorithms National Autonomous University of Mexico alumni Western Michigan University alumni Cornell University alumni University of California, Davis faculty American academics of Mexican descent Fellows of the American Mathematical Society Fellows of the Society for Industrial and Applied Mathematics Place of birth mi
https://en.wikipedia.org/wiki/American%20Review%20%28political%20magazine%29
American Review (Global Perspectives on US Affairs) was an online political magazine published twice a year by the United States Studies Centre at the University of Sydney in Australia. The magazine was established in 2006. As the magazine was based outside America, the emphasis was on a "fair and balanced" interpretation of American affairs with some focus on the US relationship with Australia and the Pacific region. It ceased publishing new material in 2015. References External links 2006 establishments in Australia 2015 disestablishments in Australia Biannual magazines published in Australia Defunct political magazines published in Australia Magazines established in 2006 Magazines disestablished in 2015 Magazines published in Sydney Online magazines
https://en.wikipedia.org/wiki/Particle%20mass%20analyser
A Particle mass analyser (PMA) is an instrument for classifying aerosol particles according to their mass-to-charge ratio using opposing electrical and centrifugal forces. This allows the classifier to select particles of a specified mass-to-charge ratio independent of particle shape. It one of the three types of monodisperse aerosol classifier, the others being the differential mobility analyser (DMA, for electrical mobility size), and the aerodynamic aerosol classifier (AAC, for relaxation time, or aerodynamic diameter). The corresponding three quantities are related by the expression τ = mB, where τ is relaxation time, m is mass and B is mobility. Further work improved the technique by engineering the centrifugal force to match the electrostatic force across the whole classification region, thus increasing the throughput. References Measuring instruments Scientific instruments Scientific techniques Aerosol measurement
https://en.wikipedia.org/wiki/Nucleus%20%28order%20theory%29
In mathematics, and especially in order theory, a nucleus is a function on a meet-semilattice such that (for every in ): Every nucleus is evidently a monotone function. Frames and locales Usually, the term nucleus is used in frames and locales theory (when the semilattice is a frame). Proposition: If is a nucleus on a frame , then the poset of fixed points of , with order inherited from , is also a frame. References Order theory
https://en.wikipedia.org/wiki/Day%20One%20%28app%29
Day One is a personal journaling app available for Android, macOS and iOS devices. Some features include: data syncing with multiple devices; end-to-end encryption, Markdown authoring of entries; location, weather, date, time, and other automatic metadata; quick entry menu bar (Mac only); and reminders. The application has been reviewed by Macworld, The Verge, Lifehacker and other tech websites. The Mac version of Day One was chosen as "Mac App of the Year" in 2012 by the Mac App Store and won an Apple Design Award in 2014 for being "super well-designed on OS X and being exactly what it needs to be and nothing more." Day One was acquired by Automattic, the owner of WordPress.com, on June 14, 2021. History Day One had a service called Publish for publishing selected entries to a Day One-hosted webpage and sharing via Facebook, Twitter, Tumblr, and other social services. Development On February 4, 2016, the developers, Bloom Built, introduced Day One 2 for Mac and iOS, citing the release as a full app rebuild. The update introduced a visual overhaul and new features such as multiple journals and photos. End-to-End Encryption On June 12, 2017, Bloom Built added end-to-end encryption to Day One 2 after two years of development. The update was a culmination of the Day One Sync services launched in 2015 as a replacement for iCloud and DropBox sync. References External links iOS App Store link Android (operating system) software IOS software MacOS software Automattic 2021 mergers and acquisitions
https://en.wikipedia.org/wiki/Kallithea%20%28software%29
Kallithea is a cross-platform free software source code management system, the primary goal of which is to provide a repository hosting service with features for collaboration, such as forking, pull requests, code review, issue tracking etc. Kallithea is a fork of RhodeCode, created after the original developer had changed the license terms. While earlier versions of RhodeCode were licensed entirely under the GNU General Public License version 3, RhodeCode version 2.0 (released in August 2013) introduced exceptions for parts of the software distribution. According to Bradley M. Kuhn of Software Freedom Conservancy, this exception statement is ambiguous and "leaves the redistributor feeling unclear about their rights". Kallithea is mostly written in Python. Kallithea is a member project of Software Freedom Conservancy. Features Kallithea supports hosting repositories of Mercurial and Git version control systems. Repositories can be grouped and thus allow to define common properties like access control. Its web interface for projects allows to fork as well as management of pull requests. It can also be used to quickly exchange code snippets by means of a revision controlled pastebin ("gists"). See also Comparison of project management software List of tools for code review Comparison of source code hosting facilities Apache Allura Apache Bloodhound Trac References External links Open-source hosted development tools Project management software Free software programmed in Python Free project management software Version control Software review Software forks
https://en.wikipedia.org/wiki/David%20G.%20Cantor
David Geoffrey Cantor (April 12, 1935 – November 19, 2012) was an American mathematician, specializing in number theory and combinatorics. The Cantor–Zassenhaus algorithm for factoring polynomials is named after him; he and Hans Zassenhaus published it in 1981. Biography Cantor was born on April 12, 1935. He completed his undergraduate studies at the California Institute of Technology, graduating in 1956, and earned his doctorate from the University of California, Los Angeles (UCLA) in 1960, where he was supervised by Basil Gordon and Ernst G. Straus. He became an assistant professor at the University of Washington in 1962, moved back to UCLA in 1964, and retired in 1991. After his retirement, he worked at the Center for Communications Research in La Jolla, California. Cantor specialized in number theory and combinatorics. The Cantor–Zassenhaus algorithm for factoring polynomials is named after him; he and Hans Zassenhaus published it in 1981. He received the National Science Foundation Postdoctoral Fellowship in 1960 and a Sloan Foundation Fellowship in 1968. In 2012, he became one of the inaugural fellows of the American Mathematical Society. At the time of his death, he had been a member of the American Mathematical Society for 54 years. Cantor lived in San Diego, California. He died on November 19, 2012, at the age of 77. See also Cantor's algorithm References 1935 births 2012 deaths 20th-century American mathematicians Combinatorialists Number theorists California Institute of Technology alumni University of California, Los Angeles alumni University of Washington faculty University of California, Los Angeles faculty Fellows of the American Mathematical Society
https://en.wikipedia.org/wiki/Ubimet
The Institute for Ubiquitous Meteorology (UBIMET) is a worldwide private provider of weather forecasts and severe weather warnings. History Ubimet was established in Vienna, Austria, by Dr. Michael Fassnauer and Manfred Spatzierer in September 2004 under the name Meteomedia GmbH. It began by providing weather services to institutions such as the German Press Agency, Austrian Federal Railways, and Uniqa Insurance Group. Ubimet went on to build the Austrian Severe Weather Center to forecast extreme weather events. In December 2008, shares of Meteomedia Group were bought back by the company's founders, and the name was changed to UBIMET GmbH. Today, companies in the insurance, media (such as Der Standard and Kronen Zeitung), sports (the International Automobile Federation, Moto GP, Red Bull, the Australian Olympic sailing team, the 2013 European Beach Volleyball Championships), infrastructure (Austrian Federal Railways, Deutsche Bahn), energy (E.on), mining (Austmine), and logistics (STRABAG) industries use UBIMET's corporate services. In 2011, the company opened an office in Melbourne, Australia, with a full-scale operation including meteorologists and business development resources, as well as marketing and infrastructure. In 2012, Ubimet provided weather data for the Red Bull Stratos space diving project during which Felix Baumgartner successfully ascended to 128,100 feet in a stratospheric balloon and free-fell at supersonic speed before parachuting to the ground. In the same year, Red Bull became a company shareholder. Ubimet now provides weather data for many of Red Bull's global sporting events (e.g., the Wings For Life World Run in 2014 and 2015). Since 2014, Ubimet has had a contract with the International Automobile Federation (FIA) to provide weather forecasts for all Formula One Grand Prix races. It did so for the Formula One race on October 5, 2014, in Suzuka, Japan, when Typhoon Phanfone approached the city. Ubimet also provided meteorological data f
https://en.wikipedia.org/wiki/Frubber
Frubber (from "flesh rubber") is a patented elastic form of rubber used in robotics. The spongy elastomer has been used by Hanson Robotics for the face of its android robots, including Einstein 3 and Sophia. References Robotics
https://en.wikipedia.org/wiki/Nine-point%20conic
In geometry, the nine-point conic of a complete quadrangle is a conic that passes through the three diagonal points and the six midpoints of sides of the complete quadrangle. The nine-point conic was described by Maxime Bôcher in 1892. The better-known nine-point circle is an instance of Bôcher's conic. The nine-point hyperbola is another instance. Bôcher used the four points of the complete quadrangle as three vertices of a triangle with one independent point: Given a triangle and a point in its plane, a conic can be drawn through the following nine points: the midpoints of the sides of , the midpoints of the lines joining to the vertices, and the points where these last named lines cut the sides of the triangle. The conic is an ellipse if lies in the interior of or in one of the regions of the plane separated from the interior by two sides of the triangle, otherwise the conic is a hyperbola. Bôcher notes that when is the orthocenter, one obtains the nine-point circle, and when is on the circumcircle of , then the conic is an equilateral hyperbola. In 1912 Maud Minthorn showed that the nine-point conic is the locus of the center of a conic through four given points. References Fanny Gates (1894) Some Considerations on the Nine-point Conic and its Reciprocal, Annals of Mathematics 8(6):185–8, link from Jstor. Eric W. Weisstein Nine-point conic from MathWorld. Michael DeVilliers (2006) The nine-point conic: a rediscovery and proof by computer from International Journal of Mathematical Education in Science and Technology, a Taylor & Francis publication. Christopher Bradley The Nine-point Conic and a Pair of Parallel Lines from University of Bath. Further reading W. G. Fraser (1906) "On relations of certain conics to a triangle", Proceedings of the Edinburgh Mathematical Society 25:38–41. Thomas F. Hogate (1894) On the Cone of Second Order which is Analogous to the Nine-point Conic, Annals of Mathematics 7:73–6. P. Pinkerton (1905) "On a nine-po
https://en.wikipedia.org/wiki/Blackman%27s%20theorem
Blackman's theorem is a general procedure for calculating the change in an impedance due to feedback in a circuit. It was published by Ralph Beebe Blackman in 1943, was connected to signal-flow analysis by John Choma, and was made popular in the extra element theorem by R. D. Middlebrook and the asymptotic gain model of Solomon Rosenstark. Blackman's approach leads to the formula for the impedance Z between two selected terminals of a negative feedback amplifier as Blackman's formula: where ZD = impedance with the feedback disabled, TSC = loop transmission with a small-signal short across the selected terminal pair, and TOC = loop transmission with an open circuit across the terminal pair. The loop transmission also is referred to as the return ratio. Blackman's formula can be compared with Middlebrook's result for the input impedance Zin of a circuit based upon the extra-element theorem: where: is the impedance of the extra element; is the input impedance with removed (or made infinite); is the impedance seen by the extra element with the input shorted (or made zero); is the impedance seen by the extra element with the input open (or made infinite). Blackman's formula also can be compared with Choma's signal-flow result: where is the value of under the condition that a selected parameter P is set to zero, return ratio is evaluated with zero excitation and is for the case of short-circuited source resistance. As with the extra-element result, differences are in the perspective leading to the formula. See also Mason's gain formula Further reading References Electronic feedback Signal processing Electronic amplifiers Control engineering
https://en.wikipedia.org/wiki/Annals%20of%20Nuclear%20Energy
Annals of Nuclear Energy is a monthly peer-reviewed scientific journal covering research on nuclear energy and nuclear science. It was established in 1975 and is published by Elsevier. The current editors-in-chief are Lynn E. Weaver (Florida Institute of Technology), S. Mostafa Ghiaasiaan (Georgia Institute of Technology) and Imre Pázsit (Chalmers University of Technology). Abstracting and indexing The journal is abstracted and indexed in: Chemical Abstracts Service Index Medicus/MEDLINE/PubMed Science Citation Index Expanded Current Contents/Engineering, Computing & Technology Scopus According to the Journal Citation Reports, the journal has a 2013 impact factor of 1.020. Former titles history Annals of Nuclear Energy is derived from the following former titles: Journal of Nuclear Energy (1954-1959) Journal of Nuclear Energy. Part A. Reactor Science (1959-1961) Journal of Nuclear Energy. Part B. Reactor Technology (1959) Journal of Nuclear Energy. Parts A/B. Reactor Science and Technology (1961-1966) Journal of Nuclear Energy (1967-1973) Annals of Nuclear Science and Engineering (1974) Annals of Nuclear Energy (1975–present) Notes References External links Energy and fuel journals Elsevier academic journals English-language journals Monthly journals Academic journals established in 1975