source
stringlengths
31
203
text
stringlengths
28
2k
https://en.wikipedia.org/wiki/Food%20Processing%20Technology%20Building
The Food Processing Technology Building is a Georgia Institute of Technology and Georgia Tech Research Institute facility. It houses the Food Processing Technology Division of GTRI, which includes the Agricultural Technology Research Program (ATRP) and Georgia’s Traditional Industries Program for Food Processing. It opened on March 1, 2005, and was dedicated on May 19, 2005. Facilities The Food Processing Technology Building contains over 36,000 square feet of office and laboratory space, including a 4,370 square foot high-bay testing and fabrication space, a 16-by-24-foot climate-controlled experiment chamber, an indoor environmental pilot area, a full-service chemical wet laboratory, and a 48-seat auditorium. The building houses five research laboratories: an automation research laboratory, an electronics lab, a systems development and integration laboratory, an environmental laboratory, and an optics laboratory. The building's lower lobby area features an interactive exhibit about the role of technology in poultry and food processing. References Georgia Tech buildings and structures 2005 establishments in Georgia (U.S. state) Education in Atlanta Food processing
https://en.wikipedia.org/wiki/Directed%20differentiation
Directed differentiation is a bioengineering methodology at the interface of stem cell biology, developmental biology and tissue engineering. It is essentially harnessing the potential of stem cells by constraining their differentiation in vitro toward a specific cell type or tissue of interest. Stem cells are by definition pluripotent, able to differentiate into several cell types such as neurons, cardiomyocytes, hepatocytes, etc. Efficient directed differentiation requires a detailed understanding of the lineage and cell fate decision, often provided by developmental biology. Conceptual frame During differentiation, pluripotent cells make a number of developmental decisions to generate first the three germ layers (ectoderm, mesoderm and endoderm) of the embryo and intermediate progenitors, followed by subsequent decisions or check points, giving rise to all the body's mature tissues. The differentiation process can be modeled as sequence of binary decisions based on probabilistic or stochastic models. Developmental biology and embryology provides the basic knowledge of the cell types' differentiation through mutation analysis, lineage tracing, embryo micro-manipulation and gene expression studies. Cell differentiation and tissue organogenesis involve a limited set of developmental signaling pathways. It is thus possible to direct cell fate by controlling cell decisions through extracellular signaling, mimicking developmental signals. Source material Directed differentiation is primarily applied to pluripotent stem cells (PSCs) of mammalian origin, in particular mouse and human cells for biomedical research applications. Since the discovery of embryonic stem (ES) cells (1981) and induced pluripotent stem (iPS) cells (2006), source material is potentially unlimited. Historically, embryonic carcinoma (EC) cells have also been used. Fibroblasts or other differentiated cell types have been used for direct reprogramming strategies. Methods Cell differentiation invo
https://en.wikipedia.org/wiki/Chemical%20reaction%20model
Chemical reaction models transform physical knowledge into a mathematical formulation that can be utilized in computational simulation of practical problems in chemical engineering. Computer simulation provides the flexibility to study chemical processes under a wide range of conditions. Modeling of a chemical reaction involves solving conservation equations describing convection, diffusion, and reaction source for each component species. Species transport equation Ri is the net rate of production of species i by chemical reaction and Si is the rate of creation by addition from the dispersed phase and the user defined source. Ji is the diffusion flux of species i, which arises due to concentration gradients and differs in both laminar and turbulent flows. In turbulent flows, computational fluid dynamics also considers the effects of turbulent diffusivity. The net source of chemical species i due to reaction, Ri which appeared as the source term in the species transport equation is computed as the sum of the reaction sources over the NR reactions among the species. Reaction models These reaction rates R can be calculated by following models: Laminar finite rate model Eddy dissipation model Eddy dissipation concept Laminar finite rate model The laminar finite rate model computes the chemical source terms using the Arrhenius expressions and ignores turbulence fluctuations. This model provides with the exact solution for laminar flames but gives inaccurate solution for turbulent flames, in which turbulence highly affects the chemistry reaction rates, due to highly non-linear Arrhenius chemical kinetics. However this model may be accurate for combustion with small turbulence fluctuations, for example supersonic flames. Eddy dissipation model The eddy dissipation model or the Magnussen model, based on the work of Magnussen and Hjertager, is a turbulent-chemistry reaction model. Most fuels are fast burning and the overall rate of reaction is controlled by t
https://en.wikipedia.org/wiki/Flajolet%E2%80%93Martin%20algorithm
The Flajolet–Martin algorithm is an algorithm for approximating the number of distinct elements in a stream with a single pass and space-consumption logarithmic in the maximal number of possible distinct elements in the stream (the count-distinct problem). The algorithm was introduced by Philippe Flajolet and G. Nigel Martin in their 1984 article "Probabilistic Counting Algorithms for Data Base Applications". Later it has been refined in "LogLog counting of large cardinalities" by Marianne Durand and Philippe Flajolet, and "HyperLogLog: The analysis of a near-optimal cardinality estimation algorithm" by Philippe Flajolet et al. In their 2010 article "An optimal algorithm for the distinct elements problem", Daniel M. Kane, Jelani Nelson and David P. Woodruff give an improved algorithm, which uses nearly optimal space and has optimal O(1) update and reporting times. The algorithm Assume that we are given a hash function that maps input to integers in the range , and where the outputs are sufficiently uniformly distributed. Note that the set of integers from 0 to corresponds to the set of binary strings of length . For any non-negative integer , define to be the -th bit in the binary representation of , such that: We then define a function that outputs the position of the least-significant set bit in the binary representation of , and if no such set bit can be found as all bits are zero: Note that with the above definition we are using 0-indexing for the positions, starting from the least significant bit. For example, , since the least significant bit is a 1 (0th position), and , since the least significant set bit is at the 3rd position. At this point, note that under the assumption that the output of our hash function is uniformly distributed, then the probability of observing a hash output ending with (a one, followed by zeroes) is , since this corresponds to flipping heads and then a tail with a fair coin. Now the Flajolet–Martin algorithm for estimat
https://en.wikipedia.org/wiki/Lea%20Verou
Lea Verou () is a computer scientist, front end web developer, speaker and author, originally from Lesbos, Greece. Verou is currently a research assistant at Massachusetts Institute of Technology (MIT) Computer Science and Artificial Intelligence Laboratory (CSAIL), an elected participant in the World Wide Web Consortium (W3C) Technical Architecture Group (TAG), and an Invited Expert in the W3C CSS Working Group. She is the author of CSS Secrets: Better Solutions to Everyday Web Design Problems (). Education Verou earned her Bachelor of Science in Computer Science from Athens University of Economics and Business. She had previously attended Aristotle University of Thessaloniki studying Electrical Engineering and Computer Science. She later earned a Master of Science in Computer Science from MIT. Since 2014, she has been a Doctor of Philosophy (PhD) candidate at MIT in Electrical Engineering and Computer Science. Her background encompasses both technical development and visual design. Career Verou started her career as a freelancer in 2005 while she was in university. In 2008, she co-founded a Greek web development company called Fresset Ltd, which managed a number of Greek community websites that she and her co-founder had created. The company was sold in 2013. In 2012, Verou joined W3C as a Developer Advocate. Open source work During the course of her studies and career, Verou has started over thirty open-source software projects, including PrismJS, a popular syntax highlighter, and Mavo, her research project at MIT. Conference speaking Verou has spoken at over a hundred conferences, including Smashing Magazine's conference SmashingConf, A List Apart's conference An Event Apart, South by Southwest (SXSW), and JQuery's conference. Her talks are known for their distinctive style involving live coding through a custom presentation framework that she built to optimize the experience for the audience. Bibliography and publications Her first book, CSS Se
https://en.wikipedia.org/wiki/Grit%2C%20not%20grass%20hypothesis
The grit, not grass hypothesis is an evolutionary hypothesis that explains the evolution of high-crowned teeth, particularly in New World mammals. The hypothesis is that the ingestion of gritty soil is the primary driver of hypsodont tooth development, not the silica-rich composition of grass, as was previously thought. Traditional co-evolution hypothesis Since the morphology of the hypsodont tooth is suited to a more abrasive diet, hypsodonty was thought to have evolved concurrently with the spread of grasslands. During the Cretaceous Period (145-66 million years ago), the Great Plains were covered by a shallow inland sea called the Western Interior Seaway which began to recede during the Late Cretaceous to the Paleocene (65-55 million years ago), leaving behind thick marine deposits and relatively flat terrain. During the Miocene and Pliocene epochs (25 million years), the continental climate became favorable to the evolution of grasslands. Existing forest biomes declined and grasslands became much more widespread. The grasslands provided a new niche for mammals, including many ungulates that switched from browsing diets to grazing diets. Grass contains silica-rich phytoliths (abrasive granules), which wear away dental tissue more quickly. So the spread of grasslands was linked to the development of high-crowned (hypsodont) teeth in grazers. Modern evolutionary hypothesis Early evidence In 2006 Strömberg examined the independent acquisition of high-crowned cheek teeth (hypsodonty) in several ungulate lineages (e.g., camelids, equids, rhinoceroses) from the early to middle Miocene of North America, which had been classically linked to the spread of grasslands. She showed habitats dominated by C3 grasses (cool-season grasses) were established in the Central Great Plains by early late Arikareean (≥21.9 Million years ago), at least 4 million years prior to the emergence of hypsodonty in Equidae. In 2008 Mendoza and Palmqvist determined the relative importance of gr
https://en.wikipedia.org/wiki/IBM%20ALP
IBM Assembly Language Processor (ALP) is an assembler written by IBM for 32-bit OS/2 Warp (OS/2 3.0), which was released in 1994. ALP accepts source programs compatible with Microsoft Macro Assembler (MASM) version 5.1, which was originally used to build many of the device drivers included with OS/2. For OS/2 versions 3 and 4 ALP was distributed, along with other tools and documentation, as part of the Device Driver Kit (DDK). The DDK was withdrawn in 2004 as part of IBM's discontinuance of OS/2. References External links ALP binaries and documentation ALP Assemblers OS/2 software Programming tools
https://en.wikipedia.org/wiki/Ore%20algebra
In computer algebra, an Ore algebra is a special kind of iterated Ore extension that can be used to represent linear functional operators, including linear differential and/or recurrence operators. The concept is named after Øystein Ore. Definition Let be a (commutative) field and be a commutative polynomial ring (with when ). The iterated skew polynomial ring is called an Ore algebra when the and commute for , and satisfy , for . Properties Ore algebras satisfy the Ore condition, and thus can be embedded in a (skew) field of fractions. The constraint of commutation in the definition makes Ore algebras have a non-commutative generalization theory of Gröbner basis for their left ideals. References Computer algebra Ring theory
https://en.wikipedia.org/wiki/Bulked%20segregant%20analysis
Bulked segregant analysis (BSA) is a technique used to identify genetic markers associated with a mutant phenotype. This allows geneticists to discover genes conferring certain traits of interest, such as disease resistance or susceptibility. This technique involves forming two groups that display opposing phenotypes for a trait of interest. For example, the individuals in one group are resistant to a disease, whereas those in the second group are not. Two bulked DNA samples are then created by pooling the DNA of all individuals in each group. These two bulked samples can then be analysed using techniques such as Restriction fragment length polymorphism or RAPD to detect similarities and differences in the various loci of the genome. The two groups will have a random distribution of alleles in all loci of the genome except for loci that are associated with the mutation. A consistent difference on a locus between the two bulked samples likely means that the locus is associated with the mutation of interest. Generation of testing groups In animals, the individuals making up the two testing groups are usually produced by a cross between two siblings heterozygous for the mutation of interest. The use of siblings is necessary to ensure that the alleles contributing to the mutation are the same among the individuals. There must be a minimum amount of heterozygosity in the various loci of the groups to allow the genes that are associated with the trait of interest to be identified. Since most laboratory strains are inbred, outcrossing of the homozygous mutated individual with a polymorphic strain is essential to generate effective testing groups. The offspring are crossed with each other to generate testing groups. Analysis techniques Bulked DNA samples can be analysed using Southern blotting. Use of restriction enzymes or PCR amplification on the DNA is required for RFLP or RAPD analysis respectively. In these techniques, the loci that are analysed are the restrict
https://en.wikipedia.org/wiki/Industrial%20and%20Mining%20Water%20Research%20Unit
The Industrial and Mining Water Research Unit (abbreviated IMWaRU) is one of several research entities based in the School of Chemical and Metallurgical Engineering at the University of the Witwatersrand, Johannesburg. It provides research as well as supervision to masters and doctorate students within the University, as well as consulting to industry. Unit Structure The unit deals with cross disciplinary water issues relating to industry and mining. As such the group includes experts in chemical engineering, microbiology and other sciences. The unit includes five NRF rated researchers and over 20 masters and doctoral level postgraduate students in the faculties of engineering and science. Members The group currently comprises 7 academics (alphabetically - Mogopoleng (Paul) Chego, Kevin Harding, Michelle Low, Craig Sheridan, Geoffrey Simate, Karl Rumbold and Lizelle van Dyk), as well as several postgraduate students. Logo The logo of the Unit is in the shape of a drop of water, with the left half representing the blue of water. The right half of the drop is modified to show grass and how water is linked to all life. Underneath the icon are the letters IMWaRU, while to the right, the name "Industrial and Mining Water Research Unit" appears. Location The unit is housed in several buildings across the University, most notably in the Richard Ward Building on East campus. Additionally, some members are located in the Biology Building on East Campus and have access to laboratories in that building. They also have access to an outdoor facility on West Campus where constructed wetland, and other outdoor, experiments take place. Research The group has a broad range of research publications in the areas as listed below: Acid mine drainage (AMD) - methods of reducing, treating and managing AMD. Algal Studies - including to clean water, and as a source of biomass for biodiesel Biorefineries - the use of biomass for values add product, including obtaining thes
https://en.wikipedia.org/wiki/Laminar%20flamelet%20model
The laminar flamelet model is a mathematical method for modelling turbulent combustion. The laminar flamelet model is formulated specifically as a model for non-premixed combustion The concept of ensemble of laminar flamelets was first introduced by Forman A. Williams in 1975, while the theoretical foundation was developed by Norbert Peters in the early 80s. Theory The flamelet concept considers the turbulent flame as an aggregate of thin, laminar (Re < 2000), locally one-dimensional flamelet structures present within the turbulent flow field. Counterflow diffusion flame is a common laminar flame which is used to represent a flamelet in a turbulent flow. Its geometry consists of opposed and axi-symmetric fuel and oxidizer jets. As the distance between the jets is decreased and/or the velocity of the jets is increased, the flame is strained and departs from its chemical equilibrium until it eventually extinguishes. The mass fraction of species and temperature fields can be measured or calculated in laminar counterflow diffusion flame experiments. When calculated, a self-similar solution exists, and the governing equations can be simplified to only one dimension i.e. along the axis of the fuel and oxidizer jets. It is in this direction where complex chemistry calculations can be performed affordably. Logic and formulae To model a non-premixed combustion, governing equations for fluid elements are required. The conservation equation for the species mass fraction is as follows:- Lek → lewis number of kth species and the above formula was derived with keeping constant heat capacity. The energy equation with variable heat capacity:- As can be seen from above formulas that the mass fraction and temperature are dependent on 1. Mixture fraction Z 2. Scalar dissipation χ 3. Time Many times we neglect the unsteady terms in above equation and assume the local flame structure having a balance between steady chemical equations and steady diffusion equation which r
https://en.wikipedia.org/wiki/What3words
What3words (stylized as what3words) is a proprietary geocode system designed to identify any location on the surface of Earth with a resolution of about . It is owned by What3words Limited, based in London, England. The system encodes geographic coordinates into three permanently fixed dictionary words. For example, the front door of 10 Downing Street in London is identified by ///slurs.this.shark. What3words differs from most location encoding systems in that it uses words rather than strings of numbers or letters, and the pattern of this mapping is not obvious; the algorithm mapping locations to words is copyrighted. The company has a website, apps for iOS and Android, and an API for bidirectional conversion between What3words addresses and latitude–longitude coordinates. History Founded by Chris Sheldrick, Jack Waley-Cohen, Mohan Ganesalingam and Michael Dent, What3words was launched in July 2013. Sheldrick and Ganesalingam conceived the idea when Sheldrick, working as an event organizer, struggled to get bands and equipment to music venues using inadequate address information. Sheldrick tried using GPS coordinates to locate the venues, but decided that words were better than numbers after a one-digit error led him to the wrong location. He credits a mathematician friend for the idea of dividing the world into squares, and the linguist Jack Waley-Cohen with using memorable words. The company was incorporated in March 2013 and a patent application for the core technology filed in April 2013. In November 2013, What3words raised US$500,000 of seed funding. What3words originally sold "OneWord" addresses, which were stored in a database for a yearly fee, but this offering was discontinued as the company switched to a business-to-business model. In 2015, the company was targeting logistics companies, post offices, and couriers. In January 2018, Mercedes-Benz bought approximately 10% of the company and announced support for What3words in future versions of their i
https://en.wikipedia.org/wiki/Enics
Enics is an Electronics Manufacturing Services (EMS) provider that specializes in industrial electronics. Enics is headquartered in Zürich, Switzerland and employs over 3,000 people in 8 manufacturing plants in Europe and Asia. Enics provides development, support, prototyping, industrialization, manufacturing, product care, spare part production, and repair and maintenance services for the entire life-cycle of its customers’ products. Enics is ranked number 30 on Manufacturing Market Insider's list of the top 50 EMS providers in the world. and is considered one of the largest EMS providers in Europe. Enics has been a member of the United Nations Global Compact since December 2011. Enics offers the following services: Engineering (electronics design support, NPI, fast prototyping, design for manufacturing, test system development), manufacturing (PCB assembly, box build and system assembly, low-volume production, RoHS services), life extension (component and product care, test system maintenance), after sales (spare parts production, electronics repair, preventive maintenance, recycling services), and supply chain (sourcing, optimization of supply chain, delivery flexibility). Operations/Locations Enics has 7 production plants in: China (Beijing and Suzhou) Malaysia (Johor Bahru) Estonia (Elva) Finland (Lohja) Slovakia (Nová Dubnica) Sweden (Västerås) History Enics was formed in 2004 as a result of a management buyout of Elcoteq's Industrial Electronics Division. During that same year the company started a new production plant in Elva, located in Southern Estonia, where Enics is the largest employer. In 2005 Enics acquired Altronix, a part of ABB Västerås’ operations, 2 Flextronics plants in Sweden, Enermet's Jyskä's operations in Finland, and assets from Elcoteq in China. In 2006 Enics’ Beijing facility was officially inaugurated, the 3rd extension in Enics Estonia was completed, and the takeover of activities from ABB Switzerland was finished. Enics
https://en.wikipedia.org/wiki/Plumbr
Plumbr was an Estonian software product company founded in late 2011 that developed performance monitoring software. The Plumbr product was built on top of a proprietary algorithm that automatically detected the root causes of performance issues by interpreting application performance data. In October 2020, Plumbr was acquired by Splunk. Products Plumbr monitored customers' JVM applications for memory leaks, garbage collection pauses and locked threads. Plumbr problem detection algorithms were based on analysis of performance data of thousands of applications. Plumbr consisted of an agent and a portal. Plumbr Agent was attached to application runtime and sent memory usage and garbage collection information to Plumbr Portal. On Plumbr Portal one could see information such as heap and permgen memory usage, garbage collection pauses' and lock contention duration. Clients that were not able to send data to third parties could order a self-hosted portal and have a full solution in-house. In case of performance incidents Plumbr provided its users with information on problem severity and problem's root cause location in source code or runtime configuration, and listed the steps needed to take to remediate the problem. Clients included NASA, NATO, Dell, HBO, Experian, EMC Corporation. References External links Software companies of Estonia Programming tools Java development tools Java virtual machine Software performance management Profilers Software companies established in 2011
https://en.wikipedia.org/wiki/Operation%20Onymous
Operation Onymous was an international law enforcement operation targeting darknet markets and other hidden services operating on the Tor network. Background Operation Onymous was formed as a joint law enforcement operation between the Federal Bureau of Investigation (FBI) and the European Union Intelligence Agency Europol. The international effort also included the United States Department of Homeland Security, Immigration and Customs Enforcement (ICE), and Eurojust. The operation was part of the international strategies that address the problems of malware, botnet schemes, and illicit markets or darknets. It was also linked with the war on drugs effort with the participation of the U.S. Drug Enforcement Administration (DEA). Raids On 5 and 6 November 2014, a number of websites, initially claimed to be over 400, were shut down including drug markets such as Silk Road 2.0, Cloud 9 and Hydra. Other sites targeted included money laundering sites and "contraband sites". The operation involved the police forces of 17 countries. In total there were 17 arrests. A 26-year-old software developer was arrested in San Francisco and accused of running Silk Road 2.0 under the pseudonym 'Defcon'. Defcon was "one of the primary targets". Within hours of the seizure a third incarnation of the site appeared, 'Silk Road 3.0'; Silk Road had previously been seized in October 2013, and then resurrected, weeks later, as 'Silk Road 2.0'. $1 million in Bitcoin was seized, along with €180,000 in cash, gold, silver and drugs. Of the "illicit services" that were initially claimed to have been shut down, few were online marketplaces like Silk Road. A complaint filed on 7 November 2014 in the United States District Court for the Southern District of New York, "seeking the forfeiture of any and all assets of the following dark market websites operating on the Tor network", referred to just 27 sites, fourteen of which were claimed to be drug markets; the others allegedly sold counterfeit c
https://en.wikipedia.org/wiki/Texas%20Instruments%20Professional%20Computer
The Texas Instruments Professional Computer (abbreviated TIPC or TI PC) and the Texas Instruments Professional Portable Computer (TIPPC) are personal computers produced by Texas Instruments that were both released on January 31, 1983, and discontinued around 1985; the TIPC is a desktop PC and the TIPPC is a portable version that is fully compatible with it. Both computers were most often used by white-collar information workers and professionals that needed to gather, manipulate and transmit information. Features and specifications The TIPC is very similar to the IBM PC both architecturally and from a user-experience perspective, with some technically superior aspects. It is based on the Intel 8088 CPU and an optional Intel 8087 floating point coprocessor. It supports MS-DOS compatible operating systems, but is not a fully IBM PC compatible computer. Alternative operating systems are CP/M-86, Concurrent CP/M-86, and the UCSD p-System. The TIPC was reviewed in Byte magazine in its December 1983 issue. The CPU clocked at 5 MHz (a bit faster than the 4.77 MHz of the IBM PC) and has 64 KB of RAM pre-installed. A RAM board can be installed in an expansion slot providing an additional 192 KB or RAM, for a maximum of 256 KB. A later version supports up to 768 KB of total memory. The computer featured 5 expansion slots and has either a 12-inch green-phosphor monochrome (CRT) monitor or a 12-inch color monitor with a color graphics resolution of 720x300 pixels. For text, the display shows 25 lines of 80 columns each. The device has a 5¼-inch floppy disk drive and can support a second floppy drive or a "Winchester" hard drive without requiring the use of an expansion slot or separate chassis, and typically features one of each. Byte said the keyboard, which has a different layout for the arrow keys and is quieter than the IBM PC, was "well, wonderful". The keyboard has 57 typewriter keys, 5 cursor control keys, 12 function keys and a separate 18-key numeric keypad area. Th
https://en.wikipedia.org/wiki/Richard%20Zach
Richard Zach is a Canadian logician, philosopher of mathematics, and historian of logic and analytic philosophy. He is currently Professor of Philosophy at the University of Calgary. Research Zach's research interests include the development of formal logic and historical figures (Hilbert, Gödel, and Carnap) associated with this development. In the philosophy of mathematics Zach has worked on Hilbert's program and the philosophical relevance of proof theory. In mathematical logic, he has made contributions to proof theory (epsilon calculus, proof complexity) and to modal and many-valued logic, especially Gödel logic. Career Zach received his undergraduate education at the Vienna University of Technology and his Ph.D. at the Group in Logic and the Methodology of Science at the University of California, Berkeley. His dissertation, Hilbert's Program: Historical, Philosophical, and Metamathematical Perspectives, was jointly supervised by Paolo Mancosu and Jack Silver. He has taught at the University of Calgary since 2001, and holds the rank of Professor. He has held visiting appointments at the University of California, Irvine and McGill University. Zach is a founding editor of the Review of Symbolic Logic and the Journal for the Study of the History of Analytic Philosophy, and is also associate editor of Studia Logica, and a subject editor for the Stanford Encyclopedia of Philosophy (History of Modern Logic). He serves on the editorial boards of the Bernays edition and the Carnap edition. He was elected to the Council of the Association for Symbolic Logic in 2008 (ASL) and he has served on the ASL Committee on Logic Education and the executive committee of the Kurt Gödel Society. References External links Official Website LogBlog: A Logic Blog Departmental information page Society for the Study of the History of Analytical Philosophy Open Logic Project Year of birth missing (living people) Living people Canadian philosophers Academic staff of the Universit
https://en.wikipedia.org/wiki/YubiKey
The YubiKey is a hardware authentication device manufactured by Yubico to protect access to computers, networks, and online services that supports one-time passwords (OTP), public-key cryptography, and authentication, and the Universal 2nd Factor (U2F) and FIDO2 protocols developed by the FIDO Alliance. It allows users to securely log into their accounts by emitting one-time passwords or using a FIDO-based public/private key pair generated by the device. YubiKey also allows for storing static passwords for use at sites that do not support one-time passwords. Google, Amazon, Microsoft, Twitter, and Facebook use YubiKey devices to secure employee accounts as well as end user accounts. Some password managers support YubiKey. Yubico also manufactures the Security Key, a similar lower cost device with only FIDO2/WebAuthn and FIDO/U2F support. The YubiKey implements the HMAC-based One-time Password Algorithm (HOTP) and the Time-based One-time Password Algorithm (TOTP), and identifies itself as a keyboard that delivers the one-time password over the USB HID protocol. A YubiKey can also present itself as an OpenPGP card using 1024, 2048, 3072 and 4096-bit RSA (for key sizes over 2048 bits, GnuPG version 2.0 or higher is required) and elliptic curve cryptography (ECC) p256, p384 and more depending on version, allowing users to sign, encrypt and decrypt messages without exposing the private keys to the outside world. Also supported is the PKCS#11 standard to emulate a PIV smart card. This feature allows for code signing of Docker images as well as certificate-based authentication for Microsoft Active Directory and SSH. Founded in 2007 by CEO Stina Ehrensvärd, Yubico is a private company with offices in Palo Alto, Seattle, and Stockholm. Yubico CTO, Jakob Ehrensvärd, is the lead author of the original strong authentication specification that became known as Universal 2nd Factor (U2F). YubiKey released the YubiKey 5 series in 2018 which adds support for FIDO2. History Yubi
https://en.wikipedia.org/wiki/TDLS
TDLS, shortened from Tunneled Direct Link Setup, is "a seamless way to stream media and other data faster between devices already on the same Wi-Fi network." Devices using it communicate directly with one another, without involving the wireless network's router. Wi-Fi Alliance added certification for TDLS in 2012. It describes this feature as technology that enables devices to link directly to one another when connected to a traditional Wi-Fi network. Wi-Fi CERTIFIED TDLS devices can set up secure links and transfer data directly between them. TDLS-linked devices benefit from an optimized connection to do things such as streaming video or synching content, without burdening the network as a whole. The IEEE has endorsed this as the IEEE 802.11z standard. Google's Cast protocol used by Chromecast utilizes TDLS to initiate screen mirroring. See also Wi-Fi Direct References Wi-Fi
https://en.wikipedia.org/wiki/UPower
UPower (previously DeviceKit-power) is a piece of middleware (an abstraction layer) for power management on Linux systems. It enumerates power sources, maintains statistics and history data on them and notifies about status changes. It consists of a daemon (upowerd), an application programming interface and a set of command line tools. The daemon provides its functionality to applications over the system bus (an instance of D-Bus, service org.freedesktop.UPower). PolicyKit restricts access to the UPower functionality for initiating hibernate mode or shutting down the operating system (freedesktop.upower.policy). The command-line client program upower can be used to query and monitor information about the power supply devices in the system. Graphical user interfaces to the functionality of UPower include the GNOME Power Manager and the Xfce Power Manager. UPower is a product of the cross-desktop freedesktop.org project. As free software it is published with its source code under the terms of version 2 or later of the GNU General Public License (GPL). It was conceived as a replacement for the corresponding features of the deprecated HAL. In 2008, David Zeuthen started a comprehensive rewrite of HAL. This resulted in a set of separate services under the new name "DeviceKit". In 2010 the included DeviceKit-power was renamed. UPower was first introduced and established as a standard in GNOME. In January 2011 the desktop environment Xfce followed (version 4.8). Sources External links Red Hat, Inc.: Red Hat Enterprise Linux 7 – Power Management Guide, sections 2.6.: UPower, 2.7.: GNOME Power Manager Servers (computing) Free system software Freedesktop.org
https://en.wikipedia.org/wiki/Rimose
Rimose is an adjective used to describe a surface that is cracked or fissured. The term is often used in describing crustose lichens. A rimose surface of a lichen is sometimes contrasted to the surface being areolate. Areolate is an extreme form of being rimose, where the cracks or fissures are so deep that they create island-like pieces called areoles, which look the "islands" of mud on the surface of a dry lake bed. Rimose and areolate are contrasted with being verrucose, or "warty". Verrucose surfaces have warty bumps which are distinct, but not separated by cracks. In mycology the term describes mushrooms whose caps crack in a radial pattern, as commonly found in the genera Inocybe and Inosperma. References Biology terminology English words Lichens
https://en.wikipedia.org/wiki/Contraves%20Cora
The Cora was a digital fire-control system designed by Hungarian-Swiss Peter Tóth and produced by the Swiss company Contraves. Development Peter Tóth started the design of Cora-1 in 1957. The system was intended for anti-aircraft fire direction with the Swiss Army. However, Cora-1 turned out to be too slow and too bulky for this application. It was programmed for other applications by programmers including Heinz Lienhard. One copy of the system was used at the École Polytechnique Fédérale de Lausanne (EPFL) for cartography, and was put on display during Expo 64. The unit was rediscovered in storage in 2011, and is now on display at the Musée Bolo, in the Computer Science department of the EPFL. Cora-1-was one of the first fully transistorized digital computers built in Switzerland according to the Von Neumann architecture. An improved version of this computer was developed subsequently by a team led by Swiss engineer Peter Blum. Cora-2 was successfully used for anti-aircraft fire direction being compact enough to fit into the corresponding mobile control unit. The Musée Bolo met with Peter Tóth and released several videos around his work on the Cora. References External links Discovery of Two Historical Computers in Switzerland: Zuse Machine M9 and Contraves Cora and Discovery of Unknown Documents on the Early History of Computing at the ETH Archives, Making the History of Computing Relevant , Springer 2013 Schweizer Transistorrechner für militärische und zivile Zwecke, IT Magazine 2012/01 Military computers Artillery components Artillery operation Applications of control engineering Ballistics Military electronics Computer-related introductions in 1978
https://en.wikipedia.org/wiki/Reinsurance%20Actuarial%20Premium
Actuarial reinsurance premium calculation uses the similar mathematical tools as actuarial insurance premium. Nevertheless, Catastrophe modeling, Systematic risk or risk aggregation statistics tools are more important. Burning cost Typically burning cost is the estimated cost of claims in the forthcoming insurance period, calculated from previous years' experience adjusted for changes in the numbers insured, the nature of cover and medical inflation. Historical (aggregate) data extraction Adjustments to obtain 'as if' data: present value adjustment using actuarial rate, prices index,... base insurance premium correction, underwriting policy evolution, clauses application 'as if' data, calcul of the 'as if' historical reinsurance indemnity, Reinsurance pure premium rate computing, add charges, taxes and reduction of treaty "As if" data involves the recalculation of prior years of loss experience to demonstrate what the underwriting results of a particular program would have been if the proposed program had been in force during that period. Probabilist methods Premium formulation Let us note the and the deductible of XS or XL, with the limite ( XS ). The premium : where XS or XL premium formulation with Pareto If and : $ if and il n'y a pas de solution. If and : If and : XS premium using Lognormal cost distribution If follows then follows Then: With deductible and without limit : Monte Carlo estimation Vulnerability curve Regression estimation This method uses data along the x-y axis to compute fitted values. It is actually based on the equation for a straight line, y=bx+a.(2) Includes reinsurances specificities Clauses Long-Term Indemnity Claims Actuarial reserves modellisation. See also Reinsurance Insurance Actuarial Science Ruin Theory References 2. [2] http://www.r-tutor.com/elementary-statistics/simple-linear-regression/estimated-simple-regression-equation Actuarial science
https://en.wikipedia.org/wiki/Induced-charge%20electrokinetics
Induced-charge electrokinetics in physics is the electrically driven fluid flow and particle motion in a liquid electrolyte. Consider a metal particle (which is neutrally charged but electrically conducting) in contact with an aqueous solution in a chamber/channel. If different voltages apply to the end of this chamber/channel, electric field will generate in this chamber/channel. This applied electric field passes through this metal particle and causes the free charges inside the particle migrate under the skin of particle. As a result of this migration, the negative charges move to the side which is close to the positive (or higher) voltage while the positive charges move to the opposite side of the particle. These charges under the skin of the conducting particle attract the counter-ions of the aqueous solution; thus, the electric double layer (EDL) forms around the particle. The EDL sign on the surface of the conducting particle changes from positive to negative and the distribution of the charges varies along the particle geometry. Due to these variations, the EDL is non-uniform and has different signs. Thus, the induced zeta potential around the particle, and consequently slip velocity on the surface of the particle, vary as a function of the local electric field. Differences in magnitude and direction of slip velocity on the surface of the conducting particle effects the flow pattern around this particle and causes micro vortices. Yasaman Daghighi and Dongqing Li, for the first time, experimentally illustrated these induced vortices around a 1.2mm diameter carbon-steel sphere under the 40V/cm direct current (DC) external electric filed. Chenhui Peng et al. also experimentally showed the patterns of electro-osmotic flow around an Au sphere when alternating current (AC) is involved (E=10mV/μm, f=1 kHz). Electrokinetics here refers to a branch of science related to the motion and reaction of charged particles to the applied electric filed and its effects on its
https://en.wikipedia.org/wiki/Chaos%20engineering
Chaos engineering is the discipline of experimenting on a system in order to build confidence in the system's capability to withstand turbulent conditions in production. Concept In software development, a given software system's ability to tolerate failures while still ensuring adequate quality of service—often generalized as resilience—is typically specified as a requirement. However, development teams often fail to meet this requirement due to factors such as short deadlines or lack of knowledge of the field. Chaos engineering is a technique to meet the resilience requirement. Chaos engineering can be used to achieve resilience against infrastructure failures, network failures, and application failures. Operational readiness using chaos engineering Calculating how much confidence we can have in the interconnected complex systems that are put into production environment requires operational readiness metrics. Operational readiness can be evaluated using chaos engineering simulations supported by Kubernetes infrastructure in big data. Solutions for operational readiness of a platform stands for strengthening the backup, restore, network file transfer, failover capabilities and overall security. Gautam Siwach et al, performed evaluation of inducing chaos to a Kubernetes environment which terminates random pods with data from edge devices in data centers while processing analytics on big data network and infer the recovery time of pods to calculate an estimated response time. History 1983 – Apple While MacWrite and MacPaint were being developed for the first Apple Macintosh computer, Steve Capps created "Monkey", a desk accessory which randomly generated user interface events at high speed, simulating a monkey frantically banging the keyboard and moving and clicking the mouse. It was promptly put to use for debugging by generating errors for programmers to fix, because automated testing was not possible; the first Macintosh had too little free memory space for
https://en.wikipedia.org/wiki/MXenes
In materials science, MXenes are a class of two-dimensional inorganic compounds , that consist of atomically thin layers of transition metal carbides, nitrides, or carbonitrides. MXenes accept a variety of hydrophilic terminations. MXenes were first reported in 2012. Structure As-synthesized MXenes prepared via HF etching have an accordion-like morphology, which can be referred to as multi-layer MXene (ML-MXene), or few-layer MXene (FL-MXene) given fewer than five layers. Because the surfaces of MXenes can be terminated by functional groups, the naming convention Mn+1XnTx can be used, where T is a functional group (e.g. O, F, OH, Cl). Mono transition MXenes adopt three structures with one metal on the M site, as inherited from the parent MAX phases: M2C, M3C2, and M4C3. They are produced by selectively etching out the A element from a MAX phase or other layered precursor (e.g., Mo2Ga2C), which has the general formula Mn+1AXn, where M is an early transition metal, A is an element from group 13 or 14 of the periodic table, X is C and/or N, and n = 1–4. MAX phases have a layered hexagonal structure with P63/mmc symmetry, where M layers are nearly closed packed and X atoms fill octahedral sites. Therefore, Mn+1Xn layers are interleaved with the A element, which is metallically bonded to the M element. Double transition Double transition metal MXenes can take two forms, ordered double transition metal MXenes or solid solution MXenes. For ordered double transition metal MXenes, they have the general formulas: M'2M"C2 or M'2M"2C3 where M' and M" are different transition metals. Double transition metal carbides that have been synthesized include Mo2TiC2, Mo2Ti2C3, Cr2TiC2, and Mo4VC4. In some of these MXenes (such as Mo2TiC2, Mo2Ti2C3, and Cr2TiC2), the Mo or Cr atoms are on outer edges of the MXene and these atoms control electrochemical properties of the MXenes. For solid-solution MXenes, they have the general formulas: (M'2−yM"y)C, (M'3−yM"y)C2, (M'4−yM"y)C3, o
https://en.wikipedia.org/wiki/Flow%20process
The region of space enclosed by open system boundaries is usually called a control volume. It may or may not correspond to physical walls. It is convenient to define the shape of the control volume so that all flow of matter, in or out, occurs perpendicular to its surface. One may consider a process in which the matter flowing into and out of the system is chemically homogeneous. Then the inflowing matter performs work as if it were driving a piston of fluid into the system. Also, the system performs work as if it were driving out a piston of fluid. Through the system walls that do not pass matter, heat () and work () transfers may be defined, including shaft work. Classical thermodynamics considers processes for a system that is initially and finally in its own internal state of thermodynamic equilibrium, with no flow. This is feasible also under some restrictions, if the system is a mass of fluid flowing at a uniform rate. Then for many purposes a process, called a flow process, may be considered in accord with classical thermodynamics as if the classical rule of no flow were effective. For the present introductory account, it is supposed that the kinetic energy of flow, and the potential energy of elevation in the gravity field, do not change, and that the walls, other than the matter inlet and outlet, are rigid and motionless. Under these conditions, the first law of thermodynamics for a flow process states: the increase in the internal energy of a system is equal to the amount of energy added to the system by matter flowing in and by heating, minus the amount lost by matter flowing out and in the form of work done by the system. Under these conditions, the first law for a flow process is written: where and respectively denote the average internal energy entering and leaving the system with the flowing matter. There are then two types of work performed: 'flow work' described above, which is performed on the fluid in the control volume (this is also often
https://en.wikipedia.org/wiki/Critical%20process%20parameters
Critical process parameters (CPP) in pharmaceutical manufacturing are key variables affecting the production process. CPPs are attributes that are monitored to detect deviations in standardized production operations and product output quality or changes in critical quality attributes. Those attributes with a higher impact on CQAs should be prioritized and held in a stricter state of control. The manufacturer should conduct tests to set acceptable range limits of the determined CPPs and define acceptable process variable variability. Operational conditions within this range are considered acceptable operational standards. Any deviation from the acceptable range will be indicative of issues within the process and the subsequent production of substandard products. Data relating to CPP should be recorded, stored, and analyzed by the manufacturer. CPP variables and ranges should be reevaluated after careful analysis of historical CPP data. Identifying CPPs is done in stage one of process validation: process design are an essential part of a manufacturing control strategy. One method of defining CPPs is to look at the effect of certain production processes on critical quality attributes. Those production parameters which have a measurable effect on those quality attributes that have been identified as critical can be considered CPPs and must always be in a state of control. References Formal methods Enterprise modelling Quality management Business process Pharmaceutical industry
https://en.wikipedia.org/wiki/Kaplan%E2%80%93Yorke%20conjecture
In applied mathematics, the Kaplan–Yorke conjecture concerns the dimension of an attractor, using Lyapunov exponents. By arranging the Lyapunov exponents in order from largest to smallest , let j be the largest index for which and Then the conjecture is that the dimension of the attractor is This idea is used for the definition of the Lyapunov dimension. Examples Especially for chaotic systems, the Kaplan–Yorke conjecture is a useful tool in order to estimate the fractal dimension and the Hausdorff dimension of the corresponding attractor. The Hénon map with parameters a = 1.4 and b = 0.3 has the ordered Lyapunov exponents and . In this case, we find j = 1 and the dimension formula reduces to The Lorenz system shows chaotic behavior at the parameter values , and . The resulting Lyapunov exponents are {2.16, 0.00, −32.4}. Noting that j = 2, we find References Dimension Dynamical systems Limit sets Conjectures
https://en.wikipedia.org/wiki/Fredrick%20Brennan
Fredrick Robert Brennan (born February 21, 1994) is an American software developer and type designer who founded the imageboard website 8chan in 2013, before going on to repudiate it in 2019. Following 8chan's surge in popularity in 2014, largely due to many Gamergate proponents migrating to the site from 4chan, Brennan moved to the Philippines to work for Jim Watkins, who provided hosting services to 8chan and later became the site's owner. Brennan cut ties with 8chan in 2016 and with Watkins in 2018. Brennan has since become an outspoken critic of both 8chan and Watkins, and has actively battled to try to take 8chan offline. Brennan is also an outspoken critic of the QAnon conspiracy theory, which centers on 8chan posts by an anonymous figure named "Q", and Brennan has researched who may be behind the Q identity. Personal life Fredrick Brennan was born on February 21, 1994, in Albany, New York. He was born with osteogenesis imperfecta, commonly known as brittle bone disease, which stunted his growth and requires him to use a wheelchair. He estimates he had broken bones 120 times by the time he was 19. Brennan comes from a multi-generational family with osteogenesis imperfecta, and his mother has the same condition. His parents divorced when Brennan was five years old. He and his siblings were in his father's custody until Brennan was 14, when they were placed into the New York State foster care system, where Brennan remained until he was 16 while his mother went through the legal proceedings to regain custody. He then lived with his mother until he was 18. After graduating high school, he chose not to attend college. In January 2014, Brennan was robbed of almost $5,000 ($ in dollars) that he was saving up for a new wheelchair. When the suspect was arrested, Brennan went to the police station for a lineup, but when a bus did not come on the way back, he was left stranded in the snow and later had to be treated for hypothermia. He received a personal apology fr
https://en.wikipedia.org/wiki/K%C3%A4ll%C3%A9n%20function
The Källén function, also known as triangle function, is a polynomial function in three variables, which appears in geometry and particle physics. In the latter field it is usually denoted by the symbol . It is named after the theoretical physicist Gunnar Källén, who introduced it as a short-hand in his textbook Elementary Particle Physics. Definition The function is given by a quadratic polynomial in three variables Applications In geometry the function describes the area of a triangle with side lengths : See also Heron's formula. The function appears naturally in the kinematics of relativistic particles, e.g. when expressing the energy and momentum components in the center of mass frame by Mandelstam variables. Properties The function is (obviously) symmetric in permutations of its arguments, as well as independent of a common sign flip of its arguments: If the polynomial factorizes into two factors If the polynomial factorizes into four factors Its most condensed form is Interesting special cases are References Kinematics (particle physics) Mathematical concepts
https://en.wikipedia.org/wiki/15-Hydroxyeicosatetraenoic%20acid
15-Hydroxyeicosatetraenoic acid (also termed 15-HETE, 15(S)-HETE, and 15S-HETE) is an eicosanoid, i.e. a metabolite of arachidonic acid. Various cell types metabolize arachidonic acid to 15(S)-hydroperoxyeicosatetraenoic acid (15(S)-HpETE). This initial hydroperoxide product is extremely short-lived in cells: if not otherwise metabolized, it is rapidly reduced to 15(S)-HETE. Both of these metabolites, depending on the cell type which forms them, can be further metabolized to 15-oxo-eicosatetraenoic acid (15-oxo-ETE), 5(S),15(S)-dihydroxy-eicosatetraenoic acid (5(S),15(S)-diHETE), 5-oxo-15(S)-hydroxyeicosatetraenoic acid (5-oxo-15(S)-HETE), a subset of specialized pro-resolving mediators viz., the lipoxins, a class of pro-inflammatory mediators, the eoxins, and other products that have less well-defined activities and functions. Thus, 15(S)-HETE and 15(S)-HpETE, in addition to having intrinsic biological activities, are key precursors to numerous biologically active derivatives. Some cell types (e.g. platelets) metabolize arachidonic acid to the stereoisomer of 15(S)-HpETE, 15(R)-HpETE. Both stereoisomers may also be formed as result of the metabolism of arachidonic acid by cellular microsomes or as a result of arachidonic acid auto-oxidation. Similar to 15(S)-HpETEs, 15(R)-HpETE may be rapidly reduced to 15(R)-HETE. These R,S stereoisomers differ only in having their hydroxy residue in opposite orientations. While the two R stereoisomers are sometimes referred to as 15-HpETE and 15-HETE, proper usage should identify them as R stereoisomers. 15(R)-HpETE and 15(R)-HETE lack some of the activity attributed to their S stereoisomers but can be further metabolized to bioactive products viz., the 15(R) class of lipoxins (also termed epi-lipoxins). 15(S)-HETE, 15(S)-HpETE, and many of their derivative metabolites are thought to have physiologically important functions. They appear to act as hormone-like autocrine and paracrine signaling agents that are involved in regula
https://en.wikipedia.org/wiki/Pointr
Pointr is a startup company based in London specialized in indoor positioning and navigation utilising iBeacons, which are Bluetooth Low Energy devices formalised by Apple Inc. Pointr have created a GPS-like experience with true position and turn-by-turn navigation that is supported by most modern smartphones operating on both Android and iOS. Analytics and messaging modules can be added on to help communicate with users and understand venue usage respectively. The features are provided through a software package (SDK) which aims to improve user experience whilst connecting the online and offline worlds. Many of the features are available without an internet connection, including sending messages between users with a form of Mesh networking, however for intelligent offers and live analytics then an internet connection is required. The markets where the technology is most frequently used are retail, exhibition centres, airports and museums, but there are a number of uses in hospitals, warehouses, offices and entertainment venues as well. The majority of software development is done in their office in Istanbul, with specialist modules created in London. The technology is commonly used in permanent installations where the SDK is offered with a license fee model, however some installations have been temporary and hence one-off payments have been used. History Pointr was founded in November 2013 by Ege Akpinar under the name Indoorz; he was then joined by co-founders Axel Katalan, Chris Charles and Can Akpinar in early 2014. The software was developed for seven months before launching, allowing time to build and test the product. In November 2014 the company adopted its current name of Pointr after receiving a client question about whether it could work outdoors as well. Pointr raised its first round of angel funding in January 2015 and has grown steadily with its first customers in retail, warehouses, offices and libraries. In February 2015, Pointr was accepted
https://en.wikipedia.org/wiki/2D%20adaptive%20filters
A two-dimensional (2D) adaptive filter is very much like a one-dimensional adaptive filter in that it is a linear system whose parameters are adaptively updated throughout the process, according to some optimization approach. The main difference between 1D and 2D adaptive filters is that the former usually take as inputs signals with respect to time, what implies in causality constraints, while the latter handles signals with 2 dimensions, like x-y coordinates in the space domain, which are usually non-causal. Moreover, just like 1D filters, most 2D adaptive filters are digital filters, because of the complex and iterative nature of the algorithms. Motivation The topic of 2D adaptive filters is very important in electrical engineering and signal processing since these filters have the ability to take into account the nonstationary statistical properties of 2D signals. Adaptive filters find applications in areas such as Noise cancellation, Signal prediction, Equalization and Echo cancellation. Examples of applications of 2D adaptive filters include Image Denoising, Motion Tracking, OFDM channel estimation, magnetic recording equalization Example Application 2D Adaptive Filters can be used to identify systems. The system function of the unknown system is given by , and is the system function of the 2D adaptive filter when its output comes to steady. The error signal between the unknown system output,, and the adaptive filter output,, is minimized if the unknown system and known 2D adaptive filter have the same input, and if the resulting outputs are similar. Then, it can be shown that can be represented by . is known as the system identification model of the unknown system. Problem Statement In digital signal processing, any linear shift invariant system can be represented by the convolution of the signal with the filter's impulse response, given by the expression: If this system is to model a desired response , the adaptive system can be obtained by contin
https://en.wikipedia.org/wiki/Honeywell%20CP-6
CP-6 is a discontinued computer operating system, developed by Honeywell, Inc. in 1976, which was a backward-compatible work-alike of the Xerox CP-V fully rewritten for Honeywell Level/66 hardware. CP-6 was a command line oriented system. A terminal emulator allowed use of PCs as CP-6 terminals. History In 1975, Xerox decided to sell the computer business which it had purchased from Scientific Data Systems in 1969. In a deal put together by Harry Sweatt, Honeywell purchased Xerox Data Systems, and took on the Xerox sales and field computer support staff to provide field service support to the existing customer base. Xerox made available all the spare equipment and supplies and the warehouses containing them. Revenues were shared 60/40 Xerox until CP-6 General Release, and 60/40 Honeywell for three years thereafter. Following that, all revenue went to Honeywell. In the early 1960s, Honeywell had built and sold a large number of H200 machines, together with software. In 1970 it had bought the computer business of General Electric. LADC and the development of CP-6 In 1976, Honeywell began developing the CP-6 system, including its operating system and program products to attract Xerox CP-V users (about 750 Sigma users) to buy and use Honeywell equipment. Honeywell employed an initial team of 60 programmers from the Xerox CP-V development team, and added another 30 programmers plus management and staff. Organized by Hank Hoagland and Shel Klee, the team was housed at an old Xerox marketing office at 5250 W. Century Blvd in Los Angeles, which became known as the Los Angeles Development Center (LADC). The new operating system was to be called CP-6. LADC reported administratively to the Honeywell computer group in Phoenix, a facility, which Honeywell had acquired from General Electric. The first beta site was installed at Carleton University in Ottawa, Canada, in June 1979, and three other sites were installed before the end of 1979. Customers worked with LADC both di
https://en.wikipedia.org/wiki/Piggyback%20attack
A Piggyback attack is an active form of wiretapping where the attacker gains access to a system via intervals of inactivity in another user's legitimate connection. It is also called a “between the line attack” or "piggyback-entry wiretapping". In security, piggybacking refers to when someone tags along with another person who is authorized to gain entry into a restricted area the term is applied to computer networks in this context. References Surveillance Internet security
https://en.wikipedia.org/wiki/Symmetry%20%28geometry%29
In geometry, an object has symmetry if there is an operation or transformation (such as translation, scaling, rotation or reflection) that maps the figure/object onto itself (i.e., the object has an invariance under the transform). Thus, a symmetry can be thought of as an immunity to change. For instance, a circle rotated about its center will have the same shape and size as the original circle, as all points before and after the transform would be indistinguishable. A circle is thus said to be symmetric under rotation or to have rotational symmetry. If the isometry is the reflection of a plane figure about a line, then the figure is said to have reflectional symmetry or line symmetry; it is also possible for a figure/object to have more than one line of symmetry. The types of symmetries that are possible for a geometric object depend on the set of geometric transforms available, and on what object properties should remain unchanged after a transformation. Because the composition of two transforms is also a transform and every transform has, by definition, an inverse transform that undoes it, the set of transforms under which an object is symmetric form a mathematical group, the symmetry group of the object. Euclidean symmetries in general The most common group of transforms applied to objects are termed the Euclidean group of "isometries", which are distance-preserving transformations in space commonly referred to as two-dimensional or three-dimensional (i.e., in plane geometry or solid geometry Euclidean spaces). These isometries consist of reflections, rotations, translations, and combinations of these basic operations. Under an isometric transformation, a geometric object is said to be symmetric if, after transformation, the object is indistinguishable from the object before the transformation. A geometric object is typically symmetric only under a subset or "subgroup" of all isometries. The kinds of isometry subgroups are described below, followed by other ki
https://en.wikipedia.org/wiki/Metal%20profiles
Metal profile sheet systems are used to build cost efficient and reliable envelopes of mostly commercial buildings. They have evolved from the single skin metal cladding often associated with agricultural buildings to multi-layer systems for industrial and leisure application. As with most construction components, the ability of the cladding to satisfy its functional requirements is dependent on its correct specification and installation. Also important is its interaction with other elements of the building envelope and structure. Metal profile sheets are metal structural members that due to the fact they can have different profiles, with different heights and different thickness, engineers and architects can use them for a variety of buildings, from a simple industrial building to a high demand design building. Trapezoidal profiles are large metal structural members, which, thanks to the profiling and thickness, retain their high load bearing capability. They have been developed from the corrugated profile. The profile programme offered by specific manufacturers covers a total of approximately 60 profile shapes with different heights. Cassettes are components that are mainly used as the inner shell in dual-shell wall constructions. They are mainly used in walls today, even though they were originally designed for use in roofs. History Trapezoidal profiles and cassettes have been known in Europe for around 100 years. Today's characteristic profile shape came to Europe from the USA in the 50s and has gained relevance since about 1960. At present the proportion of load bearing, room sealing trapezoidal profiles used in the overall area of new and slightly sloping roofs amounts to 90%. Above all else the wide acceptance has resulted from the simple constructive training, fast assembly, and the low costs of the trapezoidal profile construction. Specification of the metal profile Weathertightness The primary function of the cladding system is to provide a weathert
https://en.wikipedia.org/wiki/First%20passage%20percolation
First passage percolation is a mathematical method used to describe the paths reachable in a random medium within a given amount of time. Introduction First passage percolation is one of the most classical areas of probability theory. It was first introduced by John Hammersley and Dominic Welsh in 1965 as a model of fluid flow in a porous media. It is part of percolation theory, and classical Bernoulli percolation can be viewed as a subset of first passage percolation. Most of the beauty of the model lies in its simple definition (as a random metric space) and the property that several of its fascinating conjectures do not require much effort to be stated. Most times, the goal of first passage percolation is to understand a random distance on a graph, where weights are assigned to edges. Most questions are tied to either find the path with the least weight between two points, known as a geodesic, or to understand how the random geometry behaves in large scales. Mathematics As is the case in percolation theory in general, many of the problems related to first passage percolation involve finding optimal routes or optimal times. The model is defined as follows. Let be a graph. We place a non-negative random variable , called the passage time of the edge , at each nearest-neighbor edge of the graph . The collection is usually assumed to be independent, identically distributed but there are variants of the model. The random variable is interpreted as the time or the cost needed to traverse edge . Since each edge in first passage percolation has its own individual weight (or time) we can write the total time of a path as the summation of weights of each edge in the path. Given two vertices of one then sets where the infimum is over all finite paths that start at and end at . The function induces a random pseudo-metric on . The most famous model of first passage percolation is on the lattice . One of its most notorious questions is "What does a ball o
https://en.wikipedia.org/wiki/Neochromosome
A neochromosome is a chromosome that is not normally found in nature. Cancer-associated neochromosomes are found in some cancer cells. Neochromosomes have also been created using genetic engineering techniques. Cancer-associated neochromosomes Cancer-associated neochromosomes are giant supernumerary chromosomes. They harbor the mutations that drive certain cancers (highly amplified copies of key oncogenes, such as MDM2, CDK4, HMGA2). They may be circular or linear chromosomes. They have functional centromeres, and telomeres when linear. They are rare overall, being found in about 3% of cancers, but are common in certain rare cancers. For example, they are found in 90% of parosteal osteosarcomas. Neochromosomes from well- and de-differentiated liposarcoma have been studied at high resolution by isolation (using flow sorting) and sequencing, as well as microscopy. They consist of hundreds of fragments of DNA, often derived from multiple normal chromosomes, stitched together randomly, and contain high levels of DNA amplification (~30-60 copies of some genes). Using statistical inference and mathematical modelling, the process of how neochromosomes initially form and evolve has been made clearer. Fragments of DNA produced following chromothriptic shattering of chromosome 12 undergo DNA repair to form of a circular or ring chromosome. This undergoes hundreds of circular breakage-fusion-bridge cycles, causing random amplification and deletion of DNA with selection for the amplification of key oncogenes. DNA from additional chromosomes is somehow added during this process. Erosion of centromeres can lead to the formation of neocentromeres or the capture of new native centromeres from other chromosomes. The process ends when the neochromosome forms a linear chromosome following the capture of telomeric caps, which can be chromothriptically derived. References Oncology Genetic engineering
https://en.wikipedia.org/wiki/River%20barrier%20hypothesis
The river barrier hypothesis is a hypothesis seeking to partially explain the high species diversity in the Amazon Basin, first presented by Alfred Russel Wallace in his 1852 paper On Monkeys of the Amazon. It argues that the formation and movement of the Amazon and some of its tributaries presented a significant enough barrier to movement for wildlife populations to precipitate allopatric speciation. Facing different selection pressures and genetic drift, the divided populations diverged into separate species. There are several observable qualities that should be present if speciation has resulted from a river barrier. Divergence of species on either side of the river should increase with the size of the river, expressing weakly or not at all in the headwaters and more strongly in the wider, deeper channels further downriver. Organisms endemic to terra firme forest should be more affected than those that live in alluvial forests alongside the river, as they have a longer distance to cross before reaching appropriate habitat and lowland populations can rejoin relatively frequently when a river shifts or narrows in the early stages of oxbow lake formation. Finally, if a river barrier is the cause of speciation, sister species should exist on opposing shores more frequently than expected by chance. Mechanisms River barrier speciation occurs when a river is of sufficient size to provide a vicariance for allopatric speciation, or when the river is large enough to prevent or interfere with a genetic exchange between populations. Population division is initiated either when a river shifts into or forms within the range of a species that cannot cross it, effectively splitting the population in half, or when a small founder group is transported across an existing river through random chance. Usually a river's strength as a barrier is viewed as proportional to its width; wider rivers present a longer crossing distance and thus a greater obstacle to movement. Barrier stren
https://en.wikipedia.org/wiki/Chv%C3%A1tal%E2%80%93Sankoff%20constants
In mathematics, the Chvátal–Sankoff constants are mathematical constants that describe the lengths of longest common subsequences of random strings. Although the existence of these constants has been proven, their exact values are unknown. They are named after Václav Chvátal and David Sankoff, who began investigating them in the mid-1970s. There is one Chvátal–Sankoff constant for each positive integer k, where k is the number of characters in the alphabet from which the random strings are drawn. The sequence of these numbers grows inversely proportionally to the square root of k. However, some authors write "the Chvátal–Sankoff constant" to refer to , the constant defined in this way for the binary alphabet. Background A common subsequence of two strings S and T is a string whose characters appear in the same order (not necessarily consecutively) both in S and in T. The problem of computing a longest common subsequence has been well studied in computer science. It can be solved in polynomial time by dynamic programming; this basic algorithm has additional speedups for small alphabets (the Method of Four Russians), for strings with few differences, for strings with few matching pairs of characters, etc. This problem and its generalizations to more complex forms of edit distance have important applications in areas that include bioinformatics (in the comparison of DNA and protein sequences and the reconstruction of evolutionary trees), geology (in stratigraphy), and computer science (in data comparison and revision control). One motivation for studying the longest common subsequences of random strings, given already by Chvátal and Sankoff, is to calibrate the computations of longest common subsequences on strings that are not random. If such a computation returns a subsequence that is significantly longer than what would be obtained at random, one might infer from this result that the match is meaningful or significant. Definition and existence The Chvátal–Sanko
https://en.wikipedia.org/wiki/Loggly
SolarWinds Loggly is a cloud-based log management and analytics service provider based in San Francisco, California. Jon Gifford, Raffael Marty, and Kord Campbell founded the company in 2009, and Charlie Oppenheimer was the CEO of Loggly until its announced acquisition by SolarWinds (as part of the SolarWinds Cloud division of brands) on January 8, 2018. History In 2009, Jon Gifford, Raffael Marty, and Kord Campbell founded Loggly. App47, a mobile application management provider, partnered with Loggly in September 2012. The company chose Loggly because of its software-as-a-service (SaaS) deployment option. In September 2013, Loggly released "Generation 2", an updated version of its service. The update included log collection through standard syslog protocols and a graphical web interface that allowed users to use a point-and-click process to find log events and generate charts. That month, Loggly completed a $10.5 million funding round led by Cisco and Data Collective. Trinity Ventures, True Ventures and Matrix Partners also participated in the round. In October 2014, the company announced a $15 million series C funding round led by Harmony Partners. Matrix Partners, Trinity Ventures, Cisco, Data Collective, and True Ventures also participated. The funding round raised Loggly's total investment funding to $33.4 million. The company released Loggly Dynamic Field Explorer, a new user experience that aims to reduce the time developers spend on identifying and troubleshooting problems, that month. On January 8, 2018, the company announced that they are now part of SolarWinds. Operations Loggly is headquartered in San Francisco, California. The company had 54 employees and 10,000 customers in October 2017. Loggly records log data from any device and reports it in a real-time management platform with trend data. Technology Loggly is a cloud-based log management service provider. It does not require the use of proprietary software agents to collect log data. The serv
https://en.wikipedia.org/wiki/Ruselectronics
JSC Ruselectronics (), is a Russian state-owned holding company founded in 1997. It is fully owned by Rostec. Ruselectronics is responsible for the production of approximately 80 percent of all Russian electronics components. History Ruselectronics integrates the electronics sector companies focused on designing and producing electronic materials, equipment, semiconductor products and microwave technologies. The Holding company was established in the beginning of 2009 on the basis of the holding that was created in 1997. At the end of 2012 the Supervisory Board of the Corporation decided to integrate Sirius and Orion groups of companies into the Ruselectronics Holding. In December 2012, Rostec’s Supervisory Board transferred the assets of Sirius and Orion to Russian Electronics. Orion was founded in 2009 as a special-purpose research and industrial association to develop communication systems, subsystems and equipment for defence, special and double purposes. Its companies were located in six federal regions. Orion employed 11,000 people. It controlled 17 organizations, including JSC Omsk Research Institute for Instrument Engineering, JSC Barnaul Special Vostok Construction Bureau, and JSC Integral Research Institute for Special Communication Systems. Sirius was established in 2009. Its key products include customized and replicated software of various uses, television equipment and devices for television reception, in particular, in extreme conditions (space environment, ultrahigh temperature environment and other hostile environments), automated control system elements, hardware and software for automated special-purpose systems, comprehensive security systems for critical facilities, territories and passenger transportation systems and telecommunication equipment. The company included over 20 enterprises, such as Internavigation Research Center for Advanced Navigation Technologies; JSC Radiozavod; FGUP Solid-State Engineering Construction Bureau; JSC Solnech
https://en.wikipedia.org/wiki/Example-centric%20programming
Example-centric programming is an approach to software development that helps the user to create software by locating and modifying small examples into a larger whole. That approach can be helped by tools that allow an integrated development environment (IDE) to show code examples or API documentation related to coding behaviors occurring in the IDE. “Borrow” tactics are often employed from online sources, by programmers leaving the IDE to troubleshoot. The purpose of example-centric programming is to reduce the time spent by developers searching online. Ideally, in example-centric programming, the user interface integrates with help module examples for assistance without programmers leaving the IDE. The idea for this type of “instant documentation” is to reduce programming interruptions. The usage of this feature is not limited to experts, as some novices reap the benefits of an integrated knowledge base, without resorting to frequent web searches or browsing. Background The growth of the web has fundamentally changed the way software is built. Vast increase in information resources and the democratization of access and distribution are main factors in the development of example-centric programming for end-user development. Tutorials are available on the web in seconds thus broadening the space of who writes it: designers, scientists, or hobbyists. By 2012 13 million program as a part of their job, yet only three million of those are actual professional programmers. Prevalence of online code repositories, documentation, blogs and forums—enables programmers to build applications iteratively searching for, modifying, and combining examples. Using the web is integral to an opportunistic approach to programming when focusing on speed and ease of development over code robustness and maintainability. There is a widespread use of the web by programmers, novices and experts alike, to prototype, ideate, and discover. To develop software quickly programmers often mash up
https://en.wikipedia.org/wiki/Emerald%20Program
The SNIA Emerald Program Power Efficiency Measurement Specification, is a storage specification developed and maintained by the Storage Networking Industry Association (SNIA) and cross-referenced by the Environmental Protection Agency’s EnergyStar program. The specification consists of a storage types taxonomy, system under test workload and energy measurement method, measured metrics for active and idle operational states, and presence tests for capacity optimization technologies. The measured metric data is generated through the use of well-defined standard testing and data reduction procedures prescribed in the SNIA Emerald Specification. SNIA's ongoing collaboration with the EPA has helped to shape the Energy Star Data Center Storage (DCS) Specification. The EPA DCS specification cross-references the SNIA Emerald Specification as the test and measurement methodology. References Links SNIA Emerald™ Program homepage provides information on power efficiency, product measurement results, training, etc. for storage systems Welcome to SNIA Emerald | SNIA SNIA Green Storage Initiative homepage provides information about energy efficiency and conservations for networked storage technologies www.snia.org/gsi European Code of Conduct for Energy Efficiency in Data Centre Code of Conduct for Energy Efficiency in Data Centres The Green Grid homepage – energy efficient IT The Green Grid 80 PLUS homepage and information about power supply energy efficiency CLEAResult Plug Load Solutions Transaction Processing Performance Council – includes power efficiency data TPC-Homepage V5 Storage Performance Council – includes Energy Extension Standard Performance Evaluation Corporation – benchmarks for servers and other computer systems that include power efficiency Third Party References to SNIA Emerald EnergyStar references to SNIA Emerald Specification in Table 4 Data Center Storage Key Product Criteria Industry report reference to SNIA Emerald on Page 21 Lot 9 study fo
https://en.wikipedia.org/wiki/Multidimensional%20spectral%20estimation
Multidimension spectral estimation is a generalization of spectral estimation, normally formulated for one-dimensional signals, to multidimensional signals or multivariate data, such as wave vectors. Motivation Multidimensional spectral estimation has gained popularity because of its application in fields like medicine, aerospace, sonar, radar, bio informatics and geophysics. In the recent past, a number of methods have been suggested to design models with finite parameters to estimate the power spectrum of multidimensional signals. In this article, we will be looking into the basics of methods used to estimate the power spectrum of multidimensional signals. Applications There are many applications of spectral estimation of multi-D signals such as classification of signals as low pass, high pass, pass band and stop band. It is also used in compression and coding of audio and video signals, beam forming and direction finding in radars, Seismic data estimation and processing, array of sensors and antennas and vibrational analysis. In the field of radio astronomy, it is used to synchronize the outputs of an array of telescopes. Basic Concepts In a single dimensional case, a signal is characterized by an amplitude and a time scale. The basic concepts involved in spectral estimation include autocorrelation, multi-D Fourier transform, mean square error and entropy. When it comes to multidimensional signals, there are two main approaches: use a bank of filters or estimate the parameters of the random process in order to estimate the power spectrum. Methods Classical Estimation Theory It is a technique to estimate the power spectrum of a single dimensional or a multidimensional signal as it cannot be calculated accurately. Given are samples of a wide sense stationary random process and its second order statistics (measurements).The estimates are obtained by applying a multidimensional Fourier transform of the autocorrelation function of the random signal. The estimati
https://en.wikipedia.org/wiki/ISO/IEC%2020248
ISO/IEC 20248 Automatic Identification and Data Capture Techniques – Data Structures – Digital Signature Meta Structure is an international standard specification under development by ISO/IEC JTC 1/SC 31/WG 2. This development is an extension of SANS 1368, which is the current published specification. ISO/IEC 20248 and SANS 1368 are equivalent standard specifications. SANS 1368 is a South African national standard developed by the South African Bureau of Standards. ISO/IEC 20248 [and SANS 1368] specifies a method whereby data stored within a barcode and/or RFID tag is structured and digitally signed. The purpose of the standard is to provide an open and interoperable method, between services and data carriers, to verify data originality and data integrity in an offline use case. The ISO/IEC 20248 data structure is also called a "DigSig" which refers to a small, in bit count, digital signature. ISO/IEC 20248 also provides an effective and interoperable method to exchange data messages in the Internet of Things [IoT] and machine to machine [M2M] services allowing intelligent agents in such services to authenticate data messages and detect data tampering. Description ISO/IEC 20248 can be viewed as an X.509 application specification similar to S/MIME. Classic digital signatures are typically too big (the digital signature size is typically more than 2k bits) to fit in barcodes and RFID tags while maintaining the desired read performance. ISO/IEC 20248 digital signatures, including the data, are typically smaller than 512 bits. X.509 digital certificates within a public key infrastructure (PKI) is used for key and data description distribution. This method ensures the open verifiable decoding of data stored in a barcode and/or RFID tag into a tagged data structure; for example JSON and XML. ISO/IEC 20248 addresses the need to verify the integrity of physical documents and objects. The standard counters verification costs of online services and device to serve
https://en.wikipedia.org/wiki/Manta%20Matcher
Manta Matcher is a global online database for manta rays. Creation It is one of the Wildbook Web applications developed by Wild Me, a 501(c)(3) not-for-profit organization in the United States, and was created in partnership with Andrea Marshall of the Marine Megafauna Foundation. Manta rays have unique spot patterning on their undersides, which allows for individual identification. Scuba divers around the world can photograph mantas and upload their manta identification photographs to the Manta Matcher website, supporting global research and conservation efforts. Identification of rays Manta Matcher is a pattern-matching software that eases researcher workload; key spot pattern features are extracted using a scale-invariant feature transform (SIFT) algorithm, which can cope with complications presented by highly variable spot patterns and low contrast photographs. Purpose and research supported This citizen science tool is free to use by researchers worldwide. Manta Matcher represents a global initiative to centralize manta ray sightings and facilitate research on these vulnerable species through collaborative studies, including the cross-referencing of regional databases. Manta Matcher has already supported research that contributed to the listing of reef mantas (Manta alfredi) on Appendix 1 of the Convention on Migratory Species in November 2014. References External links Myliobatidae Online databases Biodiversity databases
https://en.wikipedia.org/wiki/Riodina%20lysippus
Riodina lysippus is a species of Neotropical metalmark butterfly. It is the type species of its genus and of the entire family Riodinidae. It is found throughout northern parts of South America and the Caribbean. Description Upperside. The wings are chocolate black. On the anterior is an orange-coloured line, which, rising about the middle of the anterior edge, crosses the wing towards the anal angle, where it suddenly bends, and terminates at the posterior edge. The posterior wings, which are angulated, have a circular orange line, rising at the anterior edge, near the corner, crossing the wings, and meeting near the anal angle. Underside. The wings are of the same colour as on the upperside, with the same orange line, whereon, in the anterior pair, are some white spots. Between this and the base are several faint, dirty grey, oblong spots, namely, four on the anterior, and about twenty on the posterior wing. The base of the anterior margin of the fore wings, and the anal margin of the posterior wings, are of a red colour. Subspecies Riodina lysippus lysippus (Venezuela) Riodina lysippus erratica Seitz, 1913 (Brazil: Amazonas) Riodina lysippus lysias Stichel, 1910 (Ecuador, Bolivia) References Riodinini Riodinidae of South America Butterflies described in 1758 Descriptions from Illustrations of Exotic Entomology Taxa named by Carl Linnaeus Fauna of Suriname
https://en.wikipedia.org/wiki/Zale%20undularis
Zale undularis, the black zale moth, is a moth of the family Noctuidae (or in some classifications, Erebidae). The species was first described by Dru Drury in 1773. It is found in the eastern United States and southern Ontario. Description Upperside: the antennae are brown and thread like. The head, body, abdomen, and wings are of a very dark brown, bordering on black, and appear somewhat glossy. All the wings are a little dentated, and on the anterior ones, from the base to the extremity, is a series of black indented lines or bars, whereof the last or outer one is strong and conspicuous, crossing the wing from the anterior to the posterior edges, about a quarter of an inch from the external margin. The posterior wings are marked exactly like the superior. The underside is of a lighter colour, with the same kind of markings, but fainter. The wingspan is about 2 inches (5 cm). References Catocalinae Moths described in 1773 Taxa named by Dru Drury Descriptions from Illustrations of Exotic Entomology
https://en.wikipedia.org/wiki/Arcplan
Arcplan is a software for business intelligence (BI), budgeting, planning & forecasting (BP&F), business analytics and collaborative Business Intelligence. It is the enhancement of the enterprise software inSight® and dynaSight of the former German provider arcplan Information Services GmbH. The company got merged with Canada-based Longview Solutions in 2015. Version 8.7. is available in a 32 and a 64 bit version at the moment. Before the merger the company was managed by CEO Roland Hölscher and COO Steffen Weissbarth. Introduction The software supports users especially in the area of business intelligence (BI) that comprises clustered processes to select, evaluate and present data electronically. Furthermore, enterprise planning and decision making processes can be executed via data analysis. Arcplan is usually employed in midsized companies and large enterprises. Public institutions as well use arcplan. In Germany arcplan is developed, distributed and maintained by arcplan Information Services GmbH (Langenfeld, Northrhine Westphalia). History and development The Arcplan software emerged from the client server based software Insight in Düsseldorf, Germany in 1993. The founders were the brothers Udo Wollschläger and Frank Hagedorn (both physicists) as well as the mathematician Dr. Hartmut Krins and the chemist Dr. Bernd Fröhlich. Initially, the software development was done by the physicists Udo Wollschläger, Frank Hagedorn, and Svend Dunkhorst. Holding 40% of the capital, the German Lindner Hotels Group was the largest shareholder, the other shares were held by the founders. In 2005 the company was acquired by Frankfurt (Main) based venture capital company Viewpoint Capital Partners GmbH. In 2015 arcplan got sold to Marlin Equity Partners and merged with their portfolio company Longview Solutions. Insight had an integrated development environment that reflected a flexible editor´s concept. With the help of the editor´s concept users were enabled to create cu
https://en.wikipedia.org/wiki/LEAP%20Science%20and%20Maths%20Schools
Langa Education Assistance Program (LEAP), also known as LEAP Science and Mathematics Schools, is a collection of six free secondary education schools located in three provinces in South Africa. The first LEAP school opened in 2004 in rented premises in Observatory, Cape Town and mainly served the township of Langa. LEAP is an independent school mainly founded by South African Corporates with limited subsidies from the Department of Basic Education. History Langa Education Assistance Program (or LEAP) John Gilmour was a teacher at Pinelands High School in Cape Town in 1987 when he decided to respond to a call from the South African business community, to contribute to the redress of the devastation of the Bantu Education Act, a segregation law imposed in the education sector by the Apartheid system in 1953. “Africa Week" was then introduced by a team led by John Gilmour to bring black learners under the "Bantu Education" system to spend a week at Pinelands High School, which was then a whites-only school. The program became the precursor for the Langa Education Assistance Program (or LEAP) which aimed at providing one hundred black students from the Langa township with support tuition from Pinelands High teachers in English, Mathematics and Science three afternoons a week. The prohibitive transport cost of bringing learners from the township schools to Pinelands High School forced the model to be revised. In 1996, it was then decided that instead of learners being bussed in to Pinelands High, teachers will be transported to meet learners in township schools. Community members in Langa perceived the new model of LEAP as an attempt of white teachers to "save" black children. Teachers from the community felt that they were as able as other teachers to provide extra lessons that would address the inadequacies in the students' educational foundations. The uneasiness within the community obliged the leadership of LEAP to change the model and include community teach
https://en.wikipedia.org/wiki/Neueuhren
NeueUhren.de is a German-language, digital special interest interest magazine aimed at those interested in mechanical watches and time measurement. It is published online. The magazine offers information about new mechanical watches with price and technical specifications. Reports on visits to watch manufacturers and background information on watch companies are also among the editorial topics. It competes with blogs such as Zeigr.com, Herrstrohmsuhrsachen.com and Chronautix.com. References External links Official website 2012 establishments in Germany German-language magazines German websites Hobby magazines Magazines established in 2012 Online magazines Technology websites
https://en.wikipedia.org/wiki/Circuit%20Scribe
Circuit Scribe is a ball-point pen containing silver conductive ink one can use to draw circuits instantly on flexible substrates like paper. Circuit Scribe made its way onto Kickstarter (an online site where people can fund projects) on November 19, 2013, with its goal of raising $85,000 for the manufacturing of the first batch of pens. By December 31, 2013, Circuit Scribe was able to raise a total of $674,425 with 12,277 'backers' or donors. Similarly to drawing a picture, users can use a Circuit Scribe pen to draw lines on a simple piece of paper. They can then attach special electrical components on the drawn lines which allows the electrical currents to run through the components. This replaces the use of breadboards and wires. Development A team of researchers in Electroninks Incorporated, a startup company located at Research Park of the University of Illinois at Urbana-Champaign, created a water-based, non-toxic conductive ink that was noted as the Invention of the Month by Popular Science. The team began by developing a prototype using pens from a different company and replacing the ink with their special silver ink. Once completed, they started a Kickstarter campaign to earn funding for a mass production of the final form of the pens. Team The researching team consists of S. Brett Walker, Jennifer A. Lewis, Michael Bell, Analisa Russo, and Nancy Beardsly. Walker is the CEO of Electroninks and the co-founder along with Lewis, Bell, and the director of product development, Russo. Bell is also the chief operating officer while Beardsley is the technical support and user experience. Prototype The prototype pens are hand-cleaned Sakura Gelly Roll Metallic pens. The ink is replaced with the researchers' silver conductive ink. In order to have the right amount of ink flow to make smooth lines, the ink is precisely tuned. Ink The ink is created by placing an aqueous solution of silver nitrate into a flask of water combined with polyacrylic acid (PAA) and
https://en.wikipedia.org/wiki/OrCam%20device
OrCam devices such as OrCam MyEye are portable, artificial vision devices that allow visually impaired people to understand text and identify objects through audio feedback, describing what they are unable to see. Reuters described an important part of how it works as "a wireless smartcamera" which, when attached outside eyeglass frames, can read and verbalize text, and also supermarket barcodes. This information is converted to spoken words and entered "into the user’s ear." Face-recognition is also part of OrCam's feature set. Devices OrCam Technologies Ltd has created three devices; OrCam MyEye 2.0, OrCam MyEye 1, and OrCam MyReader. OrCam My Eye 2.0: OrCam debuted the second-generation model, the OrCam MyEye 2.0 in December 2017. About the size of a finger, the MyEye 2.0 is battery-powered, and has been compressed into a self-contained device. The device snaps onto any eyeglass frame magnetically. Orcam 2.0 is small and light (22.5 grams/0.8 ounces) with functionality to restore independence to the visually impaired. It comes in two versions. The basic model can read text, and a more advanced one adds features such as face recognition and barcode reading. As of July 2023, the retail cost is between $4000 and $6000 (USD). Clinical Studies JAMA Ophthalmology: In 2016 JAMA Ophthalmology conducted a study involving 12 legally blind participants to evaluate the usefulness of a portable artificial vision device (OrCam) for patients with low vision. The results showed that the OrCam device improved the patient's ability to perform tasks simulating those of daily living, such as reading a message on an electronic device, a newspaper article or a menu. Wills Eye: Wills Eye was a clinical study designed to measure the impact of the OrCam device on the quality of life of patients with End-stage Glaucoma. The conclusion was that OrCam, a novel artificial vision device using a mini-camera mounted on eyeglasses
https://en.wikipedia.org/wiki/List%20of%20arbitrary-precision%20arithmetic%20software
This article lists libraries, applications, and other software which enable or support arbitrary-precision arithmetic. Libraries Stand-alone application software Software that supports arbitrary precision computations: bc the POSIX arbitrary-precision arithmetic language that comes standard on most Unix-like systems. dc: "Desktop Calculator" arbitrary-precision RPN calculator that comes standard on most Unix-like systems. KCalc, Linux based scientific calculator Maxima: a computer algebra system which bignum integers are directly inherited from its implementation language Common Lisp. In addition, it supports arbitrary-precision floating-point numbers, bigfloats. Maple, Mathematica, and several other computer algebra software include arbitrary-precision arithmetic. Mathematica employs GMP for approximate number computation. PARI/GP, an open source computer algebra system that supports arbitrary precision. Qalculate!, an open-source free software arbitrary precision calculator with autocomplete. SageMath, an open-source computer algebra system SymPy, a CAS Symbolic Math toolbox (MATLAB) Windows Calculator, since Windows 98, uses arbitrary precision for basic operations (addition, subtraction, multiplication, division) and 32 digits of precision for advanced operations (square root, transcendental functions). SmartXML, a free programming language with integrated development environment (IDE) for mathematical calculations. Variables of BigNumber type can be used, or regular numbers can be converted to big numbers using conversion operator # (e.g., #2.3^2000.1). SmartXML big numbers can have up to 100,000,000 decimal digits and up to 100,000,000 whole digits. Languages Programming languages that support arbitrary precision computations, either built-in, or in the standard library of the language: Ada: the upcoming Ada 202x revision adds the Ada.Numerics.Big_Numbers.Big_Integers and Ada.Numerics.Big_Numbers.Big_Reals packages to the standard library, providing arb
https://en.wikipedia.org/wiki/Log%20Gabor%20filter
In signal processing it is useful to simultaneously analyze the space and frequency characteristics of a signal. While the Fourier transform gives the frequency information of the signal, it is not localized. This means that we cannot determine which part of a (perhaps long) signal produced a particular frequency. It is possible to use a short time Fourier transform for this purpose, however the short time Fourier transform limits the basis functions to be sinusoidal. To provide a more flexible space-frequency signal decomposition several filters (including wavelets) have been proposed. The Log-Gabor filter is one such filter that is an improvement upon the original Gabor filter. The advantage of this filter over the many alternatives is that it better fits the statistics of natural images compared with Gabor filters and other wavelet filters. Applications The Log-Gabor filter is able to describe a signal in terms of the local frequency responses. Because this is a fundamental signal analysis technique, it has many applications in signal processing. Indeed, any application that uses Gabor filters, or other wavelet basis functions may benefit from the Log-Gabor filter. However, there may not be any benefit depending on the particulars of the design problem. Nevertheless, the Log-Gabor filter has been shown to be particularly useful in image processing applications, because it has been shown to better capture the statistics of natural images. In image processing, there are a few low-level examples of the use of Log-Gabor filters. Edge detection is one such primitive operation, where the edges of the image are labeled. Because edges appear in the frequency domain as high frequencies, it is natural to use a filter such as the Log-Gabor to pick out these edges. These detected edges can be used as the input to a segmentation algorithm or a recognition algorithm. A related problem is corner detection. In corner detection the goal is to find points in the image that are c
https://en.wikipedia.org/wiki/Ephyriades%20arcas
Ephyriades arcas, the Caribbean duskywing, is a butterfly of the family Hesperiidae. It is found in Central America and the Caribbean, the type specimen being described from Saint Kitts. Description Upper side: Thorax and abdomen black. Wings very dark brownish black, immaculate. Margins entire. Under side: Legs, breast, and abdomen dark brown, but rather lighter than on the upper side, immaculate, except a small white spot on the anterior, placed near the anterior edge towards the tip. Subspecies Ephyriades arcas arcas Ephyriades arcas philemon (Fabricius, 1775) (Cuba, Bahamas) References Erynnini Butterflies of Central America Butterflies of the Caribbean Descriptions from Illustrations of Exotic Entomology Butterflies described in 1773 Taxa named by Dru Drury
https://en.wikipedia.org/wiki/Winograd%20schema%20challenge
The Winograd schema challenge (WSC) is a test of machine intelligence proposed in 2012 by Hector Levesque, a computer scientist at the University of Toronto. Designed to be an improvement on the Turing test, it is a multiple-choice test that employs questions of a very specific structure: they are instances of what are called Winograd schemas, named after Terry Winograd, professor of computer science at Stanford University. On the surface, Winograd schema questions simply require the resolution of anaphora: the machine must identify the antecedent of an ambiguous pronoun in a statement. This makes it a task of natural language processing, but Levesque argues that for Winograd schemas, the task requires the use of knowledge and commonsense reasoning. Nuance Communications announced in July 2014 that it would sponsor an annual WSC competition, with a prize of $25,000 for the best system that could match human performance. However, the prize is no longer offered. The challenge is considered defeated in 2019 since a number of transformer-based language models achieved accuracies of over 90%. Background The Winograd Schema Challenge was proposed in the spirit of the Turing test. Proposed by Alan Turing in 1950, the Turing test plays a central role in the philosophy of artificial intelligence. Turing proposed that, instead of debating whether a machine can think, the science of AI should be concerned with demonstrating intelligent behavior, which can be tested. But the exact nature of the test Turing proposed has come under scrutiny, especially since an AI chatbot named Eugene Goostman claimed to pass it in 2014. One of the major concerns with the Turing test is that a machine could easily pass the test with brute force and/or trickery, rather than true intelligence. The Winograd schema challenge was proposed in 2012 in part to ameliorate the problems that came to light with the nature of the programs that performed well on the test. Turing's original proposal was
https://en.wikipedia.org/wiki/Evolution%20of%20biparental%20care%20in%20tropical%20frogs
The Evolution of biparental care in tropical frogs is the evolution of the behaviour of a parental care system in frogs in which both the mother and father raise their offspring. Evolution Many tropical frogs have developed a parental care system where both the mother and father partake in raising their offspring. The evolution of biparental care, which is the joint effort of both parents, is a topic that is still under investigation. Biparentalism arose in some species of tropical frogs as a result of ecological conditions, the differences between the sexes, and their natural tendencies. Male parental care could have served as the basis for the development of biparental care. Phylogenetic evidence shows that male parental care is the ancestral strategy in Dendrobates. Currently there are Dendrobates species, such as D. ventrimaculatus and D. fantasticus, that exhibit biparental care. The trend of using males to guard or brood eggs for biparental care or paternal care can be understood from the perspective of the female. After oviposition, or when the eggs are laid, the females need to replenish their bodies that have been dedicated to nurturing the eggs before they can mate again. Brooding by the females would delay the opportunity to mate by about two to four weeks. Since this outcome would cause many males to compete for a few females that are able to mate, the males are favored for the brooding. Environment The environment can have a substantial impact on the uses of parental care. Not all tropical frogs have the ability to lay their eggs plainly on land or plants. Tropical frogs can choose from a variety of water sources, such as lakes, streams, and small puddles. There is greater risk involved with reproducing in bigger bodies of water because of the higher likelihood of fish and other aquatic predators being there. Instead, frogs can choose to place eggs in phytotelmata. However, there is a trade-off that comes with electing a smaller water source. Not m
https://en.wikipedia.org/wiki/Multidimensional%20Multirate%20Systems
Multidimensional Multirate systems find applications in image compression and coding. Several applications such as conversion between progressive video signals require usage of multidimensional multirate systems. In multidimensional multirate systems, the basic building blocks are decimation matrix (M), expansion matrix(L) and Multidimensional digital filters. The decimation and expansion matrices have dimension of D x D, where D represents the dimension. To extend the one dimensional (1-D) multirate results, there are two different ways which are based on the structure of decimation and expansion matrices. If these matrices are diagonal, separable approaches can be used, which are separable operations in each dimension. Although separable approaches might serve less complexity, non-separable methods, with non-diagonal expansion and decimation matrices, provide much better performance. The difficult part in non-separable methods is to create results in MD case by extend the 1-D case. Polyphase decomposition and maximally decimated reconstruction systems are already carried out. MD decimation / interpolation filters derived from 1-D filters and maximally decimated filter banks are widely used and constitute important steps in the design of multidimensional multirate systems. Basic Building Blocks of MD Multirate Systems Decimation and interpolation are necessary steps to create multidimensional multirate systems. In the one dimensional system, decimation and interpolation can be seen in the figure. Theoretically, explanations of decimation and interpolation are: • Decimation (Down-sampling): The M times decimated version of x(n) is defined as y(n)= x(Mn), where M is a nonsingular integer matrix called decimation matrix. In the frequency domain, relation becomes where k is in the range of S which is set of all integer vectors in the form of MTx. J(M) denotes |det(M)| which is also equals to number of k in the determined range. Above expression changes
https://en.wikipedia.org/wiki/Phileurus%20didymus
Phileurus didymus is a species of wood-feeding beetle of the family Scarabaeidae. Description Head, black, small, and triangular, having three tubercles issuing from it, of which the anterior is pointed, the others blunt. Thorax black, which is the general colour of the insect, rounded, smooth, and margined, having an impression in front, with a short tubercle situated on it near the edge; from whence runs a hollow groove or channel to the posterior margin. Scutellum small. Elytra shining, margined and furrowed. Abdomen smooth and shining, without hair. Tibiae furnished with spines, as are the first joints of the middle and posterior tarsi. Length 2 inches. Distribution P. didymus is native to Central America, and is also found in Peru and French Guiana. References External links Biolib Zipcodezoo Species Identifier Catalogue of Life Dynastinae Beetles described in 1758 Descriptions from Illustrations of Exotic Entomology
https://en.wikipedia.org/wiki/Basel%20Computational%20Biology%20Conference
The Basel Computational Biology Conference (stylized as [BC]2) is a scientific meeting on the subjects of bioinformatics and computational biology. It covers a wide spectrum of disciplines, including bioinformatics, computational biology, genomics, computational structural biology, and systems biology. The conference is organized biannually by the SIB Swiss Institute of Bioinformatics in Basel, Switzerland. The next conference 2021 [BC]2 Basel Computational Biology Conference List of previous conferences 2019 [BC]2 Basel Computational Biology Conference "Big Data in Molecular Medicine" in association with BASEL LIFE 2017 [BC]2 Basel Computational Biology Conference 2015 [BC]2 Basel Computational Biology Conference. 2013 "Genetic Variation + Human Health" 2012 "ECCB'12, 11th European Conference on Computational Biology" in association with the 10th Basel Computational Biology conference. 2011 "Multiscale Modeling" 2010 "Regulation & Control in Biological Systems" 2009 "Molecular Evolution" 2008 "Computational Structural Biology" 2007 "From Euler to Computational Biology" in association with USGEB 2006 "Comparative Genomics" 2005 "Biological Systems In Silico" 2004 "From Information to Simulation" 2003 "Life Sciences Meet IT" References Computational science Bioinformatics Computer science conferences Biology conferences Science events in Switzerland
https://en.wikipedia.org/wiki/Hierarchical%20editing%20language%20for%20macromolecules
The hierarchical editing language for macromolecules (HELM) is a method of describing complex biological molecules. It is a notation that is machine readable to render the composition and structure of peptides, proteins, oligonucleotides, and related small molecule linkers. HELM was developed by a consortium of pharmaceutical companies in what is known as the Pistoia Alliance. Development began in 2008. In 2012 the notation was published openly and for free. The HELM open source project can be found on GitHub. HELM The need for HELM became obvious as researchers began working on modeling and computational projects involving molecules and engineered biomolecules of this type. There was not a language to describe the entities in an accurate manner which described both the composition and the complex branching and structure common in these entity types. Protein sequences can describe larger proteins and chemical language files such as mol files can describe simple peptides. But the complexity of new research biomolecules makes describing large complex molecules difficult with chemical formats, and peptide formats are not sufficiently flexible to describe non-natural amino acids and other chemistries. In HELM, molecules are represented at a four levels in a hierarchy: Complex polymer Simple polymer Monomer Atom Monomers are assigned short unique identifiers in internal HELM databases and can be represented by the identifier in strings. The approach is similar to that used in Simplified molecular-input line-entry system (SMILES). An exchangeable file format allows sharing of data between companies who have assigned different identifiers to monomers. In 2014 ChEMBL announced plans to adopt HELM by 2014. The informatics company BIOVIA developed a modified Molfile format called the Self-Contained Sequence Representation (SCSR) A standard which can incorporate individual attempts to solve the problem and be used universally and avoid proliferating standards is
https://en.wikipedia.org/wiki/Taxonomy%20of%20Narcissus
The taxonomy of Narcissus is complex, and still not fully resolved. Known to the ancients, the genus name appears in Graeco-Roman literature, although their interest was as much medicinal as botanical. It is unclear which species the ancients were familiar with. Although frequently mentioned in Mediaeval and Renaissance texts it was not formally described till the work of Linnaeus in 1753. By 1789 it had been grouped into a family (Narcissi) but shortly thereafter this was renamed Amaryllideae, from which comes the modern placement within Amaryllidaceae, although for a while it was considered part of Liliaceae. Many of the species now considered to be Narcissus were in separate genera during the nineteenth century, and the situation was further confused by the inclusion of many cultivated varieties. By 1875 the current circumscription was relatively settled. By 2004 phylogenetic studies had allowed the place of Narcissus within its fairly large family to be established, nested within a series of subfamilies (Amaryllidoideae) and tribes (Narcisseae). It shares its position in the latter tribe with Sternbergia. The infrageneric classification has been even more complex and many schemes of subgenera, sections, subsections and series have been proposed, although all had certain similarities. Most authorities now consider there to be 10 – 11 sections based on phylogenetic evidence. The problems have largely arisen from the diversity of the wild species, frequent natural hybridisation and extensive cultivation with escape and subsequent naturalisation. The number of species has varied anywhere from 16 to nearly 160, but is probably around 50 – 60. The genus appeared some time in the Late Oligocene to Early Miocene eras, around 24 million years ago, in the Iberian peninsula. While the exact origin of the word Narcissus is unknown it is frequently linked to its fragrance which was thought to be narcotic, and to the legend of the youth of that name who fell in love with h
https://en.wikipedia.org/wiki/Global%20cascades%20model
Global cascades models are a class of models aiming to model large and rare cascades that are triggered by exogenous perturbations which are relatively small compared with the size of the system. The phenomenon occurs ubiquitously in various systems, like information cascades in social systems, stock market crashes in economic systems, and cascading failure in physics infrastructure networks. The models capture some essential properties of such phenomenon. Model description To describe and understand global cascades, a network-based threshold model has been proposed by Duncan J. Watts in 2002. The model is motivated by considering a population of individuals who must make a decision between two alternatives, and their choices depend explicitly on other people's states or choices. The model assumes that an individual will adopt a new particular opinion (product or state) if a threshold fraction of his/her neighbors have adopted the new one, else he would keep his original state. To initiate the model, a new opinion will be randomly distributed among a small fraction of individuals in the network. If the fraction satisfies a particular condition, a large cascades can be triggered.(see Global Cascades Condition) A phase transition phenomenon has been observed: when the network of interpersonal influences is sparse, the size of the cascades exhibits a power law distribution, the most highly connected nodes are critical in triggering cascades, and if the network is relatively dense, the distribution shows a bimodal form, in which nodes with average degree show more importance by serving as triggers. Several generalizations of the Watt's threshold model have been proposed and analyzed in the following years. For example, the original model has been combined with independent interaction models to provide a generalized model of social contagion, which classifies the behavior of the system into three universal classes. It has also been generalized on modular networks deg
https://en.wikipedia.org/wiki/SIP%20extensions%20for%20the%20IP%20Multimedia%20Subsystem
The Session Initiation Protocol (SIP) is the signaling protocol selected by the 3rd Generation Partnership Project (3GPP) to create and control multimedia sessions with multiple participants in the IP Multimedia Subsystem (IMS). It is therefore a key element in the IMS framework. SIP was developed by the Internet Engineering Task Force (IETF) as a standard for controlling multimedia communication sessions in Internet Protocol (IP) networks. It is characterized by its position in the application layer of the Internet Protocol Suite. Several SIP extensions published in Request for Comments (RFC) protocol recommendations, have been added to the basic protocol for extending its functionality. The 3GPP, which is a collaboration between groups of telecommunications associations aimed at developing and maintaining the IMS, stated a series of requirements for SIP to be successfully used in the IMS. Some of them could be addressed by using existing capabilities and extensions in SIP while, in other cases, the 3GPP had to collaborate with the IETF to standardize new SIP extensions to meet the new requirements. The IETF develops SIP on a generic basis, so that the use of extensions is not restricted to the IMS framework. 3GPP requirements for SIP The 3GPP has stated several general requirements for operation of the IMS. These include an efficient use of the radio interface by minimizing the exchange of signaling messages between the mobile terminal and the network, a minimum session setup time by performing tasks prior to session establishment instead of during session establishment, a minimum support required in the terminal, the support for roaming and non-roaming scenarios with terminal mobility management (supported by the access network, not SIP), and support for IPv6 addressing. Other requirements involve protocol extensions, such as SIP header fields to exchange user or server information, and SIP methods to support new network functionality: requirement for regis
https://en.wikipedia.org/wiki/Meshedness%20coefficient
In graph theory, the meshedness coefficient is a graph invariant of planar graphs that measures the number of bounded faces of the graph, as a fraction of the possible number of faces for other planar graphs with the same number of vertices. It ranges from 0 for trees to 1 for maximal planar graphs. Definition The meshedness coefficient is used to compare the general cycle structure of a connected planar graph to two extreme relevant references. In one end, there are trees, planar graphs with no cycle. The other extreme is represented by maximal planar graphs, planar graphs with the highest possible number of edges and faces for a given number of vertices. The normalized meshedness coefficient is the ratio of available face cycles to the maximum possible number of face cycles in the graph. This ratio is 0 for a tree and 1 for any maximal planar graph. More generally, it can be shown using the Euler characteristic that all n-vertex planar graphs have at most 2n − 5 bounded faces (not counting the one unbounded face) and that if there are m edges then the number of bounded faces is m − n + 1 (the same as the circuit rank of the graph). Therefore, a normalized meshedness coefficient can be defined as the ratio of these two numbers: It varies from 0 for trees to 1 for maximal planar graphs. Applications The meshedness coefficient can be used to estimate the redundancy of a network. This parameter along with the algebraic connectivity which measures the robustness of the network, may be used to quantify the topological aspect of network resilience in water distribution networks. It has also been used to characterize the network structure of streets in urban areas. Limitations Using the definition of the average degree , one can see that in the limit of large graphs (number of edges the meshedness tends to Thus, for large graphs, the meshedness does not carry more information than the average degree. References Graph invariants Planar graphs
https://en.wikipedia.org/wiki/Ricardo%20Bressani
Cesar Ricardo Bressani Castignoli (28 September 1926 – 30 January 2015) was a Guatemalan food scientist. Born in Guatemala City, he received a bachelor of science in chemical engineering degree from the University of Dayton in 1948. In 1951, he received a master's degree from Iowa State University. In the same year, Bressani returned to Guatemala where he worked at the Institute of Nutrition of Central America and Panama, INCAP. In 1952, he received a scholarship from the Rockefeller Foundation to study biochemistry at Purdue University, where Bressani obtained his Ph.D. in 1956. Afterwards, he reincorporated to the INCAP, this time as the Head of the Division of Agricultural Sciences and Food until 1993. In 1983, Bressani became one of the 42 founding members of the Third World Academy of Sciences, known today as The World Academy of Sciences. In the 1990s, Bressani was the editor-in-chief of the journal Archivos Latinoamericanos de Nutrición. He was also an associate editor of the Food and Nutrition Bulletin. In 1992, he arrived to the Universidad del Valle de Guatemala, and in 1998, he founded the Center for the Studies of Food Science and Technology. Bressani wrote more than 300 publications in many scholarly international journals. Bressani performed an investigation on practical solutions to nutritional problems within the population of Guatemala and the rest of Central America. His experiments led to the creation and production of Incaparina, a nutritional supplement based on a mixture of corn flour, soy flour, cottonseed meal, and Torula yeast. This supplement was intended to be primarily served in the form of gruel. References Further reading External links UNICEF campaign 'Te Toca' (It's your turn) Corinne A. Pernet, “Between Entanglements and Dependencies: Food, Nutrition, and National Development at the Central American Institute of Nutrition (INCAP),” In Sönke Kunkel, Corinna Unger und Marc Frey, eds., International Organizations and Development
https://en.wikipedia.org/wiki/Pollination%20network
A pollination network is a bipartite mutualistic network in which plants and pollinators are the nodes, and the pollination interactions form the links between these nodes. The pollination network is bipartite as interactions only exist between two distinct, non-overlapping sets of species, but not within the set: a pollinator can never be pollinated, unlike in a predator-prey network where a predator can be depredated. A pollination network is two-modal, i.e., it includes only links connecting plant and animal communities. Nested structure of pollination networks A key feature of pollination networks is their nested design. A study of 52 mutualist networks (including plant-pollinator interactions and plant-seed disperser interactions) found that most of the networks were nested. This means that the core of the network is made up of highly connected generalists (a pollinator that visits many different species of plant), while specialized species interact with a subset of the species that the generalists interact with (a pollinator that visits few species of plant, which are also visited by generalist pollinators). As the number of interactions in a network increases, the degree of nestedness increases as well. One property that results from nested structure of pollination networks is an asymmetry in specialization, where specialist species are often interacting with some of the most generalized species. This is in contrast to the idea of reciprocal specialization, where specialist pollinators interact with specialist plants. Similar to the relationship between network complexity and network nestedness, the amount of asymmetry in specialization increases as the number of interactions increases. Modularity of networks Another feature that is common in pollination networks is modularity. Modularity occurs when certain groups of species within a network are much more highly connected to each other than they are with the rest of the network, with weak interactions c
https://en.wikipedia.org/wiki/Jeevan%20Pramaan
Jeevan Pramaan is a Life Certificate program affiliated with Aadhaar for people with pensions. It was started by Prime Minister Narendra Modi on 10 November 2014. The certificate was made for people who receive pensions from central or state governments or other government organisations. Jeevan Pramaan was made by the Department of Electronics and IT, Government of India. The Jeevan Pramaan software can be downloaded from https://jeevanpramaan.gov.in/ & from the Google Play Store for both PC and Android devices. This procedure can also be completed in one of the several Jeevan Pramaan Centres. A pension recipient can receive an electronic Jeevan Praaman certificate by using this software and a fingerprint or iris scan, as well as the Aadhaar platform for identification. The certificate can then be made available electronically to the Pension Disbursing Agency. Eastern Railway launched a Jeevan Pramaan Centre to facilitate the pensioners at Zonal Headquarter at Fairlie Place, Kolkata on 6 August 2015. The service will be extended to additional divisions, workshops and other major establishments under the jurisdiction of Eastern Railway. References E-government in India Pensions in India Public-key cryptography Modi administration initiatives Digital India initiatives
https://en.wikipedia.org/wiki/Nanoinjection
Nanoinjection is the process of using a microscopic lance and electrical forces to deliver DNA to a cell. It is claimed to be more effective than microinjection because the lance used is ten times smaller than a micropipette and the method uses no fluid. The nanoinjector mechanism is operated while submerged in a pH buffered solution. Then, a positive electrical charge is applied to the lance, which accumulates negatively charged DNA on its surface. The nanoinjector mechanism then penetrates the zygotic membranes, and a negative charge is applied to the lance, releasing the accumulated DNA within the cell. The lance is required to maintain a constant elevation on both entry and exit of the cell. Nanoinjection results in a long-term cell viability of 92% following the electrophoretic injection process with a 100 nm diameter nanopipette, the typical diameter of nanoinjection pipet. Single cell transfections are used to virtually transfer any type of mammalian cell into another using a syringe which creates an entry for DNA to be released. A nano needle is used as a mechanical vector for plasmid DNA. This is called Atomic Force Microscopy or AFM. The purpose is to not cause permanent damage to the cell or provoke cellular leaking of intracellular fluid. AFM is a tool of choice as it allows for precise positioning of the DNA. This is important because it allows for tip penetration into the cytosol, which is critical for viable DNA transfer into the cell. Reasons to use nanoinjection include the insertion of genetic material into the genome of a zygote. This method is a critical step in understanding and developing gene functions. Nanoinjection is also used to genetically modify animals to aid in the research of cancer, Alzheimer’s disease, and diabetes. Fabrication The lance is made using the polyMUMPs fabrication technology.  It creates a gold layer, and two structural layers that are 2.0 and 1.5 μm thick respectively.  It is a simple process, which makes it good
https://en.wikipedia.org/wiki/Climate-adaptive%20building%20shell
In building engineering, a climate-adaptive building shell (CABS) is a facade or roof that interacts with the variability of its environment in a dynamic way. Conventional structures have static building envelopes and therefore cannot act in response to changing weather conditions and occupant requirements. Well-designed CABS have two main functions: they contribute to energy-saving for heating, cooling, ventilation, and lighting, and they induce a positive impact on the indoor environmental quality of buildings. Definition The description of CABS made by Loonen et al. says that:A climate adaptive building shell has the ability to repeatedly and reversibly change some of its functions, features or behavior over time in response to changing performance requirements and variable boundary conditions, and does this with the aim of improving overall building performance. This definition shows several components that conform CABS, and are addressed in this article. The first part of the definition is related to its fundamental characteristic; being adaptive envelopes, or in other words, having skins that could adjust to new circumstances. This means that envelopes should be able to "alter slightly as to achieve the desired result", "become used to a new situation", and even return to their original stage if needed. Although occupants’ desired conditions are indoors, they are affected by the outdoor surroundings. While these outcomes can be broadly defined, there is a consensus that the purpose of CABS is to provide shelter, protection, and a comfortable indoor environmental quality by consuming the minimum amount of energy needed. Therefore, the objective is to improve the well-being and productivity of people inside the building by making it sensitive to its surroundings. CABS must satisfy different demands that compete or even conflict with each other. For example, they must find the compromise between daylight and glare, fresh air and draft, ventilation and excess
https://en.wikipedia.org/wiki/International%20Building%20Performance%20Simulation%20Association
The International Building Performance Simulation Association (IBPSA), is a non-profit international society of building performance simulation researchers, developers and practitioners, dedicated to improving the built environment. IBPSA aims to provide a forum for researchers, developers and practitioners to review building model developments, encourage the use of software programs, address standardization, accelerate integration and technology transfer, via exchange of knowledge and organization of (inter)national conferences. Organization IBPSA is an international organization with regional affiliate organizations around the world. IBPSA is governed by a board of directors elected by the membership of all the regional affiliates. In addition to the president, vice-president, secretary, and treasurer, the board is made up of members-at-large and representatives sent by the regional affiliates. Publications Newsletter ibpsaNEWS, IBPSA's online newsletter is published twice per year. The current edition and past issues are available at the IBPSA website. Conference proceedings IBPSA is organizer of the bi-annual international IBPSA Building Simulation Conference and Exhibition. Building Simulation is the premier international event in the field of building performance simulation. In addition to the international conferences, some regional affiliates organize local conferences, as well. All papers presented in the proceedings of these conferences are available at IBPSA's website. Journal The Journal of Building Performance Simulation (JBPS) is the official peer-reviewed scientific journal of the International Building Performance Simulation Association. JBPS publishes articles of the highest quality that are original, cutting-edge, well-researched and of significance to the international community. The journal also offers a forum for original review papers and researched case studies. JBPS is published by Taylor & Francis Group, and co-edited by Dr. Jan Hen
https://en.wikipedia.org/wiki/Individual%20mobility
Individual human mobility is the study that describes how individual humans move within a network or system. The concept has been studied in a number of fields originating in the study of demographics. Understanding human mobility has many applications in diverse areas, including spread of diseases, mobile viruses, city planning, traffic engineering, financial market forecasting, and nowcasting of economic well-being. Data In recent years, there has been a surge in large data sets available on human movements. These data sets are usually obtained from cell phone or GPS data, with varying degrees of accuracy. For example, cell phone data is usually recorded whenever a call or a text message has been made or received by the user, and contains the location of the tower that the phone has connected to as well as the time stamp. In urban areas, user and the telecommunication tower might be only a few hundred meters away from each other, while in rural areas this distance might well be in region of a few kilometers. Therefore, there is varying degree of accuracy when it comes to locating a person using cell phone data. These datasets are anonymized by the phone companies so as to hide and protect the identity of actual users. As example of its usage, researchers used the trajectory of 100,000 cell phone users within a period of six months, while in much larger scale trajectories of three million cell phone users were analyzed. GPS data are usually much more accurate even though they usually are, because of privacy concerns, much harder to acquire. Massive amounts of GPS data describing human mobility are produced, for example, by on-board GPS devices on private vehicles. The GPS device automatically turns on when the vehicle starts, and the sequence of GPS points the device produces every few seconds forms a detailed mobility trajectory of the vehicle. Some recent scientific studies compared the mobility patterns emerged from mobile phone data with those emerged from
https://en.wikipedia.org/wiki/Evolution%20of%20snake%20venom
Venom in snakes and some lizards is a form of saliva that has been modified into venom over its evolutionary history. In snakes, venom has evolved to kill or subdue prey, as well as to perform other diet-related functions. While snakes occasionally use their venom in self defense, this is not believed to have had a strong effect on venom evolution. The evolution of venom is thought to be responsible for the enormous expansion of snakes across the globe. The evolutionary history of snake venom is a matter of debate. Historically, snake venom was believed to have evolved once, at the base of the Caenophidia, or derived snakes. Molecular studies published beginning in 2006 suggested that venom originated just once among a putative clade of reptiles, called Toxicofera, approximately 170 million years ago. Under this hypothesis, the original toxicoferan venom was a very simple set of proteins that were assembled in a pair of glands. Subsequently, this set of proteins diversified in the various lineages of toxicoferans, including Serpentes, Anguimorpha, and Iguania: several snake lineages also lost the ability to produce venom. The Toxicoferan hypothesis was challenged by studies in the mid-2010s, including a 2015 study which found that venom proteins had homologs in many other tissues in the Burmese python. The study therefore suggested that venom had evolved independently in different reptile lineages, including once in the Caenophid snakes. Venom containing most extant toxin families is believed to have been present in the last common ancestor of the Caenophidia: these toxins subsequently underwent tremendous diversification, accompanied by changes in the morphology of venom glands and delivery systems. Snake venom evolution is thought to be driven by an evolutionary arms race between venom proteins and prey physiology. The common mechanism of evolution is thought to be gene duplication followed by natural selection for adaptive traits. The adaptations produced by t
https://en.wikipedia.org/wiki/High-level%20language%20computer%20architecture
A high-level language computer architecture (HLLCA) is a computer architecture designed to be targeted by a specific high-level programming language (HLL), rather than the architecture being dictated by hardware considerations. It is accordingly also termed language-directed computer design, coined in and primarily used in the 1960s and 1970s. HLLCAs were popular in the 1960s and 1970s, but largely disappeared in the 1980s. This followed the dramatic failure of the Intel 432 (1981) and the emergence of optimizing compilers and reduced instruction set computer (RISC) architectures and RISC-like complex instruction set computer (CISC) architectures, and the later development of just-in-time compilation (JIT) for HLLs. A detailed survey and critique can be found in . HLLCAs date almost to the beginning of HLLs, in the Burroughs large systems (1961), which were designed for ALGOL 60 (1960), one of the first HLLs. The best known HLLCAs may be the Lisp machines of the 1970s and 1980s, for the language Lisp (1959). At present the most popular HLLCAs are Java processors, for the language Java (1995), and these are a qualified success, being used for certain applications. A recent architecture in this vein is the Heterogeneous System Architecture (2012), which HSA Intermediate Layer (HSAIL) provides instruction set support for HLL features such as exceptions and virtual functions; this uses JIT to ensure performance. Definition There are a wide variety of systems under this heading. The most extreme example is a Directly Executed Language (DEL), where the instruction set architecture (ISA) of the computer equals the instructions of the HLL, and the source code is directly executable with minimal processing. In extreme cases, the only compiling needed is tokenizing the source code and feeding the tokens directly to the processor; this is found in stack-oriented programming languages running on a stack machine. For more conventional languages, the HLL statements are grouped
https://en.wikipedia.org/wiki/Design%20smell
In computer programming, design smells are "structures in the design that indicate violation of fundamental design principles and negatively impact design quality". The origin of the term "design smell" can be traced to the term "code smell" which was featured in the book Refactoring: Improving the Design of Existing Code by Martin Fowler. Details Different authors have defined the word "smell" in different ways: N. Moha et al.: "Code and design smells are poor solutions to recurring implementation and design problems." R. C. Martin: "Design smells are the odors of rotting software." Fowler: "Smells are certain structures in the code that suggest (sometimes they scream for) the possibility of refactoring." Design smells indicate the accumulated design debt (one of the prominent dimensions of technical debt). Bugs or unimplemented features are not accounted as design smells. Design smells arise from the poor design decisions that make the design fragile and difficult to maintain. It is a good practice to identify design smells in a software system and apply appropriate refactoring to eliminate it to avoid accumulation of technical debt. The context (characterized by various factors such as the problem at hand, design eco-system, and platform) plays an important role to decide whether a certain structure or decision should be considered as a design smell. Generally, it is appropriate to live with design smells due to constraints imposed by the context. Nevertheless design smells should be tracked and managed as technical debt because they degrade the overall system quality over time. Common design smells Missing abstraction when clumps of data or encoded strings are used instead of creating an abstraction. Also known as "primitive obsession" and "data clumps". Multifaceted abstraction when an abstraction has multiple responsibilities assigned to it. Also known as "conceptualization abuse". Duplicate abstraction when two or more abstractions have identical
https://en.wikipedia.org/wiki/Galerita%20americana
Galerita americana is a species of beetles in the family Carabidae. It is native to Central and South America (Bolivia, Brazil, Colombia, Costa Rica, French Guiana, Guatemala, Panama, Paraguay, Suriname and Venezuela). Description Head very long, black, with a red brown spot on the middle. Antennae dark brown, the basal joint being longest, thickest, and lightest coloured; the others are nearly of equal length; the whole being a little longer than the elytra. Neck distinct and black. Thorax light red brown, and almost oval, about the length of the head, and a little broader; it is also a little margined, and next the body truncate. Scutellum minute, black, and triangular. Elytra black, margined and furrowed, oval next the thorax, but more square at their extremities, and not covering the anus. Abdomen black. Breast light red brown; as are all the legs. The basal joints of the posterior tarsi are very long. Length of body ¾ inch (19 mm). References Dryptinae Descriptions from Illustrations of Exotic Entomology Beetles described in 1758 Taxa named by Carl Linnaeus
https://en.wikipedia.org/wiki/Mediterranean%20Mathematics%20Competition
The Mediterranean Mathematics Competition (also: Peter O’Halloran Memorial) is a mathematics competition for school students, taking place annually since 1998. All countries bordering the Mediterranean Sea are allowed to participate, as well as, if invited, their neighbouring countries. Motto The Mediterranean Competition's goals are: Discovery, development and challenge of mathematically gifted students Establishment of friendly and cooperative relationships between students and teachers of various mediterranean countries Creation of a possibility for international exchange about school practices Support for the engagement in solving mathematical olympiad problems, as well as the dealing with other mathematical problems, also in non participating countries Rules The contest is conducted separately in every country. Each participating country can let an unrestricted number of students write the contest, but only the results of the top ten, according to national evaluation, can be submitted for international ranking. Every of these is awarded a certificate either of participation or merit, whereas the levels of merit - Gold, Silver, Bronze and Honorable Mention – are awarded similarly to the International Mathematical Olympiad. The participants have to be less than 20 years of age and may not have enrolled in a university study or a comparable educative scheme. History The Mediterranean Competition initially took place in 1998, created and until today organized by Spanish Francisco Bellot Rosado. In the first year, only three problems were to be solved. However from the second year on, the contest consisted of four problems each, with an overall contest time of four hours. References External links Contest papers from 1998 to 2005 on the official webpage of the International Mathematical Olympiad Contest papers from 1998 to 2011 Official rule document Annual events European student competitions Mathematics competitions Recurring events established in
https://en.wikipedia.org/wiki/Ind-scheme
In algebraic geometry, an ind-scheme is a set-valued functor that can be written (represented) as a direct limit (i.e., inductive limit) of closed embedding of schemes. Examples is an ind-scheme. Perhaps the most famous example of an ind-scheme is an infinite grassmannian (which is a quotient of the loop group of an algebraic group G.) See also formal scheme References A. Beilinson, Vladimir Drinfel'd, Quantization of Hitchin’s integrable system and Hecke eigensheaves on Hitchin system, preliminary version V.Drinfeld, Infinite-dimensional vector bundles in algebraic geometry, notes of the talk at the `Unity of Mathematics' conference. Expanded version http://ncatlab.org/nlab/show/ind-scheme Algebraic geometry
https://en.wikipedia.org/wiki/Dirtbox%20%28cell%20phone%29
A dirtbox (or DRT box) is a cell site simulator, a phone device mimicking a cell phone tower, that creates a signal strong enough to cause nearby dormant mobile phones to switch to it. Mounted on aircraft, it is used by the United States Marshals Service to locate and collect information from cell phones believed to be connected with criminal activity. It can also be used to jam phones. The device's name comes from the company that developed it, Digital Receiver Technology, Inc. (DRT), owned by the Boeing company. Boeing describes the device as a hybrid of "jamming, managed access and detection". A similar device with a smaller range, the controversial StingRay phone tracker, has been widely used by U.S. federal entities, including the Federal Bureau of Investigation (FBI). History It is not known when DRT first manufactured the dirtbox. The company does not publicly advertise it, stating on its web site: "Due to the sensitive nature of our work, we are unable to publicly advertise many of our products." The Wall Street Journal wrote that the U.S. Marshals Service program utilizing the device had "fully matured by 2007". Boeing bought DRT in 2008. Similar devices from the Harris Corporation, like the Stingray phone tracker, have been sold around the same time. Since 2008, their airborne mounting kit for cell phone surveillance has been said to cost $9,000. On June 11, 2010, the Boeing Company asked the National Telecommunications and Information Administration to advise the United States Congress that the "... Communications Act of 1934 be modified to allow prison officials and state and local law enforcement to use appropriate cell phone management", and suggested that special weapons and tactics (SWAT) teams and other paramilitary tactical units could use their devices to control wireless communications during raids. The Chicago Police Department bought dirtboxes to eavesdrop on demonstrators during the 2012 NATO summit, and used them during the 2014 Black L
https://en.wikipedia.org/wiki/ISO%205428
ISO 5428:1984, Greek alphabet coded character set for bibliographic information interchange, is an ISO standard for an 8-bit character encoding for the modern Greek language. It contains a set of 73 graphic characters and is available through UNIMARC. In practice it is now superseded by Unicode. Character set See also Greek Alphabet ISO/IEC 8859-7 Notes References ISO 5428:1984 "Greek alphabet coded character set for bibliographic information interchange" MARC 21 Specifications for Record Structure, Character Sets, and Exchange Media > Character Sets > Part 3: Code Tables (Character Sets) > Code Table 8: Greek "Greek Character Tables: ISO 5428-1980" "Greek alphabet coded character set for bibliographic information interchange" (June 1, 1982) "Greek alphabet character set for bibliographic use" (August 13, 1976; older standard) 05428 Character sets
https://en.wikipedia.org/wiki/Single-cell%20variability
In cell biology, single-cell variability occurs when individual cells in an otherwise similar population differ in shape, size, position in the cell cycle, or molecular-level characteristics. Such differences can be detected using modern single-cell analysis techniques. Investigation of variability within a population of cells contributes to understanding of developmental and pathological processes, Single-cell analysis A sample of cells may appear similar, but the cells can vary in their individual characteristics, such as shape and size, mRNA expression levels, genome, or individual counts of metabolites. In the past, the only methods available for investigating such properties required a population of cells and provided an estimate of the characteristic of interest, averaged over the population, which could obscure important differences among the cells. Single-cell analysis allows scientists to study the properties of a single cell of interest with high accuracy, revealing individual differences among populations and offering new insights in molecular biology. These individual differences are important in fields such as developmental biology, where individual cells can take on different "fates" - become specialized cells such as neurons or organ tissue - during the growth of an embryo; in cancer research, where individual malignant cells can vary in their response to therapy; or in infectious disease, where only a subset of cells in a population become infected by a pathogen. Population-level views of cells can offer a distorted view of the data by averaging out the properties of distinct subsets of cells. For example, if half the cells of a particular group are expressing high levels of a given gene, and the rest are expressing low levels, results from a population-wide analysis may appear as if all cells are expressing a medium level of the given gene. Thus, single-cell analysis allows researchers to study biological processes in finer detail and answer ques
https://en.wikipedia.org/wiki/Bertie%20the%20Brain
Bertie the Brain is one of the first games developed in the early history of video games. It was built in Toronto by Josef Kates for the 1950 Canadian National Exhibition. The four meter (13 foot) tall computer allowed exhibition attendees to play a game of tic-tac-toe against an artificial intelligence. The player entered a move on a keypad in the form of a three-by-three grid, and the game played out on a grid of lights overhead. The machine had an adjustable difficulty level. After two weeks on display by Rogers Majestic, the machine was disassembled at the end of the exhibition and largely forgotten as a curiosity. Kates built the game to showcase his additron tube, a miniature version of the vacuum tube, though the transistor overtook it in computer development shortly thereafter. Patent issues prevented the additron tube from being used in computers besides Bertie before it was no longer useful. Bertie the Brain is a candidate for the first video game, as it was potentially the first computer game to have any sort of visual display of the game. It appeared only three years after the 1947 invention of the cathode-ray tube amusement device, the earliest known interactive electronic game to use an electronic display. Berties use of light bulbs rather than a screen with real-time visual graphics, however, much less moving graphics, does not meet some definitions of a video game. History Bertie the Brain was a video game version of tic-tac-toe, built by Dr. Josef Kates for the 1950 Canadian National Exhibition. Kates had previously worked at Rogers Majestic designing and building radar tubes during World War II, then after the war pursued graduate studies in the computing center at the University of Toronto while continuing to work at Rogers Majestic. While there, he helped build the University of Toronto Electronic Computer (UTEC), one of the first working computers in the world. He also designed his own miniature version of the vacuum tube, called the additron
https://en.wikipedia.org/wiki/Process%20qualification
Process qualification is the qualification of manufacturing and production processes to confirm they are able to operate at a certain standard during sustained commercial manufacturing. Data covering critical process parameters must be recorded and analyzed to ensure critical quality attributes can be guaranteed throughout production. This may include testing equipment at maximum operating capacity to show quantity demands can be met. Once all processes have been qualified the manufacturer should have a complete understanding of the process design and have a framework in place to routinely monitor operations. Only after process qualification has been completed can the manufacturing process begin production for commercial use. Equally important as qualifying processes and equipment is qualifying software and personnel. A well trained staff and accurate, thorough records helps ensure ongoing protection from process faults and quick recovery from otherwise costly process malfunctions. In many countries qualification measures are also required, especially in the pharmaceutical manufacturing field. Process qualification should cover the following aspects of manufacturing: Facility Utilities Equipment Personnel End-to-end manufacturing Control protocols and monitoring software. Process qualification is the second stage of process validation. A vital component of process qualification is process performance qualification protocol. PPQ protocol is essential in defining and maintaining production standards within an organization. See also Installation qualification Design qualification Performance qualification Process validation References External links Drugregulations.org Formal methods Enterprise modelling Business process management
https://en.wikipedia.org/wiki/Rapid%20antigen%20test
A rapid antigen test (RAT), sometimes called a rapid antigen detection test (RADT), antigen rapid test (ART), or loosely just a rapid test, is a rapid diagnostic test suitable for point-of-care testing that directly detects the presence or absence of an antigen. RATs are a type of lateral flow test detecting antigens, rather than antibodies (antibody tests) or nucleic acid (nucleic acid tests). Rapid tests generally give a result in 5 to 30 minutes, require minimal training or infrastructure, and have significant cost advantages. Rapid antigen tests for the detection of SARS-CoV-2, the virus that causes COVID-19, have been commonly used during the COVID-19 pandemic. For many years, an early and major class of RATs—the rapid strep tests for streptococci—were so often the referent when RATs or RADTs were mentioned that the two latter terms were often loosely treated as synonymous with those. Since the COVID-19 pandemic, awareness of RATs is no longer limited to health professionals and COVID-19 has become the expected referent, so more precise usage is required in other circumstances. RATs are based on the principle of antigen-antibody interaction. They detect antigens (generally a protein on the surface of a virus). A linear chromatography substrate (a porous piece of material) bears an indicator line, onto which antibodies directed against the target antigen are fixed. Antibodies are also fixed to a visualisation marker (generally a dye, though sometimes these antibodies are modified to fluoresce), to which the sample is added. Any virus particles present will bind to these markers. This mix then travels through the substrate through capillarity. When it reaches the indicator line, virus particles are immobilised by the antibodies fixed there, along with the visualisation marker, allowing concentration and thus visual detection of significant levels of virus in a sample. A positive result with an antigen test should generally be confirmed by RT-qPCR or some ot
https://en.wikipedia.org/wiki/Virtual%20metrology
In semiconductor manufacturing, virtual metrology refers to methods to predict the properties of a wafer based on machine parameters and sensor data in the production equipment, without performing the (costly) physical measurement of the wafer properties. Statistical methods such as classification and regression are used to perform such a task. Depending on the accuracy of this virtual data, it can be used in modelling for other purposes, such as predicting yield, preventative analysis, etc. This virtual data is helpful for modelling techniques that are adversely affected by missing data. Another option to handle missing data is to use imputation techniques on the dataset, but virtual metrology in many cases, can be a more accurate method. Examples of virtual metrology include: the prediction of the silicon nitride () layer thickness in the chemical vapor deposition process (CVD), using multivariate regression methods; the prediction of critical dimension in photolithography, using multi-level and regularization approaches; the prediction of layer width in etching. References Semiconductor device fabrication Automation Lithography (microfabrication) Metrology
https://en.wikipedia.org/wiki/Ingredient-flavor%20network
In network science, ingredient-flavor networks are networks describing the sharing of flavor compounds of culinary ingredients. In the bipartite form, an ingredient-flavor network consist of two different types of nodes: the ingredients used in the recipes and the flavor compounds that contributes to the flavor of each ingredients. The links connecting different types of nodes are undirected, represent certain compound occur in each ingredients. The ingredient-flavor network can also be projected in the ingredient or compound space where nodes are ingredients or compounds, links represents the sharing of the same compounds to different ingredients or the coexistence in the same ingredient of different compounds. History In 2011, Yong-Yeol Ahn, Sebastian E. Ahnert, James P. Bagrow and Albert-László Barabási investigated the ingredient-flavor networks of North American, Latin American, Western European, Southern European and East Asian cuisines. Based on culinary repository epicurious.com, allrecipes.com and menupan.com, 56,498 recipes were included in the survey. The efforts to apply network analysis on foods also occurred in the work of Kinouchi and Chun-Yuen Teng, with the former examined the relationship between ingredients and recipes, and the latter derived the ingredient-ingredient networks of both compliments and substitutions. Yet Ahn's ingredient-flavor network was constructed based on the molecular level understanding of culinary networks and received wide attention Properties According to Ahn, in the total number of 56,498 recipes studied, 381 ingredients and 1021 flavor compounds were identified. On average, each ingredient connected to 51 flavor compounds. It was found that in comparison with random pairing of ingredients and flavor compounds, North American cuisines tend to share more compounds while East Asian cuisines tend to share fewer compounds. It was also shown that this tendency was mostly generated by the frequently used ingredients in e
https://en.wikipedia.org/wiki/Circular%20thresholding
Circular thresholding is an algorithm for automatic image threshold selection in image processing. Most threshold selection algorithms assume that the values (e.g. intensities) lie on a linear scale. However, some quantities such as hue and orientation are a circular quantity, and therefore require circular thresholding algorithms. The example shows that the standard linear version of Otsu's method when applied to the hue channel of an image of blood cells fails to correctly segment the large white blood cells (leukocytes). In contrast the white blood cells are correctly segmented by the circular version of Otsu's method. Methods There are a relatively small number of circular image threshold selection algorithms. The following examples are all based on Otsu's method for linear histograms: (Tseng, Li and Tung 1995) smooth the circular histogram, and apply Otsu's method. The histogram is cyclically rotated so that the selected threshold is shifted to zero. Otsu's method and histogram rotation are applied iteratively until several heuristics involving class size, threshold location, and class variance are satisfied. (Wu et al. 2006) smooth the circular histogram until it contains only two peaks. The histogram is cyclically rotated so that the midpoint between the peaks is shifted to zero. Otsu's method and histogram rotation are applied iteratively until convergence of the threshold. (Lai and Rosin 2014) applied Otsu's method to the circular histogram. For the two class circular thresholding task they showed that, for a histogram with an even number of bins, the optimal solution for Otsu's criterion of within-class variance is obtained when the histogram is split into two halves. Therefore the optimal solution can be efficiently obtained in linear rather than quadratic time. References and further reading D.-C. Tseng, Y.-F. Li, and C.-T. Tung, Circular histogram thresholding for color image segmentation in Proc. Int. Conf. Document Anal. Recognit., 1995, pp. 6
https://en.wikipedia.org/wiki/Radio%20Science
Radio Science is a quarterly peer-reviewed scientific journal published by Wiley-Blackwell on behalf of the American Geophysical Union and co-sponsored by the International Union of Radio Science. It contains original scientific contributions on radio-frequency electromagnetic propagation and its applications (radio science). Its full aims and scope read: Volumes for the years 1966 through 1968 were issued by the Environmental Science Services Administration (ESSA), the precursor of the National Oceanic and Atmospheric Administration (NOAA), in cooperation with the United States National Committee of the International Scientific Radio Union. See also Advances in Radio Science References External links Academic journals established in 1966 English-language journals American Geophysical Union academic journals Wiley (publisher) academic journals Wiley-Blackwell academic journals Engineering journals Physics journals Electromagnetism journals
https://en.wikipedia.org/wiki/Banisia%20myrtaea
Banisia myrtaea is a species of moth of the family Thyrididae. It was first described by Dru Drury in 1773 from Madras. Description Upper side: antennæ brown and setaceous. Palpi, head, neck, thorax, abdomen, and wings reddish flesh-coloured; the latter having some very faint waved lines crossing them. Cilia dark brown. Under side: breast, sides, legs, and abdomen coloured as on the upper side. Wings yellowish, with many small narrow streaks. On the external edges of the anterior wings is a dark brown patch, near the tips. Cilia dark brown. Wing-span 1¼ inches (32 mm). References Thyrididae Descriptions from Illustrations of Exotic Entomology Moths described in 1773 Taxa named by Dru Drury
https://en.wikipedia.org/wiki/Yramea%20cytheris
Yramea cytheris is a species of butterfly in the family Nymphalidae. It was first described by Dru Drury in 1773 from the Falkland Islands. In some systems it is included in genus Issoria. Description Upperside: head, eyes, thorax, and abdomen dark brown. Anterior wings brown orange, with a number of small black spots thereon (not less than twenty) of various shapes and sizes. Next to the body these wings are darker and pilose (covered with fine hair). Posterior wings are the same colour as the superior, and spotted with many small black spots of different shapes dispersed all over the wings; they are also darker next to the body and hairy. Underside: palpi reddish. Anterior wings sandy orange coloured, rather paler than the upperside. Near the tips is a white streak placed next the anterior margin, and close thereto is a cloud of a dark red. Most of the small black spots are seen on this side, but less distinctly. Posterior wings dark red, with several faint clouds. A narrow white streak, about a quarter of an inch (6 mm) in length, is placed near the middle of each of these wings; and another much shorter is placed on the anterior edge, near the upper corner. All the wings are entire. Wingspan inches (32 mm). Subspecies Yramea cytheris cytheris (Falkland Islands) Yramea cytheris siga (Geyer, 1832) (Chile, Argentina) References Argynnini Descriptions from Illustrations of Exotic Entomology Butterflies described in 1773 Nymphalidae of South America Taxa named by Dru Drury
https://en.wikipedia.org/wiki/Advances%20in%20Radio%20Science
Advances in Radio Science is a peer-reviewed open access scientific journal published by the German National Committee of the International Union of Radio Science. It covers radio science and radio engineering. It is abstracted and indexed in Scopus. See also Radio Science External links Electrical and electronic engineering journals Academic journals established in 2003 Multilingual journals Copernicus Publications academic journals Creative Commons Attribution-licensed journals Electromagnetism journals
https://en.wikipedia.org/wiki/Host%20adaptation
When considering pathogens, host adaptation can have varying descriptions. For example, in the case of Salmonella, host adaptation is used to describe the "ability of a pathogen to circulate and cause disease in a particular host population." Another usage of host adaptation, still considering the case of Salmonella, refers to the evolution of a pathogen such that it can infect, cause disease, and circulate in another host species. Description While there might be pathogens that can infect other hosts and cause disease, the inability to pervade, or spread, throughout the infected host species indicates that the pathogen is not adapted to that host species. In this case, the ability or lack thereof of a pathogen to adapt to its host environment is an indicator of the pathogen's fitness or virulence. If a pathogen has high fitness in the host environment, or is virulent, it will be able to grow and spread quickly within its host. Conversely, if the pathogen is not well adapted to its host environment, then it will not spread or infect the way a well adapted pathogen would. Pathogens like Salmonella, which is a food borne pathogen, are able to adapt to the host environment and maintain virulence via several pathways. In a paper by Baumler et al. 1998, characters of Salmonella, such as its ability to cause intestinal infection were attributed to virulence factors like its ability to invade intestinal epithelial cells, induce neutrophil recruitment and interfere with the secretion of intestinal fluid. Phylogenetic analysis also revealed that many strains or lineages of Salmonella exist, which is advantageous for the pathogen because its genetic diversity can acts as fodder for natural selection to tinder with. For instance, if a particular Salmonella strain is more fit in the host stomach environment, compared to other Salmonella strains, then the former will be positively selected for and increase in prevalence. Eventually this strain will colonize and infect the st
https://en.wikipedia.org/wiki/Display%20%28social%20network%29
Display Inc., stylized as display, is an online social media and networking service based in Norwalk, Connecticut. The display app officially launched in May 2021. History In September 2019, display (formerly Tsū) announced its planned relaunch. According to Chief Executive Officer John Acunto, display would share ad revenue with users, a 50% payout of ad revenue. display also incorporated protections against spam and empowers users to access data, analytics and insights related to their content. In an interview with Fox Business (in response to the October 29th NCAA Board of Governors unanimous vote to allow student-athletes to be paid for the use of their name, image and likeness), Tiki Barber, former NY Giants running back, cited display as a “great platform” to put the new NCAA rules to use, adding that display enables influencers of all kinds to have the ability to monetize their own content and brands.  Barber went on to say that display is creating a platform that gives all users access to brand partnership, by sharing in ad revenue, providing storefronts and more.  display’s John Acunto echoed Barber’s sentiment saying that college athletes were just one example of those who could benefit from display: "I see this as an opportunity for all kinds of categories of people who are influencers [and] who have brands to engage with us." display rebranded from Tsū in April 2021. Core pillars "display, the ‘Social That Pays’, put the creators on the focus. Our platform revolves around the people who create valuable content. We believe that they deserve the reward for all the efforts they put in." Awarding content creators based on the ad revenue they generate on the platform Enabling commerce at the point of discovery via a personalized storefront Providing tools that inspire content creation and empower creators to post efficiently displayFest Beginning on May 3, 2021, displayFest began. It featured a series of performances including sets from Snoop Dog
https://en.wikipedia.org/wiki/Partial%20dominance%20hypothesis
The partial dominance hypothesis in genetics states that inbreeding depression is the result of the frequency increase of homozygous deleterious recessive or partially recessive alleles. The partial dominance hypothesis can be explained by looking at a population that is divided into a large number of separately inbred lines. Deleterious alleles will eventually be eliminated from some lines and become fixed in other lines, while some lines disappear because of fixation of deleterious alleles. This will cause an overall decline in population and trait value, but then increase to a trait value that is equal to or greater than the trait value in the original population. Crossing inbred lines restores fitness in the overdominance hypothesis and a fitness increase in the partial dominance hypothesis. References Genetics
https://en.wikipedia.org/wiki/Woeseian%20revolution
The Woeseian revolution was the progression of the phylogenetic tree of life concept from two main divisions, known as the Prokarya and Eukarya, into three domains now classified as Bacteria, Archaea, and Eukaryotes. The discovery of the new domain stemmed from the work of biophysicist Carl Woese in 1977 from a principle of evolutionary biology designated as Woese's dogma. It states that the evolution of ribosomal RNA (rRNA) was a necessary precursor to the evolution of modern life forms. Although the three-domain system has been widely accepted, the initial introduction of Woese’s discovery received criticism from the scientific community. Phylogenetic implications The basis of phylogenetics was limited by the technology of the time, which led to a greater dependence on phenotypic classification before advances that would allow for molecular organization methods. This was a major reason why the dichotomy of all living things, being either animal or plant in nature, was deemed an acceptable theory. Without truly understanding the genetic implication of each organismal classification in phylogenies via nucleic acid sequencing of shared molecular material, the phylogenetic tree of life and other such phylogenies would no doubt be incorrect. Woese’s advances in molecular sequencing and phylogenetic organization allowed for a better understanding of the three domains of life - the Bacteria, Archaea, and Eukaryotes. Regarding their varying types of shared rRNA, the small subunit rRNA was deemed as the best molecule to sequence to distinguish phylogenetic relationships because of its relatively small size, ease of isolation, and universal distribution. Controversy This reorganization caused an initial pushback: it wasn't accepted until nearly a decade after its publication. Possible factors that led to initial criticisms of his discovery included Woese's oligonucleotide cataloging, of which he was one of "only two or three people in the world" to be able to execute th
https://en.wikipedia.org/wiki/Pitthea%20famula
Pitthea famula is a species of moth in the family Geometridae. It was first described by Dru Drury in 1773 from Calabar, in what is now Nigeria. It is found in Angola, Benin, Cameroon, the Republic of the Congo, the Democratic Republic of the Congo, Equatorial Guinea (Bioko), Nigeria, Sierra Leone and Zambia. Description Upperside: antennae long and pectinated (comb like). Thorax spiral. Neck orange. Thorax and abdomen dusky grey. Anterior wings about halfway from the tips black, but at the base are of a pellucid (transparent) white; being surrounded along the anterior edge and part of the posterior with black; an oblong white spot is placed near the tips on the black part. Posterior wings black and white; the white entirely surrounded by the black, which on the anterior and abdominal edges is very narrow. Underside: palpi orange, black at the tips. Neck, breast, and sides orange. Feet black. Thighs white. Abdomen white, annulated with dusky grey. Anterior wings as on the upperside, the black parts being of a russet hue. Posterior wings differ a little, the white part running down to the middle of the external edges, with a white spot at the upper corners. Margins of the wings entire. Wingspan 2 inches (50 mm). References Ennominae Moths described in 1773 Taxa named by Dru Drury Descriptions from Illustrations of Exotic Entomology