source
stringlengths
31
227
text
stringlengths
9
2k
https://en.wikipedia.org/wiki/SAP%20NetWeaver%20Business%20Warehouse
SAP Business Warehouse (SAP BW) is SAP’s Enterprise Data Warehouse product. It can transform and consolidate business information from virtually any source system. It ran on industry standard RDBMS until version 7.3 at which point it began to transition onto SAP's HANA in-memory DBMS, particularly with the release of version 7.4. Latterly, it evolved into a product called BW/4HANA so as to align with SAP's sister ERP Product called S/4HANA. This strategy allowed SAP to engineer the database to use the HANA in-Memory database. Consequently, this facilitates the push down of complex OLAP based functions to the database as opposed to NetWeaver ABAP Application Server to improve performance. The product is also more open and can incorporate SAP and Non-SAP data more easily. History In 1998 SAP released the first version of SAP BW, providing a model-driven approach to EDW that made data warehousing easier and more efficient, particularly for SAP R/3 data. Since then, SAP BW has evolved to become a key component for thousands of companies. An article, provided the history of SAP BW from inception to the newer releases powered by SAP HANA.
https://en.wikipedia.org/wiki/Quantum%20boomerang%20effect
The quantum boomerang effect is a quantum mechanical phenomenon whereby wavepackets launched through disordered media return, on average, to their starting points, as a consequence of Anderson localization and the inherent symmetries of the system. At early times, the initial parity asymmetry of the nonzero momentum leads to asymmetric behavior: nonzero displacement of the wavepackets from their origin. At long times, inherent time-reversal symmetry and the confining effects of Anderson localization lead to correspondingly symmetric behavior: both zero final velocity and zero final displacement. History In 1958, Philip W. Anderson introduced the eponymous model of disordered lattices which exhibits localization, the confinement of the electrons' probability distributions within some small volume. In other words, if a wavepacket were dropped into a disordered medium, it would spread out initially but then approach some maximum range. On the macroscopic scale, the transport properties of the lattice are reduced as a result of localization, turning what might have been a conductor into an insulator. Modern condensed matter models continue to study disorder as an important feature of real, imperfect materials. In 2019, theorists considered the behavior of a wavepacket not merely dropped, but actively launched through a disordered medium with some initial nonzero momentum, predicting that the wavepacket's center of mass would asymptotically return to the origin at long times — the quantum boomerang effect. Shortly after, quantum simulation experiments in cold atom settings confirmed this prediction by simulating the quantum kicked rotor, a model that maps to the Anderson model of disordered lattices. Description Consider a wavepacket with initial momentum which evolves in the general Hamiltonian of a Gaussian, uncorrelated, disordered medium: where and , and the overbar notation indicates an average over all possible realizations of the disorder. The classical
https://en.wikipedia.org/wiki/Autoantibody
An autoantibody is an antibody (a type of protein) produced by the immune system that is directed against one or more of the individual's own proteins. Many autoimmune diseases (notably lupus erythematosus) are associated with such antibodies. Production Antibodies are produced by B cells in two ways: (i) randomly, and (ii) in response to a foreign protein or substance within the body. Initially, one B cell produces one specific kind of antibody. In either case, the B cell is allowed to proliferate or is killed off through a process called clonal deletion. Normally, the immune system is able to recognize and ignore the body's own healthy proteins, cells, and tissues, and to not overreact to non-threatening substances in the environment, such as foods. Sometimes, the immune system ceases to recognize one or more of the body's normal constituents as "self", leading to production of pathological autoantibodies. Autoantibodies may also play a nonpathological role; for instance they may help the body to destroy cancers and to eliminate waste products. The role of autoantibodies in normal immune function is also a subject of scientific research. Cause The causes of autoantibody production are varied and not well understood. It is thought that some autoantibody production is due to a genetic predisposition combined with an environmental trigger, such as a viral illness or a prolonged exposure to certain toxic chemicals. There is generally not a direct genetic link however. While families may be susceptible to autoimmune conditions, individual family members may have different autoimmune disorders, or may never develop an autoimmune condition. Researchers believe that there may also be a hormonal component as many of the autoimmune conditions are much more prevalent in women of childbearing age. While the initial event that leads to the production of autoantibodies is still unknown, there is a body of evidence that autoantibodies may have the capacity to maintain thei
https://en.wikipedia.org/wiki/Random%20forest
Random forests or random decision forests is an ensemble learning method for classification, regression and other tasks that operates by constructing a multitude of decision trees at training time. For classification tasks, the output of the random forest is the class selected by most trees. For regression tasks, the mean or average prediction of the individual trees is returned. Random decision forests correct for decision trees' habit of overfitting to their training set. Random forests generally outperform decision trees, but their accuracy is lower than gradient boosted trees. However, data characteristics can affect their performance. The first algorithm for random decision forests was created in 1995 by Tin Kam Ho using the random subspace method, which, in Ho's formulation, is a way to implement the "stochastic discrimination" approach to classification proposed by Eugene Kleinberg. An extension of the algorithm was developed by Leo Breiman and Adele Cutler, who registered "Random Forests" as a trademark in 2006 (, owned by Minitab, Inc.). The extension combines Breiman's "bagging" idea and random selection of features, introduced first by Ho and later independently by Amit and Geman in order to construct a collection of decision trees with controlled variance. History The general method of random decision forests was first proposed by Ho in 1995. Ho established that forests of trees splitting with oblique hyperplanes can gain accuracy as they grow without suffering from overtraining, as long as the forests are randomly restricted to be sensitive to only selected feature dimensions. A subsequent work along the same lines concluded that other splitting methods behave similarly, as long as they are randomly forced to be insensitive to some feature dimensions. Note that this observation of a more complex classifier (a larger forest) getting more accurate nearly monotonically is in sharp contrast to the common belief that the complexity of a classifier can
https://en.wikipedia.org/wiki/Virtex%20%28FPGA%29
Virtex is the flagship family of FPGA products currently developed by AMD, originally Xilinx before being acquired by the former. Other current product lines include Kintex (mid-range) and Artix (low-cost), each including configurations and models optimized for different applications. In addition, AMD offers the Spartan low-cost series, which continues to be updated and is nearing production utilizing the same underlying architecture and process node as the larger 7-series devices. Virtex FPGAs are typically programmed in hardware description languages such as VHDL or Verilog, using the Xilinx ISE or Vivado computer software. Xilinx FPGA products have been recognized by EE Times, EDN and others for innovation and market impact. Architecture The Virtex series of FPGAs are based on Configurable Logic Blocks (CLBs), where each CLB is equivalent to multiple ASIC gates. Each CLB is composed of multiple slices, that differ in construction between Virtex families. Virtex FPGAs include an I/O Block for controlling input/output pins on the Virtex chip, that support a variety of signalling standards. All pins default to "input" mode (high impedance). I/O pins are grouped into I/O Banks, where each Bank can support a different voltage. In addition to configurable FPGA logic, Virtex FPGAs include fixed-function hardware for multipliers, memories, microprocessor cores, FIFO and ECC logic, DSP blocks, PCI Express controllers, Ethernet MAC blocks, and high-speed serial transceivers. Some Virtex family members (such as the Virtex-5QX) are available in radiation-hardened packages, for outer-space applications. Families Virtex-E The Virtex-E family was introduced in September 1999 on a 180 nm process technology. Virtex-E includes a two million system gate device and supports twice the system-gate density and has a 50 percent higher I/O performance than the original Virtex FPGAs. Virtex-II Xilinx introduced Virtex-II family in January 2001 on 150 nm process technology, and V
https://en.wikipedia.org/wiki/Elastic%20scattering
Elastic scattering is a form of particle scattering in scattering theory, nuclear physics and particle physics. In this process, the kinetic energy of a particle is conserved in the center-of-mass frame, but its direction of propagation is modified (by interaction with other particles and/or potentials) meaning the two particles in the collision do not lose energy. Furthermore, while the particle's kinetic energy in the center-of-mass frame is constant, its energy in the lab frame is not. Generally, elastic scattering describes a process in which the total kinetic energy of the system is conserved. During elastic scattering of high-energy subatomic particles, linear energy transfer (LET) takes place until the incident particle's energy and speed has been reduced to the same as its surroundings, at which point the particle is "stopped". Rutherford scattering When the incident particle, such as an alpha particle or electron, is diffracted in the Coulomb potential of atoms and molecules, the elastic scattering process is called Rutherford scattering. In many electron diffraction techniques like reflection high energy electron diffraction (RHEED), transmission electron diffraction (TED), and gas electron diffraction (GED), where the incident electrons have sufficiently high energy (>10 keV), the elastic electron scattering becomes the main component of the scattering process and the scattering intensity is expressed as a function of the momentum transfer defined as the difference between the momentum vector of the incident electron and that of the scattered electron. Optical elastic scattering In Thomson scattering light interacts with electrons (this is the low-energy limit of Compton scattering). In Rayleigh scattering a medium composed of particles whose sizes are much smaller than the wavelength scatters light sideways. In this scattering process, the energy (and therefore the wavelength) of the incident light is conserved and only its direction is changed. In
https://en.wikipedia.org/wiki/Equatorial%20Rossby%20wave
Equatorial Rossby waves, often called planetary waves, are very long, low frequency water waves found near the equator and are derived using the equatorial beta plane approximation. Mathematics Using the equatorial beta plane approximation, , where β is the variation of the Coriolis parameter with latitude, . With this approximation, the primitive equations become the following: the continuity equation (accounting for the effects of horizontal convergence and divergence and written with geopotential height): the U-momentum equation (zonal component): the V-momentum equation (meridional component): In order to fully linearize the primitive equations, one must assume the following solution: Upon linearization, the primitive equations yield the following dispersion relation: , where c is the phase speed of an equatorial Kelvin wave (). Their frequencies are much lower than that of gravity waves and represent motion that occurs as a result of the undisturbed potential vorticity varying (not constant) with latitude on the curved surface of the earth. For very long waves (as the zonal wavenumber approaches zero), the non-dispersive phase speed is approximately: , which indicates that these long equatorial Rossby waves move in the opposite direction (westward) of Kelvin waves (which move eastward) with speeds reduced by factors of 3, 5, 7, etc. To illustrate, suppose c = 2.8 m/s for the first baroclinic mode in the Pacific; then the Rossby wave speed would correspond to ~0.9 m/s, requiring a 6-month time frame to cross the Pacific basin from east to west. For very short waves (as the zonal wavenumber increases), the group velocity (energy packet) is eastward and opposite to the phase speed, both of which are given by the following relations: Frequency relation: Group velocity: Thus, the phase and group speeds are equal in magnitude but opposite in direction (phase speed is westward and group velocity is eastward); note that is often useful to
https://en.wikipedia.org/wiki/Large%20cardinal
In the mathematical field of set theory, a large cardinal property is a certain kind of property of transfinite cardinal numbers. Cardinals with such properties are, as the name suggests, generally very "large" (for example, bigger than the least α such that α=ωα). The proposition that such cardinals exist cannot be proved in the most common axiomatization of set theory, namely ZFC, and such propositions can be viewed as ways of measuring how "much", beyond ZFC, one needs to assume to be able to prove certain desired results. In other words, they can be seen, in Dana Scott's phrase, as quantifying the fact "that if you want more you have to assume more". There is a rough convention that results provable from ZFC alone may be stated without hypotheses, but that if the proof requires other assumptions (such as the existence of large cardinals), these should be stated. Whether this is simply a linguistic convention, or something more, is a controversial point among distinct philosophical schools (see Motivations and epistemic status below). A is an axiom stating that there exists a cardinal (or perhaps many of them) with some specified large cardinal property. Most working set theorists believe that the large cardinal axioms that are currently being considered are consistent with ZFC. These axioms are strong enough to imply the consistency of ZFC. This has the consequence (via Gödel's second incompleteness theorem) that their consistency with ZFC cannot be proven in ZFC (assuming ZFC is consistent). There is no generally agreed precise definition of what a large cardinal property is, though essentially everyone agrees that those in the list of large cardinal properties are large cardinal properties. Partial definition A necessary condition for a property of cardinal numbers to be a large cardinal property is that the existence of such a cardinal is not known to be inconsistent with ZF and that such a cardinal Κ would be an uncountable initial ordinal for which LΚ
https://en.wikipedia.org/wiki/Ingo%20Alth%C3%B6fer
Ingo Althöfer (born 1961) is a German mathematician at the University of Jena, where he holds the chair of operations research. Althöfer earned his PhD in 1986 at Bielefeld University. His dissertation, Asymptotic Properties of Certain Competition Systems in Artificial Intelligence and Ecology, was supervised by Rudolf Ahlswede. Contributions Topics in Althöfer's professional research include the realization of finite metric spaces by shortest path metrics in graphs and their approximation by greedy spanners, algorithmic game theory and combinatorial game theory, and heuristic search algorithms for optimization problems. Althöfer is also known for his inventions of games and puzzles, including dice game EinStein würfelt nicht!, for his experiments with self-assembly of Lego building blocks by running them through a washing machine, and for his innovations in computer-human chess playing. In the 1990s he tested his "drei hirn" ["3-brains"] system, in which a human decides between the choices of two computer chess players, against strong human players including grandmaster David Bronstein and woman grandmaster Sofia Polgar. In 2004 he and Timo Klaustermeyer introduced freestyle chess, a style of human chess playing allowing arbitrary consultation with computers or other people. Books He has also self-published other books through his personal publishing company, 3-Hirn Verlag, and is one of the editors of the multi-volume book series Rudolf Ahlswede’s Lectures on Information Theory. Selected Papers Selected Board Games EinStein würfelt nicht! (2005) Galtoni (2012), a mixture of Connect Four and the Galton board San Jego (2015), a variant of Clobber
https://en.wikipedia.org/wiki/Injective%20function
In mathematics, an injective function (also known as injection, or one-to-one function) is a function that maps distinct elements of its domain to distinct elements; that is, implies . (Equivalently, implies in the equivalent contrapositive statement.) In other words, every element of the function's codomain is the image of one element of its domain. The term must not be confused with that refers to bijective functions, which are functions such that each element in the codomain is an image of exactly one element in the domain. A homomorphism between algebraic structures is a function that is compatible with the operations of the structures. For all common algebraic structures, and, in particular for vector spaces, an is also called a . However, in the more general context of category theory, the definition of a monomorphism differs from that of an injective homomorphism. This is thus a theorem that they are equivalent for algebraic structures; see for more details. A function that is not injective is sometimes called many-to-one. Definition Let be a function whose domain is a set The function is said to be injective provided that for all and in if then ; that is, implies Equivalently, if then in the contrapositive statement. Symbolically, which is logically equivalent to the contrapositive, Examples For visual examples, readers are directed to the gallery section. For any set and any subset the inclusion map (which sends any element to itself) is injective. In particular, the identity function is always injective (and in fact bijective). If the domain of a function is the empty set, then the function is the empty function, which is injective. If the domain of a function has one element (that is, it is a singleton set), then the function is always injective. The function defined by is injective. The function defined by is injective, because (for example) However, if is redefined so that its domain is the non-negative re
https://en.wikipedia.org/wiki/Scaleway
Scaleway (previously Online SAS or Online.net) is a French Cloud computing and web hosting company, founded by Xavier Niel in 1999 and a majority owned subsidiary of the Iliad group. The company provides physical dedicated servers and cloud computing architectures through Scaleway Dedibox and Scaleway Elements brands, domain registration services through the BookMyName brand, and colocation services in its datacenters through the brand Scaleway Datacenters. The company is the second player in France with over Tb/s of Internet traffic. History In 1999, Online started its activities in web hosting and domain name registration services In August 2002, the domain name registrar BookMyName has been bought by Iliad from the concurrent LDCom. In May 2006, rental of dedicated servers through the Dedibox brand was launched. In December 2008, Iliad bought Alice ADSL: They also took over construction and operation of Datacenters, launched in 1999 by ISDnet, bought by Cable & Wireless in January 2000 acquired by Tiscali France in June 2003 and finally renamed as Iliad Datacenter. In April 2010, Online merges with Dedibox, another subsidiary of Iliad, bringing together different hosting activities under a single brand. In 2012, the company has opened its third datacenter of 11800 m² in Vitry-sur-Seine after 11 months of construction works. The site received the first Tier-III certification in France by Uptime Institute in January 2014. Since 2012, the company publishes in real time the PUE of its datacenters on pue.online.net, in an effort of transparency. In 2013, Online launched labs.online.net in preview. An infrastructure as a service offer, based on dedicated hardware and without virtualization, based on ARM CPUs. The hardware is made in a factory near Laval in France. In April 2015 the service left its beta status and has been renamed as Scaleway. As the popularity of the platform grows, Online added servers with x86_64 based CPUs in March 2016. Repeat softwa
https://en.wikipedia.org/wiki/Invasive%20Species%20Act
Invasive Species Act may refer to: National Invasive Species Act, a 1996 United States federal law North Texas Invasive Species Barrier Act of 2014, a 2014 Texas state law in the United States Invasive Species Act (Ontario), a 2015 Ontario provincial law in Canada See also Alien Species Prevention and Enforcement Act of 1992 - a United States federal law British Columbia Weed Control Act Hazardous Substances and New Organisms Act 1996 of New Zealand Invasive species
https://en.wikipedia.org/wiki/International%20Society%20of%20Biomechanics
The International Society of Biomechanics, commonly known as the ISB, is a society dedicated to promoting biomechanics in its various forms. It promotes the study of all areas of biomechanics at the international level, although special emphasis is given to the biomechanics of human movement. The Society encourages international contacts amongst scientists, promotes the dissemination of knowledge, and forms liaisons with national organizations. The Society's membership includes scientists from a variety of disciplines including anatomy, physiology, engineering (mechanical, industrial aerospace, etc.), orthopedics, rehabilitation medicine, sport science and medicine, ergonomics, electro-physiological kinesiology and others. History The decision to establish the society was made at the 3rd International Seminar on Biomechanics held in Rome in 1971. This meeting was organized by the “Working Group on Biomechanics” which was part of the International Council of Sport and Physical Education, which itself was part of the United Nations Educational, Scientific, and Cultural Organization (UNESCO). At this meeting on September 29 it was voted to form the ISB at the next meeting. The 4th International Seminar on Biomechanics was held at Penn State University from August 26 until August 31, 1973. The constitution was voted on and approved on August 29. Two hundred and fifty of those present became charter members of the society. Executive Council The ISB is governed by its Executive Council. This council is elected every two years, by ballot, and is composed of officers and council members that represent countries from throughout the world and scientific areas that span all facets of biomechanics. The council, which meets annually, provides leadership for the continued development of the Society. Many on-going activities are performed by Council appointed sub-committees. The council also publishes a quarterly newsletter, known as ISB NOW, to inform members of
https://en.wikipedia.org/wiki/Action%20%28UML%29
In the Unified Modeling Language, an action is a named element that is the fundamental unit of executable functionality. Actions take a set of inputs, which may be empty, and convert them to a set of outputs, which in turn may also be empty. The execution of an action represents some transformation or processing in the modeled system. An action execution represents the run-time behavior of executing an action within a specific behavior execution. All action executions will be executions of specific kinds of actions because action is an abstract class. When the action executes, and what its actual inputs are, is determined by the concrete action and the behaviors in which it is used. An action is the specification of an executable statement and is the fundamental unit of processing or behavior in an activity node that represents some transformation in the modeled system. An action forms an abstraction of a computational procedure which is an atomic execution and therefore completes without interruption. An action is considered to take zero time and cannot be interrupted. In contrast, an activity is a more complex collection of behavior that may run for a long duration. An activity may be interrupted by events, in which case it does not run to completion. An action is a result of a system state change and is realized by sending a message to an object or modifying a link or a value of an attribute. An action may receive inputs in the form of control flows and object flows (the latter via input pins) and passes the results of its processing or transformations to one or more outgoing control flows or object flows (the latter via output pins) and onto downstream nodes. Execution of the action cannot begin until all its prerequisites are satisfied. All incoming control flows have control tokens and all input pins have object tokens. An action refers to the suite of rules and policies associated with a state machine state, and is represented as an object method.
https://en.wikipedia.org/wiki/Split-flap%20display
A split-flap display, or sometimes simply a flap display, is a digital electromechanical display device that presents changeable alphanumeric text, and occasionally fixed graphics. Often used as a public transport timetable in airports or railway stations, as such they are often called Solari boards after Italian display manufacturer Solari di Udine, or in Central European countries they are called Pragotron after the Czech manufacturer. Split-flap displays were once commonly used in consumer digital clocks known as flip clocks. Description Each character position or graphic position has a collection of flaps on which the characters or graphics are painted or silkscreened. These flaps are precisely rotated to show the desired character or graphic. These displays are often found in railway stations and airports, where they serve as flight information display system and typically display departure or arrival information. Sometimes the flaps are large and display whole words, and in other installations there are several smaller flaps, each displaying a single character. Flip-dot displays and LED display boards may be used instead of split-flap displays in most applications. Their output can be changed by reprogramming instead of replacement of physical parts but they suffer from lower readability. They also can refresh more quickly, as a split-flap display often must cycle through many states. Advantages to these displays include: high visibility and wide viewing angle in most lighting conditions little or no power consumption while the display remains static Distinct metallic flapping sound draws attention when the information is updated. The Massachusetts Bay Transportation Authority has designed the new LED replacements for its aging Solari boards at North Station and South Station to emit an electronically generated flapping noise to cue passengers to train boarding updates. Many game shows of the 1970s used this type of display for the contestant podiu
https://en.wikipedia.org/wiki/Frank%20Kelly%20%28mathematician%29
Francis Patrick Kelly, CBE, FRS (born 28 December 1950) is Professor of the Mathematics of Systems at the Statistical Laboratory, University of Cambridge. He served as Master of Christ's College, Cambridge from 2006 to 2016. Kelly's research interests are in random processes, networks and optimisation, especially in very large-scale systems such as telecommunication or transportation networks. In the 1980s, he worked with colleagues in Cambridge and at British Telecom's Research Labs on Dynamic Alternative Routing in telephone networks, which was implemented in BT's main digital telephone network. He has also worked on the economic theory of pricing to congestion control and fair resource allocation in the internet. From 2003 to 2006 he served as Chief Scientific Advisor to the United Kingdom Department for Transport. Kelly was elected a Fellow of the Royal Society in 1989. In December 2006 he was elected 37th Master of Christ's College, Cambridge. He was appointed Commander of the Order of the British Empire (CBE) in the 2013 New Year Honours for services to mathematical science. Awards 1979 Davidson Prize of the University of Cambridge 1989 Guy Medal in Silver of the Royal Statistical Society 1989 Fellow of the Royal Society 1992 Lanchester Prize of the Institute for Operations Research and the Management Sciences 1997 Naylor Prize of the London Mathematical Society 2001 Honorary D.Sc. from Heriot-Watt University 2005 IEEE Koji Kobayashi Computers and Communications Award 2006 Companionship of OR by the Operational Research Society 2008 John von Neumann Theory Prize of the Institute for Operations Research and the Management Sciences 2009 SIGMETRICS Achievement Award 2009 EURO Gold Medal from European Operational Research Society 2013 Commander of the Order of the British Empire in the New Years Honours List for "services to mathematical sciences" 2015 IEEE Alexander Graham Bell Medal, for "Creating principled mathematical foundations for the d
https://en.wikipedia.org/wiki/Zinc%20in%20biology
Zinc is an essential trace element for humans and other animals, for plants and for microorganisms. Zinc is required for the function of over 300 enzymes and 1000 transcription factors, and is stored and transferred in metallothioneins. It is the second most abundant trace metal in humans after iron and it is the only metal which appears in all enzyme classes. In proteins, zinc ions are often coordinated to the amino acid side chains of aspartic acid, glutamic acid, cysteine and histidine. The theoretical and computational description of this zinc binding in proteins (as well as that of other transition metals) is difficult. Roughly  grams of zinc are distributed throughout the human body. Most zinc is in the brain, muscle, bones, kidney, and liver, with the highest concentrations in the prostate and parts of the eye. Semen is particularly rich in zinc, a key factor in prostate gland function and reproductive organ growth. Zinc homeostasis of the body is mainly controlled by the intestine. Here, ZIP4 and especially TRPM7 were linked to intestinal zinc uptake essential for postnatal survival. In humans, the biological roles of zinc are ubiquitous. It interacts with "a wide range of organic ligands", and has roles in the metabolism of RNA and DNA, signal transduction, and gene expression. It also regulates apoptosis. A review from 2015 indicated that about 10% of human proteins (~3000) bind zinc, in addition to hundreds more that transport and traffic zinc; a similar in silico study in the plant Arabidopsis thaliana found 2367 zinc-related proteins. In the brain, zinc is stored in specific synaptic vesicles by glutamatergic neurons and can modulate neuronal excitability. It plays a key role in synaptic plasticity and so in learning. Zinc homeostasis also plays a critical role in the functional regulation of the central nervous system. Dysregulation of zinc homeostasis in the central nervous system that results in excessive synaptic zinc concentrations is believed
https://en.wikipedia.org/wiki/Nokia%20Business%20Center
Nokia Business Center (NBC) was a mobile email solution by Nokia, providing push e-mail and (through a paid-for client upgrade) calendar and contact availability to mobile devices. The server runs on Red Hat Enterprise Linux. It was discontinued in 2014. External links Press Release about support for IBM Lotus Notes and Domino addition to NBC Nokia services Mobile web
https://en.wikipedia.org/wiki/Nesfatin-1
Nesfatin-1 is a neuropeptide produced in the hypothalamus of mammals. It participates in the regulation of hunger and fat storage. Increased nesfatin-1 in the hypothalamus contributes to diminished hunger, a 'sense of fullness', and a potential loss of body fat and weight. A study of metabolic effects of nesfatin-1 in rats was done in which subjects administered nesfatin-1 ate less, used more stored fat and became more active. Nesfatin-1-induced inhibition of feeding may be mediated through the inhibition of orexigenic neurons. In addition, the protein stimulated insulin secretion from the pancreatic beta cells of both rats and mice. Biochemistry Nesfatin-1 is a polypeptide encoded in the N-terminal region of the protein precursor, Nucleobindin-2 (NUCB2). Recombinant human Nesfatin-1 is a 9.7 kDa protein containing 82 amino acid residues. Nesfatin-1 is expressed in the hypothalamus, in other areas of the brain, and in pancreatic islets, gastric endocrine cells and adipocytes. Satiety Nesfatin/NUCB2 is expressed in the appetite-control hypothalamic nuclei such as paraventricular nucleus (PVN), arcuate nucleus (ARC), supraoptic nucleus (SON) of hypothalamus, lateral hypothalamic area (LHA), and zona incerta in rats. Nesfatin-1 immunoreactivity was also found in the brainstem nuclei such as nucleus of the solitary tract (NTS) and Dorsal nucleus of vagus nerve. Brain Nesfatin-1 can cross the blood–brain barrier without saturation. The receptors within the brain are in the hypothalamus and the solitary nucleus, where nesfatin-1 is believed to be produced via peroxisome proliferator-activated receptors (PPARs). It appears there is a relationship between nesfatin-1 and cannabinoid receptors. Nesfatin-1-induced inhibition of feeding may be mediated through the inhibition of orexigenic NPY neurons. Nesfatin/NUCB2 expression has been reported to be modulated by starvation and re-feeding in the Paraventricular nucleus (PVN) and supraoptic nucleus (SON) of the brain. Ne
https://en.wikipedia.org/wiki/Trichosanthin
Trichosanthin is a ribosome-inactivating protein. It is derived from Trichosanthes kirilowii. It is also an abortifacient.
https://en.wikipedia.org/wiki/Roy%20Radner
Roy Radner (June 29, 1927 – October 6, 2022) was Leonard N. Stern School Professor of Business at New York University. He was a micro-economic theorist. Radner's research interests included strategic analysis of climate change, bounded rationality, game-theoretic models of corruption, pricing of information goods and statistical theory of data mining. Previously he was a faculty member at the University of California, Berkeley, and a Distinguished Member of Technical Staff at AT&T Bell Laboratories. Life and career Roy Radner received his Ph.B. in the liberal arts from the University of Chicago in 1945. Continuing his education at the University of Chicago, Radner went on to receive a B.S. in mathematics in 1950, an M.S. in mathematics in 1951, and his Ph.D. in mathematical statistics in 1956. He died on October 6, 2022, at Pennswood Village in Newtown, Bucks County, Pennsylvania, aged 95. Radner equilibrium Among Radner's various contributions, the one that bears his name, Radner equilibrium (1968), is a model of financial markets. In the traditional approach if the value of an asset or a contingent claim is affordable then it can be achieved. Not so with incomplete market as the payoff has to be replicable by trading of available assets that are now part of the definition of the economy. The first consequence of such a requirement is that budget sets do not fill the available space and are typically smaller than hyperplanes. Because the dimension of vectors orthogonal to the budget set is larger than one there is no reason for the price systems supporting an equilibrium to be unique up to scaling, likewise the first order conditions no longer implies that gradient of agents are collinear at equilibrium. Both happen to fail to hold generically: the first theorem of welfare economics is hence the first victim of incompleteness. Pareto-optimality of equilibria generally does not hold. In traditional complete markets any policy would be undone through trading of
https://en.wikipedia.org/wiki/Self-confrontation%20method
In psychology, the self-confrontation method (SCM), developed by Hubert Hermans, is a technique for examining people's behavior modification. It relies on people's inconsistent knowledge and dissatisfaction with their own values, motivation, behaviors, or with their personal meaning systems and those of significant others. Self-confrontation psychology is based on two theories which are valuation theory and dialogical-self theory. Background The self-confrontation method is a specific evaluation and intervention tool guided which focuses on the individual's feelings and motivation. SCM is influenced by James's work (1890) and Merleau-Pony (1945,1962), as well as Bruner(1986) and Sarbin(1986). SCM is organized into a narrative structure when Hermans developed the Valuation Theory as a framework for research of personal experience over time (1987-1989). Self-Confrontation is not limited to one field of practice it can be used in different purposes such as psychotherapy and psychology. Process The self-confrontation method is a relational process that involves researching and organising the client or subject's valuations, including the ways in which these valuations are reorganized over time. The process has been separated into three parts: 1) valuation elicitation; 2) affective rating; 3) discussion of the results. Firstly, the construction of the client's valuation system is happening in the consultation between the client and psychologist. The consultation starts with open questions which refer to the client's past, present and future. In the second phase of the SCM, a list of affect items about the basic motives (Self and Other) is provided to the client who is invited to evaluate the emotional quality of the different valuations. The S-motive is expressed in self-esteem, strength, self-confidence, and pride, whereas the O-motive is expressed in tenderness, caring, love, and intimacy. In the last phase, there is a deep discussion about the content and organi
https://en.wikipedia.org/wiki/BKS%20theory
The Bohr–Kramers–Slater theory (BKS theory) was perhaps the final attempt at understanding the interaction of matter and electromagnetic radiation on the basis of the so-called old quantum theory, in which quantum phenomena are treated by imposing quantum restrictions on classically describable behaviour. It was advanced in 1924, and sticks to a classical wave description of the electromagnetic field. It was perhaps more a research program than a full physical theory, the ideas that are developed not being worked out in a quantitative way. The purpose of BKS Theory was to disprove Einstein's hypothesis of the light quantum. One aspect, the idea of modelling atomic behaviour under incident electromagnetic radiation using "virtual oscillators" at the absorption and emission frequencies, rather than the (different) apparent frequencies of the Bohr orbits, significantly led Max Born, Werner Heisenberg and Hendrik Kramers to explore mathematics that strongly inspired the subsequent development of matrix mechanics, the first form of modern quantum mechanics. The provocativeness of the theory also generated great discussion and renewed attention to the difficulties in the foundations of the old quantum theory. However, physically the most provocative element of the theory, that momentum and energy would not necessarily be conserved in each interaction but only overall, statistically, was soon shown to be in conflict with experiment. Origins When Einstein introduced the light quantum (photon) in 1905, there was much resistance from the scientific community. However, when in 1923, the Compton effect showed the results could be explained by assuming the light beam behaves as light-quanta and that energy and momentum are conserved, Bohr was still resistant against quantized light, even repudiating it in his 1922 Nobel Prize lecture. So Bohr found a way of using Einstein's approach without also using the light-quantum hypothesis by reinterpreting the principles of energy an
https://en.wikipedia.org/wiki/Portevin%E2%80%93Le%20Chatelier%20effect
The Portevin–Le Chatelier (PLC) effect describes a serrated stress–strain curve or jerky flow, which some materials exhibit as they undergo plastic deformation, specifically inhomogeneous deformation. This effect has been long associated with dynamic strain aging or the competition between diffusing solutes pinning dislocations and dislocations breaking free of this stoppage. The onset of the PLC effect occurs when the strain rate sensitivity becomes negative and inhomogeneous deformation starts. This effect also can appear on the specimen's surface and in bands of plastic deformation. This process starts at a so-called critical strain, which is the minimum strain needed for the onset of the serrations in the stress–strain curve. The critical strain is both temperature and strain rate dependent. The existence of a critical strain is attributed to better solute diffusivity due to the deformation created vacancies and increased mobile dislocation density. Both of these contribute to the instability in substitutional alloys, while interstitial alloys are only affected by the increase in mobile dislocation densities. History While the effect is named after Albert Portevin and François Le Chatelier, they were not the first to discover it. Félix Savart made the discovery when he observed non-homogeneous deformation during a tensile test of copper strips. He documented the physical serrations in his samples that are currently known as Portevin–Le Chatelier bands. A student of Savart, Antoine Masson, repeated the experiment while controlling for loading rate. Masson observed that under a constant loading rate, the samples would experience sudden large changes in elongation (as large as a few millimeters). Underlying physics Much of the underlying physics of the Portevin-Le Chatelier effect lies in a specific case of solute drag creep. Adding solute atoms to a pure crystal introduces a size misfit into the system. This size misfit leads to restriction of dislocation mo
https://en.wikipedia.org/wiki/Expand%20Networks
Expand Networks, Ltd. was a Tel Aviv, Israel based provider of WAN optimization technology founded in 1998 and liquidated in 2011. About Expand Networks was a privately held company, co-founded by Talmon Marco in 1998; initial financing was provided by Discount Investment Corporation Ltd., The Eurocom Group, Ophir Holdings, and a private group of investors, including Memco Software founder Israel Mezin. Additional investors joined in subsequent rounds of funding. The company raised over $95 million. Expand Networks headquarters was in Tel-Aviv, Israel with sales in the United States and Europe, New Jersey, Australia, China, Singapore, and South Africa. The company manufactured accelerators in physical, virtual and mobile deployment options. Liquidation In mid October 2011, following the requests of Plenus, one of the company's lenders, an Israeli court appointed a liquidator - Paz Rimer. The liquidator gradually terminated the company's employees and eventually, on 11 January 2012 sold most of the assets of the company to Riverbed Technology, which immediately terminated all the company's products and ceased support. External links Expand Networks Home Page Expand Networks reassures partners it's business as usual
https://en.wikipedia.org/wiki/Slow-scan%20television
Slow-scan television (SSTV) is a picture transmission method, used mainly by amateur radio operators, to transmit and receive static pictures via radio in monochrome or color. A literal term for SSTV is narrowband television. Analog broadcast television requires at least 6 MHz wide channels, because it transmits 25 or 30 picture frames per second (see ITU analog broadcast standards), but SSTV usually only takes up to a maximum of 3 kHz of bandwidth. It is a much slower method of still picture transmission, usually taking from about eight seconds to a couple of minutes, depending on the mode used, to transmit one image frame. Since SSTV systems operate on voice frequencies, amateurs use it on shortwave (also known as HF by amateur radio operators), VHF and UHF radio. History Concept The concept of SSTV was introduced by Copthorne Macdonald in 1957–58. He developed the first SSTV system using an electrostatic monitor and a vidicon tube. It was deemed sufficient to use 120 lines and about 120 pixels per line to transmit a black-and-white still picture within a 3 kHz telephone channel. First live tests were performed on the 11-meter ham band which was later given to the CB service in the US. In the 1970s, two forms of paper printout receivers were invented by hams. Early usage in space exploration SSTV was used to transmit images of the far side of the Moon from Luna 3. The first space television system was called Seliger-Tral-D and was used aboard Vostok. Vostok was based on an earlier videophone project which used two cameras, with persistent LI-23 iconoscope tubes. Its output was 10 frames per second at 100 lines per frame video signal. The Seliger system was tested during the 1960 launches of the Vostok capsule, including Sputnik 5, containing the space dogs Belka and Strelka, whose images are often mistaken for the dog Laika, and the 1961 flight of Yuri Gagarin, the first man in space on Vostok 1. Vostok 2 and thereafter used an improved 400-line televisi
https://en.wikipedia.org/wiki/Conjugated%20oligoelectrolytes
Conjugated oligoelectrolytes, or COEs, are a class of synthetic antimicrobials designed to prevent and circumvent antimicrobial resistance via different mechanism of action than traditional antibiotics. COEs insert into cell membranes and can function as electron transporters, but were found to inhibit bacterial growth. They can also be used for tracking the progress of tumor growth.
https://en.wikipedia.org/wiki/Population%20reconstruction
Population reconstruction is a method used by historical demographers. Using records such as church registries the size and composition of families living in a given region in a given past time is determined. This allows the identification and analysis of patterns of family formation, fertility, mortality, and migration, and of consequent trends such as population growth. See also One-place study
https://en.wikipedia.org/wiki/Nukernel
NuKernel is a microkernel which was developed at Apple Computer during the early 1990s. Written from scratch and designed using concepts from the Mach 3.0 microkernel, with extensive additions for soft real-time scheduling to improve multimedia performance, it was the basis for the Copland operating system. Only one NuKernel version was released, with a Copland alpha release. Development ended in 1996 with the cancellation of Copland. The External Reference Specification (ERS) for NuKernel is contained in its entirety in its patent. The one-time technical lead for NuKernel, Jeff Robbin, was one of the leaders of iTunes and the iPod. Apple's NuKernel is not the microkernel in BeOS, nukernel. See also XNU, the microkernel in Mac OS X
https://en.wikipedia.org/wiki/Flag%20of%20Kiribati
The flag of Kiribati (Gilbertese: buraki ni Kiribati) is red in the upper half with a gold frigatebird (Fregata minor, in Gilbertese: te eitei) flying over a gold rising sun (otintaai), and the lower half is blue with three horizontal wavy white stripes to represent the ocean and the three archipelagoes (Gilbert, Phoenix and Line Islands). The 17 rays of the sun represent the 16 Gilbert Islands and Banaba (former Ocean Island). The yellow frigatebird symbolises command over the sea, freedom, and dance patterns. The blue and white wavy bands represent the Pacific Ocean, which surrounds Kiribati, and the sun refers to Kiribati's position astride the Equator. The flag is derived from a badge designed by Sir Arthur Grimble in 1931 for the flag of the Gilbert and Ellice Islands British colony and granted in 1937. Official description Following the schedule 2 (section 8) of the National Identity Act of 1989, the official description is this one: "The bird is a frigate bird which represents power, freedom and Kiribati cultural dance patterns". "The rising sun is the tropical sun as Kiribati lies astride the Equator". "The sea is the Pacific Ocean which surrounds Kiribati". History Kiribati's flag is an armorial banner, a flag having a design corresponding exactly to that of the shield in the coat of arms, the former badge of the flag of the British colony of the Gilbert and Ellice Islands. In a letter on 15 August 1931 to the Western Pacific High Commissioner, Arthur Grimble has written (concerning his badge proposal): "Sea. Represents the space of the ocean in which the colony is isolated". "Sun. The islands lie close to the 180 degree meridian. Sun is either rising or setting". "Bird. Frigate bird (Fregat Aquilla (sic)) symbolises power, poise and freedom. To the natives it is a sign of sovereignty and kingly birth, and highly regarded as such". The coat of arms dates was then granted by the College of Arms in May 1937 when it was officially granted to the colony
https://en.wikipedia.org/wiki/Holmium%20phosphide
Holmium phosphide is a binary inorganic compound of holmium and phosphorus with the chemical formula HoP. The compound forms dark crystals and does not dissolve in water. Synthesis Heating powdered holmium and red phosphorus in an inert atmosphere or vacuum: 4Ho + P4 -> 4HoP Properties Holmium phosphide forms dark crystals of a cubic system, stable in air, does not dissolve in water. HoP belongs to the large class of NaCl-structured rare earth monopnictides. Ferromagnetic at low temperatures. HoP actively reacts with nitric acid. Uses The compound is a semiconductor used in high power, high frequency applications and in laser diodes.
https://en.wikipedia.org/wiki/Two-point%20discrimination
Two-point discrimination (2PD) is the ability to discern that two nearby objects touching the skin are truly two distinct points, not one. It is often tested with two sharp points during a neurological examination and is assumed to reflect how finely innervated an area of skin is. In clinical settings, two-point discrimination is a widely used technique for assessing tactile perception. It relies on the ability and/or willingness of the patient to subjectively report what they are feeling and should be completed with the patient’s eyes closed. The therapist may use calipers or simply a reshaped paperclip to do the testing. The therapist may alternate randomly between touching the patient with one point or with two points on the area being tested (e.g. finger, arm, leg, toe). The patient is asked to report whether one or two points was felt. The smallest distance between two points that still results in the perception of two distinct stimuli is recorded as the patient's two-point threshold. Performance on the two extremities can be compared for discrepancies. Although the test is still commonly used clinically, it has been roundly criticized by many researchers as providing an invalid measure of tactile spatial acuity, and several highly regarded alternative tests have been proposed to replace it. Normal and impaired performance Body areas differ both in tactile receptor density and somatosensory cortical representation. Normally, a person should be able to recognize two points separated by 2 to 8 mm on fingertips. On the lips, it is 2 to 4 mm, and on the palms, it is 8 to 12 mm and 30–40 mm on the shins or back (assuming the points are at the same dermatome). The posterior column-medial lemniscus pathway is responsible for carrying information involving fine, discriminative touch. Therefore, two-point discrimination can be impaired by damage to this pathway or to a peripheral nerve. Criticisms Although two-point testing is commonly used clinically, evidence acc
https://en.wikipedia.org/wiki/Optical%20molasses
Optical molasses is a laser cooling technique that can cool neutral atoms to temperatures lower than a magneto-optical trap (MOT). An optical molasses consists of 3 pairs of counter-propagating circularly polarized laser beams intersecting in the region where the atoms are present. The main difference between optical molasses and a MOT is the absence of magnetic field in the former. Therefore, unlike a MOT, an optical molasses provides only cooling and no trapping. While a typical Sodium MOT can cool atoms down to 300μK, optical molasses can cool the atoms down to 40μK, an order of magnitude colder. History When laser cooling was proposed in 1975, a theoretical limit on the lowest possible temperature was predicted. Known as the Doppler limit, , this was given by the lowest possible temperature attainable considering the cooling of two-level atoms by Doppler cooling and the heating of atoms due to momentum diffusion from the scattering of laser photons. Here, , is the natural line-width of the atomic transition, , is the reduced Planck constant and, , is the Boltzmann constant. Experiments at the National Institute of Standards and Technology, Gaithersburg, found the temperature of cooled atoms to be well below the theoretical limit. Initially, it was a surprise to theorists, until the full explanation came out. In 1985, Chu et al. directed a thermal beam of sodium atoms through an optical molasses region formed by counter-propagating light from a frequency-doubled Neodymium:YAlG laser. They measured the average temperature by quickly shutting off the beams and measuring the fluorescence of the released atoms. They measured an average temperature of , near the Doppler cooling limit of sodium. Theory The best explanation of the phenomenon of optical molasses is based on the principle of polarization gradient cooling. Counterpropagating beams of circularly polarized light cause a standing wave, where the light polarization is linear but the direction rotates alo
https://en.wikipedia.org/wiki/ALGOL%2068C
ALGOL 68C is an imperative computer programming language, a dialect of ALGOL 68, that was developed by Stephen R. Bourne and Michael Guy to program the Cambridge Algebra System (CAMAL). The initial compiler was written in the Princeton Syntax Compiler (PSYCO, by Edgar T. Irons) that was implemented by J. H. Mathewman at Cambridge. ALGOL 68C was later used for the CHAOS OS for the capability-based security CAP computer at University of Cambridge in 1971. Other early contributors were Andrew D. Birrell and Ian Walker. Subsequent work was done on the compiler after Bourne left Cambridge University in 1975. Garbage collection was added, and the code base is still running on an emulated OS/MVT using Hercules. The ALGOL 68C compiler generated output in ZCODE, a register-based intermediate language, which could then be either interpreted or compiled to a native executable. This ability to interpret or compile ZCODE encouraged the porting of ALGOL 68C to many different computing platforms. Aside from the CAP computer, the compiler was ported to systems including Conversational Monitor System (CMS), TOPS-10, and Zilog Z80. Popular culture A very early predecessor of this compiler was used by Guy and Bourne to write the first Game of Life programs on the PDP-7 with a DEC 340 display. Various Liverpool Software Gazette issues detail the Z80 implementation. The compiler required about 120 KB of memory to run; hence the Z80's 64 KB memory is actually too small to run the compiler. So ALGOL 68C programs for the Z80 had to be cross-compiled from the larger CAP computer, or an IBM System/370 mainframe computer. Algol 68C and Unix Stephen Bourne subsequently reused ALGOL 68's if ~ then ~ else ~ fi, case ~ in ~ out ~ esac and for ~ while ~ do ~ od clauses in the common Unix Bourne shell, but with in's syntax changed, out removed, and od replaced with done (to avoid conflict with the od utility). After Cambridge, Bourne spent nine years at Bell Labs with the Version 7 Unix (Se
https://en.wikipedia.org/wiki/Ally%20Sloper
Alexander "Ally" Sloper is the eponymous fictional character of the British comic strip Ally Sloper. First appearing in 1867, he is considered one of the earliest comic strip characters and he is regarded as the first recurring character in comics. Red-nosed and blustery, an archetypal lazy schemer often found "sloping" through alleys to avoid his landlord and other creditors, he was created for the British magazine Judy by writer and fledgling artist Charles H. Ross, and inked and later fully illustrated by his French wife Émilie de Tessier under the pseudonym "Marie Duval" (or "Marie Du Val"; sources differ). The strips, which used text narrative beneath unbordered panels, premiered in the 14 August 1867 issue of Judy, a humour-magazine rival of the famous Punch. The highly popular character was spun off into his own comic, Ally Sloper's Half Holiday, in 1884. Artists The first illustrations were by Ross, then Tessier took over. When publisher Gilbert Dalziel re-launched the cartoon as Ally Sloper's Half Holiday, in 1884, Sloper was illustrated by William Baxter until his death in 1888. He was succeeded by W. Fletcher Thomas, who continued the illustrations until approximately 1899, when the publisher invited C. H. Chapman to illustrate the series until it ended in 1916. Films The highly popular character was spun off into his own comic, Ally Sloper's Half Holiday in 1884. Sloper appeared in three feature films and a wide array of merchandising from pocket watches to door stops. His popularity and influence led to his being used on occasion as a propaganda tool for the British government's policies. Sloper has also been cited as an influence on W. C. Fields and Charlie Chaplin's "little tramp" character and its imitators. The arrival of the First World War in 1914 saw severe paper rationing, and in 1916 the Half Holiday comic ceased production. Attempts after the war to revive Sloper proved short-lived, as Sloper was a somewhat stereotypical Victorian and Ed
https://en.wikipedia.org/wiki/Client-side%20persistent%20data
Client-side persistent data or CSPD is a term used in computing for storing data required by web applications to complete internet tasks on the client-side as needed rather than exclusively on the server. As a framework it is one solution to the needs of Occasionally connected computing or OCC. A major challenge for HTTP as a stateless protocol has been asynchronous tasks. The AJAX pattern using XMLHttpRequest was first introduced by Microsoft in the context of the Outlook e-mail product. The first CSPD were the 'cookies' introduced by the Netscape Navigator. ActiveX components which have entries in the Windows registry can also be viewed as a form of client-side persistence. See also Occasionally connected computing Curl (programming_language) AJAX HTTP Web storage External links CSPD Safari preview Netscape on persistent client state Clients (computing) Data management Web applications
https://en.wikipedia.org/wiki/Image%20analysis
Image analysis or imagery analysis is the extraction of meaningful information from images; mainly from digital images by means of digital image processing techniques. Image analysis tasks can be as simple as reading bar coded tags or as sophisticated as identifying a person from their face. Computers are indispensable for the analysis of large amounts of data, for tasks that require complex computation, or for the extraction of quantitative information. On the other hand, the human visual cortex is an excellent image analysis apparatus, especially for extracting higher-level information, and for many applications — including medicine, security, and remote sensing — human analysts still cannot be replaced by computers. For this reason, many important image analysis tools such as edge detectors and neural networks are inspired by human visual perception models. Digital Digital Image Analysis or Computer Image Analysis is when a computer or electrical device automatically studies an image to obtain useful information from it. Note that the device is often a computer but may also be an electrical circuit, a digital camera or a mobile phone. It involves the fields of computer or machine vision, and medical imaging, and makes heavy use of pattern recognition, digital geometry, and signal processing. This field of computer science developed in the 1950s at academic institutions such as the MIT A.I. Lab, originally as a branch of artificial intelligence and robotics. It is the quantitative or qualitative characterization of two-dimensional (2D) or three-dimensional (3D) digital images. 2D images are, for example, to be analyzed in computer vision, and 3D images in medical imaging. The field was established in the 1950s—1970s, for example with pioneering contributions by Azriel Rosenfeld, Herbert Freeman, Jack E. Bresenham, or King-Sun Fu. Techniques There are many different techniques used in automatically analysing images. Each technique may be useful for a small
https://en.wikipedia.org/wiki/Father%20Time
Father Time is a personification of time. In recent centuries he is usually depicted as an elderly bearded man, sometimes with wings, dressed in a robe and carrying a scythe and an hourglass or other timekeeping device. As an image, "Father Time's origins are curious." The ancient Greeks themselves began to associate chronos, their word for time, with the god Chronos, who had the attribute of a harvester's sickle. The Romans equated Cronos with Saturn, who also had a sickle, and was treated as an old man, often with a crutch. The wings and hourglass were early Renaissance additions and he eventually became a companion of the Grim Reaper, personification of Death, often taking his scythe. He may have as an attribute a snake with its tail in its mouth, an ancient Egyptian symbol of eternity. New Year Around New Year's Eve, the media (in particular editorial cartoons) use the convenient trope of Father Time as the personification of the previous year (or "the Old Year") who typically "hands over" the duties of time to the equally allegorical Baby New Year (or "the New Year") or who otherwise characterizes the preceding year. In these depictions, Father Time is usually depicted wearing a sash with the old year's date on it. Time (in his allegorical form) is often depicted revealing or unveiling the allegorical Truth, sometimes at the expense of a personification of Falsehood, Fraud, or Envy. This theme is related to the idea of veritas filia temporis (Time is the father of Truth). In the arts Father Time is an established symbol in numerous cultures and appears in a variety of art and media. In some cases, they appear specifically as Father Time while in other cases they may have another name (such as Saturn), but the characters demonstrate the attributes which Father Time has acquired over the centuries. Art Paintings Venus, Cupid, Folly, and Time, a c.1545 painting by Agnolo Bronzino, National Gallery, London. An Allegory of Truth and Time, a 1584–85 painting
https://en.wikipedia.org/wiki/Automation%20in%20construction
Automation in construction is the combination of methods, processes, and systems that allow for greater machine autonomy in construction activities. Construction automation may have multiple goals, including but not limited to, reducing jobsite injuries, decreasing activity completion times, and assisting with quality control and quality assurance. Some systems may be fielded as a direct response to increasing skilled labor shortages in some countries. Opponents claim that increased automation may lead to less construction jobs and that software leaves heavy equipment vulnerable to hackers. Transportation Construction Kratos Defense & Security Solutions fielded the world’s first Autonomous Truck-Mounted Attenuator (ATMA) in 2017, in conjunction with Royal Truck & Equipment. Uses of Automation in Construction Equipment control and management: Automation can be used to control and monitor construction equipment, such as cranes, excavators, and bulldozers. Material handling: Automated systems can be used to handle, transport, and place materials such as concrete, bricks, and stones. Surveying: Automated survey equipment and drones can be used to collect and analyze data on construction sites. Quality control: Automated systems can be used to monitor and control the quality of materials and construction processes. Safety management: Automated systems can be used to monitor and control safety conditions on construction sites. Scheduling and planning: Automated systems can be used to manage schedules, resources, and costs. Waste management: Automated systems can be used to manage and dispose of waste materials generated during construction. 3D printing: Automated 3D printing can be used to create prototypes, models, and even full-scale building components. Benefits of Automation in Construction The use of automation in construction has become increasingly prevalent in recent years due to its numerous benefits. Automation in construction refers to the use of
https://en.wikipedia.org/wiki/Tutte%20graph
In the mathematical field of graph theory, the Tutte graph is a 3-regular graph with 46 vertices and 69 edges named after W. T. Tutte. It has chromatic number 3, chromatic index 3, girth 4 and diameter 8. The Tutte graph is a cubic polyhedral graph, but is non-hamiltonian. Therefore, it is a counterexample to Tait's conjecture that every 3-regular polyhedron has a Hamiltonian cycle. Published by Tutte in 1946, it is the first counterexample constructed for this conjecture. Other counterexamples were found later, in many cases based on Grinberg's theorem. Construction From a small planar graph called the Tutte fragment, W. T. Tutte constructed a non-Hamiltonian polyhedron, by putting together three such fragments. The "compulsory" edges of the fragments, that must be part of any Hamiltonian path through the fragment, are connected at the central vertex; because any cycle can use only two of these three edges, there can be no Hamiltonian cycle. The resulting graph is 3-connected and planar, so by Steinitz' theorem it is the graph of a polyhedron. It has 25 faces. It can be realized geometrically from a tetrahedron (the faces of which correspond to the four large nine-sided faces in the drawing, three of which are between pairs of fragments and the fourth of which forms the exterior) by multiply truncating three of its vertices. Algebraic properties The automorphism group of the Tutte graph is Z/3Z, the cyclic group of order 3. The characteristic polynomial of the Tutte graph is : Related graphs Although the Tutte graph is the first 3-regular non-Hamiltonian polyhedral graph to be discovered, it is not the smallest such graph. In 1965 Lederberg found the Barnette–Bosák–Lederberg graph on 38 vertices. In 1968, Grinberg constructed additional small counterexamples to the Tait's conjecture – the Grinberg graphs on 42, 44 and 46 vertices. In 1974 Faulkner and Younger published two more graphs – the Faulkner–Younger graphs on 42 and 44 vertices. Finally Holton an
https://en.wikipedia.org/wiki/Confrontation%20analysis
Confrontation analysis (also known as dilemma analysis) is an operational analysis technique used to structure, understand and think through multi-party interactions such as negotiations. It is the underpinning mathematical basis of drama theory. It is derived from game theory but considers that instead of resolving the game, the players often redefine the game when interacting. Emotions triggered from the potential interaction play a large part in this redefinition. So whereas game theory looks on an interaction as a single decision matrix and resolves that, confrontation analysis looks on the interaction as a sequence of linked interactions, where the decision matrix changes under the influence of precisely defined emotional dilemmas. Derivation and use Confrontation analysis was devised by Professor Nigel Howard in the early 1990s drawing from his work on game theory and metagame analysis. It has been turned to defence, political, legal, financial and commercial applications. Much of the theoretical background to General Rupert Smith's book The Utility of Force drew its inspiration from the theory of confrontation analysis. Confrontation analysis can also be used in a decision workshop as structure to support role-playing for training, analysis and decision rehearsal. Method Confrontation analysis looks on an interaction as a sequence of confrontations. During each confrontation the parties communicate until they have made their positions clear to one another. These positions can be expressed as a card table (also known as an options board) of yes/no decisions. For each decision each party communicates what they would like to happen (their position) and what will happen if they cannot agree (the threatened future). These interactions produce dilemmas and the card table changes as players attempt to eliminate these. Consider the example on the right (Initial Card Table), taken from the 1995 Bosnian Conflict. This represents an interaction between the B
https://en.wikipedia.org/wiki/Linear%20elasticity
Linear elasticity is a mathematical model of how solid objects deform and become internally stressed due to prescribed loading conditions. It is a simplification of the more general nonlinear theory of elasticity and a branch of continuum mechanics. The fundamental "linearizing" assumptions of linear elasticity are: infinitesimal strains or "small" deformations (or strains) and linear relationships between the components of stress and strain. In addition linear elasticity is valid only for stress states that do not produce yielding. These assumptions are reasonable for many engineering materials and engineering design scenarios. Linear elasticity is therefore used extensively in structural analysis and engineering design, often with the aid of finite element analysis. Mathematical formulation Equations governing a linear elastic boundary value problem are based on three tensor partial differential equations for the balance of linear momentum and six infinitesimal strain-displacement relations. The system of differential equations is completed by a set of linear algebraic constitutive relations. Direct tensor form In direct tensor form that is independent of the choice of coordinate system, these governing equations are: Equation of motion, which is an expression of Newton's second law: Strain-displacement equations: Constitutive equations. For elastic materials, Hooke's law represents the material behavior and relates the unknown stresses and strains. The general equation for Hooke's law is where is the Cauchy stress tensor, is the infinitesimal strain tensor, is the displacement vector, is the fourth-order stiffness tensor, is the body force per unit volume, is the mass density, represents the nabla operator, represents a transpose, represents the second derivative with respect to time, and is the inner product of two second-order tensors (summation over repeated indices is implied). Cartesian coordinate form Expressed in terms of compon
https://en.wikipedia.org/wiki/DG/UX
DG/UX is a discontinued Unix operating system developed by Data General for its Eclipse MV minicomputer line, and later the AViiON workstation and server line (both Motorola 88000 and Intel IA-32-based variants). Overview DG/UX 1.00, released in March, 1985, was based on UNIX System V Release 2 with additions from 4.1BSD. By 1987, DG/UX 3.10 had been released, with 4.2BSD TCP/IP networking, NFS and the X Window System included. DG/UX 4.00, in 1988, was a comprehensive re-design of the system, based on System V Release 3, and supporting symmetric multiprocessing on the Eclipse MV. The 4.00 filesystem was based on the AOS/VS II filesystem and, using the logical disk facility, could span multiple disks. DG/UX 5.4, released around 1991, replaced the legacy Unix file buffer cache with unified, demand paged virtual memory management. Later versions were based on System V Release 4. On the AViiON, DG/UX supported multiprocessor machines at a time when most variants of Unix did not. The operating system was also more complete than some other Unix variants; for example, the operating system included a full C compiler (gcc) and also a logical volume manager. The OS was small and compact, but rich in features. It was simple and easy to install and did not require vast resources of memory or processing power. For example, a six-way Pentium Pro-based AViiON would support several hundred users using text terminals. The volume manager built into the OS was simple, but very powerful. All disk administration could be performed online, without taking any file system offline. This included extending, relocating, mirroring or shrinking. The same functions could be performed on the swap area, allowing in-place migrations of disk storage without downtime. DG/UX 5.4 supported filesystem shrinking, "split mirror" online backup, filesystems up to 2 TB, and filesystem journaling in 1991. Few vendors offered similar features at that time. DG/UX had a high-performance and stable clustered
https://en.wikipedia.org/wiki/Circular%20thresholding
Circular thresholding is an algorithm for automatic image threshold selection in image processing. Most threshold selection algorithms assume that the values (e.g. intensities) lie on a linear scale. However, some quantities such as hue and orientation are a circular quantity, and therefore require circular thresholding algorithms. The example shows that the standard linear version of Otsu's method when applied to the hue channel of an image of blood cells fails to correctly segment the large white blood cells (leukocytes). In contrast the white blood cells are correctly segmented by the circular version of Otsu's method. Methods There are a relatively small number of circular image threshold selection algorithms. The following examples are all based on Otsu's method for linear histograms: (Tseng, Li and Tung 1995) smooth the circular histogram, and apply Otsu's method. The histogram is cyclically rotated so that the selected threshold is shifted to zero. Otsu's method and histogram rotation are applied iteratively until several heuristics involving class size, threshold location, and class variance are satisfied. (Wu et al. 2006) smooth the circular histogram until it contains only two peaks. The histogram is cyclically rotated so that the midpoint between the peaks is shifted to zero. Otsu's method and histogram rotation are applied iteratively until convergence of the threshold. (Lai and Rosin 2014) applied Otsu's method to the circular histogram. For the two class circular thresholding task they showed that, for a histogram with an even number of bins, the optimal solution for Otsu's criterion of within-class variance is obtained when the histogram is split into two halves. Therefore the optimal solution can be efficiently obtained in linear rather than quadratic time.
https://en.wikipedia.org/wiki/Formal%20criteria%20for%20adjoint%20functors
In category theory, a branch of mathematics, the formal criteria for adjoint functors are criteria for the existence of a left or right adjoint of a given functor. One criterion is the following, which first appeared in Peter J. Freyd's 1964 book Abelian Categories, an Introduction to the Theory of Functors: Another criterion is:
https://en.wikipedia.org/wiki/Microsoft%20reaction%20card%20method
The Microsoft Reaction Card, developed by Microsoft in 2002 by Joey Benedek and Trish Miner, is a method used to check the emotional response and desirability of a design or product. This method is commonly used in the field of software design. Using this method involves a participant describing a design / product based on a list of 118 words. Each word is placed on a separate card. After viewing a design or product, the participant is asked to pick out the words they feel are relevant. The moderator then asks the participant to explain the reason for their selection.
https://en.wikipedia.org/wiki/JIC%20Germplasm%20Resources%20Unit
The Germplasm Resources Unit is part of the John Innes Centre. Located in the Norwich Research Park, Norwich, England, is a germplasm conservation unit and National Capability supported by the Biotechnology and Biological Sciences Research Council. This unit houses a number of internationally recognised reference- and working-collections for wheat, oats, barley and peas, which serves UK and non-UK based academic, industrial and non-industrial groups. History The collections from the Germplasm Resources Unit were brought together in the mid-1980s from working collection from several research institutes from around the UK that worked with small grain cereals and legumes, including the extinct Plant Breeding Institute. This centralisation effort was supported by the John Innes Institute, and was designed to act as an open collection that would provide access to important resources for ongoing research and breeding. The connection of the unit and the JIC has the advantage of placing germplasm material on sites of active research where a higher level of interaction with the scientific community is possible. This two-way interaction ensures that scientists and students are exposed to, and have greater opportunities to view and discuss genetic variability, while affording staff involvement in research objectives and priorities within both basic and strategic applied science. In 2012 the unit became a National Capability supported by the Biotechnology and Biological Sciences Research Council as part of its new funding arrangements. Today, the cereal collections have been successfully screened for many traits leading to the identification of new sources of disease resistance to a range of diseases as well as tolerance to drought, salinity and aluminium. Collections The Germplasm Resources Unit houses a diverse range of seed collections, accounting for more than 20,000 accessions. The seeds are stored in a special low temperature, low humidity facility and a complete li
https://en.wikipedia.org/wiki/Debugfs
debugfs is a special file system available in the Linux kernel since version 2.6.10-rc3. It was written by Greg Kroah-Hartman. debugfs is a simple-to-use RAM-based file system specially designed for debugging purposes. It exists as a simple way for kernel developers to make information available to user space. Unlike , which is only meant for information about a process, or sysfs, which has strict one-value-per-file rules, debugfs has no rules at all. Developers can put any information they want there. Use To compile a Linux kernel with the debugfs facility, the option must be set to yes. It is typically mounted at with a command such as: mount -t debugfs none /sys/kernel/debug It can be manipulated using several calls from the C header file , which include: for creating a file in the debug filesystem. for creating a directory inside the debug filesystem. for creating a symbolic link inside the debug filesystem. for removing a debugfs entry from the debug filesystem.
https://en.wikipedia.org/wiki/Plus%20and%20minus%20signs
The plus sign and the minus sign are mathematical symbols used to represent the notions of positive and negative, respectively. In addition, represents the operation of addition, which results in a sum, while represents subtraction, resulting in a difference. Their use has been extended to many other meanings, more or less analogous. and are Latin terms meaning "more" and "less", respectively. History Though the signs now seem as familiar as the alphabet or the Hindu-Arabic numerals, they are not of great antiquity. The Egyptian hieroglyphic sign for addition, for example, resembled a pair of legs walking in the direction in which the text was written (Egyptian could be written either from right to left or left to right), with the reverse sign indicating subtraction: Nicole Oresme's manuscripts from the 14th century show what may be one of the earliest uses of as a sign for plus. In early 15th century Europe, the letters "P" and "M" were generally used. The symbols (P with overline, , for (more), i.e., plus, and M with overline, , for (less), i.e., minus) appeared for the first time in Luca Pacioli's mathematics compendium, , first printed and published in Venice in 1494. The sign is a simplification of the (comparable to the evolution of the ampersand ). The may be derived from a tilde written over when used to indicate subtraction; or it may come from a shorthand version of the letter itself. In his 1489 treatise, Johannes Widmann referred to the symbols and as minus and mer (Modern German ; "more"): They weren't used for addition and subtraction in the treatise, but were used to indicate surplus and deficit; usage in the modern sense is attested in a 1518 book by Henricus Grammateus. Robert Recorde, the designer of the equals sign, introduced plus and minus to Britain in 1557 in The Whetstone of Witte: "There be other 2 signes in often use of which the first is made thus + and betokeneth more: the other is thus made − and betokeneth lesse.
https://en.wikipedia.org/wiki/Schedule
A schedule or a timetable, as a basic time-management tool, consists of a list of times at which possible tasks, events, or actions are intended to take place, or of a sequence of events in the chronological order in which such things are intended to take place. The process of creating a schedule — deciding how to order these tasks and how to commit resources between the variety of possible tasks — is called scheduling, and a person responsible for making a particular schedule may be called a scheduler. Making and following schedules is an ancient human activity. Some scenarios associate this kind of planning with learning life skills. Schedules are necessary, or at least useful, in situations where individuals need to know what time they must be at a specific location to receive a specific service, and where people need to accomplish a set of goals within a set time. Schedules can usefully span both short periods, such as a daily or weekly schedule, and long-term planning for periods of several months or years. They are often made using a calendar, where the person making the schedule can note the dates and times at which various events are planned to occur. Schedules that do not set forth specific times for events to occur may instead list algorithmically an expected order in which events either can or must take place. In some situations, schedules can be uncertain, such as where the conduct of daily life relies on environmental factors outside human control. People who are vacationing or otherwise seeking to reduce stress and achieve relaxation may intentionally avoid having a schedule for a certain period of time. Kinds of schedules Publicly available schedules Certain kinds of schedules reflect information that is generally made available to the public, so that members of the public can plan certain activities around them. These may include things like: Hours of operation of businesses, tourist attractions, and government offices, which allow consumers of
https://en.wikipedia.org/wiki/Lami%27s%20theorem
In physics, Lami's theorem is an equation relating the magnitudes of three coplanar, concurrent and non-collinear vectors, which keeps an object in static equilibrium, with the angles directly opposite to the corresponding vectors. According to the theorem, where A, B and C are the magnitudes of the three coplanar, concurrent and non-collinear vectors, , which keep the object in static equilibrium, and α, β and γ are the angles directly opposite to the vectors. Lami's theorem is applied in static analysis of mechanical and structural systems. The theorem is named after Bernard Lamy. Proof As the vectors must balance , hence by making all the vectors touch its tip and tail the result is a triangle with sides A, B, C and angles By the law of sines then Then by applying that for any angle , , and the result is See also Mechanical equilibrium Parallelogram of force Tutte embedding
https://en.wikipedia.org/wiki/Kruskal%E2%80%93Szekeres%20coordinates
In general relativity, Kruskal–Szekeres coordinates, named after Martin Kruskal and George Szekeres, are a coordinate system for the Schwarzschild geometry for a black hole. These coordinates have the advantage that they cover the entire spacetime manifold of the maximally extended Schwarzschild solution and are well-behaved everywhere outside the physical singularity. There is no misleading coordinate singularity at the horizon. The Kruskal–Szekeres coordinates also apply to space-time around a spherical object, but in that case do not give a description of space-time inside the radius of the object. Space-time in a region where a star is collapsing into a black hole is approximated by the Kruskal–Szekeres coordinates (or by the Schwarzschild coordinates). The surface of the star remains outside the event horizon in the Schwarzschild coordinates, but crosses it in the Kruskal–Szekeres coordinates. (In any "black hole" which we observe, we see it at a time when its matter has not yet finished collapsing, so it is not really a black hole yet.) Similarly, objects falling into a black hole remain outside the event horizon in Schwarzschild coordinates, but cross it in Kruskal–Szekeres coordinates. Definition Kruskal–Szekeres coordinates on a black hole geometry are defined, from the Schwarzschild coordinates , by replacing t and r by a new timelike coordinate T and a new spacelike coordinate : for the exterior region outside the event horizon and: for the interior region . Here is the gravitational constant multiplied by the Schwarzschild mass parameter, and this article is using units where = 1. It follows that on the union of the exterior region, the event horizon and the interior region the Schwarzschild radial coordinate (not to be confused with the Schwarzschild radius ), is determined in terms of Kruskal–Szekeres coordinates as the (unique) solution of the equation: Using the Lambert W function the solution is written as: . Moreover one sees immediately
https://en.wikipedia.org/wiki/Diskless%20node
A diskless node (or diskless workstation) is a workstation or personal computer without disk drives, which employs network booting to load its operating system from a server. (A computer may also be said to act as a diskless node, if its disks are unused and network booting is used.) Diskless nodes (or computers acting as such) are sometimes known as network computers or hybrid clients. Hybrid client may either just mean diskless node, or it may be used in a more particular sense to mean a diskless node which runs some, but not all, applications remotely, as in the thin client computing architecture. Advantages of diskless nodes can include lower production cost, lower running costs, quieter operation, and manageability advantages (for example, centrally managed software installation). In many universities and in some large organizations, PCs are used in a similar configuration, with some or all applications stored remotely but executed locally—again, for manageability reasons. However, these are not diskless nodes if they still boot from a local hard drive. Distinction between diskless nodes and centralized computing Diskless nodes process data, thus using their own CPU and RAM to run software, but do not store data persistently—that task is handed off to a server. This is distinct from thin clients, in which all significant processing happens remotely, on the server—the only software that runs on a thin client is the "thin" (i.e. relatively small and simple) client software, which handles simple input/output tasks to communicate with the user, such as drawing a dialog box on the display or waiting for user input. A collective term encompassing both thin client computing, and its technological predecessor, text terminals (which are text-only), is centralized computing. Thin clients and text terminals can both require powerful central processing facilities in the servers, in order to perform all significant processing tasks for all of the clients. Diskless no
https://en.wikipedia.org/wiki/ES%20PEVM
ES PEVM (ЕС ПЭВМ) was a Soviet clone of the IBM PC in the 1980s. The ES PEVM models lineup also included analogues of IBM PC XT, IBM PC AT, IBM XT/370. The computers and software were adapted in Minsk, Belarus, at the Scientific Research Institute of Electronic Computer Machines (НИИ ЭВМ). They were manufactured in Minsk as well, at Minsk Production Group for Computing Machinery (Минское производственное объединение вычислительной техники (МПО ВТ)). Description The first models of ES PEVM (ES-1840, ES-1841, ES-1842), unlike the IBM PC, had two units: a system unit and a floppy drives unit. These models used a backplane instead of the main board. Although the system bus was compatible with ISA bus, it used a different type of connector, so the IBM PC expansion cards could not be installed in ES PEVM. Later models (ES-1843, ES-1849 etc.) were fully compatible with IBM PC XT and IBM PC AT. Unlike IBM PC using the Intel 8088 processor, the early ES PEVM models used K1810VM86 processor with a 16-bit bus and a clock frequency of 5 MHz. The processor was placed on a separate board. Early versions of the board did not have a socket for the floating point coprocessor. The following boards were produced for ES-1840 and ES-1841: CPU board (containing 54 chips). It was installed in every configuration of the computer. 128 KiB RAM board. It was installed in early models of ES PEVM. 512 KiB RAM board (containing 110 chips). Variants with 256 KiB or 128 Kib were available. MDA adapter board (containing 91 chips). It was built using the Bulgarian CM607 chip (MC6845 clone). CGA adapter board (containing 94 chips). Floppy-drive adapter board (containing 32 chips). It was built using the Bulgarian CM609 chip (Intel 8272 clone). Serial interface adapter board (containing 56 chips). It was electrically incompatible with the IBM PC COM port and was not supported in the BIOS. Volume of production for a number of models: Software The computers were shipped with AlphaDOS,
https://en.wikipedia.org/wiki/Chain%20loading
Chain loading is a method used by computer programs to replace the currently executing program with a new program, using a common data area to pass information from the current program to the new program. It occurs in several areas of computing. Chain loading is similar to the use of overlays. Unlike overlays, however, chain loading replaces the currently executing program in its entirety. Overlays usually replace only a portion of the running program. Like the use of overlays, the use of chain loading increases the I/O load of an application. Chain loading in boot manager programs In operating system boot manager programs, chain loading is used to pass control from the boot manager to a boot sector. The target boot sector is loaded in from disk, replacing the in-memory boot sector from which the boot manager itself was bootstrapped, and executed. Chain loading in Unix In Unix (and in Unix-like operating systems), the exec() system call is used to perform chain loading. The program image of the current process is replaced with an entirely new image, and the current thread begins execution of that image. The common data area comprises the process' environment variables, which are preserved across the system call. Chain loading in Linux In addition to the process level chain loading Linux supports the system call to replace the entire operating system kernel with a different version. The new kernel boots as if it were started from power up and no running processes are preserved. Chain loading in BASIC programs In BASIC programs, chain loading is the purview of the CHAIN statement (or, in Commodore BASIC, the LOAD statement), which causes the current program to be terminated and the chained-to program to be loaded and invoked (with, on those dialects of BASIC that support it, an optional parameter specifying the line number from which execution is to commence, rather than the default of the first line of the new program). The common data area varies
https://en.wikipedia.org/wiki/Wlan%20slovenija
The wlan slovenija open wireless network of Slovenia is an initiative for deployment of open and free community wireless network across all Slovenia through collaboration and with use of off-the-shelf widely available equipment. The main idea is that if everyone deploys a node of the open network in their home and share part of their Internet connectivity with others around them, then whole Slovenia will be covered with the wireless network which will allow everyone free basic connectivity to Internet. With this idea it operates since 2009 onward. The network is based on open source software and hardware which has been developed by participants as well. It is one of largest Slovenian open source projects and has participated multiple times in Google Summer of Code. General description wlan slovenija open wireless network is an example of a wireless community network. It uses common and widespread wireless technologies based on IEEE 802.11 standards, open source systems and because of its wireless and innovative organic nature its spread and access to the network is limited only with participation of users in it: more users there is more the network is spread and useful, for more purposes users can use it, more and richer content it provides. Around the network, which is in the first place an initiative, a community of people and organizations is forming and thus ideas, motivations and concepts around the network are various as participants in the network are various. But mostly all agree about common reasons for the network: free access and sharing of knowledge and information, building an alternative media and because it is interesting, educational and fun. More people put up nodes of the wireless networks, more people share their Internet uplinks with others, more everybody gains. wlan slovenija tries to attract also people without technical skills but who still share common philosophy. Because of this wlan slovenija focuses on ease to use, proven good pract
https://en.wikipedia.org/wiki/Anderson%20localization
In condensed matter physics, Anderson localization (also known as strong localization) is the absence of diffusion of waves in a disordered medium. This phenomenon is named after the American physicist P. W. Anderson, who was the first to suggest that electron localization is possible in a lattice potential, provided that the degree of randomness (disorder) in the lattice is sufficiently large, as can be realized for example in a semiconductor with impurities or defects. Anderson localization is a general wave phenomenon that applies to the transport of electromagnetic waves, acoustic waves, quantum waves, spin waves, etc. This phenomenon is to be distinguished from weak localization, which is the precursor effect of Anderson localization (see below), and from Mott localization, named after Sir Nevill Mott, where the transition from metallic to insulating behaviour is not due to disorder, but to a strong mutual Coulomb repulsion of electrons. Introduction In the original Anderson tight-binding model, the evolution of the wave function ψ on the d-dimensional lattice Zd is given by the Schrödinger equation where the Hamiltonian H is given by with Ej random and independent, and potential V(r) falling off faster than r−3 at infinity. For example, one may take Ej uniformly distributed in [−W,   +W], and Starting with ψ0 localised at the origin, one is interested in how fast the probability distribution diffuses. Anderson's analysis shows the following: if d is 1 or 2 and W is arbitrary, or if d ≥ 3 and W/ħ is sufficiently large, then the probability distribution remains localized: uniformly in t. This phenomenon is called Anderson localization. if d ≥ 3 and W/ħ is small, where D is the diffusion constant. Analysis The phenomenon of Anderson localization, particularly that of weak localization, finds its origin in the wave interference between multiple-scattering paths. In the strong scattering limit, the severe interferences can completely halt the wav
https://en.wikipedia.org/wiki/Exercise%20and%20Sport%20Sciences%20Reviews
Exercise and Sport Sciences Reviews is a quarterly peer-reviewed review journal covering sports medicine and exercise science. It was established in 1973 as a hardcover book series, and became a quarterly peer-reviewed journal in January 2000. It is published by Lippincott Williams & Wilkins, and is an official journal of the American College of Sports Medicine. The editor-in-chief is Sandra K. Hunter, Ph.D., FACSM (Marquette University). According to the Journal Citation Reports, the journal has a 2021 impact factor of 6.642.
https://en.wikipedia.org/wiki/Fluxapyroxad
Fluxapyroxad is a broad-spectrum pyrazole-carboxamide fungicide used on a large variety of commercial crops. It stunts fungus growth by inhibiting the succinate dehydrogenase (SQR) enzyme. Application of fluxapyroxad helps prevent many wilts and other fungal infections from taking hold. As with other systemic pesticides that have a long chemical half-life, there are concerns about keeping fluxapyroxad out of the groundwater, especially when combined with pyraclostrobin. There is also concern that some fungi may develop resistance to fluxapyroxad. Chemical structure The compound is an amide of 3-(difluoromethyl)-1-methyl-1H-pyrazole-4-carboxylic acid combined with an aniline having an ortho-substituted triflurorobenzene group. Biological action Fluxapyroxad is a succinate dehydrogenase inhibitor (SDHI). It interferes with a number of key fungal life functions, including spore germination, germ tube growth, appresoria formation and mycelium growth. Specifically it interferes with the production of succinate dehydrogenase, the complex II in the mitochondrial respiration chain, which in turn interferes with the tricarboxylic cycle and mitochondrial electron transport.SDHIs are not specific to fungi, they inhibit SDH, the CII of the mitochondrial respiratory chain. As a result, they act on all organisms that have mitochondria, it below to the large familly of SDHI (PMID: 31697708) . Crops Fluxapyroxad is commonly used as a fungicide for grains, row crops, vegetable crops, and fruit trees (pome and prunus), including: Fungal diseases Fluxapyroxad provides protection against many fungal diseases. Studies have shown specific efficacy against diseases such as black point, Botrytis gray mold, early blight, and powdery mildew; however, fluxapyroxad was found to have no efficacy against anthracnose on lentils. Toxicity Fluxapyroxad has a low toxicity for humans, slightly toxic after a single ingestion, and relatively non-toxic after single inhalation or topical skin cont
https://en.wikipedia.org/wiki/Species%E2%80%93area%20relationship
The species–area relationship or species–area curve describes the relationship between the area of a habitat, or of part of a habitat, and the number of species found within that area. Larger areas tend to contain larger numbers of species, and empirically, the relative numbers seem to follow systematic mathematical relationships. The species–area relationship is usually constructed for a single type of organism, such as all vascular plants or all species of a specific trophic level within a particular site. It is rarely if ever, constructed for all types of organisms if simply because of the prodigious data requirements. It is related but not identical to the species discovery curve. Ecologists have proposed a wide range of factors determining the slope and elevation of the species–area relationship. These factors include the relative balance between immigration and extinction, rate and magnitude of disturbance on small vs. large areas, predator-prey dynamics, and clustering of individuals of the same species as a result of dispersal limitation or habitat heterogeneity. The species–area relationship has been reputed to follow from the 2nd law of thermodynamics. In contrast to these "mechanistic" explanations, others assert the need to test whether the pattern is simply the result of a random sampling process. Species–area relationships are often evaluated in conservation science in order to predict extinction rates in the case of habitat loss and habitat fragmentation. Authors have classified the species–area relationship according to the type of habitats being sampled and the census design used. Frank W. Preston, an early investigator of the theory of the species–area relationship, divided it into two types: samples (a census of a contiguous habitat that grows in the census area, also called "mainland" species–area relationships), and isolates (a census of discontiguous habitats, such as islands, also called "island" species–area relationships). Michael Rosenzwe
https://en.wikipedia.org/wiki/Prompt%20engineering
Prompt engineering is the process of structuring text that can be interpreted and understood by a generative AI model. A prompt is natural language text describing the task that an AI should perform. A prompt for a text-to-text model can be a query such as "what is Fermat's little theorem?", a command such as "write a poem about leaves falling", a short statement of feedback (for example, "too verbose", "too formal", "rephrase again", "omit this word") or a longer statement including context, instructions, and input data. Prompt engineering may involve phrasing a query, specifying a style, providing relevant context or assigning a role to the AI such as "Act as a native French speaker". A prompt may include a few examples for a model to learn from, such as "maison -> house, chat -> cat, chien ->", an approach called few-shot learning. When communicating with a text-to-image or a text-to-audio model, a typical prompt is a description of a desired output such as "a high-quality photo of an astronaut riding a horse" or "Lo-fi slow BPM electro chill with organic samples". Prompting a text-to-image model may involve adding, removing, emphasizing and re-ordering words to achieve a desired subject, style, layout, lighting, and aesthetic. In-context learning Prompt engineering is enabled by in-context learning, defined as a model's ability to temporarily learn from prompts. The ability for in-context learning is an emergent ability of large language models. In-context learning itself is an emergent property of model scale, meaning breaks in downstream scaling laws occur such that its efficacy increases at a different rate in larger models than in smaller models. In contrast to training and fine tuning for each specific task, which are not temporary, what has been learnt during in-context learning is of a temporary nature. It does not carry the temporary contexts or biases, except the ones already present in the (pre)training dataset, from one conversation to the other.
https://en.wikipedia.org/wiki/Magnetic%20dip
Magnetic dip, dip angle, or magnetic inclination is the angle made with the horizontal by the Earth's magnetic field lines. This angle varies at different points on the Earth's surface. Positive values of inclination indicate that the magnetic field of the Earth is pointing downward, into the Earth, at the point of measurement, and negative values indicate that it is pointing upward. The dip angle is in principle the angle made by the needle of a vertically held compass, though in practice ordinary compass needles may be weighted against dip or may be unable to move freely in the correct plane. The value can be measured more reliably with a special instrument typically known as a dip circle. Dip angle was discovered by the German engineer Georg Hartmann in 1544. A method of measuring it with a dip circle was described by Robert Norman in England in 1581. Explanation Magnetic dip results from the tendency of a magnet to align itself with lines of magnetic field. As the Earth's magnetic field lines are not parallel to the surface, the north end of a compass needle will point upward in the southern hemisphere (negative dip) or downward in the northern hemisphere (positive dip) . The range of dip is from -90 degrees (at the South Magnetic Pole) to +90 degrees (at the North Magnetic Pole). Contour lines along which the dip measured at the Earth's surface is equal are referred to as isoclinic lines. The locus of the points having zero dip is called the magnetic equator or aclinic line. Calculation for a given latitude The inclination is defined locally for the magnetic field due to the Earth's core, and has a positive value if the field points below the horizontal (ie into the Earth). Here we show how to determine the value of at a given latitude, following the treatment given by Fowler. Outside Earth's core we consider Maxwell's equations in a vacuum, and where and the subscript denotes the core as the origin of these fields. The first means we can introduc
https://en.wikipedia.org/wiki/OpenEpi
OpenEpi is a free, web-based, open source, operating system-independent series of programs for use in epidemiology, biostatistics, public health, and medicine, providing a number of epidemiologic and statistical tools for summary data. OpenEpi was developed in JavaScript and HTML, and can be run in modern web browsers. The program can be run from the OpenEpi website or downloaded and run without a web connection. The source code and documentation is downloadable and freely available for use by other investigators. OpenEpi has been reviewed, both by media organizations and in research journals. The OpenEpi developers have had extensive experience in the development and testing of Epi Info, a program developed by the Centers for Disease Control and Prevention (CDC) and widely used around the world for data entry and analysis. OpenEpi was developed to perform analyses found in the DOS version of Epi Info modules StatCalc and EpiTable, to improve upon the types of analyses provided by these modules, and to provide a number of tools and calculations not currently available in Epi Info. It is the first step toward an entirely web-based set of epidemiologic software tools. OpenEpi can be thought of as an important companion to Epi Info and to other programs such as SAS, PSPP, SPSS, Stata, SYSTAT, Minitab, Epidata, and R (see the R programming language). Another functionally similar Windows-based program is Winpepi. See also list of statistical packages and comparison of statistical packages. Both OpenEpi and Epi Info were developed with the goal of providing tools for low and moderate resource areas of the world. The initial development of OpenEpi was supported by a grant from the Bill and Melinda Gates Foundation to Emory University. Types The types of calculations currently performed by OpenEpi include: Various confidence intervals for proportions, rates, standardized mortality ratio, mean, median, percentiles 2x2 crude and stratified tables for count and rate da
https://en.wikipedia.org/wiki/Heterogeneous%20ribonucleoprotein%20particle
Heterogeneous nuclear ribonucleoproteins (hnRNPs) are complexes of RNA and protein present in the cell nucleus during gene transcription and subsequent post-transcriptional modification of the newly synthesized RNA (pre-mRNA). The presence of the proteins bound to a pre-mRNA molecule serves as a signal that the pre-mRNA is not yet fully processed and therefore not ready for export to the cytoplasm. Since most mature RNA is exported from the nucleus relatively quickly, most RNA-binding protein in the nucleus exist as heterogeneous ribonucleoprotein particles. After splicing has occurred, the proteins remain bound to spliced introns and target them for degradation. hnRNPs are also integral to the 40s subunit of the ribosome and therefore important for the translation of mRNA in the cytoplasm. However, hnRNPs also have their own nuclear localization sequences (NLS) and are therefore found mainly in the nucleus. Though it is known that a few hnRNPs shuttle between the cytoplasm and nucleus, immunofluorescence microscopy with hnRNP-specific antibodies shows nucleoplasmic localization of these proteins with little staining in the nucleolus or cytoplasm. This is likely because of its major role in binding to newly transcribed RNAs. High-resolution immunoelectron microscopy has shown that hnRNPs localize predominantly to the border regions of chromatin, where it has access to these nascent RNAs. The proteins involved in the hnRNP complexes are collectively known as heterogeneous ribonucleoproteins. They include protein K and polypyrimidine tract-binding protein (PTB), which is regulated by phosphorylation catalyzed by protein kinase A and is responsible for suppressing RNA splicing at a particular exon by blocking access of the spliceosome to the polypyrimidine tract. hnRNPs are also responsible for strengthening and inhibiting splice sites by making such sites more or less accessible to the spliceosome. Cooperative interactions between attached hnRNPs may encourage cert
https://en.wikipedia.org/wiki/Veronica%20bucket
The Veronica bucket is a mechanism for hand washing originating in Ghana which consists of a bucket of water with a tap fixed at the bottom, mounted at hand height, and a bowl at the bottom to collect waste water. The Veronica bucket was developed by Veronica Bekoe. The Veronica bucket serves as a simple way to encourage proper hand washing using flowing water. Bekoe in an interview stated that the bucket was originally made to help her and her colleagues wash their hands under running water after each lab session. She said, "We are used to washing hands in a bowl with others washing in the same water, which will do more harm than good." These colleagues were contaminating their hands rather than decontaminating them. In addition to the COVID benefit of hand washing, the Veronica bucket is also essential for areas where potable water is not readily available. Uses The bucket is also used in other African countries. It is common in places such as schools, hospitals, churches and areas with no running taps. It has become very popular in Ghana following the outbreak of the novel coronavirus (COVID-19) as citizens engage in frequent hand washing to stem its spread. In Ekiti State, Nigeria, the governor Kayode Fayemi directed all public places to provide running tap water or Veronica buckets "to encourage frequent handwashing" as part of the measures to contain COVID-19. Before the COVID-19 outbreak, the invention was used in some schools and hospitals but now it is in high demand due to its role in curbing the outbreak. Now, the setup could be spotted in places like malls, hospitals, corporate institutions and government offices. It was invented by a Ghanaian, Veronica Bekoe, whom the invention was named after. She claimed the bucket was named after her in 1993 by Joan Hetrick. Bekoe is a biologist who has worked at the Public Health and Reference Laboratory of the Ghana Health Service from 1972 to 2008. Production The invention was initially produced by local artisa
https://en.wikipedia.org/wiki/Polar%20%28star%29
In astronomy, a polar is a highly magnetic type of cataclysmic variable (CV) binary star system, originally known as an AM Herculis star after the prototype member AM Herculis. Like other CVs, polars contain two stars: an accreting white dwarf (WD), and a low-mass donor star (usually a red dwarf) which is transferring mass to the WD as a result of the WD's gravitational pull, overflowing its Roche lobe. Polars are distinguished from other CVs by the presence of a very strong magnetic field in the WD. Typical magnetic field strengths of polar systems are 10 million to 80 million gauss (1000–8000 teslas). The WD in the polar AN Ursae Majoris has the strongest known magnetic field among cataclysmic variables, with a field strength of 230 million gauss (23 kT). Accretion mechanism One of the most critical consequences of the WD's magnetism is that it synchronizes the rotational period of the WD with the orbital period of the binary; to first order, this means that the same side of the WD always faces the donor star. This synchronous rotation is considered a defining feature of polars. Additionally, the WD's magnetic field captures the accretion stream from the donor star before it can develop into an accretion disk. The capture of the accretion stream is known as threading, and it occurs when the magnetic pressure from the WD matches the stream's ram pressure. The captured material flows along the WD's magnetic field lines until it violently accretes onto the WD in a shock near one or more of the star's magnetic poles. This accretion region covers only a fraction of the WD's surface, but it can contribute half of the system's optical light. In addition to optical and near-infrared cyclotron radiation, the accretion region also produces X-rays due to the high temperature of gas within the shock, so polars are frequently brighter in X-rays than non-magnetic CVs. Whereas accretion in a non-magnetic system is governed by viscosity within the accretion disk, accretion
https://en.wikipedia.org/wiki/Local%20diffeomorphism
In mathematics, more specifically differential topology, a local diffeomorphism is intuitively a map between Smooth manifolds that preserves the local differentiable structure. The formal definition of a local diffeomorphism is given below. Formal definition Let and be differentiable manifolds. A function is a local diffeomorphism, if for each point there exists an open set containing such that is open in and is a diffeomorphism. A local diffeomorphism is a special case of an immersion where the image of under locally has the differentiable structure of a submanifold of Then and may have a lower dimension than Characterizations A map is a local diffeomorphism if and only if it is a smooth immersion (smooth local embedding) and an open map. The inverse function theorem implies that a smooth map is a local diffeomorphism if and only if the derivative is a linear isomorphism for all points This implies that and must have the same dimension. A map between two connected manifolds of equal dimension () is a local diffeomorphism if and only if it is a smooth immersion (smooth local embedding), or equivalently, if and only if it is a smooth submersion. This is because every smooth immersion is a locally injective function while invariance of domain guarantees that any continuous injective function between manifolds of equal dimensions is necessarily an open map. Discussion For instance, even though all manifolds look locally the same (as for some ) in the topological sense, it is natural to ask whether their differentiable structures behave in the same manner locally. For example, one can impose two different differentiable structures on that make into a differentiable manifold, but both structures are not locally diffeomorphic (see below). Although local diffeomorphisms preserve differentiable structure locally, one must be able to "patch up" these (local) diffeomorphisms to ensure that the domain is the entire (smooth) manifold. Fo
https://en.wikipedia.org/wiki/Tetrad%20%28meiosis%29
The tetrad is the four spores produced after meiosis of a yeast or other Ascomycota, Chlamydomonas or other alga, or a plant. After parent haploids mate, they produce diploids. Under appropriate environmental conditions, diploids sporulate and undergo meiosis. The meiotic products, spores, remain packaged in the parental cell body to produce the tetrad. Genetic typification If the two parents have a mutation in two different genes, the tetrad can segregate these genes as the parental ditype (PD), the non-parental ditype (NPD) or as the tetratype (TT). Parental ditype is a tetrad type containing two different genotypes, both of which are parental. A spore arrangement in ascomycetes that contains only the two non-recombinant-type ascospores. Non-parental ditype (NPD) is a spore that contains only the two recombinant-type ascospores (assuming two segregating loci). A tetrad type containing two different genotypes, both of which are recombinant. Tetratype is a tetrad containing four different genotypes, two parental and two recombinant. A spore arrangement in ascomycetes that consists of two parental and two recombinant spores indicates a single crossover between two linked loci. Linkage analysis The ratio between the different segregation types arising after the sporulation is a measure of the linkage between the two genes. Tetrad dissection Tetrad dissection has become a powerful tool of yeast geneticists, and is used in conjunction with the many established procedures utilizing the versatility of yeasts as model organisms. Use of modern microscopy and micromanipulation techniques allows the four haploid spores of a yeast tetrad to be separated and germinated individually to form isolated spore colonies. Uses Tetrad analysis can be used to confirm whether a phenotype is caused by a specific mutation, construction of strains, and for investigating gene interaction. Since the frequency of tetrad segregation types is influenced by the recombination frequency for
https://en.wikipedia.org/wiki/Vacuum%20Rabi%20oscillation
A vacuum Rabi oscillation is a damped oscillation of an initially excited atom coupled to an electromagnetic resonator or cavity in which the atom alternately emits photon(s) into a single-mode electromagnetic cavity and reabsorbs them. The atom interacts with a single-mode field confined to a limited volume V in an optical cavity. Spontaneous emission is a consequence of coupling between the atom and the vacuum fluctuations of the cavity field. Mathematical treatment A mathematical description of vacuum Rabi oscillation begins with the Jaynes–Cummings model, which describes the interaction between a single mode of a quantized field and a two level system inside an optical cavity. The Hamiltonian for this model in the rotating wave approximation is where is the Pauli z spin operator for the two eigenstates and of the isolated two level system separated in energy by ; and are the raising and lowering operators of the two level system; and are the creation and annihilation operators for photons of energy in the cavity mode; and is the strength of the coupling between the dipole moment of the two level system and the cavity mode with volume and electric field polarized along . The energy eigenvalues and eigenstates for this model are where is the detuning, and the angle is defined as Given the eigenstates of the system, the time evolution operator can be written down in the form If the system starts in the state , where the atom is in the ground state of the two level system and there are photons in the cavity mode, the application of the time evolution operator yields The probability that the two level system is in the excited state as a function of time is then where is identified as the Rabi frequency. For the case that there is no electric field in the cavity, that is, the photon number is zero, the Rabi frequency becomes . Then, the probability that the two level system goes from its ground state to its excited state as a function of tim
https://en.wikipedia.org/wiki/Ubicom
Ubicom was a company which developed communications and media processor (CMP) and software platforms for real-time interactive applications and multimedia content delivery in the digital home. The company provided optimized system-level solutions to OEMs for a wide range of products including wireless routers, access points, VoIP gateways, streaming media devices, print servers and other network devices. Ubicom was a venture-backed, privately held company with corporate headquarters in San Jose, California. History Ubicom was founded as Scenix Semiconductor in 1996. The company operated under that name until 1999. In 2000, Scenix became "Ubicom," a word derived from "ubiquitous communications". April 1999: Mayfield Fund leads $10 million equity investment in Scenix. November 2000: Scenix changes its name to Ubicom. November 2002: Intersil and Ubicom demonstrate world's first 802.11g wireless access point. March 2006: Ubicom secures $20 million in Series 3 funding, led by Investcorp Technology Ventures. March 2012: Ubicom is taken over by Qualcomm Atheros. Products As Scenix and Ubicom, the company designed several families of microcontrollers, including: The SX Series of 8-bit microcontrollers, a product line which was partially compatible with Arizona Microchip devices and ran at up to 100 MHz, single cycle. This product was eventually sold to Parallax, who continued its production. The IP series of high performance media and Internet processors. These devices were designed to act as gateways for streaming media and data over wired and wireless links. The Scenix/Ubicom processors relied on very high speed and low latency processing to emulate hardware interfaces in software such as interrupt-polled soft-UARTS. This reduced the size of the silicon chip and therefore the cost, but increased the complexity of the software required on the chip. Ubicom developed its own architecture, the Ubicom32, and a real-time operating system (RTOS) for it. For exam
https://en.wikipedia.org/wiki/Fibre%20Channel%20switch
In the computer storage field, a Fibre Channel switch is a network switch compatible with the Fibre Channel (FC) protocol. It allows the creation of a Fibre Channel fabric, that is the core component of a storage area network (SAN). The fabric is a network of Fibre Channel devices which allows many-to-many communication, device name lookup, security, and redundancy. FC switches implement zoning, a mechanism that disables unwanted traffic between certain fabric nodes. Fibre Channel switches may be deployed one at a time or in larger multi-switch configurations. SAN administrators typically add new switches as their server and storage needs grow, connecting switches together via fiber optic cable using the standard device ports. Some switch vendors offer dedicated high-speed stacking ports to handle inter-switch connections (similar to existing stackable Ethernet switches), allowing high-performance multi-switch configurations to be created using fewer switches overall. Major manufacturers of Fibre Channel switches include Brocade(Broadcom), Cisco Systems, and QLogic(Marvell). Fibre Channel Director A special variety of a FC switch is the Fibre Channel Director, a switch meant to provide backbone infrastructure in a fabric usually featuring at least 128 ports and high-availability attributes, however the term is loose and varies among to manufacturers. It does not differ from a switch in core FC protocol functionality. The director term itself is derived from legacy ESCON Directors such as the IBM 9032-005. See also List of Fibre Channel switches Host Bus Adapter (HBA)
https://en.wikipedia.org/wiki/Message%20switching
In telecommunications, message switching involves messages routed in their entirety, one hop at a time. It evolved from circuit switching and was the precursor of packet switching. An example of message switching is email in which the message is sent through different intermediate servers to reach the mail server for storing. Unlike packet switching, the message is not divided into smaller units and sent independently over the network. History Western Union operated a message switching system, Plan 55-A, for processing telegrams in the 1950s. Leonard Kleinrock wrote a doctoral thesis at the Massachusetts Institute of Technology in 1962 that analyzed queueing delays in this system. Message switching was built by Collins Radio Company, Newport Beach, California, during the period 1959–1963 for sale to large airlines, banks and railroads. The original design for the ARPANET was Wesley Clark's April 1967 proposal for using Interface Message Processors to create a message switching network. After the seminal meeting at the first ACM Symposium on Operating Systems Principles in October 1967, where Roger Scantlebury presented Donald Davies work and mentioned the work of Paul Baran, Larry Roberts incorporated packet switching into the design. The SITA High-Level Network (HLN) became operational in 1969, handling data traffic for airlines in real time via a message-switched network over common carrier leased lines. It was organised to act like a packet-switching network. Message switching systems are nowadays mostly implemented over packet-switched or circuit-switched data networks. Each message is treated as a separate entity. Each message contains addressing information, and at each switch this information is read and the transfer path to the next switch is decided. Depending on network conditions, a conversation of several messages may not be transferred over the same path. Each message is stored (usually on hard drive due to RAM limitations) before being transmi
https://en.wikipedia.org/wiki/Anne%20Richards
Dame Anne Helen Richards (born 1964) is the chief executive officer of Fidelity International. Previously, she served as chief investment officer of Aberdeen Asset Management and chief executive officer of M&G Investments. Early career Scottish-born, she was educated at the Royal High School, Edinburgh. Richards began her career with a research fellowship at CERN, after graduating from the University of Edinburgh with a First-Class Honours degree in Electronics and Electrical Engineering. She subsequently worked for Cambridge Consultants, before pursuing further studies in France at INSEAD graduating as MBA. City career Richards worked initially as an analyst with Alliance Capital, before moving to JP Morgan, working in portfolio management, and then moving on to Mercury Asset Management, and later MLIM. In 2002 she took up the post of Chief Investment Officer and Joint Managing Director of Edinburgh Fund Managers plc. When EFM was taken over by Aberdeen Asset Management in 2003, Richards continued in her role as Chief Investment Officer. It was announced in February 2016 that she would be taking over from Michael McLintock as chief executive of M&G Investments from June 2016. In July 2018, it was announced that she would leave M&G Investments, following the merger to become M&G Prudential (and subsequent demerger from Prudential UK), to join Fidelity International as chief executive officer. In 2021, Richards led development of a range of flexible working options at Fidelity International, following the COVID-19 lockdown. Non-executive positions Richards was Vice-Convener of the Court of the University of Edinburgh and is a former director of both the Esure Group plc and the Scottish Chamber Orchestra. She is a member of the US-based Board of Leaders of 2020 Women on Boards, which works to increasing the proportion of women on corporate boards. Richards chaired the CERN and Society Foundation Board from 2015 to at least 2019. Fellowships and honours Dam
https://en.wikipedia.org/wiki/Geoffrey%20Horrocks%20%28mathematician%29
Geoffrey Horrocks (1932/33 Leicester – 12 September 2012) was a British mathematician working on vector bundles, who introduced the Horrocks construction used in the ADHM construction, and the Horrocks–Mumford bundle and monads. He was a professor at Newcastle University until his retirement in 1998. Publications
https://en.wikipedia.org/wiki/AUTOart
AUTOart is a Hong Kong-based scale model car line manufactured by Gateway Autoart Ltd. and sold by AA Collection Ltd. History AUTOart was established in 1998. Other lines of diecast vehicles formerly associated with AUTOart were Gateway, Gate and UT Models. The latter was originally a German company with diecast cars made in China and associated with Paul's Model Art which produces scale models under the Minichamps brand. AUTOart currently makes composite (ABS + diecast) model cars, switching to the material after formerly producing diecast models. The car models have been divided over time into various series, including the Millennium series, Performance series and Signature series. In addition to scale models, the brand also produces automobilia like 'shock absorber' pens, carbon fiber tissue boxes, household items, and clocks in the form of a disc brake rotor with caliper. Models AUTOart has produced scale models of more than 45 different car marques in nine different scales. Sizes generally range from 1:64 scale (2 to 3 inches long) to 1:12 scale (about 12 to 14 inches). Fine details such as carpeting, seat belts, door handles, engines, suspension, sun visors and door/hood/trunk lid dampers are reproduced. Many AUTOart models are of racing, sports, or performance vehicles, but commuter vehicles like the New Beetle or the Chrysler PT Cruiser are also produced. The AUTOart line consists of, but is not limited to, European vehicles. One example reviewed in Hemmings Sports & Exotic Car was the company's 1970 Lamborghini Espada 2+2 coupe which the magazine saw as a good model choice – distinct from the plethora of scale replicas made of the Lamborghini Countach or Diablo. AUTOart features both new vehicles, like the Chevrolet Corvette C7 or Lamborghini Veneno, as well as vintage racing cars like the 1965 Formula 1 Honda raced at Monaco. Historic racing cars have also included a number of other Can Am and Grand Prix cars from the 1960s and 1970s. The earliest year
https://en.wikipedia.org/wiki/Immunohistochemistry
Immunohistochemistry (IHC) is the most common application of immunostaining. It involves the process of selectively identifying antigens (proteins) in cells of a tissue section by exploiting the principle of antibodies binding specifically to antigens in biological tissues. IHC takes its name from the roots "immuno", in reference to antibodies used in the procedure, and "histo", meaning tissue (compare to immunocytochemistry). Albert Coons conceptualized and first implemented the procedure in 1941. Visualising an antibody-antigen interaction can be accomplished in a number of ways, mainly either of the following: Chromogenic immunohistochemistry (CIH), wherein an antibody is conjugated to an enzyme, such as peroxidase (the combination being termed immunoperoxidase), that can catalyse a colour-producing reaction. Immunofluorescence, where the antibody is tagged to a fluorophore, such as fluorescein or rhodamine. Immunohistochemical staining is widely used in the diagnosis of abnormal cells such as those found in cancerous tumors. Specific molecular markers are characteristic of particular cellular events such as proliferation or cell death (apoptosis). Immunohistochemistry is also widely used in basic research to understand the distribution and localization of biomarkers and differentially expressed proteins in different parts of a biological tissue. Sample preparation Preparation of the sample is critical to maintaining cell morphology, tissue architecture and the antigenicity of target epitopes. This requires proper tissue collection, fixation and sectioning. A solution of formalin is often used to fix tissue, but other methods may be used. Preparing tissue slices The tissue may then be sliced or used whole, dependent upon the purpose of the experiment or the tissue itself. Before sectioning, the tissue sample may be embedded in a medium, like paraffin wax or cryomedia. Sections can be sliced on a variety of instruments, most commonly a microtome, cryosta
https://en.wikipedia.org/wiki/NamePros
NamePros is an online community for domain name investors. Its services include forums and domain name auctions. The forums implement a freemium business model, whereby membership is free, but paid subscriptions offer additional features. History Ron "RJ" James publicly launched NamePros in February, 2003. He created NamePros to fill the hole left by Afternic, a popular online community that was quickly failing. NamePros started to see success around June, 2003, four months after its launch. Ron James dismissed the idea of adopting a subscription business model, favoring free services. Acquisition by Bodis Bodis, a domain parking company, acquired NamePros in January, 2012. Speculation circulating around blogs and other communities point to a sale price in the range of $200,000 to $300,000 USD. Rumors of the sale began as early as January 11. By January 19, Matt Wegrzyn, owner of Bodis, had publicly confirmed the acquisition. Matt Wegrzyn hinted that improvements to the website and its services would follow and stated that Bodis continue to keep NamePros an open community. Former owner Ron James noted that Bodis had better resources and would be capable of supporting NamePros' continued growth. In years leading up to the acquisition, the domain investor community noted some issues with website reliability. On May 14, 2013, NamePros released a series of updates to its service, including a new layout and the addition of a domain sales history tool. The new tool, developed by a community member, scraped domain sale information from other websites and RSS feeds. The author claimed to be indexing over 80,000 sales that members could query. Third administration NamePros changed ownership again in July, 2013. In May, 2014, NamePros released another series of updates. They migrated from vBulletin to XenForo, resulting in significant changes. DNForum, a competing website, made a similar update a month prior. Services NamePros' services center aro
https://en.wikipedia.org/wiki/Infectious%20tolerance
Infectious tolerance is a term referring to a phenomenon where a tolerance-inducing state is transferred from one cell population to another. It can be induced in many ways; although it is often artificially induced, it is a natural in vivo process. A number of research deal with the development of a strategy utilizing this phenomenon in transplantation immunology. The goal is to achieve long-term tolerance of the transplant through short-term therapy. History The term "infectious tolerance" was originally used by Gershon and Kondo in 1970 for suppression of naive lymphocyte populations by cells with regulatory function and for the ability to transfer a state of unresponsiveness from one animal to another. Gershon and Kondo discovered that T cells can not only amplify but also diminish immune responses. The T cell population causing this down-regulation was called suppressor T cells and was intensively studied for the following years (nowadays they are called regulatory T cells and are again a very attractive for research). These and other research in the 1970s showed greater complexity of immune regulation, unfortunately these experiments were largely disregarded, as methodological difficulties prevented clear evidence. Later developed new tolerogenic strategies have provided strong evidence to re-evaluate the phenomenon of T cell mediated suppression, in particular the use of non-depleting anti-CD4 monoclonal antibodies, demonstrating that neither thymus nor clonal deletion is necessary to induce tolerance. In 1989 was successfully induced classical transplantation tolerance to skin grafts in adult mice using antibodies blocking T cell coreceptors in CD4+ populations. Later was shown that the effect of monoclonal antibodies is formation of regulatory T lymphocytes. It has been shown that transfer of tolerance to other recipients can be made without further manipulation and that this tolerance transfer depends only on CD4+ T-lymphocytes. Because second-generation
https://en.wikipedia.org/wiki/Sequence%20graph
Sequence graph, also called an alignment graph, breakpoint graph, or adjacency graph, are bidirected graphs used in comparative genomics. The structure consists of multiple graphs or genomes with a series of edges and vertices represented as adjacencies between segments in a genome and DNA segments respectively. Traversing a connected component of segments and adjacency edges (called a thread) yields a sequence, which typically represents a genome or a section of a genome. The segments can be thought of as synteny blocks, with the edges dictating how to arrange these blocks in a particular genome, and the labelling of the adjacency edges representing bases that are not contained in synteny blocks. Construction Before constructing a sequence graph, there must be at least two genomes represented as directed graphs with edges as threads (adjacency edges) and vertices as DNA segments. The genomes should be labeled P and Q, while the sequence graph is labeled as BreakpointGraph(P, Q). The directional vertices of Q and their edges are arranged in the order of P. Once completed, the edges of Q are reconnected to their original vertices. After all edges have been matched the vertex directions are removed and instead each vertex is labeled as vh (vertex head) and vt (vertex tail). Similarity between genomes is represented by the number of cycles (independent systems) within the sequence graph. The number of cycles is equal to cycles (P, Q). The max number of cycles possible is equal to the number of vertices in the sequence graph. Example Figure example. Upon receiving genomes P (+a +b -c) and Q (+a +b -c), Q should be realigned to follow the direction edges (red) of P. The vertices should be renamed from a, b, c to ah at, bh bt, ch ct and the edges of P and Q should be connected to their original vertices (P edges = black, Q edges = green). Remove the directional edges (red). The number of cycles in G(P, Q) is 1 while the max possible is 3. Applications Breakpoint
https://en.wikipedia.org/wiki/Scapular%20line
The scapular line, also known as the linea scapularis, is a vertical line passing through the inferior angle of the scapula. It has been used in the evaluation of brachial plexus birth palsy.
https://en.wikipedia.org/wiki/Lawrence%20Blair
Lawrence Blair is an English anthropologist, author, explorer and filmmaker. He is the writer, presenter and co-producer of the TV series Ring of Fire, an Emmy award nominee and winner of the 1989 National Educational Film and Video Festival Silver Apple awards. Born in England, he has been a resident of Bali, Indonesia for the past 35 years. Life and career Lawrence Blair emigrated from England to Mexico with his parents and his brother Lorne in his early years. He has been a diver in Mexico and Indonesia, a fisherman in Alaska, an actor, a model, a photographer and an interpreter. He earned his PhD at Lancaster University, England with a doctoral thesis exploring and defining the field of psycho-anthropology. He wrote the book Rhythms of Vision: The Changing Patterns of Belief in 1976 in which he discussed sacred geometry, subtle energy, chakras, spiritual planes of existence and many other topics, the book has been compared to the work of the occultist Corinne Heline and the theosophist Alice Bailey. The book is most well known for first discussing the Hundredth monkey effect. Lyall Watson was a friend of Blair's and wrote a foreword for the book. Blair starred in the documentary Ring of Fire which discussed the varieties of volcanism and earthquake activity around the Pacific Rim. He wrote, co-produced and starred in the five film series documentary called Ring of Fire: An Indonesian Odyssey which highlights a 10-year exploration of Indonesian Islands, an adventure of Lawrence and his brother Lorne. The series was later developed into a book of the same name. Lawrence Blair went on to write and host Myth, Magic and Monsters a four-Part documentary series, in which he explores the Indonesian archipelago for rare and mysterious creatures, which include reptiles, ocean dwellers and domesticated animals. According to Blair: “There are still undiscovered species and tribes of people out there.” In 2009, Blair appeared in the documentary film Oh My God in which
https://en.wikipedia.org/wiki/Andrew%20Troelsen
Andrew W. Troelsen is currently a technology manager at Thomson Reuters in the Enterprise Content Platform (ECP - Big Data) division. He is an author of several books in the Microsoft technology space including books on Microsoft (D)COM, ATL, .NET, C#, VB (4.0 - modern) and COM & .NET Interoperability. His latest edition of his C# book covers the .NET Core platform and each C# 7.0 update. He has over 18 years experience authoring software development (3-5 day) workshops for engineers on MS platform technologies. Books Developer's Workshop to Com and Atl 3.0 (Andrew Troelsen (2000)) C# and the .NET Platform (Andrew Troelsen (2001)) COM and .NET Interoperability (Andrew Troelsen (2002)) Visual Basic .NET and the .NET Platform: An Advanced Guide (Andrew Troelsen (2002)) C# and the .NET Platform, Second Edition (Andrew Troelsen (2003)) Exploring .Net (Andrew Troelsen, Jason Bock (2003)) Pro C# 2005 and the .NET 2.0 Platform, Third Edition (Andrew Troelsen (2005)) Pro VB 2005 and the .NET 2.0 Platform, Second Edition (Andrew Troelsen (2006)) Expert ASP.NET 2.0: Advanced Application Design (Dominic Selly, Andrew Troelsen, Tom Barnaby (2006)) Pro C# 2008 and the .NET 3.5 Platform, Fourth Edition (Andrew Troelsen (2007)) Pro C# with .NET 3.0, Special Edition (Andrew Troelsen (2007)) Pro VB 2008 and the .NET 3.5 Platform (Andrew Troelsen (2008)) Pro VB 2010 and the .NET 4 Platform (Andrew Troelsen, Vidya Vrat Agarwal (2010)) Pro C# 2010 and the .NET 4 Platform, Fifth Edition (Andrew Troelsen (2010)) Pro Expression Blend 4 (Andrew Troelsen (2011)) Pro C# 5.0 and the .NET 4.5 Framework, Sixth Edition (Andrew Troelsen (2012)) C# 6.0 and the .NET 4.6 Framework, Seventh Edition (Andrew Troelsen, Philip Japikse (2015)) Pro C# 7: With .NET and .NET Core, Eighth Edition (Andrew Troelsen, Philip Japikse (2017)) Pro C# 8 with .NET Core 3: Foundational Principles and Practices in Programming, Ninth Edition (Andrew Troelsen, Phil Japikse (2020)) Pro C# 9 with .NET
https://en.wikipedia.org/wiki/Matrox%20Simple%20Interface
Matrox Simple Interface (in short MSI) is the name of a proprietary DOS and Windows 95 application programming interface for Matrox Mystique graphics cards made by Matrox. MSI API supported a maximum of 640x480x16 resolution with z-buffer and no bilinear filtering. It used color look up tables to save memory. When Matrox released the Matrox m3D (using the PowerVR PCX2 chipset), MSI was completely abandoned. Sources https://web.archive.org/web/20110807162742/http://www.forums.murc.ws/archive/index.php/t-38427.html Legacy Matrox-enhanced games on g200/g400 [Archive] - MURC http://www.geocities.com/k_lupinsky/Mystique.htm Kanajana's MechWarrior 2 3D Page - Matrox Mystique Edition http://www.unet.univie.ac.at/~a0503736/php/drdoswiki/index.php?n=Main.Hardware Club Dr-DOS Wiki Graphics cards 3D graphics APIs
https://en.wikipedia.org/wiki/Tally%20stick
A tally stick (or simply tally) was an ancient memory aid device used to record and document numbers, quantities and messages. Tally sticks first appear as animal bones carved with notches during the Upper Palaeolithic; a notable example is the Ishango Bone. Historical reference is made by Pliny the Elder (AD 23–79) about the best wood to use for tallies, and by Marco Polo (1254–1324) who mentions the use of the tally in China. Tallies have been used for numerous purposes such as messaging and scheduling, and especially in financial and legal transactions, to the point of being currency. Kinds of tallies Principally, there are two different kinds of tally sticks: the single tally and the split tally. A common form of the same kind of primitive counting device is seen in various kinds of prayer beads. Possible palaeolithic tally sticks A number of anthropological artefacts have been conjectured to be tally sticks: The Lebombo bone, dated between 44,200 and 43,000 years old, is a baboon's fibula with 29 distinct notches, discovered within the Border Cave in the Lebombo Mountains of Eswatini. The so-called Wolf bone (cs) is a prehistoric artefact discovered in 1937 in Czechoslovakia during excavations at Dolní Věstonice, Moravia, led by Karl Absolon. Dated to the Aurignacian, approximately 30,000 years ago, the bone is marked with 55 marks which some believe to be tally marks. The head of an ivory Venus figurine was excavated close to the bone. The Ishango bone is a bone tool, dated to the Upper Palaeolithic era, around 18,000 to 20,000 BC. It is a dark brown length of bone. It has a series of possible tally marks carved in three columns running the length of the tool. It was found in 1950 in Ishango (east Belgian Congo). Single tally The single tally stick was an elongated piece of bone, ivory, wood, or stone which is marked with a system of notches (see: Tally marks). The single tally stick serves predominantly mnemonic purposes. Related to the single tally con
https://en.wikipedia.org/wiki/RDH13
Retinol dehydrogenase 13 (all-trans/9-cis) is a protein that in humans is encoded by the RDH13 gene. This gene encodes a mitochondrial short-chain dehydrogenase/reductase, which catalyzes the reduction and oxidation of retinoids. The encoded enzyme may function in retinoic acid production and may also protect the mitochondria against oxidative stress. Alternatively spliced transcript variants have been described. Gene The human RDH13 gene is on the 19th chromosome, with its specific localization being 19q13.42. The gene contains 12 exons in total. Structure The analysis of the submitochondrial localization of RDH13 indicates its association with the inner mitochondrial membrane. The primary structure of RDH13 contains two hydrophobic segments, 2–21 and 242–261, which are sufficiently long to serve as transmembrane segments; however, as shown in the present study, alkaline extraction completely removes the protein from the membrane, indicating that RDH13 is a peripheral membrane protein. The peripheral association of RDH13 with the membrane further distinguishes this protein from the microsomal retinaldehyde reductases, which are integral membrane proteins that appear to be anchored in the membrane via their N-terminal hydrophobic segments. Function RDH13 is most closely related to the NADP+-dependent microsomal enzymes RDH11, RDH12 and RDH14. Purified RDH13 acts on retinoids in an oxidative reductive manner, and strongly prefers the cofactor NADPH over NADH. Moreover, RDH13 is much has much more efficient reductase activity than dehydrogenase activity. RDH13 as a retinaldehyde reductase is significantly less active than that of a related protein RDH11, primarily because of the much higher Km value for retinaldehyde. However, the kcat value of RDH13 for retinaldehyde reduction. arable with that of RDH11, and the Km values of the two enzymes for NADPH are also very similar. Thus, consistent with its sequence similarity to RDH11, RDH12 and RDH14, RDH13 acts as
https://en.wikipedia.org/wiki/Linksys%20WRT54G%20series
The Linksys WRT54G Wi-Fi series is a series of Wi-Fi–capable residential gateways marketed by Linksys, a subsidiary of Cisco, from 2003 until acquired by Belkin in 2013. A residential gateway connects a local area network (such as a home network) to a wide area network (such as the Internet). Models in this series use one of various 32-bit MIPS processors. All WRT54G models support Fast Ethernet for wired data links, and 802.11b/g for wireless data links. Hardware and revisions WRT54G The original WRT54G was first released in December 2002. It has a 4+1 port network switch (the Internet/WAN port is part of the same internal network switch, but on a different VLAN). The devices have two removable antennas connected through Reverse Polarity TNC connectors. The WRT54GC router is an exception and has an internal antenna with optional external antenna. As a cost-cutting measure, as well as to satisfy FCC rules that prohibit fitting external antennas with higher gain, the design of the latest version of the WRT54G no longer has detachable antennas or TNC connectors. Instead, version 8 routers simply route thin wires into antenna 'shells' eliminating the connector. As a result, Linksys HGA7T and similar external antennas are no longer compatible with this model. Until version 5, WRT54G shipped with Linux-based firmware. WRT54GS The WRT54GS is nearly identical to the WRT54G except for additional RAM, flash memory, and SpeedBooster software. Versions 1 to 3 of this router have 8 MB of flash memory. Since most third parties' firmware only use up to 4 MB flash, a JFFS2-based read/write filesystem can be created and used on the remaining 4 MB free flash. This allows for greater flexibility of configurations and scripting, enabling this small router to both load-balance multiple ADSL lines (multi-homed) or to be run as a hardware layer-2 load balancer (with appropriate third party firmware). WRT54GL Linksys released the WRT54GL (the best-selling router of all time)
https://en.wikipedia.org/wiki/Grady%20Booch
Grady Booch (born February 27, 1955) is an American software engineer, best known for developing the Unified Modeling Language (UML) with Ivar Jacobson and James Rumbaugh. He is recognized internationally for his innovative work in software architecture, software engineering, and collaborative development environments. Education Booch earned his bachelor's degree in 1977 from the United States Air Force Academy and a master's degree in electrical engineering in 1979 from the University of California, Santa Barbara. Career and research Booch worked at Vandenberg Air Force Base after he graduated. He started as a project engineer and later managed ground-support missions for the space shuttle and other projects. After he gained his master's degree he became an instructor at the Air Force Academy. Booch served as Chief Scientist of Rational Software Corporation from its founding in 1981 through its acquisition by IBM in 2003, where he continued to work until March 2008. After this he became Chief Scientist, Software Engineering in IBM Research and series editor for Benjamin Cummings. Booch has devoted his life's work to improving the art and the science of software development. In the 1980s, he wrote one of the more popular books on programming in Ada. He is best known for developing the Unified Modeling Language with Ivar Jacobson and James Rumbaugh in the 1990s. IBM 1130 Booch got his first exposure to programming on an IBM 1130. ... I pounded the doors at the local IBM sales office until a salesman took pity on me. After we chatted for a while, he handed me a Fortran [manual]. I'm sure he gave it to me thinking, "I'll never hear from this kid again." I returned the following week saying, "This is really cool. I've read the whole thing and have written a small program. Where can I find a computer?" The fellow, to my delight, found me programming time on an IBM 1130 on weekends and late-evening hours. That was my first programming experience, and I must thank
https://en.wikipedia.org/wiki/MIR9-3HG
MIR9-3 host gene is a protein that in humans is encoded by the MIR9-3HG gene.
https://en.wikipedia.org/wiki/Compartmentalized%20ciliogenesis
Compartmentalized ciliogenesis is the most common type of ciliogenesis where the cilium axoneme is formed separated from the cytoplasm by the ciliary membrane and a ciliary gate known as the transition zone.
https://en.wikipedia.org/wiki/QualNet
QualNet is a testing and simulation tool owned and provided by Scalable Network Technologies, Inc. As network simulation software, it acts as a planning, testing, and training tool which mimics the behavior of a physical communications network. See also Network simulation Wireless networking Computer network analysis Computer networking Simulation software
https://en.wikipedia.org/wiki/Translation%20%28biology%29
In biology, translation is the process in living cells in which proteins are produced using RNA molecules as templates. The generated protein is a sequence of amino acids. This sequence is determined by the sequence of nucleotides in the RNA. The nucleotides are considered three at a time. Each such triple results in addition of one specific amino acid to the protein being generated. The matching from nucleotide triple to amino acid is called the genetic code. The translation is performed by a large complex of functional RNA and proteins called ribosomes. The entire process is called gene expression. In translation, messenger RNA (mRNA) is decoded in a ribosome, outside the nucleus, to produce a specific amino acid chain, or polypeptide. The polypeptide later folds into an active protein and performs its functions in the cell. The ribosome facilitates decoding by inducing the binding of complementary tRNA anticodon sequences to mRNA codons. The tRNAs carry specific amino acids that are chained together into a polypeptide as the mRNA passes through and is "read" by the ribosome. Translation proceeds in three phases: Initiation: The ribosome assembles around the target mRNA. The first tRNA is attached at the start codon. Elongation: The last tRNA validated by the small ribosomal subunit (accommodation) transfers the amino acid. It carries to the large ribosomal subunit which binds it to the one of the preceding admitted tRNA (transpeptidation). The ribosome then moves to the next mRNA codon to continue the process (translocation), creating an amino acid chain. Termination: When a stop codon is reached, the ribosome releases the polypeptide. The ribosomal complex remains intact and moves on to the next mRNA to be translated. In prokaryotes (bacteria and archaea), translation occurs in the cytosol, where the large and small subunits of the ribosome bind to the mRNA. In eukaryotes, translation occurs in the cytoplasm or across the membrane of the endoplasmic ret
https://en.wikipedia.org/wiki/Anatomy
Anatomy () is the branch of biology concerned with the study of the structure of organisms and their parts. Anatomy is a branch of natural science that deals with the structural organization of living things. It is an old science, having its beginnings in prehistoric times. Anatomy is inherently tied to developmental biology, embryology, comparative anatomy, evolutionary biology, and phylogeny, as these are the processes by which anatomy is generated, both over immediate and long-term timescales. Anatomy and physiology, which study the structure and function of organisms and their parts respectively, make a natural pair of related disciplines, and are often studied together. Human anatomy is one of the essential basic sciences that are applied in medicine. Anatomy is a complex and dynamic field that is constantly evolving as new discoveries are made. In recent years, there has been a significant increase in the use of advanced imaging techniques, such as MRI and CT scans, which allow for more detailed and accurate visualizations of the body's structures. The discipline of anatomy is divided into macroscopic and microscopic parts. Macroscopic anatomy, or gross anatomy, is the examination of an animal's body parts using unaided eyesight. Gross anatomy also includes the branch of superficial anatomy. Microscopic anatomy involves the use of optical instruments in the study of the tissues of various structures, known as histology, and also in the study of cells. The history of anatomy is characterized by a progressive understanding of the functions of the organs and structures of the human body. Methods have also improved dramatically, advancing from the examination of animals by dissection of carcasses and cadavers (corpses) to 20th-century medical imaging techniques, including X-ray, ultrasound, and magnetic resonance imaging. Etymology and definition Derived from the Greek anatomē "dissection" (from anatémnō "I cut up, cut open" from ἀνά aná "up", and τέμνω té
https://en.wikipedia.org/wiki/Mystery%20meat
Mystery meat is a disparaging term for meat products that have an unidentifiable source, typically ground or otherwise ultra-processed foods such as burger patties, chicken nuggets, Salisbury steaks, sausages and hot dogs. Most often the term is used in reference to food served in institutional cafeterias, such as prison food or a North American school lunch. The term is also sometimes applied to meat products where the species from which the meat has come from is known, but the cuts of meat used are unknown. This is often the case where the cuts of meat used include offal and mechanically separated meat, or when non-meat substitutes such as textured vegetable protein are used to stretch the meat, where explicitly stating the type of meat used might diminish the perceived palatability of the product to some purchasers. Common products The most common mystery meat products sold in The United States include Spam and sometimes sausages. It is also disputed whether or not bologna/baloney is a 'mystery meat' product. Use in marketing In 2016, Nissin, a Japanese food company that produces Cup Noodles, started to call their ingredients as self-deprecating Nazoniku (literally Mystery Meat) as part of their official marketing campaign. Nazoniku, or formally known as Daisuminchi (literally Minced Meat Dice), are made from pork, soybeans and other ingredients. See also Pink slime Mystery meat navigation Chicken McNuggets
https://en.wikipedia.org/wiki/Gadolinium%20phosphide
Gadolinium phosphide is an inorganic compound of gadolinium and phosphorus with the chemical formula GdP. Synthesis Gadolinium phosphide can be obtained by reacting gadolinium and phosphorus at high temperature, and single crystals can be obtained by mineralization. 4 Gd + P4 → 4 GdP Physical properties GdP has a NaCl-structure and transforms to a CsCl-structure at 40 GPa. GdP forms crystals of a cubic system, space group Fm3m. Gadolinium phosphide is antiferromagnetic. Uses The compound is a semiconductor used in high power, high frequency applications and in laser diodes.
https://en.wikipedia.org/wiki/Cascade%20algorithm
In the mathematical topic of wavelet theory, the cascade algorithm is a numerical method for calculating function values of the basic scaling and wavelet functions of a discrete wavelet transform using an iterative algorithm. It starts from values on a coarse sequence of sampling points and produces values for successively more densely spaced sequences of sampling points. Because it applies the same operation over and over to the output of the previous application, it is known as the cascade algorithm. Successive approximation The iterative algorithm generates successive approximations to ψ(t) or φ(t) from {h} and {g} filter coefficients. If the algorithm converges to a fixed point, then that fixed point is the basic scaling function or wavelet. The iterations are defined by For the kth iteration, where an initial φ(0)(t) must be given. The frequency domain estimates of the basic scaling function is given by and the limit can be viewed as an infinite product in the form If such a limit exists, the spectrum of the scaling function is The limit does not depends on the initial shape assume for φ(0)(t). This algorithm converges reliably to φ(t), even if it is discontinuous. From this scaling function, the wavelet can be generated from Successive approximation can also be derived in the frequency domain.
https://en.wikipedia.org/wiki/EMC%20Elastic%20Cloud%20Storage
EMC Elastic Cloud Storage (ECS), formerly Project Nile, is an object storage software product marketed by EMC Corporation. It is marketed as software-defined storage that follows several principles of object storage, such as scalability, data resilience, and cost efficiency. Applications EMC Elastic Cloud Storage has a number of applications, including the Internet of things and financial services, where it was determined by Cohasset Associates Inc. to meet "the relevant storage requirements of SEC Rule 17a-4(f) and CFTC Rule 1.31(b)-(c)" when "Compliance is enabled for a Namespace and when properly configured and utilized to store and retain records in non-erasable and non-rewriteable format." Its use of object storage and flat namespace, according to the Edison Group, "allows for multiple types of data to be stored side by side. Regardless of the data, it is all viewed as object, their globally unique IDs, and metadata." This approach allows multiple data types from multiple sources to be stored alongside one another, including: Large data sets: Financial, pharmaceutical, geospatial, biotech, and legal Public data sets: Weather, government Security, imagery, and social media: Images, videos, blogs Revenue chain data: Sensors, devices, Internet of Things. ECS was mentioned in a marketing vendor assessment for object storage in 2014.
https://en.wikipedia.org/wiki/Midpoint%20circle%20algorithm
In computer graphics, the midpoint circle algorithm is an algorithm used to determine the points needed for rasterizing a circle. It's a generalization of Bresenham's line algorithm. The algorithm can be further generalized to conic sections. and Van Aken. In machining (CNC), it is known as circular interpolation. Summary This algorithm draws all eight octants simultaneously, starting from each cardinal direction (0°, 90°, 180°, 270°) and extends both ways to reach the nearest multiple of 45° (45°, 135°, 225°, 315°). It can determine where to stop because when = , it has reached 45°. The reason for using these angles is shown in the above picture: As increases, it does not skip nor repeat any value until reaching 45°. So during the while loop, increments by 1 each iteration, and decrements by 1 on occasion, never exceeding 1 in one iteration. This changes at 45° because that is the point where the tangent is rise=run. Whereas rise>run before and rise<run after. The second part of the problem, the determinant, is far trickier. This determines when to decrement . It usually comes after drawing the pixels in each iteration, because it never goes below the radius on the first pixel. Because in a continuous function, the function for a sphere is the function for a circle with the radius dependent on (or whatever the third variable is), it stands to reason that the algorithm for a discrete(voxel) sphere would also rely on this Midpoint circle algorithm. But when looking at a sphere, the integer radius of some adjacent circles is the same, but it is not expected to have the same exact circle adjacent to itself in the same hemisphere. Instead, a circle of the same radius needs a different determinant, to allow the curve to come in slightly closer to the center or extend out farther. Algorithm The objective of the algorithm is to approximate the curve using pixels; in layman's terms every pixel should be approximately the same distance from the center.
https://en.wikipedia.org/wiki/Hyperbolic%20geometric%20graph
A hyperbolic geometric graph (HGG) or hyperbolic geometric network (HGN) is a special type of spatial network where (1) latent coordinates of nodes are sprinkled according to a probability density function into a hyperbolic space of constant negative curvature and (2) an edge between two nodes is present if they are close according to a function of the metric (typically either a Heaviside step function resulting in deterministic connections between vertices closer than a certain threshold distance, or a decaying function of hyperbolic distance yielding the connection probability). A HGG generalizes a random geometric graph (RGG) whose embedding space is Euclidean. Mathematical formulation Mathematically, a HGG is a graph with a vertex set V (cardinality ) and an edge set E constructed by considering the nodes as points placed onto a 2-dimensional hyperbolic space of constant negative Gaussian curvature, and cut-off radius , i.e. the radius of the Poincaré disk which can be visualized using a hyperboloid model. Each point has hyperbolic polar coordinates with and . The hyperbolic law of cosines allows to measure the distance between two points and , The angle  is the (smallest) angle between the two position vectors. In the simplest case, an edge is established iff (if and only if) two nodes are within a certain neighborhood radius , , this corresponds to an influence threshold. Connectivity decay function In general, a link will be established with a probability depending on the distance . A connectivity decay function represents the probability of assigning an edge to a pair of nodes at distance . In this framework, the simple case of hard-code neighborhood like in random geometric graphs is referred to as truncation decay function. Generating hyperbolic geometric graphs Krioukov et al. describe how to generate hyperbolic geometric graphs with uniformly random node distribution (as well as generalized versions) on a disk of radius in . These
https://en.wikipedia.org/wiki/Zero-Force%20Evolutionary%20Law
The Zero-Force Evolutionary Law (ZFEL) is a theory proposed by Daniel McShea and Robert Brandon regarding the evolution of diversity and complexity. Under the ZFEL, diversity is understood as the variation among organisms and complexity as the variation among the parts within an organism. A part is understood as a system that is to some degree internally integrated and isolated from its surroundings. In a multicellular organism, for example, a cell is a part, and therefore complexity is the number of different cell types. Like the theory of relativity, the theory has a special and general formulation. The special formulation states that in the absence of natural selection, an evolutionary system with variation and heredity will tend spontaneously to diversify and complexify. The general formulation states that evolutionary systems have a tendency to diversify and complexify, but that these processes may be amplified or constrained by other forces, including natural selection. The mechanism of the ZFEL is the inherently error-prone process of replication and reproduction. In the absence of selection, errors tend to accumulate, with the result that individuals within a population tend to become more different from each other (diversity) and parts within an individual tend to become more different from each other (complexity). Both of these tendencies can be overcome by selection, including stabilizing or negative selection, with the result that diversity or complexity often does not change, or even decreases. What the ZFEL offers is not so much a prediction as a null expectation, telling us what will happen in evolution when selection is absent. It is the analogue of Newton's law of momentum, which tells us the trajectory of a moving object in the absence of forces (a straight line). See also Constructive neutral evolution Evolution of biological complexity Neutral theory of molecular evolution
https://en.wikipedia.org/wiki/Graphmatica
Graphmatica is a graphing program created by Keith Hertzer, a graduate of the University of California, Berkeley. It runs on Microsoft Windows (all versions), Mac OS X 10.5 and higher, and iOS 5.0 and higher. Graphmatica for Windows and Macs is distributed free of charge for evaluation purposes. After one month, non-commercial users are asked to pay a $25 licensing fee. Other licensing plans are available for commercial users. Graphmatica for iOS is distributed via the Apple App Store. Capabilities Graphmatica can graph Cartesian functions, relations, and inequalities, plus polar, parametric and ordinary differential equations. See also C.a.R. KmPlot