source
stringlengths
31
227
text
stringlengths
9
2k
https://en.wikipedia.org/wiki/Nitroxoline
Nitroxoline is an antibiotic that has been in use in Europe for about fifty years, and has proven to be very effective at combating biofilm infections. Nitroxoline was shown to cause a decrease in the biofilm density of P. aeruginosa infections, which would allow access to the infection by the immune system in vivo. It was shown that nitroxoline functions by chelating Fe2+ and Zn2+ ions from the biofilm matrix; when Fe2+ and Zn2+ were reintroduced into the system, biofilm formation was reconstituted. The activity of biofilm degradation is comparable to EDTA, but has a history of human use in clinical settings and therefore has a precedent with which to allow its use against “slimy” biofilm infections. Anticancer activity The chelating activities of nitroxoline have also been used in an anticancer setting. Nitroxoline has been shown to be more cytotoxic to HL60, DHL-4, PANC-1, and cells lines than clioquinol and other 8-hydroxyquinoline derivatives. It also demonstrated an increase in reactive oxygen species (ROS) production over controls, especially when Cu2+ was added. The ROS levels reached over 350% of the controls with addition of CuCl2. The cytotoxicity production was markedly decreased with addition of ZnCl2, indicating, based on this model, that nitroxoline is not a zinc chelator. Because the zinc chelating action of clioquinol has been associated with subacute myelo-optic neuropathy, the use of nitroxoline as a cytotoxic drug in the treatment of cancers should not exhibit neurotoxic effects in humans, and in vivo trials on tumour xenografts in mice have not yielded any negative neurodegenerative effects. Nitroxoline has been shown to inhibit the enzymatic activity of cathepsin B. Cathepsin B degrades extra-cellular membrane proteins in tumor cells, allowing them to proliferate more freely, and metastasize throughout the body. Nitroxoline was shown to be a noncompetitive, reversible inhibitor of these actions in MCF-10A neoT cells. The Ki (dissoc
https://en.wikipedia.org/wiki/Clofoctol
Clofoctol is a bacteriostatic antibiotic. It is used in the treatment of respiratory tract and ear, nose and throat infections caused by Gram-positive bacteria. It has been marketed in France till 2005 under the trade name Octofene and in Italy as Gramplus. It is only functional against Gram-positive bacteria. It penetrates into human lung tissue. Apteeus is developing clofoctol as a potential therapy against SARS-CoV-2.
https://en.wikipedia.org/wiki/Xibornol
Xibornol is a lipophilic substance with antiseptic properties, mainly used in Italy and Spain. It is primarily administered to the throat as a spray mouthwash. As of 2007, all approved forms are water-based suspensions. The drug was discovered in the 1970s.
https://en.wikipedia.org/wiki/Fosfomycin
Fosfomycin, sold under the brand name Monurol among others, is an antibiotic primarily used to treat lower urinary tract infections. It is not indicated for kidney infections. Occasionally it is used for prostate infections. It is generally taken by mouth. Common side effects include diarrhea, nausea, headache, and vaginal yeast infections. Severe side effects may include anaphylaxis and Clostridioides difficile-associated diarrhea. While use during pregnancy has not been found to be harmful, such use is not recommended. A single dose when breastfeeding appears safe. Fosfomycin works by interfering with the production of the bacterial cell wall. Fosfomycin was discovered in 1969 and approved for medical use in the United States in 1996 . It is on the World Health Organization's List of Essential Medicines. The World Health Organization classifies fosfomycin as critically important for human medicine. It is available as a generic medication. It was originally produced by certain types of Streptomyces, although it is now made chemically. Medical uses Fosfomycin is used to treat bladder infections, where it is usually given as a single dose by mouth. Oral fosfomycin is not recommended for children under 12 years old. Additional uses have been proposed. The global problem of advancing antimicrobial resistance has led to a renewed interest in its use more recently. Fosfomycin can be used as an efficacious treatment for both UTIs and complicated UTIs including acute pyelonephritis. The standard regimen for complicated UTIs is an oral 3 g dose administered once every 48 or 72 hours for a total of 3 doses or a 6 g dose every 8 hours for 7–14 days when fosfomycin is given in IV form. Intravenous fosfomycin is being increasingly used for treating infections caused by multidrug-resistant bacteria, mostly as a partner drug in order to avoid the occurrence of resistances and to take advantage of its synergistic activity with several other antimicrobials. Daily adult dose
https://en.wikipedia.org/wiki/Nuclear%20bodies
Nuclear bodies (also known as nuclear domains, or nuclear dots) are membraneless structures found in the cell nuclei of eukaryotic cells. Nuclear bodies include Cajal bodies, the nucleolus, and promyelocytic leukemia protein (PML) nuclear bodies (also called PML oncogenic dots). Nuclear bodies also include ND10s. ND stands for nuclear domain, and 10 refers to the number of dots seen. Nuclear bodies were first seen as prominent interchromatin structures in the nuclei of malignant or hyperstimulated animal cells identified using anti-sp100 autoantibodies from primary biliary cirrhosis and subsequently the promyelocytic leukemia (PML) factor, but appear also to be elevated in many autoimmune and cancerous diseases. Nuclear dots are metabolically stable and resistant to nuclease digestion and salt extraction. A nuclear body subtype is a clastosome suggested to be a site of protein degradation. Structure Simple nuclear bodies (types I and II) and the shells of complex nuclear bodies (types III, IVa and V) consist of a non-chromatinic fibrillar material which is most likely proteinaceous. That nuclear bodies co-isolated with the nuclear matrix, and were linked to the fibrogranular nuclear matrix component by projections from the surface of the nuclear bodies. The primary components of the nuclear dots are the proteins sp100 nuclear antigen, LYSP100(a homolog of sp100), ISG20, PML antigen, NDP55 and 53kDa protein associated with the nuclear matrix. Other proteins, such as PIC1/SUMO-1, which are associated with nuclear pore complex also associate with nuclear dots. The proteins can reorganize in the nucleus, by increasing number of dispersion in response to different stress (stimulation or heat shock, respectively). Function One of the nuclear body proteins appears to be involved in transcriptional active regions. Expression of PML antigen and sp100 is responsive to interferons. Sp100 seems to have transcriptional transactivating properties. PML protein was reporte
https://en.wikipedia.org/wiki/Cholesterolosis%20of%20gallbladder
In surgical pathology, strawberry gallbladder, more formally cholesterolosis of the gallbladder and gallbladder cholesterolosis, is a change in the gallbladder wall due to excess cholesterol. The name strawberry gallbladder comes from the typically stippled appearance of the mucosal surface on gross examination, which resembles a strawberry. Cholesterolosis results from abnormal deposits of cholesterol esters in macrophages within the lamina propria (foam cells) and in mucosal epithelium. The gallbladder may be affected in a patchy localized form or in a diffuse form. The diffuse form macroscopically appears as a bright red mucosa with yellow mottling (due to lipid), hence the term strawberry gallbladder. It is not tied to cholelithiasis (gallstones) or cholecystitis (inflammation of the gallbladder). Additional images See also Cholecystectomy Rokitansky-Aschoff sinuses
https://en.wikipedia.org/wiki/Ski%20complex
The Ski complex is a multi-protein complex involved in the 3' end degradation of messenger RNAs in yeast. Structure The complex consists of three main proteins, the RNA helicase Ski2 and the proteins Ski3 and Ski8. This tetramer contains a 370 kDa core complex, containing N-terminal arms and C-terminal arms from Ski3. The helicase core of Ski2 is positioned by both the C-terminal of Ski3 and two subunits of Ski8. Mechanism Helicase activities are initiated by the N-terminal arm and the Ski2 insertion domain. In yeast, the complex guides RNA molecules to the exosome complex for degradation via a fourth protein, called Ski7, which contains a GTPase-like protein. Ski7 involves the 3’ to 5’ degradation of RNA through two different pathways, 3’ poly(A) tail shortening and the binding of the Ski2, Ski3, and Ski8 tetramer and the exosome. Degradation of the 3' mRNA overhang occurs by association with the 80s ribosome. The 3' end of the mRNA is threaded through the ribosome to Ski2, preparing it for the degradation process. Biochemical studies also show that the Ski complex can thread RNA through the exosome complex, thereby coupling the Ski2 protein helicase function with the exoribonuclease activity, leading to degradation of the RNA strand. See also TRAMP complex Non-stop decay Nonsense mediated decay
https://en.wikipedia.org/wiki/MeshBox
MeshBox is an item of computer hardware which is used to provide large scale wireless broadband networks. Manufactured by LocustWorld, the devices are designed to co-operate with other MeshBoxes within range, passing the internet service from one box to the next, over the air, until it reaches the final destination. The coverage area of a mesh is typically measured in square miles or square kilometres. Originally released as a bootable CD-ROM called MeshAP based on the OpenAP open source software, the system is now implemented as system image which can fit within a small 32MB CompactFlash card. The system functions have expanded beyond creating wireless networks to provide set-top box services, MP3 audio and video streaming, connection to remote windows terminal servers, other PCs, web browsing, connection to peer-to-peer networks, instant messaging and file exchange. See also Wireless mesh network Wireless LAN Wireless access point Wireless community network IEEE 802.11 External links CommunityWireless.org Wireless Internet Assigned Numbers Authority Hardware routers Linux-based devices IEEE 802.11 Mesh networking
https://en.wikipedia.org/wiki/Golomb%E2%80%93Dickman%20constant
In mathematics, the Golomb–Dickman constant arises in the theory of random permutations and in number theory. Its value is It is not known whether this constant is rational or irrational. Definitions Let an be the average — taken over all permutations of a set of size n — of the length of the longest cycle in each permutation. Then the Golomb–Dickman constant is In the language of probability theory, is asymptotically the expected length of the longest cycle in a uniformly distributed random permutation of a set of size n. In number theory, the Golomb–Dickman constant appears in connection with the average size of the largest prime factor of an integer. More precisely, where is the largest prime factor of k . So if k is a d digit integer, then is the asymptotic average number of digits of the largest prime factor of k. The Golomb–Dickman constant appears in number theory in a different way. What is the probability that second largest prime factor of n is smaller than the square root of the largest prime factor of n? Asymptotically, this probability is . More precisely, where is the second largest prime factor n. The Golomb-Dickman constant also arises when we consider the average length of the largest cycle of any function from a finite set to itself. If X is a finite set, if we repeatedly apply a function f: X → X to any element x of this set, it eventually enters a cycle, meaning that for some k we have for sufficiently large n; the smallest k with this property is the length of the cycle. Let bn be the average, taken over all functions from a set of size n to itself, of the length of the largest cycle. Then Purdom and Williams proved that Formulae There are several expressions for . These include: where is the logarithmic integral, where is the exponential integral, and and where is the Dickman function. See also Random permutation Random permutation statistics External links
https://en.wikipedia.org/wiki/Heart%20in%20hand
The Heart in Hand or Heart-in-Hand is a symbol of a heart in an open palm, and is symbolic of charity, given from the heart. It is an easily recognizable symbol in the Northeastern United States and used by the Shakers as a pictorial reminder of the words of Mother Ann Lee, the founder of the Shaker sect, who promoted a simple life of hard work and spirituality, "Put your hands to work, and your hearts to God." The image is typical of the Shaker attitude, and also implies a loving welcome. A heart in hand has also been a symbol of the Independent Order of Odd Fellows, a fraternal organization derived from English Oddfellows orders of the mid-18th century. These commonly display three linked rings representing friendship, love, and truth. The symbol originated as the seal of the Protestant Reformer John Calvin, and thus appears in the official seal of Calvin College with Calvin's motto, "My heart I offer to you LORD, promptly and sincerely." The symbol is also frequently associated with Amish communities. The symbol was also used as a chop mark in the woodblock prints of Gustave Baumann (Germ./Amer. 1881-1971). See also Religious symbolism Notes Religious symbols Heart symbols
https://en.wikipedia.org/wiki/Programmable%20sound%20generator
A programmable sound generator (PSG) is a sound chip that generates (or synthesizes) audio wave signals built from one or more basic waveforms, and often some kind of noise. PSGs use a relatively simple method of creating sound compared to other methods such as frequency modulation synthesis or pulse-code modulation. Technical details PSGs are controlled by writing data to dedicated registers on the chip via an external CPU; hence the name programmable sound generator. One or more basic waveforms are generated (typically a square, triangle or saw-tooth wave) and often a noise signal. The waveforms' frequency and volume (and noise's tone and volume) are typically shaped using an envelope and or mixed before being sent to the audio output stage. Many PSGs feature three tone channels and one noise channel including the AY-3-8910, SN76489 and MOS Technology 6581. History In the late 1970s, more electronic consumer devices began to be designed with audio features. PSG were partly developed as a way of incorporating relatively complex sounds at a low cost. PSGs were in many arcade games, game consoles, and home computers of the 1980s and 90s. In 1978, General Instrument released the AY-3-8910, the design of which was later licensed by Yamaha Corporation for their YM2149. These chips were used as the standard for the MSX computer standards 1 and 2, respectively. The features of this chip were also incorporated into other Yamaha sound chips including the YM2203 and YM2608 chips, these were also capable of FM synthesis. In the same year Atari designed the POKEY chip for its home computers and game systems. It incorporated a PSG. In 1979, Texas Instruments SN76489 was produced for the TI-99/4 computer. This was also used in the Tandy 1000 and IBM PCjr. In 1982, MOS Technology 6581 was produced for the Commodore 64. The main chip in the Nintendo Entertainment System, the Ricoh 2A03, included a PSG. List of PSGs See also List of sound card standards List of Yamaha so
https://en.wikipedia.org/wiki/V%C3%A1clav%20E.%20Bene%C5%A1
Václav Edvard "Vic" Beneš (born January 1, 1931) is a Czech-American, a mathematician known for his contributions to the theory of stochastic processes, queueing theory and control theory, as well as the design of telecommunications switches. He studied under John Kemeny and gained a doctorate in mathematics at Princeton University (1953) on a treatise on Mathematical logic. He then worked for Bell Labs until 1986, contributing to Kalman filter theory as well as the Beneš network, a permutation network of the Clos network type. In the 1980s he held a position at Columbia University as well. He has continued to publish independently since 1989. He was elected IEEE Fellow (1991) for "contributions to the structure of telephone connecting networks, stochastic control, and nonlinear filtering". The Benesfest was celebrated at Columbia University (2001) to honor his 70th birthday. He resides in Millburn, New Jersey (since 1985) where he has been a long-time mountain climber and member of the American Alpine Club, and currently heads the local historical society. He is a relative of the former President of Czechoslovakia Edvard Beneš and politician Vojta Beneš. Emilie Benes Brzezinski, a sculptor, was his sister. His first wife Janet was the daughter of Philip Franklin and niece of Norbert Wiener. Books General stochastic processes in the theory of queues (Addison-Wesley, 1963) Mathematical Theory of Connecting Networks and Telephone Traffic (Academic Press, 1965) Journal publications Mr. Mayo on “Rules” of language, Philosophical Studies, 3(2):33–36, April 1951 (review). A Partial Model for Quine's "New Foundations", The Journal of Symbolic Logic, Vol. 19, No. 3, pp. 197–200, September 1954 On the Consistency of an Axiom of Enumerability, J. Symb. Log. 20(1):29–30, 1955 On queues with Poisson arrivals, Ann. Math. Statist., vol. 28, pp. 670–677, 1956 On Trunks with Negative Exponential Holding Times Serving a Renewal Process, Bell System Technical Journal, 37,
https://en.wikipedia.org/wiki/Geodesic%20convexity
In mathematics — specifically, in Riemannian geometry — geodesic convexity is a natural generalization of convexity for sets and functions to Riemannian manifolds. It is common to drop the prefix "geodesic" and refer simply to "convexity" of a set or function. Definitions Let (M, g) be a Riemannian manifold. A subset C of M is said to be a geodesically convex set if, given any two points in C, there is a unique minimizing geodesic contained within C that joins those two points. Let C be a geodesically convex subset of M. A function is said to be a (strictly) geodesically convex function if the composition is a (strictly) convex function in the usual sense for every unit speed geodesic arc γ : [0, T] → M contained within C. Properties A geodesically convex (subset of a) Riemannian manifold is also a convex metric space with respect to the geodesic distance. Examples A subset of n-dimensional Euclidean space En with its usual flat metric is geodesically convex if and only if it is convex in the usual sense, and similarly for functions. The "northern hemisphere" of the 2-dimensional sphere S2 with its usual metric is geodesically convex. However, the subset A of S2 consisting of those points with latitude further north than 45° south is not geodesically convex, since the minimizing geodesic (great circle) arc joining two distinct points on the southern boundary of A leaves A (e.g. in the case of two points 180° apart in longitude, the geodesic arc passes over the south pole).
https://en.wikipedia.org/wiki/Liouville%27s%20equation
For Liouville's equation in dynamical systems, see Liouville's theorem (Hamiltonian). For Liouville's equation in quantum mechanics, see Von Neumann equation. For Liouville's equation in Euclidean space, see Liouville–Bratu–Gelfand equation. In differential geometry, Liouville's equation, named after Joseph Liouville, is the nonlinear partial differential equation satisfied by the conformal factor of a metric on a surface of constant Gaussian curvature : where is the flat Laplace operator Liouville's equation appears in the study of isothermal coordinates in differential geometry: the independent variables are the coordinates, while can be described as the conformal factor with respect to the flat metric. Occasionally it is the square that is referred to as the conformal factor, instead of itself. Liouville's equation was also taken as an example by David Hilbert in the formulation of his nineteenth problem. Other common forms of Liouville's equation By using the change of variables , another commonly found form of Liouville's equation is obtained: Other two forms of the equation, commonly found in the literature, are obtained by using the slight variant of the previous change of variables and Wirtinger calculus: Note that it is exactly in the first one of the preceding two forms that Liouville's equation was cited by David Hilbert in the formulation of his nineteenth problem. A formulation using the Laplace–Beltrami operator In a more invariant fashion, the equation can be written in terms of the intrinsic Laplace–Beltrami operator as follows: Properties Relation to Gauss–Codazzi equations Liouville's equation is equivalent to the Gauss–Codazzi equations for minimal immersions into the 3-space, when the metric is written in isothermal coordinates such that the Hopf differential is . General solution of the equation In a simply connected domain , the general solution of Liouville's equation can be found by using Wirtinger calculus. Its for
https://en.wikipedia.org/wiki/Microsoft%20Network%20Monitor
Microsoft Network Monitor is a deprecated packet analyzer. It enables capturing, viewing, and analyzing network data and deciphering network protocols. It can be used to troubleshoot network problems and applications on the network. Microsoft Network Monitor 1.0 (codenamed Bloodhound) was originally designed and developed by Raymond Patch, a transport protocol and network adapter device driver engineer on the Microsoft LAN Manager development team. Network Monitor was replaced by Microsoft Message Analyzer (MMA was discontinued in 2019). History The LAN Manager development team had one shared hardware-based analyzer at the time. Netmon was conceived when the hardware analyzer was taken during a test to reproduce a networking bug, and the first Windows prototype was coded over the Christmas holiday. The first 4 bytes of the Netmon capture file format were used to validate the file. The values were 'RTSS' for Ray, Tom, Steve, and Steve - the first four members of the team. The code was originally written for OS/2 and had no user interface; a symbol was placed in the device driver where the packet buffers were kept so received data could be dumped in hex from within the kernel debugger. Netmon caused a bit of a stir for Microsoft IT since networks and e-mail were not encrypted at the time. Only a few software engineers had access to hardware analyzers due to their cost, but with Netmon many engineers around the company had access to network traffic for free. At the request of Microsoft IT, two simple identification features were added - a non-cryptographic password and an identification protocol named the Bloodhound-Oriented Network Entity (BONE) (created and named by Raymond Patch as a play on the codename Bloodhound). Network Monitor 3 is a complete overhaul of the earlier Network Monitor 2.x version. Originally versions of Network Monitor were only available through other Microsoft products, such as Systems Management Server (SMS). But now the fully featured
https://en.wikipedia.org/wiki/Fawwaz%20T.%20Ulaby
Fawwaz T. Ulaby () is Arthur F. Thurnau Professor of Electrical Engineering and Computer Science at the University of Michigan in Ann Arbor and formerly the Founding Provost and Executive Vice President of the King Abdullah University of Science and Technology (KAUST) and R. Jamieson and Betty Williams Professor of Electrical Engineering and Computer Science at the University of Michigan. Life Ulaby was born in Damascus, Syria, and grew up in Lebanon. He attended the American University of Beirut, from which he received a B.S. degree in physics in 1964. He later received a Ph.D. in Electrical Engineering from the University of Texas at Austin in 1968. After teaching at the University of Kansas he moved to the University of Michigan in Ann Arbor in the mid 1980s. He served as the R. Jamieson and Betty Williams Professor of Electrical Engineering and Computer Science, and has also served as the Vice President for Research. Ulaby has done extensive work outside of academia as well, giving testimony to the House Science Committee of the US congress and serving on the board of directors for The Arab Community Center for Economic and Social Services (ACCESS). In March 2008, Ulaby was named Founding Provost of the King Abdullah University of Science and Technology (KAUST). His daughter, Neda Ulaby, is a reporter at the NPR culture desk. Research areas He is most famous for the development of micro-electronics for a suite of circuits and antennae for THz sensors and communication systems. Today, THz technology is an enabling technology in various types of industrial sensor applications. Honors Professor Ulaby is a member of the U.S. National Academy of Engineering, Fellow of the American Association for the Advancement of Science (AAAS), and Fellow of the Institute of Electrical and Electronics Engineers (IEEE). IEEE GRSS Outstanding Service Award, 1982. IEEE GRSS Distinguished Achievement Award, 1983. IEEE Centennial Medal, 1984. NASA Group Achievement Award
https://en.wikipedia.org/wiki/Norfolk%20Open%20Link
Norfolk Open Link was a free wireless service offered by Norfolk County Council in Norwich city centre, and areas of South Norfolk. Norfolk Open Link was the first community wireless network in the UK to provide free internet access for the public sector, the business community and the general public. The network - that was operational since July 2006 - was part of a pilot scheme that terminated in June 2008. In November 2007, the County Council ran a short survey to assess the impact of the service; after the service ends, a report will be compiled into the impact it has had. The project was not allowed to compete with commercial WiFi services, so the access speed for businesses and the public has been restricted to 256 kilobits per second, with sessions limited to an hour. A welcome screen displayed to users upon connecting to the network allows public sector staff to log in and access the internet at 1mb per second.
https://en.wikipedia.org/wiki/Activating%20protein%202
Activating Protein 2 (AP-2) is a family of closely related transcription factors which plays a critical role in regulating gene expression during early development.
https://en.wikipedia.org/wiki/Tunnel%20Setup%20Protocol
In computer networking, the Tunnel Setup Protocol (TSP) is an experimental networking control protocol used to negotiate IP tunnel setup parameters between a tunnel client host and a tunnel broker server, the tunnel end-points. A major use of TSP is in IPv6 transition mechanisms. Parameter negotiation The TSP protocol performs negotiation of the following parameters: User authentication using the Simple Authentication and Security Layer (SASL) protocol Tunnel encapsulation for a variety of tunneling scenarios: IPv6 over IPv4 tunnels IPv4 over IPv6 tunnels IPv6 over UDP/IPv4 tunnels for built-in traversal of network address translators (NAT) IP address assignment for both tunnel endpoints Domain Name System (DNS) registration of end point addresses and reverse DNS Tunnel keep-alive mechanism as needed IPv6 address prefix assignment for routers Routing protocols TSP Session A TSP session is initiated by the TSP client in the goal of establishing an end-to-end tunnel with the TSP server (tunnel broker). The session consists of a basic exchange of XML-encoded data using TCP or UDP. After the negotiation of tunnel setup parameters, the session is terminated and the client undertakes the task of configuring its local tunnel endpoint. See also Anything In Anything (AYIYA)
https://en.wikipedia.org/wiki/Lambda%20point%20refrigerator
A lambda point refrigerator is a device used to cool liquid helium, typically around a superconducting magnet or for low temperature measurements, from approximately 4.2 K to temperatures near the lambda point of helium (approximately 2.17 K), the temperature at which normal fluid helium (helium I) transitions to the superfluid helium II. Cooling is achieved by pumping the liquid helium in the bath through a cooling coil via a needle valve and vacuum pump. The reduced pressure in the coil causes some of the helium to evaporate, creating a two-phase system within the cooling coil. The heat removed via evaporation lowers the temperature of the cooling coil closer to the lambda point. Since the cooling coil is immersed in the liquid helium bath, liquid surrounding the coil is also cooled. The colder, higher density liquid sinks away from the coil toward the bottom of the bath while the warmer, lower density liquid helium rises to the top. Liquid helium typically has poor thermal conductivity, so convective currents associated with a temperature gradient in the bath provide a constant flow of this colder liquid helium toward the bottom of the bath, allowing temperatures below 4.2 K to be realized in the helium bath, typically close to 2.2 K. Cryogenics
https://en.wikipedia.org/wiki/Internalnet
An internalnet is a computer network composed of devices inside and on the human body. Such a system could be used to link nanochondria, bionic implants, wearable computers, and other devices. See also Nanomedicine Personal area network External links PC Magazine definition Smart computing Bionics Computer networks by scale Implants (medicine)
https://en.wikipedia.org/wiki/Velocity%20receiver
A velocity receiver (velocity sensor) is a sensor that responds to velocity rather than absolute position. For example, dynamic microphones are velocity receivers. Likewise, many electronic keyboards used for music are velocity sensitive, and may be said to possess a velocity receiver in each key. Most of these function by measuring the time difference between switch closures at two different positions along the travel of each key. There are two types of velocity receivers, moving coil and piezoelectric. The former contains a coil supported by springs and a permanently fixed magnet and require no output signal amplifiers. Movement causes the coil to move relative to the magnet, which in turn generates a voltage that is proportional to the velocity of that movement. Piezoelectric sensor velocity receivers are similar to a piezoelectric accelerometer, except that the output of the device is proportional to the velocity of the transducer. Unlike the moving coil variety, piezoelectric sensors will likely require an amplifier due to the small generated signal. Transducers
https://en.wikipedia.org/wiki/Bone%20seeker
A bone seeker is an element, often a radioisotope, that tends to accumulate in the bones of humans and other animals when it is introduced into the body. An example is strontium-90, which behaves chemically like calcium and can replace the calcium in bones. Other bone seekers include radium, and plutonium. An important thing to keep in mind is that - much like for toxic heavy metals the chemical state of the element may complicate such classifications. For example, while strontium, its oxides and hydroxides are all water-soluble and absorbed in the digestive tract or lungs (from whence they enter the bloodstream and ultimately accumulate in the bones), compounds such as strontium titanate are not, which explains their use in radioisotope thermoelectric generators. As such, while dissolved ions of the element or common compounds of the element (including its native form) may present the associated health risks, there are often forms which present much lower risks even as orphan sources or to non-target organs when administered internally in radiotherapy. Bone-seeking elements are health risks but have uses in oncology. See also ATC code V10B Pain palliation (bone seeking agents), a group of pharmaceutical bone seekers
https://en.wikipedia.org/wiki/Saporin
Saporin is a protein that is useful in biological research applications, especially studies of behavior. Saporins are so-called ribosome inactivating proteins (RIPs), due to its N-glycosidase activity, from the seeds of Saponaria officinalis (common name: soapwort). It was first described by Fiorenzo Stirpe and his colleagues in 1983 in an article that illustrated the unusual stability of the protein. Among the RIPs are some of the most toxic molecules known, such as ricin and abrin. Each of these toxins contain a second protein subunit, which inserts the RIP into a cell, enabling it to enzymatically inactivate the ribosomes, shutting down protein synthesis, stopping basic cell functions, resulting in cell death, and eventually causing death of the victim. Saporin has no chain capable of inserting it into the cell. Thus it and the soapwort plant are safe to handle. This has aided its use in research. If given a method of entry into the cell, saporin becomes a very potent toxin, since its enzymatic activity is among the highest of all RIPs. The enzymatic activity of RIPs is unusually specific: a single adenine base is removed from the ribosomal RNA of the large subunit of the ribosome. This is the Achilles’ heel of the ribosome; the removal of this base completely inhibits the ability of that ribosome to participate in protein synthesis. The fungal toxin alpha-sarcin cuts the ribosomal RNA at the adjacent base, also causing protein synthesis inhibition. The conversion of saporin into a toxin has been used to create a series of research molecules. Attachment of saporin to something that enters the cell will convert it into a toxin for that cell. If the agent is specific for a single cell type, by being an antibody specific for some molecule that is only presented on the surface of the target cell type, then a set group of cells can be removed. This has many applications, some more successful than others. Saporin is not the only molecule that is used in this way; t
https://en.wikipedia.org/wiki/Elastic%20instability
Elastic instability is a form of instability occurring in elastic systems, such as buckling of beams and plates subject to large compressive loads. There are a lot of ways to study this kind of instability. One of them is to use the method of incremental deformations based on superposing a small perturbation on an equilibrium solution. Single degree of freedom-systems Consider as a simple example a rigid beam of length L, hinged in one end and free in the other, and having an angular spring attached to the hinged end. The beam is loaded in the free end by a force F acting in the compressive axial direction of the beam, see the figure to the right. Moment equilibrium condition Assuming a clockwise angular deflection , the clockwise moment exerted by the force becomes . The moment equilibrium equation is given by where is the spring constant of the angular spring (Nm/radian). Assuming is small enough, implementing the Taylor expansion of the sine function and keeping the two first terms yields which has three solutions, the trivial , and which is imaginary (i.e. not physical) for and real otherwise. This implies that for small compressive forces, the only equilibrium state is given by , while if the force exceeds the value there is suddenly another mode of deformation possible. Energy method The same result can be obtained by considering energy relations. The energy stored in the angular spring is and the work done by the force is simply the force multiplied by the vertical displacement of the beam end, which is . Thus, The energy equilibrium condition now yields as before (besides from the trivial ). Stability of the solutions Any solution is stable iff a small change in the deformation angle results in a reaction moment trying to restore the original angle of deformation. The net clockwise moment acting on the beam is An infinitesimal clockwise change of the deformation angle results in a moment which can be rewritten as since due t
https://en.wikipedia.org/wiki/Internal%20heating
Internal heat is the heat source from the interior of celestial objects, such as stars, brown dwarfs, planets, moons, dwarf planets, and (in the early history of the Solar System) even asteroids such as Vesta, resulting from contraction caused by gravity (the Kelvin–Helmholtz mechanism), nuclear fusion, tidal heating, core solidification (heat of fusion released as molten core material solidifies), and radioactive decay. The amount of internal heating depends on mass; the more massive the object, the more internal heat it has; also, for a given density, the more massive the object, the greater the ratio of mass to surface area, and thus the greater the retention of internal heat. The internal heating keeps celestial objects warm and active. Small celestial objects In the early history of the Solar System, radioactive isotopes having a half-life on the order of a few million years (such as aluminium-26 and iron-60) were sufficiently abundant to produce enough heat to cause internal melting of some moons and even some asteroids, such as Vesta noted above. After these radioactive isotopes had decayed to insignificant levels, the heat generated by longer-lived radioactive isotopes (such as potassium-40, thorium-232, and uranium-235 and uranium-238) was insufficient to keep these bodies molten unless they had an alternative source of internal heating, such as tidal heating. Thus, Earth's Moon, which has no alternative source of internal heating is now geologically dead, whereas a moon as small as Enceladus that has sufficient tidal heating (or at least had it recently) and some remaining radioactive heating, is able to maintain an active and directly detectable cryovolcanism. Planets Terrestrial planets The internal heating within terrestrial planets powers tectonic and volcanic activities. Of the terrestrial planets in the Solar System, Earth has the most internal heating because it is the largest. Mercury and Mars have no ongoing visible surface effects of inte
https://en.wikipedia.org/wiki/Hitchhiker%27s%20thumb
Hitchhiker's thumb, also known as distal hyperextensibility of the thumb, is the condition of having a thumb that has a distal phalange that bends backwards in an angle of 90°. This condition is benign (when isolated) and does not cause pain or affect the thumb with the trait negatively. If a person has the genes for this condition, it might present bilaterally, unilaterally, or on none of the thumbs. Signs and symptoms Symptoms vary among people with hitchhiker's thumbs, those who have it as an isolated trait are not affected by any symptoms, however, this trait is sometimes a symptom of other conditions, such as a hypermobility spectrum disorder. The signs of this trait are having a thumb with a distal phalange that is able to bend backwards beyond the normal range of motion. Some people with hitchhiker's thumb also have a Z-shaped deformity, named after the appearance of this deformity, which is that of the letter Z. This is a photo of a hitchhiker's thumb. Causes The exact cause of isolated hitchhiker's thumb is not known, it is thought to be inherited in an autosomal recessive manner. However, this trait is often a symptom of an underlying condition, such as Diastrophic dysplasia Hypermobility spectrum disorder Rheumatoid arthritis Treatment It is not associated with any condition, and does not need treatment. Incidence Hitchhiker's thumbs affect 24–36% of the U.S. population. See also Hypermobility (joints) Ligamentous laxity Hypermobility spectrum disorder Ehlers–Danlos syndromes Diastrophic dysplasia
https://en.wikipedia.org/wiki/Fisher%E2%80%93Yates%20shuffle
The Fisher–Yates shuffle is an algorithm for shuffling a finite sequence. The algorithm takes a list of all the elements of the sequence, and continually determines the next element in the shuffled sequence by randomly drawing an element from the list until no elements remain. The algorithm produces an unbiased permutation: every permutation is equally likely. The modern version of the algorithm takes time proportional to the number of items being shuffled and shuffles them in place. The Fisher–Yates shuffle is named after Ronald Fisher and Frank Yates, who first described it, and is also known as the Knuth shuffle after Donald Knuth. A variant of the Fisher–Yates shuffle, known as Sattolo's algorithm, may be used to generate random cyclic permutations of length n instead of random permutations. Fisher and Yates' original method The Fisher–Yates shuffle, in its original form, was described in 1938 by Ronald Fisher and Frank Yates in their book Statistical tables for biological, agricultural and medical research. Their description of the algorithm used pencil and paper; a table of random numbers provided the randomness. The basic method given for generating a random permutation of the numbers 1 through N goes as follows: Write down the numbers from 1 through N. Pick a random number k between one and the number of unstruck numbers remaining (inclusive). Counting from the low end, strike out the kth number not yet struck out, and write it down at the end of a separate list. Repeat from step 2 until all the numbers have been struck out. The sequence of numbers written down in step 3 is now a random permutation of the original numbers. Provided that the random numbers picked in step 2 above are truly random and unbiased, so will be the resulting permutation. Fisher and Yates took care to describe how to obtain such random numbers in any desired range from the supplied tables in a manner which avoids any bias. They also suggested the possibility of using a
https://en.wikipedia.org/wiki/Pwdump
pwdump is the name of various Windows programs that outputs the LM and NTLM password hashes of local user accounts from the Security Account Manager (SAM) database and from the Active Directory domain's users cache on the operating system. It is widely used, to perform both the famous pass-the-hash attack, or also can be used to brute-force users' password directly. In order to work, it must be run under an Administrator account, or be able to access an Administrator account on the computer where the hashes are to be dumped. Pwdump could be said to compromise security because it could allow a malicious administrator to access user's passwords. History The initial program called pwdump was written by Jeremy Allison. He published the source code in 1997 (see open-source). Since then there have been further developments by other programmers: pwdump (1997) — original program by Jeremy Allison. pwdump2 (2000) — by Todd Sabin of Bindview (GPL), uses DLL injection. pwdump3 — by Phil Staubs (GPL), works over the network. pwdump3e — by Phil Staubs (GPL), sends encrypted over network. pwdump4 — by bingle (GPL), improvement on pwdump3 and pwdump2. pwdump5 — by AntonYo! (freeware). pwdump6 (c. 2006) — by fizzgig (GPL), improvement of pwdump3e. No source code. fgdump (2007) — by fizzgig, improvement of pwdump6 w/ addons. No source code. pwdump7 — by Andres Tarasco (freeware), uses own filesystem drivers. No source code. pwdump8 — by Fulvio Zanetti and Andrea Petralia, supports AES128 encrypted hashes (Windows 10 and later). No source code. Notes
https://en.wikipedia.org/wiki/Line%20fitting
Line fitting is the process of constructing a straight line that has the best fit to a series of data points. Several methods exist, considering: Vertical distance: Simple linear regression Resistance to outliers: Robust simple linear regression Perpendicular distance: Orthogonal regression Weighted geometric distance: Deming regression Scale invariance: Major axis regression See also Linear least squares Linear segmented regression Linear trend estimation Polynomial regression Regression dilution Further reading "Fitting lines", chap.1 in LN. Chernov (2010), Circular and linear regression: Fitting circles and lines by least squares, Chapman & Hall/CRC, Monographs on Statistics and Applied Probability, Volume 117 (256 pp.). Regression analysis Geometric algorithms
https://en.wikipedia.org/wiki/Chain%20fern
Chain fern is a common name for several ferns and may refer to: Woodwardia, a fern genus of the Northern Hemisphere Tmesipteris, a fern genus of the South Pacific, also called "hanging fork fern" Ferns
https://en.wikipedia.org/wiki/Zonal%20and%20poloidal
In magnetic confinement fusion the zonal direction primarily connotes the poloidal direction (i.e. the short way around the torus), the corresponding coordinate being denoted by y in the slab approximation or θ in magnetic coordinates. However, in the fusion context, usage is restricted to the context of zonal plasma flows and there will in general be a toroidal component in such flows as well. Thus, although the term zonal has come into use in plasma physics to emphasize an analogy with zonal flows in geophysics, it does not uniquely identify the direction of flow, unlike the case in geophysics. See also Toroidal and poloidal Zonal and meridional Zonal flow (plasma) Zonal flow Orientation (geometry) Magnetic confinement fusion
https://en.wikipedia.org/wiki/Subnet%20%28mathematics%29
In topology and related areas of mathematics, a subnet is a generalization of the concept of subsequence to the case of nets. The analogue of "subsequence" for nets is the notion of a "subnet". The definition is not completely straightforward, but is designed to allow as many theorems about subsequences to generalize to nets as possible. There are three non-equivalent definitions of "subnet". The first definition of a subnet was introduced by John L. Kelley in 1955 and later, Stephen Willard introduced his own (non-equivalent) variant of Kelley's definition in 1970. Subnets in the sense of Willard and subnets in the sense of Kelley are the most commonly used definitions of "subnet" but they are each equivalent to the concept of "subordinate filter", which is the analog of "subsequence" for filters (they are not equivalent in the sense that there exist subordinate filters on whose filter/subordinate–filter relationship cannot be described in terms of the corresponding net/subnet relationship). A third definition of "subnet" (not equivalent to those given by Kelley or Willard) that equivalent to the concept of "subordinate filter" was introduced independently by Smiley (1957), Aarnes and Andenaes (1972), Murdeshwar (1983), and possibly others, although it is not often used. This article discusses the definition due to Willard (the other definitions are described in the article Filters in topology#Subnets). Definitions There are several different non-equivalent definitions of "subnet" and this article will use the definition introduced in 1970 by Stephen Willard, which is as follows: If and are nets in a set from directed sets and respectively, then is said to be a of ( or a ) if there exists a monotone final function such that A function is , , and an if whenever then and it is called if its image is cofinal in The set being in means that for every there exists some such that that is, for every there exists an such that Sin
https://en.wikipedia.org/wiki/Esther%20M.%20Conwell
Esther Marley Conwell (May 23, 1922 – November 16, 2014) was a pioneering American chemist and physicist, best known for the Conwell-Weisskopf theory that describes how electrons travel through semiconductors, a breakthrough that helped revolutionize modern computing. During her life, she was described as one of the most important women in science. Conwell studied properties of semiconductors and organic conductors, especially electron transport. In 1990, she became an adjunct professor at the University of Rochester while still working at Xerox. In 1998, she joined the University of Rochester faculty full-time as a professor of chemistry, focused on the flow of electrons through DNA. Conwell held four patents and published more than 270 papers and multiple textbooks over the course of her career. Her textbook, High Field Transport in Semiconductors, became the authoritative text in the field. She received numerous honors, including the National Medal of Science in 2009. Education Conwell obtained a physics B.A. from Brooklyn College in 1942. She then went to the University of Rochester to complete a M.S. in physics in 1945 with Victor Weisskopf. She initially planned to do a Ph.D. at Rochester, but since her adviser left to work at Los Alamos after her first year there, she completed her masters and obtained a Ph.D. at a later point in time. Conwell collaborated with Karl Lark-Horovitz and Vivian Johnson at Purdue University on silicon and germanium semiconductor physics. Her masters was initially classified then finally declassified in 1945 and subsequently her M.S. was awarded in which she determined the Conwell-Weisskopf theory. Conwell received her physics Ph.D. in 1948, from the University of Chicago under the advisement of Nobel Laureate Subrahmanyan Chandrasekhar at Yerkes Observatory and was also an assistant to Enrico Fermi. She was a teaching assistant at Chicago and graded the work of Nobel Laureates such as Chen-Ning Yang and Owen Chamberlain. Care
https://en.wikipedia.org/wiki/Zonal%20flow%20%28plasma%29
In toroidally confined fusion plasma experiments the term zonal flow means a plasma flow within a magnetic surface primarily in the poloidal direction. This usage is inspired by the analogy between the quasi-two-dimensional nature of large-scale atmospheric and oceanic flows, where zonal means latitudinal, and the similarly quasi-two-dimensional nature of low-frequency flows in a strongly magnetized plasma. Zonal flows in the toroidal plasma context are further characterized by being localized in their radial extent transverse to the magnetic surfaces (in contrast to global plasma rotation), having little or no variation in either the poloidal or toroidal direction—they are m = n = 0 modes (where m and n are the poloidal and toroidal mode numbers, respectively), having zero real frequency when analyzed by linearization around an unperturbed toroidal equilibrium state (in contrast to the geodesic acoustic mode branch, which has finite frequency). Arising via a self-organization phenomenon driven by low-frequency drift-type modes, in which energy is transferred to longer wavelengths by modulational instability or turbulent inverse cascade. See also Zonal and poloidal Zonal flow List of plasma (physics) articles
https://en.wikipedia.org/wiki/Dark%20%28broadcasting%29
In broadcasting, a dark television station or silent radio station is one that has gone off the air for an indefinite period of time. Usually unlike dead air (broadcasting only silence), a station that is dark or silent does not even transmit a carrier signal. U.S. law Transmitter operations According to the U.S. Federal Communications Commission (FCC), a radio or television station is considered to have gone dark or silent if it is to be off the air for thirty days or longer. Prior to the Telecommunications Act of 1996, a "dark" station was required to surrender its broadcast license to the FCC, leaving it vulnerable to another party applying for it while its current owner was making efforts to get it back on the air. Following the 1996 landmark legislation, a licensee is no longer required to surrender the license while dark. Instead, the licensee may apply for a "Notification of Suspension of Operations/Request for Silent STA" (FCC Form 0386), stating the reason why the station has gone silent. A service can go dark for any number of reasons, including financial resources being too drained to continue effective operation of the service as being of benefit to its community of license; abandonment for a different channel or to go cable-only; complicated technical adjustments involving radio antenna repair, requiring the broadcast tower to be de-energized for the work to be done; structure fire or natural disaster that has rendered the facility inoperable; if unowned by the station, the loss of a leasehold on either the tower or the land for the transmitter, usually by sale to another party; or technical adjustments that would make it prohibitively expensive to perform the work and carry on the normal operations of the station in question. The service is not required to notify the FCC of silence if the period of silence is less than ten days. If the period of silence is to last at least ten days but less than thirty days, the licensee must notify the FCC in
https://en.wikipedia.org/wiki/Bacillus%20odysseyi
Bacillus odysseyi is a Gram-positive, aerobic, rod-shaped, round-spore- and endospore-forming eubacterium of the genus Bacillus. This novel species was discovered by scientist Myron T. La Duc of NASA’s Biotechnology and Planetary Protection Group, a unit whose purpose is to clean and sterilize spacecraft so as not to have microorganisms contaminate other celestial bodies or foreign microorganisms contaminate Earth, on the surface of the Mars Odyssey in a clean room at the Jet Propulsion Laboratory in La Cañada Flintridge before the spacecraft was launched to space. La Duc named the bacterium Bacillus odysseyi sp. nov. after the Odyssey mission. It had apparently evolved to live in the sparse environment of a clean room, and its secondary spore coat makes it especially resistant to radiation. B. odysseyi consists of an exosporium, spore coat, cortex, and core. In a test performed by the Planetary Protection unit, its spores were the most consistently resistant, and it survived exposure to all of the challenges posed against it: desiccation (100% survival), Hydrogen peroxide (H2O2, 26% survival), ultraviolet radiation (10% survival at 660 J ∙ m−2), and gamma radiation (0.4% survival). B. odysseyi shares many DNA similarities with Bacillus fusiformis and Solibacillus silvestris. The type strain for B. odysseyi is 34hs-1T (=ATCC PTA-4993T=NRRL B-30641T=NBRC 100172T).
https://en.wikipedia.org/wiki/KASCADE
KASCADE was a European physics experiment started in 1996 at Forschungszentrum Karlsruhe, Germany (now Karlsruher Institut für Technologie), an extensive air shower experiment array to study the cosmic ray primary composition and the hadronic interactions, measuring simultaneously the electronic, muonic and hadronic components. KASCADE-Grande was a further extension of the previous project by reassembling 37 detectors of the former EAS-TOP experiment running between 1987 and 2000 at Campo Imperatore, Gran Sasso Laboratories, Italy. By this Grande extension of KASCADE the energy range was extended to 1014–1018 eV. The experiment contributed significantly to the development of the CORSIKA simulation program which is use heavily in astroparticle physics. Co-located with KASCADE-Grande is the LOPES experiment. LOPES consists of radio antennas and measures the radio emission of extensive air showers. KASCADE (including all extensions) stopped operation in 2013, but a part of the detectors is still used in other experiments for cosmic-ray air showers, e.g., LOFAR or Tunka. The data acquired by KASCADE-Grande has meanwhile been made accessible to the public in the KASCADE Cosmic-Ray Data Center (KCDC). Results KASCADE studied heavier components of cosmic rays, finding a "knee" near 80 PeV in 2011, and extending the spectrum measurements to 200PeV. Later, a knee-like feature in the heavy component and an ankle-like feature in the light component of cosmic rays was discovered at an energy of about 1017 eV. Participants Institut für Kernphysik and Institut für Experimentelle Kernphysik of Karlsruher Institut für Technologie (KIT), Germany Dipartimento di Fisica Generale dell' Università and Istituto di Fisica dello Spazio Interplanetario of Istituto Nazionale di Astrofisica Torino, Italy Universität Siegen, Germany Universität Wuppertal, Germany Soltan Institute for Nuclear Studies, Łódź, Poland Institute of Physics and Nuclear Engineering, Bucharest, Romania
https://en.wikipedia.org/wiki/Fatty%20acid-binding%20protein
The fatty-acid-binding proteins (FABPs) are a family of transport proteins for fatty acids and other lipophilic substances such as eicosanoids and retinoids. These proteins are thought to facilitate the transfer of fatty acids between extra- and intracellular membranes. Some family members are also believed to transport lipophilic molecules from outer cell membrane to certain intracellular receptors such as PPAR. The FABPs are intracellular carriers that “solubilize” the endocannabinoid anandamide (AEA), transporting AEA to the breakdown by FAAH, and compounds that bind to FABPs block AEA breakdown, raising its level. The cannabinoids (THC and CBD) are also discovered to bind human FABPs (1, 3, 5, and 7) that function as intracellular carriers, as THC and CBD inhibit the cellular uptake and catabolism of AEA by targeting FABPs. Competition for FABPs may in part or wholly explain the increased circulating levels of endocannabinoids reported after consumption of cannabinoids. Levels of fatty-acid-binding protein have been shown to decline with ageing in the mouse brain, possibly contributing to age-associated decline in synaptic activity. Family members Members of this family include: Pseudogenes
https://en.wikipedia.org/wiki/CORSIKA
CORSIKA (COsmic Ray SImulations for KAscade) is a physics computer software for simulation of extensive air showers induced by high energy cosmic rays, i.e. protons and atomic nuclei, as well as Gamma rays (photons), electrons, and neutrinos. It may be used up to and beyond the highest energies of 100 EeV. In the current version the program utilizes the hadronic interaction models EPOS, QGSJET, and DPMJET, which are based on Gribov-Regge theory, and SIBYLL based on a minijet model for high energies. Hadronic interactions at lower energies are described either by the GHEISHA module, by FLUKA, or by the UrQMD model. Electromagnetic interactions are treated by an adapted version of the EGS4 code, customized by including the Landau–Pomeranchuk–Migdal effect relevant at higher energies. It can be used to simulate the generation of Cherenkov radiation, radio emission (Askaryan radiation), and atmospheric neutrinos. A complete rewrite of CORSIKA in C++ named CORSIKA 8 is currently work in progress.
https://en.wikipedia.org/wiki/Analytic%20function%20of%20a%20matrix
In mathematics, every analytic function can be used for defining a matrix function that maps square matrices with complex entries to square matrices of the same size. This is used for defining the exponential of a matrix, which is involved in the closed-form solution of systems of linear differential equations. Extending scalar function to matrix functions There are several techniques for lifting a real function to a square matrix function such that interesting properties are maintained. All of the following techniques yield the same matrix function, but the domains on which the function is defined may differ. Power series If the analytic function has the Taylor expansion then a matrix function can be defined by substituting by a square matrix: powers become matrix powers, additions become matrix sums and multiplications by coefficients become scalar multiplications. If the series converges for , then the corresponding matrix series converges for matrices such that for some matrix norm that satisfies . Diagonalizable matrices A square matrix is diagonalizable, if there is an invertible matrix such that is a diagonal matrix, that is, has the shape As it is natural to set It can be verified that the matrix does not depend on a particular choice of . For example, suppose one is seeking for One has for Application of the formula then simply yields Likewise, Jordan decomposition All complex matrices, whether they are diagonalizable or not, have a Jordan normal form , where the matrix J consists of Jordan blocks. Consider these blocks separately and apply the power series to a Jordan block: This definition can be used to extend the domain of the matrix function beyond the set of matrices with spectral radius smaller than the radius of convergence of the power series. Note that there is also a connection to divided differences. A related notion is the Jordan–Chevalley decomposition which expresses a matrix as a sum of a diagonalizable and a
https://en.wikipedia.org/wiki/Radeon%20HD%205000%20series
The Evergreen series is a family of GPUs developed by Advanced Micro Devices for its Radeon line under the ATI brand name. It was employed in Radeon HD 5000 graphics card series and competed directly with Nvidia's GeForce 400 Series. Release The existence was spotted on a presentation slide from AMD Technology Analyst Day July 2007 as "R8xx". AMD held a press event in the USS Hornet Museum on September 10, 2009 and announced ATI Eyefinity multi-display technology and specifications of the Radeon HD 5800 series' variants. The first variants of the Radeon HD 5800 series were launched September 23, 2009, with the HD 5700 series launching October 12 and HD 5970 launching on November 18 The HD 5670, was launched on January 14, 2010, and the HD 5500 and 5400 series were launched in February 2010, completing what has appeared to be most of AMD's Evergreen GPU lineup. Demand so greatly outweighed supply that more than two months after launch, many online retailers were still having trouble keeping the 5800 and 5900 series in stock. Architecture This article is about all products under the Radeon HD 5000 Series brand. TeraScale 2 was introduced with this. Radeon HD 5830 and above products,have the capability to calculate double-precision floating-point format. Radeon HD 5770 and below products,have the capability to calculate only single-precision floating-point format. OpenGL 4.x compliance requires supporting FP64 shaders. These are implemented by emulation on some TeraScale (microarchitecture) GPUs. Multi-monitor support The on-die display controllers with the new brand name ATI Eyefinity were introduced with the Radeon HD 5000 Series. The entire HD 5000 series products have Eyefinity capabilities supporting three outputs. The Radeon HD 5870 Eyefinity Edition, however, supports six mini DisplayPort outputs, all of which can be simultaneously active. Display pipeline supports xvYCC gamut and 12-bit per component output via HDMI. HDMI 1.3a output. The previous gener
https://en.wikipedia.org/wiki/Jet%20quenching
In high-energy physics, jet quenching is a phenomenon that can occur in the collision of ultra-high-energy particles. In general, the collision of high-energy particles can produce jets of elementary particles that emerge from these collisions. Collisions of ultra-relativistic heavy-ion particle beams create a hot and dense medium comparable to the conditions in the early universe, and then these jets interact strongly with the medium, leading to a marked reduction of their energy. This energy reduction is called "jet quenching". Physics background In the context of high-energy hadron collisions, quarks and gluons are collectively called partons. The jets emerging from the collisions originally consist of partons, which quickly combine to form hadrons, a process called hadronization. Only the resulting hadrons can be directly observed. The hot, dense medium produced in the collisions is also composed of partons; it is known as a quark–gluon plasma (QGP). In this realm, the laws of physics that apply are those of quantum chromodynamics (QCD). High-energy nucleus-nucleus collisions make it possible to study the properties of the QGP medium through the observed changes in the jet fragmentation functions as compared to the unquenched case. According to QCD, high-momentum partons produced in the initial stage of a nucleus-nucleus collision will undergo multiple interactions inside the collision region prior to hadronization. In these interactions, the energy of the partons is reduced through collisional energy loss and medium-induced gluon radiation, the latter being the dominant mechanism in a QGP. The effect of jet quenching in QGP is the main motivation for studying jets as well as high-momentum particle spectra and particle correlations in heavy-ion collisions. Accurate jet reconstruction will allow measurements of the jet fragmentation functions and consequently the degree of quenching and therefore provide insight on the properties of the hot dense QGP medium cre
https://en.wikipedia.org/wiki/Crown%20%28botany%29
The crown of a plant refers to the total of an individual plant's aboveground parts, including stems, leaves, and reproductive structures. A plant community canopy consists of one or more plant crowns growing in a given area. The crown of a woody plant (tree, shrub, liana) is the branches, leaves, and reproductive structures extending from the trunk or main stems. Shapes of crowns are highly variable. The major types for trees are the excurrent branching habit resulting in conoid shapes and decurrent (deliquescent) branching habit, resulting in round shapes. Crowns are also characterized by their width, depth, surface area, volume, and density. Measurements of crowns are important in quantifying and qualifying plant health, growth stage, and efficiency. Major functions of the crown include light energy assimilation, carbon dioxide absorption and release of oxygen via photosynthesis, energy release by respiration, and movement of water to the atmosphere by transpiration. These functions are performed by the leaves. Crown classes Trees can be described as fitting different crown classes. Commonly used are Kraft's classes. Kraft designated these social classes based on temperate and boreal forests in Central Europe, so they do not necessarily work with every forest type in the world. Kraft wrote in German so here are his classes with translations: 1 v vorherrschend (predominant) 2 h herrschend (dominant) 3 m mitherrschend (co-dominant) 4 b beherrscht (dominated / suppressed) 5 u unterständig (inferior) this is then split into 2 subclasses 5a (shade tolerant trees) and 5b (dying crowns / dying trees) Often it has been simplified to Dominant, Co-dominant and Suppressed. Also IUFRO developed a tree classification it is based on three components with numbers that then aggregate to give a coded classification thus: Ecological criteria Height component (Stand layer / Height class): 100 Overstorey / Overlayer 200 Middlestorey / Middlelayer 300 Understorey
https://en.wikipedia.org/wiki/Pluteus%20leoninus
Pluteus leoninus, commonly known as lion shield, can occasionally be found growing on dead wood in Europe and North Africa. The underside of the cap is typical of the genus Pluteus — the gills are pale, soon becoming pink when the spores ripen. But the upper surface is a bright tawny or olivaceous yellow. The species name leoninus (meaning leonine) refers to this cap colour. Description This description is combined from several references. The golden to olive-yellow convex cap is 3–7 cm in diameter, is hygrophanous, and usually has a grooved edge. The darker central disc has a slight velvety tomentum. The gills are yellowish at first, then salmon pink (the colour of the spore powder). The stipe is up to about 7 cm, often striate, being white to cream, and often darker near the base. The mushroom grows on stumps and wood debris of broad-leaved trees and sometimes of conifers. At the microscopic level, the filamentous cap cuticle is a trichoderm. The gills have scanty bladder-shaped pleurocystidia, and abundant fusiform cheilocystidia. The spores are smooth, almost globular, approximately 7×6 μm. Many authorities consider Pluteus fayodii to be a synonym of P. leoninus, but according to Species Fungorum, they are distinct. Edibility According to some sources, it is edible but has little to no taste. See also List of Pluteus species
https://en.wikipedia.org/wiki/National%20Institute%20for%20Astrophysics
The National Institute for Astrophysics (, or INAF) is an Italian research institute in astronomy and astrophysics, founded in 1999. INAF funds and operates twenty separate research facilities, which in turn employ scientists, engineers and technical staff. The research they perform covers most areas of astronomy, ranging from planetary science to cosmology. Research facilities INAF coordinates the activities of twenty research units, nineteen in Italy and one in Spain: Bologna Observatory Istituto di Astrofisica Spaziale e Fisica cosmica di Bologna Istituto di Radioastronomia di Bologna Cagliari Observatory Catania Observatory Arcetri Observatory (Florence) Brera Observatory (Milan) Istituto di Astrofisica Spaziale e Fisica cosmica di Milano Capodimonte Observatory (Naples) Osservatorio Astronomico di Padova Palermo Observatory Istituto di Astrofisica Spaziale e Fisica cosmica di Palermo Rome Observatory Istituto di Astrofisica Spaziale e Fisica cosmica di Roma Istituto di Fisica dello Spazio Interplanetario di Roma Collurania-Teramo Observatory Turin Observatory Istituto di Fisica dello Spazio Interplanetario di Torino Trieste Observatory Telescopio Nazionale Galileo (Canary Islands, Spain) Sardinia Radio Telescope (San Basilio, Sardinia) Noto Radio Observatory (Noto, Sicily) International partnerships INAF is involved in scientific collaborations with several international institutions, including: the European Southern Observatory (Italy has been an ESO member since 1982) the astronomical observatories located in Canary Islands (Teide Observatory and Roque de los Muchachos Observatory) the Large Binocular Telescope, in partnership with the United States and Germany the Very Long Baseline Interferometry consortium the European Space Agency (ESA) the American National Aeronautic and Space Administration (NASA) Notable figures Giampaolo Vettolani, scientific director Stefano Cristiani, board member Nicolò D'Amico, president in 2015-2020 Projects Projects i
https://en.wikipedia.org/wiki/Victoria%20%283D%20figure%29
Victoria is an articulated 3D female figure developed and sold by Daz 3D. There have been several "generations" of the figure, all bearing the same name. The figure was originally created as one of two standard characters which also included the male character "Michael" ("Stephanie" being a full body female morph of Michael). Besides being poseable, using preset morph dials the figure can be manipulated into a variety of different aesthetic and ethnic female forms. The figure comes pre-rigged with standardized features, and earlier versions can be exported to a variety of 3D modeling/animation applications. History Victoria 1 was released by then-Zygote in February 1999 and was the first named medium-resolution figure for use with Poser. She was originally released as 'The Millennium Woman', but the resulting colloquial shortening to "Millie" led to Zygote renaming her as Victoria. Zygote also sold an anime-morphed character for Victoria named Aiko, and each successive generation of Victoria has so far had an accompanying Aiko (whose facial morphs can vary between anime and a less exaggerated Asian). Victoria 2 was the next step in the figure development, refining the rigging and coming with as a reduced polygon version. Released in 2001, V2 contained more body shaping features and a wider range of facial expressions but was rigged to be compatible with V1's clothing and accessories. From this point forward, the Stephanie character (following the Generation numbering) was no longer a remorphed Michael but a character based on (and requiring) Victoria. Victoria 2's eyebrows were separately transmapped objects. Victoria 3, released in 2002 yet again raised the poly count and resculpted the face, leaving out transmapped eyebrows in favor of painting them directly onto the face texture, but was no longer backwards compatible to Victoria 2's rigging, textures or clothing. This generation was the first to have both the male and female figures be modeled from the sa
https://en.wikipedia.org/wiki/Interior%20radiation%20control%20coating
Interior Radiation Control Coating Systems (IRCCS), sometimes referred to as radiant barrier coatings, are paints designed to provide thermal insulation to buildings. Standards The American Society for Testing and Materials (ASTM) and the Reflective Insulation Manufacturer's Association (RIMA) have established an industry standard for evaluating paints claiming to have insulating characteristics. The energy conserving property has been defined as thermal emittance (the ability of a surface to release radiant energy that it has absorbed). Those coatings qualified as Interior Radiation Control Coatings must show a thermal emittance of 0.25 or less. This means that an IRCCS will block 75% or more of the radiant heat transfer. These low "E" coatings were originally developed in 1978 at the Solar Energy Corporation (SOLEC) in Princeton, New Jersey for use in tubular evacuated solar collectors. The developer, Robert Aresty, designed them to be used as low emissivity surfaces on glass to replace vacuum deposited surfaces. While SOLEC was doing collaborative work with the Florida Solar Energy Center (FSEC), Phillip Fairey, research director at FSEC and world-renowned researcher in radiant barriers discovered the availability of these coatings in the SOLEC labs. He immediately grasped that they might be used as a replacement for foil radiant barriers, and proceeded to perform experiments verifying their viability for this use. In 1986 these coatings were applied for the first commercial application in homes built by Centex Corporation. Uses Uses of IRCCS includes residential and commercial building insulation, as well as industrial and automotive applications.
https://en.wikipedia.org/wiki/Hop-by-hop%20transport
Hop-by-hop transport is a principle of controlling the flow of data in a network. With hop-by-hop transport, chunks of data are forwarded from node to node in a store-and-forward manner. As hop-by-hop transport involves not only the source and destination node, but rather some or all of the intermediate nodes as well, it allows data to be forwarded even if the path between source and destination is not permanently connected during communication. However, the End-to-end principle claims that transport control should be implemented end-to-end unless implementing hop-by-hop transport achieves considerably better performance. Moreover, hop-by-hop transport requires per-flow state information at intermediate nodes, which limits its scalability. This is one of the reasons why almost all communication today is controlled by end-to-end transport protocols such as TCP. Current research in the area of sparse mobile networks is considering hop-by-hop transport for application scenarios where end-to-end connectivity is only available intermittently, as under such conditions, hop-by-hop transport can achieve substantial performance gains. See also Disruption Tolerant Networking End-to-end principle External links On Leveraging Partial Paths in Connected Networks End-to-end vs. hop-by-hop transport under intermittent connectivity The Transport Layer Revisited CAT: Contention Aware Transport Protocol for IEEE 802.11 MANETs Transparent IP Proxy for Tactical Ad hoc Networks Network protocols Network architecture
https://en.wikipedia.org/wiki/Uncoupling%20protein
An uncoupling protein (UCP) is a mitochondrial inner membrane protein that is a regulated proton channel or transporter. An uncoupling protein is thus capable of dissipating the proton gradient generated by NADH-powered pumping of protons from the mitochondrial matrix to the mitochondrial intermembrane space. The energy lost in dissipating the proton gradient via UCPs is not used to do biochemical work. Instead, heat is generated. This is what links UCP to thermogenesis. However, not every type of UCPs are related to thermogenesis. Although UCP2 and UCP3 are closely related to UCP1, UCP2 and UCP3 do not affect thermoregulatory abilities of vertebrates. UCPs are positioned in the same membrane as the ATP synthase, which is also a proton channel. The two proteins thus work in parallel with one generating heat and the other generating ATP from ADP and inorganic phosphate, the last step in oxidative phosphorylation. Mitochondria respiration is coupled to ATP synthesis (ADP phosphorylation), but is regulated by UCPs. UCPs belong to the mitochondrial carrier (SLC25) family. Uncoupling proteins play a role in normal physiology, as in cold exposure or hibernation, because the energy is used to generate heat (see thermogenesis) instead of producing ATP. Some plants species use the heat generated by uncoupling proteins for special purposes. Eastern skunk cabbage, for example, keeps the temperature of its spikes as much as 20 °C higher than the environment, spreading odor and attracting insects that fertilize the flowers. However, other substances, such as 2,4-dinitrophenol and carbonyl cyanide m-chlorophenyl hydrazone, also serve the same uncoupling function. Salicylic acid is also an uncoupling agent (chiefly in plants) and will decrease production of ATP and increase body temperature if taken in extreme excess. Uncoupling proteins are increased by thyroid hormone, norepinephrine, epinephrine, and leptin. History Scientists observed the thermogenic activity in brown adip
https://en.wikipedia.org/wiki/Ingolfiella%20longipes
Ingolfiella longipes is a species of amphipod crustacean in the family Ingolfiellidae. It is known from a single specimen held at the Naturalis Biodiversity Center. It was collected from Walsingham Sink Cave, an anchialine cave in Hamilton Parish, Bermuda, and is thus considered to be critically endangered.
https://en.wikipedia.org/wiki/Turndown%20ratio
Turndown ratio refers to the width of the operational range of a device, and is defined as the ratio of the maximum capacity to minimum capacity. For example, a device with a maximum output of 10 units and a minimum output of 2 units has a turndown ratio of 5. The term is commonly used with measurement devices and combustion plant like boilers and gasifiers. Flow meters In flow measurement, the turndown ratio indicates the range of flow that a flow meter is able to measure with acceptable accuracy. It is also known as rangeability. It is important when choosing a flow meter technology for a specific application. If a gas flow to be measured is expected to vary between 100,000 m³ per day and 1,000,000 m³ per day, the specific application has a turndown ratio of at least 10:1. Therefore the meter requires a turndown ratio of at least 10:1. For example: if the meter had an advertised maximum flow of 2,000,000 m³ per day then the required turndown ratio would be 20:1. The turndown ratio of each type of meter is limited by theoretical considerations and by practical considerations. For example, orifice meters create a pressure drop in the measured fluid proportional to the square of the velocity. Therefore the range of differential pressure can become too large and compromise accuracy. It can also create process problems such as hydrate formation, and in the case of measuring the discharge of a compressor, there is a limit to how much pressure loss is acceptable. Typical turndown ratio of various meter types The examples are here for gas flow, but the same meter types can be used on liquids as well, with similar turndown ratios. Note that meter manufacturers state their products' turndown ratios—a specific product may have a turndown ratio that varies from the list below. A thermal mass flow meter has a turndown ratio of 1000:1. An orifice plate meter has a practical turndown ratio of 3:1. A turbine meter has a turndown ratio of 10:1. Rotary positive dis
https://en.wikipedia.org/wiki/Gas%20composition
The Gas composition of any gas can be characterised by listing the pure substances it contains, and stating for each substance its proportion of the gas mixture's molecule count.Nitrogen 78.084 Oxygen 20.9476 Argon Ar 0.934 Carbon Dioxide 0.0314 Gas composition of air To give a familiar example, air has a composition of: Standard Dry Air is the agreed-upon gas composition for air from which all water vapour has been removed. There are various standards bodies which publish documents that define a dry air gas composition. Each standard provides a list of constituent concentrations, a gas density at standard conditions and a molar mass. It is extremely unlikely that the actual composition of any specific sample of air will completely agree with any definition for standard dry air. While the various definitions for standard dry air all attempt to provide realistic information about the constituents of air, the definitions are important in and of themselves because they establish a standard which can be cited in legal contracts and publications documenting measurement calculation methodologies or equations of state. The standards below are two examples of commonly used and cited publications that provide a composition for standard dry air: ISO TR 29922-2017 provides a definition for standard dry air which specifies an air molar mass of 28,965 46 ± 0,000 17 kg·kmol-1. GPA 2145:2009 is published by the Gas Processors Association. It provides a molar mass for air of 28.9625 g/mol, and provides a composition for standard dry air as a footnote.
https://en.wikipedia.org/wiki/Sarnoff%20Fellowship
The Sarnoff Cardiovascular Research Fellowship is a program for US medical students that offers one year of monetary support and a lifetime of academic support for those medical students interested in cardiovascular research. Stanley Sarnoff established the fellowship for medical students in 1979 and it would serve as a prototype for later programs such as the American Heart Association Student Research Scholarship program and the Howard Hughes Medical Institute NIH Research Scholars program, or cloisters program, which would not begin until 1984. Stanley Sarnoff Dr. Sarnoff, a graduate of Princeton University and The Johns Hopkins University School of Medicine, was a scientist, entrepreneur and scholar who early in his career served as Assistant Professor of Physiology at the Harvard University School of Public Health. For a decade he was Chief of the Laboratory of Cardiovascular Physiology at the National Heart Institute. He later founded Rodana Research Corporation and its successor, Survival Technology, Inc., a company today known as Meridian Medical Technologies, which manufactures medical devices. At Dr. Sarnoff's death in 1990, his will named The Sarnoff Foundation the principal beneficiary of his Estate, valued shortly after his death at nearly $30 million. Goals of the fellowship Sarnoff based the program on his experience with Myron Weisfeldt, a young medical student in his lab, from which he established a set of simple rules for the fellowship: the medical student would leave school for one year, work in the laboratory of a prominent cardiovascular scientist conducting his own research and participate as if he were an independent researcher. At the conclusion of his year, the student would report his findings to his mentors. As the fellowship grew, students would attend a yearly meeting and present their findings to their peers and to the Sarnoff Cardiovascular research community of mentors, advisors and past fellows. A major emphasis of these meeting
https://en.wikipedia.org/wiki/Kho%20%28cooking%20technique%29
Kho (, meaning "to braise", "to stew", or "to simmer") or kha () is a cooking technique in Vietnamese and Cambodian cuisine, where a protein source such as fish, shrimp, poultry, pork, beef, or fried tofu is simmered on low or medium heat in a mixture of sugar, and water or a water substitute such as young coconut juice and seasoned with fish sauce or soy sauce and aromatics such as pepper, garlic, shallots and ginger. The resulting dish is salty and savory, and meant to be eaten with rice noodles, baguette, or steamed rice. Particular dishes In Vietnamese, beef stew is called bò kho or thịt bò kho and fish stew is called cá kho or cá kho tộ (tộ referring to the clay pot in which the dish is cooked). For fish stew, catfish is preferred, particularly in southern Vietnam. Chicken stew, called gà kho or gà kho gừng (gừng meaning "ginger"), is less popular. Vegetarian stew may also be prepared. See also Brining Jorim List of stews
https://en.wikipedia.org/wiki/Call%20signs%20in%20New%20Zealand
Call signs in New Zealand are no longer generally used to identify broadcast stations. However, New Zealand's radio stations were once known by their call signs and would usually broadcast their call signs as a number followed by X, Y, or Z, and another letter (e.g. 1YA). Call signs are regulated internationally by the ITU and nationally by the Ministry of Business, Innovation and Employment (MBIE), formerly the Ministry of Economic Development. The ministry is also responsible for providing policy advice to Government on the allocation of New Zealand's radio spectrum to support, efficient, reliable and responsive wireless telecommunications and broadcasting infrastructure. In 1924, New Zealand was granted the prefix 'Z', and in 1925 the number of licensed amateur(?) reached 100. In 1927, the International Telecommunication Union Conference in Washington (D.C., USA) established internally agreed upon call sign prefixes – New Zealand was assigned 'OZ'. In 1929 this was expanded to the ZK–ZM letter block, with New Zealand opting for the ZL prefix for land based stations. 'OZ' by 1927 was reassigned to Denmark. In 1969 the ZM prefix was allowed to celebrate the Captain James Cook bicentenary. In 1974 the prefix was allowed again to celebrate the Commonwealth Games, as well as in 1989 when the Games returned. In 1981 the ZL0 prefix was allowed for visitors to New Zealand. Allocations and assignments The International Telecommunication Union has assigned New Zealand the following call sign blocks for all radio communication, broadcasting or transmission: While not directly related to call signs, the International Telecommunication Union (ITU) further has divided all countries assigned amateur radio prefixes into three regions; New Zealand is located in ITU Region 3, within ITU Zone 60. There are 4 possible 2-letter prefixes and 40 2-letter/1-number prefixes available to New Zealand operators based on the ITU blocks (ZK, ZL, ZM and E5). This provides for about 7
https://en.wikipedia.org/wiki/MOSH%20%28Nokia%29
MOSH was a user defined distribution channel for mobile content initiated by Nokia. The name "MOSH" comes from "Mobilize and Share". The channel could have been used to both download and upload various content for mobile phones or other platforms. File types that were handled were: audio, images, applications, games, videos, documents. A person could create an account on MOSH that comes with the following stats: creator (for content created by that person), collector (for the number of objects added to the collection in the account) and sharer (for uploading files that are not necessarily the work of that person). MOSH was known for having been filled with explicit and copyright-infringing content as it had no screening process. MOSH was launched by Nokia in August 2007 and remained in Beta until its closure in 2009, by which time it had 137 million downloads. It was replaced by Ovi.
https://en.wikipedia.org/wiki/EtherNet/IP
EtherNet/IP (IP = Industrial Protocol) is an industrial network protocol that adapts the Common Industrial Protocol (CIP) to standard Ethernet. EtherNet/IP is one of the leading industrial protocols in the United States and is widely used in a range of industries including factory, hybrid and process. The EtherNet/IP and CIP technologies are managed by ODVA, Inc., a global trade and standards development organization founded in 1995 with over 300 corporate members. EtherNet/IP uses both of the most widely deployed collections of Ethernet standards –the Internet Protocol suite and IEEE 802.3 – to define the features and functions for its transport, network, data link and physical layers. EtherNet/IP performs at level session and above (level 5, 6 and 7) of the OSI model. CIP uses its object-oriented design to provide EtherNet/IP with the services and device profiles needed for real-time control applications and to promote consistent implementation of automation functions across a diverse ecosystem of products. In addition, EtherNet/IP adapts key elements of Ethernet’s standard capabilities and services to the CIP object model framework, such as the User Datagram Protocol (UDP), which EtherNet/IP uses to transport I/O messages. Ethernet/IP was estimated to have about 30% share of the industrial ethernet market in 2010 and 2018. History Development of EtherNet/IP began in the 1990s within a technical working group of ControlNet International, Ltd.(CI), another trade and standards development organization. In 2000, ODVA and CI formed a joint technology agreement (JTA) for the development of EtherNet/IP. In 2009, the JTA was terminated and EtherNet/IP became under the sole control of ODVA and its members. Today, EtherNet/IP is one of four networks that adapt CIP to an industrial network along with DeviceNet, ControlNet and CompoNet. All of these networks are managed by ODVA, Inc. Technical Detail EtherNet/IP classifies Ethernet nodes into predefined device t
https://en.wikipedia.org/wiki/Magnetic-activated%20cell%20sorting
Magnetic-activated cell sorting (MACS) is a method for separation of various cell populations depending on their surface antigens (CD molecules) invented by Miltenyi Biotec. The name MACS is a registered trademark of the company. The method was developed with Miltenyi Biotec's MACS system, which uses superparamagnetic nanoparticles and columns. The superparamagnetic nanoparticles are of the order of 100 nm. They are used to tag the targeted cells in order to capture them inside the column. The column is placed between permanent magnets so that when the magnetic particle-cell complex passes through it, the tagged cells can be captured. The column consists of steel wool which increases the magnetic field gradient to maximize separation efficiency when the column is placed between the permanent magnets. Magnetic-activated cell sorting is a commonly used method in areas like immunology, cancer research, neuroscience, and stem cell research. Miltenyi sells microbeads which are magnetic nanoparticles conjugated to antibodies which can be used to target specific cells. Procedure The MACS method allows cells to be separated by using magnetic nanoparticles coated with antibodies against a particular surface antigen. This causes the cells expressing this antigen to attach to the magnetic nanoparticles. After incubating the beads and cells, the solution is transferred to a column in a strong magnetic field. In this step, the cells attached to the nanoparticles (expressing the antigen) stay on the column, while other cells (not expressing the antigen) flow through. With this method, the cells can be separated positively or negatively with respect to the particular antigen(s). Positive and negative selection With positive selection, the cells expressing the antigen(s) of interest, which attached to the magnetic column, are washed out to a separate vessel, after removing the column from the magnetic field. This method is useful for isolation of a particular cell type, for in
https://en.wikipedia.org/wiki/S5%20%28modal%20logic%29
In logic and philosophy, S5 is one of five systems of modal logic proposed by Clarence Irving Lewis and Cooper Harold Langford in their 1932 book Symbolic Logic. It is a normal modal logic, and one of the oldest systems of modal logic of any kind. It is formed with propositional calculus formulas and tautologies, and inference apparatus with substitution and modus ponens, but extending the syntax with the modal operator necessarily and its dual possibly . The axioms of S5 The following makes use of the modal operators ("necessarily") and ("possibly"). S5 is characterized by the axioms: K: ; T: , and either: 5: ; or both of the following: 4: , and B: . The (5) axiom restricts the accessibility relation of the Kripke frame to be Euclidean, i.e. . Kripke semantics In terms of Kripke semantics, S5 is characterized by frames where the accessibility relation is an equivalence relation: it is reflexive, transitive, and symmetric. Determining the satisfiability of an S5 formula is an NP-complete problem. The hardness proof is trivial, as S5 includes the propositional logic. Membership is proved by showing that any satisfiable formula has a Kripke model where the number of worlds is at most linear in the size of the formula. Applications S5 is useful because it avoids superfluous iteration of qualifiers of different kinds. For example, under S5, if X is necessarily, possibly, necessarily, possibly true, then X is possibly true. Unbolded qualifiers before the final "possibly" are pruned in S5. While this is useful for keeping propositions reasonably short, it also might appear counter-intuitive in that, under S5, if something is possibly necessary, then it is necessary. Alvin Plantinga has argued that this feature of S5 is not, in fact, counter-intuitive. To justify, he reasons that if X is possibly necessary, it is necessary in at least one possible world; hence it is necessary in all possible worlds and thus is true in all possible worlds. Such reasoning
https://en.wikipedia.org/wiki/Procambarus%20angustatus
Procambarus angustatus was a species of crayfish in the family Cambaridae. It was only known from the type specimen, described by John Lawrence LeConte in 1856. He reported that it "lives in lesser Georgia, in the rivulets of pure water which flow between little sand hills". It was endemic to the U. S. state of Georgia, but is now believed to be extinct.
https://en.wikipedia.org/wiki/Klopsteg%20Memorial%20Award
The Klopsteg Memorial Award is an annual prize given to a notable physicist in memory of Paul E. Klopsteg. Established in 1990, it is awarded by the American Association of Physics Teachers. The Klopsteg Memorial Award recipient is asked to make a major presentation at an AAPT Summer Meeting on a topic of current significance suitable for non-specialists. Award Winners See also List of physics awards List of prizes named after people
https://en.wikipedia.org/wiki/Anti-symmetric%20operator
In quantum mechanics, a raising or lowering operator (collectively known as ladder operators) is an operator that increases or decreases the eigenvalue of another operator. In quantum mechanics, the raising operator is sometimes called the creation operator, and the lowering operator the annihilation operator. Well-known applications of ladder operators in quantum mechanics are in the formalisms of the quantum harmonic oscillator and angular momentum. Introduction Another type of operator in quantum field theory, discovered in the early 1970s, is known as the anti-symmetric operator. This operator, similar to spin in non-relativistic quantum mechanics is a ladder operator that can create two fermions of opposite spin out of a boson or a boson from two fermions. A Fermion, named after Enrico Fermi, is a particle with a half-integer spin, such as electrons and protons. This is a matter particle. A boson, named after S. N. Bose, is a particle with full integer spin, such as photons and W's. This is a force carrying particle. Spin First, we will review spin for non-relativistic quantum mechanics. Spin, an intrinsic property similar to angular momentum, is defined by a spin operator S that plays a role on a system similar to the operator L for orbital angular momentum. The operators and whose eigenvalues are and respectively. These formalisms also obey the usual commutation relations for angular momentum , , and . The raising and lowering operators, and , are defined as and respectively. These ladder operators act on the state in the following and respectively. The operators S_x and S_y can be determined using the ladder method. In the case of the spin 1/2 case (fermion), the operator acting on a state produces and . Likewise, the operator acting on a state produces and . The matrix representations of these operators are constructed as follows: Therefore, and can be represented by the matrix representations: Recalling the genera
https://en.wikipedia.org/wiki/Hibernaculum%20%28botany%29
Hibernaculum (plural hibernacula) is the term often applied to a winter bud of certain aquatic plants, such as the bladderworts (Utricularia). The buds are heavier than water, and, being developed at the approach of cold weather, they become detached, sink to the bottom of the pond, and thus survive the winter. In the spring, they enlarge, developing air spaces, rise to the surface, and reproduce their species. Certain terrestrial plants also form hibernacula. These include some temperate sundews (Drosera) such as D. anglica, D. filiformis, D. intermedia, D. rotundifolia; and some temperate butterworts (Pinguicula) such as P. balcanica, P. grandiflora, P. longifolia, and P. vulgaris. See also Hibernaculum Turion (botany)
https://en.wikipedia.org/wiki/Stanford%20University%20Mathematics%20Camp
Stanford University Mathematics Camp, or SUMaC, is a competitive summer mathematics program for rising high school juniors and seniors around the world. The camp lasts for 4 weeks, usually from mid-July to mid-August. It is based on the campus of Stanford University. Like the Ross Program at Ohio State and the PROMYS program at Boston University, SUMaC does not put emphasis on competition-math preparations but focuses instead on advanced undergraduate math topics. History SUMaC was founded in 1995 by Professors Rafe Mazzeo and Ralph Cohen of the Stanford Mathematics Department and has been directed from the beginning by Prof. Mazzeo, and Dr. Rick Sommer. Dr. Sommer was an assistant professor in the Stanford Mathematics Department and is currently a deputy director of the Education Program for Gifted Youth (EPGY), at Stanford. He designed the Program I course and has been teaching versions of it since the first SUMaC in 1995. The Program II course was designed and has been taught by Prof. Rafe Mazzeo. (In recent years, the course was cotaught by Dr. Pierre Albin, a former Stanford graduate student who currently teaches at MIT, and is currently taught by Dr. Simon Rubinstein-Salzedo, a postdoctoral fellow in statistics at Stanford.) Programs Program I investigates non-constructibility in geometry, classification of patterns in two dimensions, error-correcting codes, cryptography, and the analysis of the Rubik's Cube. The mathematics that is central to solving these problems comes from the areas of abstract algebra and number theory. Program II contains an introduction to selected topics in combinatorial, differential, and algebraic topology. The program emphasizes developing ideas from, and problems in, geometric topology where methods from abstract algebra and calculus have proven to be effective tools. Other activities During the camp, there are frequent guest lectures given by internationally renowned mathematicians. These talks are in the areas of current ma
https://en.wikipedia.org/wiki/Elective%20surgery
Elective surgery or elective procedure is surgery that is scheduled in advance because it does not involve a medical emergency. Semi-elective surgery is a surgery that must be done to preserve the patient's life, but does not need to be performed immediately. Most surgeries are elective, scheduled at a time to suit the surgeon, hospital, and patient. By contrast, an urgent surgery is one that can wait until the patient is medically stable, but should generally be done within 2 days, and an emergency surgery is one that must be performed without delay; the patient has no choice other than immediate surgery if permanent disability or death is to be avoided. Many surgeries can be performed as either elective or emergency surgeries, depending on the patient's needs. Description An elective surgery or elective procedure (from the , meaning to choose) is surgery that is scheduled in advance because it does not involve a medical emergency. Semi-elective surgery is a surgery that must be done to preserve the patient's life, but does not need to be performed immediately. Types Elective Most surgical medical treatments are elective, that is, scheduled at a time to suit the surgeon, hospital, and patient. These include inguinal hernia surgery, cataract surgery, mastectomy for breast cancer, and the donation of a kidney by a living donor. Elective surgeries include all optional surgeries performed for non-medical reasons. This includes cosmetic surgery, such as facelifts, breast implants, liposuction, and breast reduction, which aim to subjectively improve a patient's physical appearance. Another optional surgery is LASIK—currently the top elective surgery in the United States—where a patient weighs the risks against increased quality of life expectations. Semi-elective When a condition is worsening but has not yet reached the point of a true emergency, surgeons speak of semi-elective surgery: the problem must be dealt with, but a brief delay is not expected to affec
https://en.wikipedia.org/wiki/Schedules%20Direct
Schedules Direct is a non-profit organization that provides a low-cost television program listing service for open source and freeware digital video recorders. Developers from several different projects including MythTV, XMLTV, and GB-PVR founded Schedules Direct in response to Tribune Media Services's (TMS's) decision to shut down its free Data Direct program listing service as of September 1, 2007. MythTV and other such software use the data to display an on-screen electronic program guide (EPG), and to schedule upcoming recordings. Schedules Direct contracts with TMS to purchase a license to redistribute its data—which TiVo and other commercial digital video recorders also use—to Schedules Direct members. Individuals may become Schedules Direct members for $35 a year, and members may use the listings service with approved open source and freeware applications.
https://en.wikipedia.org/wiki/Marine%20pharmacognosy
Marine pharmacognosy is the investigation and identification of medically important plants and animals in the marine environment. It is a sub branch of terrestrial pharmacognosy. Generally the drugs are obtained from the marine species of bacteria, virus, algae, fungi and sponges. It is a relatively new field of study in western medicine, although many marine organisms were used in traditional Chinese medicine. It was not until 2004 that the first FDA approval of a drug came directly from the sea: ziconotide, which was isolated from a marine cone snail. With 79% of the earth's surface covered by water, research into the chemistry of marine organisms is relatively unexplored and represents a vast resource for new medicines to combat major diseases such as cancer, AIDS or malaria. Research typically focuses on sessile organisms or slow moving animals because of their inherent need for chemical defenses. Standard research involves an extraction of the organism in a suitable solvent followed by either an assay of this crude extract for a particular disease target or a rationally guided isolation of new chemical compounds using standard chromatography techniques. Marine organisms as sources of natural products Over 70% of the earth's surface is covered by oceans which contain 95% of the earth's biosphere. It was over 3500 million years ago that organisms first appeared in the sea. Over time, they have evolved many different mechanisms to survive the various harsh environments which include extreme temperatures, salinity, pressure, different levels of aeration and radiation, overcoming effects of mutation, and combating infection, fouling and overgrowth by other organisms. Adaptations to survive the different environments could be by physical or chemical adaptation. Organisms with no apparent physical defense, like sessile organisms, are believed to have evolved chemical defenses to protect themselves. It is also believed that the compounds would have to be extremely
https://en.wikipedia.org/wiki/Amadeus%20CRS
Amadeus is a computer reservation system (or global distribution system, since it sells tickets for multiple airlines) owned by the Amadeus IT Group with headquarters in Madrid, Spain. The central database is located at Erding, Germany. The major development centres are located in Sophia Antipolis (France), Bangalore (India), London (UK), and Boston (United States). In addition to airlines, the CRS is also used to book train travel, cruises, car rental, ferry reservations, and hotel rooms. Amadeus also provides New Generation departure control systems to airlines. Amadeus IT Group is a transaction processor for the global travel and tourism industry. The company is structured around two key related areas—its global distribution system and its "IT Solutions" business area. Amadeus is a member of IATA, OTA and SITA. Its IATA airline designator code is 1A. Other major reservation systems AirCore Galileo iFlyRes Navitaire (also owned by Amadeus) Sabre TravelSky Worldspan See also Amadeus IT Group Code sharing Passenger Name Record
https://en.wikipedia.org/wiki/European%20Physical%20Society
The European Physical Society (EPS) is a non-profit organisation whose purpose is to promote physics and physicists in Europe through methods such as physics outreach. Formally established in 1968, its membership includes the national physical societies of 42 countries, and some 3200 individual members. The Deutsche Physikalische Gesellschaft, the world's largest and oldest organisation of physicists, is a major member. Conferences One of its main activities is organizing international conferences. The EPS sponsors conferences other than the Europhysics Conference, like the International Conference of Physics Students in 2011. Divisions and groups The scientific activities of EPS are organised through Divisions and Groups, who organise topical conferences, seminars, and workshops. The Divisions and Groups are governed by boards elected from members. The current Divisions of the EPS are: Atomic, Molecular and Optical Physics Division Condensed Matter Division Environmental Physics Division Gravitational Physics Division High Energy Particle Physics Division Nuclear Physics Division Division of Physics in Life Sciences Physics Education Division Plasma Physics Division Quantum Electronics and Optics Division European Solar Physics Division Statistical & Nonlinear Physics Division And the current Groups of the EPS are: Accelerator Group Computational Physics Group Energy Group History of Physics Group Physics for Development Group Technology and Innovation Group Prizes The EPS awards a number of prizes, including the Edison Volta Prize, the EPS Europhysics Prize, the EPS Statistical and Nonlinear Physics Prizes, the High Energy and Particle Physics Prize and the Rolf Wideroe Prize. It also recognises sites which are historically important for advances to physics, such as the Blackett Laboratory (UK) in 2014, and the Residencia de Estudiantes (Spain) in 2015. Publications Its letters journal is EPL; its other publications include Europhysics Ne
https://en.wikipedia.org/wiki/Louis%20Narens
Louis Narens was the Graduate Director of Mathematical Behavioral Science, Professor in the Department of Cognitive Sciences and the Department of Logic and the Philosophy of Science at the University of California, Irvine. He was one of the major exponents of measurement theory in mathematical psychology. He died on October 5, 2022. Bibliography Abstract Measurement Theory. (1985) The MIT Press, Theories of Meaningfulness. (2002) Lawrence Erlbaum Associates, Introduction to the Theories of Measurement and Meaningfulness and the Use of Invariance in Science. (2007) Lawrence Erlbaum Associates, Theories of Probability: An Examination of Logical and Qualitative Foundations, (2007) World Scientific Publishing Company, See also American philosophy List of American philosophers
https://en.wikipedia.org/wiki/Native%20American%20mascot%20controversy
Since the 1960s, the issue of Native American and First Nations names and images being used by sports teams as mascots has been the subject of increasing public controversy in the United States and Canada. This has been a period of rising Indigenous civil rights movements, and Native Americans and their supporters object to the use of images and names in a manner and context they consider derogatory. They have conducted numerous protests and tried to educate the public on this issue. In response since the 1970s, an increasing number of secondary schools have retired such Native American names and mascots. Changes accelerated in 2020, following public awareness of institutional racism prompted by nationally covered cases of police misconduct. National attention was focused on the prominent use of names and images by professional franchises including the Washington Commanders (Redskins until July 2020) and the Cleveland Guardians (Indians until November 2021). In Canada, the Edmonton Eskimos became the Edmonton Elks in 2021. Each such change at the professional level has been followed by changes of school teams; for instance, 29 changed their names between August and December 2020. A National Congress of American Indians (NCAI) database tracks some 1,900 K-12 schools in 970 school districts with Native “themed” school mascots. The issue has often been reported in the media only in terms of Native American individuals being affected by the offensiveness of certain terms, images, and performances. This reduces the problem to one of feelings and personal opinions. It prevents a more comprehensive understanding of the history and context of the use of Native American names and images, and the reasons why sports teams should eliminate such practices. Social science research has shown that sports mascots and images are important symbols with deeper psychological and social effects in society. A 2020 analysis of this research indicates only negative effects; those psycholo
https://en.wikipedia.org/wiki/Normalisation%20by%20evaluation
In programming language semantics, normalisation by evaluation (NBE) is a method of obtaining the normal form of terms in the λ-calculus by appealing to their denotational semantics. A term is first interpreted into a denotational model of the λ-term structure, and then a canonical (β-normal and η-long) representative is extracted by reifying the denotation. Such an essentially semantic, reduction-free, approach differs from the more traditional syntactic, reduction-based, description of normalisation as reductions in a term rewrite system where β-reductions are allowed deep inside λ-terms. NBE was first described for the simply typed lambda calculus. It has since been extended both to weaker type systems such as the untyped lambda calculus using a domain theoretic approach, and to richer type systems such as several variants of Martin-Löf type theory. Outline Consider the simply typed lambda calculus, where types τ can be basic types (α), function types (→), or products (×), given by the following Backus–Naur form grammar (→ associating to the right, as usual): (Types) τ ::= α | τ1 → τ2 | τ1 × τ2 These can be implemented as a datatype in the meta-language; for example, for Standard ML, we might use: datatype ty = Basic of string | Arrow of ty * ty | Prod of ty * ty Terms are defined at two levels. The lower syntactic level (sometimes called the dynamic level) is the representation that one intends to normalise. (Syntax Terms) s,t,… ::= var x | lam (x, t) | app (s, t) | pair (s, t) | fst t | snd t Here lam/app (resp. pair/fst,snd) are the intro/elim forms for → (resp. ×), and x are variables. These terms are intended to be implemented as a first-order datatype in the meta-language: datatype tm = var of string | lam of string * tm | app of tm * tm | pair of tm * tm | fst of tm | snd of tm The denotational semantics of (closed) terms in the meta-language interprets the constructs of the syntax in terms o
https://en.wikipedia.org/wiki/The%20Complexity%20of%20Cooperation
The Complexity of Cooperation, by Robert Axelrod, is the sequel to The Evolution of Cooperation. It is a compendium of seven articles that previously appeared in journals on a variety of subjects. The book extends Axelrod's method of applying the results of game theory, in particular that derived from analysis of the Prisoner's Dilemma (IPD) problem, to real world situations. Prisoner's Dilemma findings Axelrod explains the Tit for tat (TFT or T4T) strategy emerged as the most robust option in early IPD tournaments on computer. This strategy combines a willingness to cooperate with a determination to punish non-cooperation. In these articles, however, he shows, that under more complex circumstances, such as the possibility of error, strategies that are a little more cooperative or a little less punitive do even better than TFT. Generous TFT, or GTFT, cooperates a bit more often than TFT, while Contrite TFT or CTFT defects less frequently. Applications Axelrod applies various models related to IPD to a variety of situations, drawing conclusions from these simulations about the ways in which groups form, adhere, oppose or join other groups, and other topics in the fields of genetic evolution, business, political science, military alliances, wars, and more. He has added introductions to these articles explaining what real-world issues drove his research. Critical response Philosopher and political economist Francis Fukuyama, writing for the Foreign Affairs, praises the book for showing that realist models, which assume that in situations lacking a single sovereign actor that anarchy will necessarily result, are too simplistic. Fukuyuma, expresses concern, however, that the game theory approaches aren't sufficiently complex to model real international relations, because they assume a world with large numbers of simple actors. Fukuyama holds that, instead, the real world consists of a small number of highly complex actors, thus potentially limiting the applicability
https://en.wikipedia.org/wiki/Proteolix
Proteolix, Inc., was a private biotechnology company with headquarters in South San Francisco, California. Proteolix was founded in 2003 based on technology developed by co-founders Dr. Craig Crews (Yale University) and Dr. Raymond J. Deshaies (California Institute of Technology). Drs. Susan Molineaux and Phil Whitcome (deceased) joined Drs. Crews and Deshaies as co-founders. Proteolix was launched based on an $18.2 million A round comprising investments by Latterell Venture Partners, US Venture Partners, Advanced Technology Ventures, and The Vertical Group. Proteolix focused primarily on the proteasome as a therapeutic target. Its lead product candidate, carfilzomib (PR-171), is a tetrapeptide epoxyketone. At the time of its sale (see below), the company had two earlier-stage programs, an orally-bioavailable proteasome inhibitor for oncology (PR-047), and an agent preferentially targeting the immuno form of the proteasome (PR-957), with potential utility in areas such as rheumatoid arthritis. At the time of sale, Carfilzomib's route of administration was intravenous, and the company was exploring its potential utility in multiple myeloma, Non-Hodgkin lymphoma (NHL) and other cancers. Proteolix was acquired by Onyx Pharmaceuticals in 2009 for $810 million (nominal value). Onyx renamed PR-047 to "ONX 0912" and PR-957 to "ONX 0914". External links Proteolix's corporate website Biotechnology companies of the United States Companies based in South San Francisco, California Life sciences industry
https://en.wikipedia.org/wiki/External%20water%20spray%20system
An external water spray system (EWSS) is a domestic external fire sprinkler system designed to protect homes from bushfires and wildfires. While external spray systems have long been used in fire protection for buildings and facilities, EWSS refers to domestic bushfire/wildfire systems. Usage The types of systems vary greatly from a single impact sprinkler placed on a roof, systems installed during construction with sprays on all windows and doors, and small sprays damping gutters. Some are installed using copper piping and sprays while others use common PVC piping. Usage depends on type of risk and belief of effectiveness. Copper piping is used to withstand high temperatures that may be experienced during a fire front or for higher reliability for in ceiling installations. External PVC piping is used where failure with exposure to high radiant heat is acceptable, as at the time the system is considered have "done its job". This is a matter of personal judgment. Issues There is a lack of scientific research regarding EWSS. Other issues that may affect the effectiveness of an EWSS: Sprinkler performance in high wind conditions typical in a bushfire Home design High vulnerability areas. e.g. decking Garden beds near walls Window sills Large windows (heat radiation) Internal furnishings (e.g. curtains) Home location Slope Surrounding vegetation If defended by occupiers Preparation (if prepared for defense) If intended for remote (unattended) defense Available water supply Ember attack Fire front (radiant heat) Adjacent building or vegetation fire Types Roof impact sprinkler Gutter sprays Window spray / deluge Commercial sprays designed for EWSS are available, but most systems use commonly available irrigation sprinklers/sprays. Research External spray systems for buildings are well documented for protection from fires in adjacent buildings. However, up there is little published scientific research on scientific information pertaining to t
https://en.wikipedia.org/wiki/Replicating%20portfolio
In mathematical finance, a replicating portfolio for a given asset or series of cash flows is a portfolio of assets with the same properties (especially cash flows). This is meant in two distinct senses: static replication, where the portfolio has the same cash flows as the reference asset (and no changes need to be made to maintain this), and dynamic replication, where the portfolio does not have the same cash flows, but has the same "Greeks" as the reference asset, meaning that for small (properly, infinitesimal) changes to underlying market parameters, the price of the asset and the price of the portfolio change in the same way. Dynamic replication requires continual adjustment, as the asset and portfolio are only assumed to behave similarly at a single point (mathematically, their partial derivatives are equal at a single point). Given an asset or liability, an offsetting replicating portfolio (a "hedge") is called a static hedge or dynamic hedge, and constructing such a portfolio (by selling or purchasing) is called static hedging or dynamic hedging. The notion of a replicating portfolio is fundamental to rational pricing, which assumes that market prices are arbitrage-free – concretely, arbitrage opportunities are exploited by constructing a replicating portfolio. In practice, replicating portfolios are seldom, if ever, exact replications. Most significantly, unless they are claims against the same counterparties, there is credit risk. Further, dynamic replication is invariably imperfect, since actual price movements are not infinitesimal – they may in fact be large – and transaction costs to change the hedge are not zero. Applications Derivatives pricing Dynamic replication is fundamental to the Black–Scholes model of derivatives pricing, which assumes that derivatives can be replicated by portfolios of other securities, and thus their prices determined. See explication under Rational pricing #The replicating portfolio. In limited cases static replicati
https://en.wikipedia.org/wiki/TSS/8
TSS/8 is a discontinued time-sharing operating system co-written by Don Witcraft and John Everett at Digital Equipment Corporation in 1967. DEC also referred to it as Timeshared-8 and later the EduSystem 50. The operating system runs on the 12-bit PDP-8 computer starting with the PDP-8I model and was released in 1968. Authorship TSS/8 was designed at Carnegie Mellon University with graduate student Adrian van de Goor, in reaction to the cost, performance, reliability, and complexity of IBM's TSS/360 (for their Model 67). Don Witcraft wrote the TSS/8 scheduler, command decoder and UUO (Unimplemented User Operations) handler. John Everett wrote the disk handler, file system, TTY (teletypewriter) handler and 680-I service routine for TSS/8. Roger Pyle and John Everett wrote the PDP-8 Disk Monitor System, and John Everett adapted PAL-III to make PAL-D for DMS. Bob Bowering, author of MACRO for the PDP-6 and PDP-10, wrote an expanded version, PAL-X, for TSS/8. Architecture This timesharing system is based on a protection architecture proposed by Adrian Van Der Goor, a grad student of Gordon Bell's at Carnegie-Mellon. It requires a minimum of 12K words of memory (8K for the operating system and 4K for the user swap area) and a swapping device; The standard swapping device, called a drum, was a disk drive with a head assigned to each track so there was no delay waiting for a read/write head to be repositioned on the drive. On a 24K word machine, it can give good support for its maximum of 16 users. Each user gets a virtual 4K PDP-8; many of the utilities users run on these virtual machines are modified versions of utilities from the Disk Monitor System or paper-tape environments. Internally, TSS/8 consists of RMON, the resident monitor, DMON, the disk monitor (file system), and KMON, the keyboard monitor (command shell). BASIC is well supported, while restricted (4K) versions of FORTRAN D and Algol are available. Like IBM's CALL/OS, it implements language variants
https://en.wikipedia.org/wiki/Monodora%20myristica
Monodora myristica, the calabash nutmeg, is a tropical tree of the family Annonaceae or custard apple family of flowering plants. It is native to Angola, Benin, Cameroon, the Central African Republic, the Democratic Republic of the Congo, Equatorial Guinea, Gabon, Ghana, Guinea-Bissau, Ivory Coast, Kenya, Liberia, Nigeria, the Republic of the Congo, Sierra Leone, Sudan, Tanzania, Togo and Uganda. In former times, its seeds were widely sold as an inexpensive nutmeg substitute. This is now less common outside its region of production. Other names of calabash nutmeg include Jamaican nutmeg, African nutmeg, ehuru, ariwo, awerewa, ehiri, airama, African orchid nutmeg, muscadier de Calabash and lubushi. Cultivation and history The calabash nutmeg tree grows naturally in evergreen forests from Liberia to Nigeria and Cameroon, Ghana, Angola and also Uganda and west Kenya. Due to the slave trade in the 18th century, the tree was introduced to the Caribbean islands where it was established and become known as Jamaican nutmeg. In 1897, Monodora myristica was introduced to Bogor Botanical Gardens, Indonesia, where the trees flower on a regular basis but no fruit could yet be collected. Due to its large and orchid-like flowers, the tree is also grown as an ornamental. Botany Tree and leaves Monodora myristica can reach a height of and in diameter at breast height (DBH). It has a clear trunk and branches horizontally. The leaves are alternately arranged and drooping with the leaf blade being elliptical, oblong or broadest towards the apex and tapering to the stalk. They are petiolate and can reach a size of up to . Flower The flower appears at the base of new shoots and is singular, pendant, large and fragrant. The pedicel bears a leaf-like bract and can reach in length. The flower’s sepals are red-spotted, crisped and long. The corolla is formed of six petals of which the three outer reach a length of and show curled margins and red, green and yellow spots. The t
https://en.wikipedia.org/wiki/Reference%20frame%20%28video%29
Reference frames are frames of a compressed video that are used to define future frames. As such, they are only used in inter-frame compression techniques. In older video encoding standards, such as MPEG-2, only one reference frame – the previous frame – was used for P-frames. Two reference frames (one past and one future) were used for B-frames. Multiple reference frames Some modern video encoding standards, such as H.264/AVC, allow the use of multiple reference frames. This allows the video encoder to choose among more than one previously decoded frame on which to base each macroblock in the next frame. While the best frame for this purpose is usually the previous frame, the extra reference frames can improve compression efficiency and/or video quality. Note that different reference frames can be chosen for different macroblocks in the same frame. The maximum number of concurrent reference frames supported by H.264 is 16. Different reference frames can be chosen for each 8x8 partition of a macroblock. Another video format that supports multiple reference frames is Snow, which can handle up to eight. The Theora codec provides a limited form of multiple reference frames, allowing references to both the preceding frame and the most recent intra frame. Disadvantages Encoding Multiple reference frames can considerably increase encoding time because many of the decisions, such as motion estimation, that are ordinarily carried out only on one reference frame have to be repeated on all of the reference frames. Heuristics can be used to reduce this speed cost at the cost of quality. Very high numbers of reference frames are rarely useful in terms of quality for live-action material because frames from farther back in time generally have less and less correlation with the current frame. This is not as true for animated sources, where repetitive motion can make high numbers of reference frames more useful. Decoding When decoding, reference frames must be stored in m
https://en.wikipedia.org/wiki/Waldhausen%20category
In mathematics, a Waldhausen category is a category C equipped with some additional data, which makes it possible to construct the K-theory spectrum of C using a so-called S-construction. It's named after Friedhelm Waldhausen, who introduced this notion (under the term category with cofibrations and weak equivalences) to extend the methods of algebraic K-theory to categories not necessarily of algebraic origin, for example the category of topological spaces. Definition Let C be a category, co(C) and we(C) two classes of morphisms in C, called cofibrations and weak equivalences respectively. The triple (C, co(C), we(C)) is called a Waldhausen category if it satisfies the following axioms, motivated by the similar properties for the notions of cofibrations and weak homotopy equivalences of topological spaces: C has a zero object, denoted by 0; isomorphisms are included in both co(C) and we(C); co(C) and we(C) are closed under composition; for each object A ∈ C the unique map 0 → A is a cofibration, i.e. is an element of co(C); co(C) and we(C) are compatible with pushouts in a certain sense. For example, if is a cofibration and is any map, then there must exist a pushout , and the natural map should be cofibration: Relations with other notions In algebraic K-theory and homotopy theory there are several notions of categories equipped with some specified classes of morphisms. If C has a structure of an exact category, then by defining we(C) to be isomorphisms, co(C) to be admissible monomorphisms, one obtains a structure of a Waldhausen category on C. Both kinds of structure may be used to define K-theory of C, using the Q-construction for an exact structure and S-construction for a Waldhausen structure. An important fact is that the resulting K-theory spaces are homotopy equivalent. If C is a model category with a zero object, then the full subcategory of cofibrant objects in C may be given a Waldhausen structure. S-construction The Waldhausen S-construc
https://en.wikipedia.org/wiki/Frozen%20noodles
Frozen noodles and chilled noodles are types of instantly prepared Asian (or European) noodles that are sold frozen or chilled. These products differ from prepackaged dehydrated noodles in a number of ways: in flavor, in texture, and in that they normally come packaged with ingredients besides the noodles, such as vegetables, meat, and soup stock. Types Chilled or frozen applications are applied to udon and Chinese-style noodles. For either type, the idea is that just 20–60 seconds of immersion in boiling water is necessary to reach a ready-to-eat state. Frozen noodles typically take less than two minutes to thaw and cook when placed in boiling water. Boiled and raw frozen noodles are the most commonly produced varieties, with raw varieties being produced less than boiled ones due to problems with dehydration that may occur when raw noodles are stored frozen. Soba (buckwheat) noodles are also manufactured as frozen noodles. Production The production of both chilled and frozen noodles starts with boiled noodles cooked to an optimum state of doneness, generally considered as having a moisture gradient at the surface of the noodles of 80% moisture absorption, and at the core of the noodles of 50% moisture absorption. After boiling, chilled noodles are placed into packaging followed by refrigeration at between . Frozen noodles, by contrast, are flash frozen using either air blast technology, a contact freezer or a combination of both, usually at for 30 minutes. Both processes induce a swelling of the starch that reduces deterioration and thus extends the noodles' shelf-life. In production, the boiling time for frozen noodles can be determined from a calculation that subtracts the cooking time of frozen boiled noodles from the total boiling time of fresh noodles. Mass-produced frozen noodles are sometimes packaged first, and then flash frozen. Flash freezing noodles can retain their quality of freshness for up to one year when kept properly frozen. Sales As of 20
https://en.wikipedia.org/wiki/Soil%20respiration
Soil respiration refers to the production of carbon dioxide when soil organisms respire. This includes respiration of plant roots, the rhizosphere, microbes and fauna. Soil respiration is a key ecosystem process that releases carbon from the soil in the form of CO2. CO2 is acquired by plants from the atmosphere and converted into organic compounds in the process of photosynthesis. Plants use these organic compounds to build structural components or respire them to release energy. When plant respiration occurs below-ground in the roots, it adds to soil respiration. Over time, plant structural components are consumed by heterotrophs. This heterotrophic consumption releases CO2 and when this CO2 is released by below-ground organisms, it is considered soil respiration. The amount of soil respiration that occurs in an ecosystem is controlled by several factors. The temperature, moisture, nutrient content and level of oxygen in the soil can produce extremely disparate rates of respiration. These rates of respiration can be measured in a variety of methods. Other methods can be used to separate the source components, in this case the type of photosynthetic pathway (C3/C4), of the respired plant structures. Soil respiration rates can be largely affected by human activity. This is because humans have the ability to and have been changing the various controlling factors of soil respiration for numerous years. Global climate change is composed of numerous changing factors including rising atmospheric CO2, increasing temperature and shifting precipitation patterns. All of these factors can affect the rate of global soil respiration. Increased nitrogen fertilization by humans also has the potential to affect rates over the entire planet. Soil respiration and its rate across ecosystems is extremely important to understand. This is because soil respiration plays a large role in global carbon cycling as well as other nutrient cycles. The respiration of plant structures releases
https://en.wikipedia.org/wiki/Behavioral%20momentum
Behavioral momentum is a theory in quantitative analysis of behavior and is a behavioral metaphor based on physical momentum. It describes the general relation between resistance to change (persistence of behavior) and the rate of reinforcement obtained in a given situation. B. F. Skinner (1938) proposed that all behavior is based on a fundamental unit of behavior called the discriminated operant. The discriminated operant, also known as the three-term contingency, has three components: an antecedent discriminative stimulus, a response, and a reinforcing or punishing consequence. The organism responds in the presence of the stimulus because past responses in the presence of that stimulus have produced reinforcement. Resistance to change According to behavioral momentum theory, there are two separable factors that independently govern the rate with which a discriminated operant occurs and the persistence of that response in the face of disruptions such as punishment, extinction, or the differential reinforcement of alternative behaviors. (see Nevin & Grace, 2000, for a review). First, the positive contingency between the response and a reinforcing consequence controls response rates (i.e., a response–reinforcer relation) by shaping a particular pattern of responding. This is governed by the relative law of effect (i.e., the matching law; Herrnstein, 1970). Secondly, the Pavlovian relation between surrounding, or context, stimuli and the rate or magnitude (but not both) of reinforcement obtained in the context (i.e., a stimulus–reinforcer relation) governs the resistance of the behavior to operations such as extinction. Resistance to change is assessed by measuring responding during operations such as extinction or satiation that tend to disrupt the behavior and comparing these measurements to stable, pre-disruption response rates. Resistance to disruption has been considered a better measure of response strength than a simple measure of response rate.(Nevin, 1974)
https://en.wikipedia.org/wiki/Metaweb
Metaweb Technologies, Inc. was a San Franciscobased company that developed Freebase, described as an "open, shared database of the world's knowledge". The company was co-founded by Danny Hillis, Veda Hlubinka-Cook and John Giannandrea in 2005. Metaweb was acquired by Google in 2010. Google shut down Freebase in 2016, transferring some of the data that met the required notability criteria to Wikidata. Funding On March 14, 2006, Metaweb received $15 million in funding. Investors included Benchmark Capital, Millennium Technology Ventures, and Omidyar Network. On January 15, 2008, Metaweb announced a $42.5 million Series B round led by Goldman Sachs and Benchmark Capital. Kevin Harvey of Benchmark Capital was a member of Metaweb's board of directors. Acquisition On July 16, 2010, Google acquired Metaweb for an undisclosed sum.
https://en.wikipedia.org/wiki/Trust%20seal
A trust seal is a seal granted by an entity to websites or businesses for display. Often the purpose is to demonstrate to customers that this business is concerned with security and their business identity. The requirements for the displaying merchant vary, but typically involve a dedication to good security practices, or the use of secure methods for transactions, or most importantly verified existence of the company. Trust seals can come in a variety of forms, including data security seals, business verified seals and privacy seals and are available from a variety of companies, for a fee. A trust seal can be either active or passive. Most seals are validated when they are created and remain so for a specific duration of time, post expiry of which the business/process has to be re-validated. Kinds of trust seals Privacy seal A privacy seal outfits a company with a privacy statement suited to its business practices. It also helps the company identify potential privacy threats that would otherwise go unnoticed. TRUSTe is an example of a privacy seal. Business practice seals These are seals that endorse an operational practice of a business. For example, an endorsement of the manufacturing quality practices of the company. Privacy seals are a subset of this category but popular enough, specifically with online retailers, to be mentioned separately. The Better Business Bureau seal is an example of a business practice seal. Business identity seal A business identity seal, also known as a Verified Existence Seal, is one which verifies the legal, physical and actual existence of the business by verifying multiple parameters such as statutory details, contact details, management details, etc. Verified existence trust seals add weight to the profiles of the deployers and boost confidence of prospective clients. A major benefit of a verified trust seal is that it represents due diligence by the grantor before granting a certificate for the business. Security seals Se
https://en.wikipedia.org/wiki/Nectar%20robbing
Nectar robbing is a foraging behavior utilized by some organisms that feed on floral nectar, carried out by feeding from holes bitten in flowers, rather than by entering through the flowers' natural openings. "Nectar robbers" usually feed in this way, avoiding contact with the floral reproductive structures, and therefore do not facilitate plant reproduction via pollination. Because many species that act as pollinators also act as nectar robbers, nectar robbing is considered to be a form of exploitation of plant-pollinator mutualism. While there is variation in the dependency on nectar for robber species, most species rob facultatively. Nectar robbers vary greatly in species diversity and include species of carpenter bees, bumblebees, stingless Trigona bees, solitary bees, wasps, ants, hummingbirds, and some passerine birds, including flowerpiercers. Nectar robbing mammals include a fruit bat and Swinhoe's striped squirrel, which robs nectar from the ginger plant. History Records of nectar robbing in nature date back at least to 1793, when German naturalist Christian Konrad Sprengel observed bumblebees perforating flowers. This was recorded in his book, The Secret of Nature in the Form and Fertilization of Flowers Discovered, which was written in Berlin. Charles Darwin observed bumblebees stealing nectar from flowers in 1859. These observations were published in his book The Origin of Species. Forms of floral larceny Nectar robbing is specifically the behavior of consuming nectar from a perforation (robbing hole) in the floral tissue rather than from the floral opening. There are two main types of nectar robbing: primary robbing, which requires that the nectar forager perforates the floral tissues itself, and secondary robbing, which is foraging from a robbing hole created by a primary robber. The former is most often on flowers whose nectar is concealed or hard to reach. Long flowers with tubular corollas are prone to robbing. Primary robbing is often performe
https://en.wikipedia.org/wiki/Dirty%20bit
A dirty bit or modified bit is a bit that is associated with a block of computer memory and indicates whether the corresponding block of memory has been modified. The dirty bit is set when the processor writes to (modifies) this memory. The bit indicates that its associated block of memory has been modified and has not been saved to storage yet. When a block of memory is to be replaced, its corresponding dirty bit is checked to see if the block needs to be written back to secondary memory before being replaced or if it can simply be removed. Dirty bits are used by the CPU cache and in the page replacement algorithms of an operating system. Dirty bits can also be used in Incremental computing by marking segments of data that need to be processed or have yet to be processed. This technique can be used with delayed computing to avoid unnecessary processing of objects or states that have not changed. When the model is updated (usually by multiple sources), only the segments that need to be reprocessed will be marked dirty. Afterwards, an algorithm will scan the model for dirty segments and process them, marking them as clean. This ensures the unchanged segments are not recalculated and saves processor time. Page replacement When speaking about page replacement, each page may have a modify bit associated with it in the hardware. The dirty bit for a page is set by the hardware whenever any word or byte in the page is written into, indicating that the page has been modified. When a page is selected for replacement, the modify bit is examined. If the bit is set, the page has been modified since it was read in from the disk. In this case, the page must be written to the disk. If the dirty bit is not set, however, the page has not been modified since it was read into memory. Therefore, if the copy of the page on the disk has not been overwritten (by some other page, for example), then there is no need to write the memory page to the disk: it is already there.
https://en.wikipedia.org/wiki/Bachmann%E2%80%93Howard%20ordinal
In mathematics, the Bachmann–Howard ordinal (also known as the Howard ordinal, or Howard-Bachmann ordinal) is a large countable ordinal. It is the proof-theoretic ordinal of several mathematical theories, such as Kripke–Platek set theory (with the axiom of infinity) and the system CZF of constructive set theory. It was introduced by and . Definition The Bachmann–Howard ordinal is defined using an ordinal collapsing function: εα enumerates the epsilon numbers, the ordinals ε such that ωε = ε. Ω = ω1 is the first uncountable ordinal. εΩ+1 is the first epsilon number after Ω = εΩ. ψ(α) is defined to be the smallest ordinal that cannot be constructed by starting with 0, 1, ω and Ω, and repeatedly applying ordinal addition, multiplication and exponentiation, and ψ to previously constructed ordinals (except that ψ can only be applied to arguments less than α, to ensure that it is well defined). The Bachmann–Howard ordinal is ψ(εΩ+1). The Bachmann–Howard ordinal can also be defined as for an extension of the Veblen functions φα to certain functions α of ordinals; this extension was carried out by Heinz Bachmann and is not completely straightforward.
https://en.wikipedia.org/wiki/Inpainting
Inpainting is a conservation process where damaged, deteriorated, or missing parts of an artwork are filled in to present a complete image. This process is commonly used in image restoration. It can be applied to both physical and digital art mediums such as oil or acrylic paintings, chemical photographic prints, sculptures, or digital images and video. With its roots in physical artwork, such as painting and sculpture, traditional inpainting is performed by a trained art conservator who has carefully studied the artwork to determine the mediums and techniques used in the piece, potential risks of treatments, and ethical appropriateness of treatment. History The modern use of inpainting can be traced back to Pietro Edwards (1744–1821), Director of the Restoration of the Public Pictures in Venice, Italy. Using a scientific approach, Edwards focused his restoration efforts on the intentions of the artist. It was during the 1930 International Conference for the Study of Scientific Methods for the Examination and Preservation of Works of Art, that the modern approach to inpainting was established. Helmut Ruhemann (1891–1973), a German restorer and conservator, led the discussions on the use of inpainting in conservation. Helmut Ruhemann was a leading figure in modernizing restoration and conservation. His greatest contribution to the field of conservation "was his insistence on following the methods of the original painter exactly, and on understanding the painter's artistic intention". After his career of over 40 years as a conservator, Ruhemann published his treatise The Cleaning of Paintings: Problems & Potentialities in 1968. In describing his method, Ruhemann states that "The surface [of the fill] should be slightly lower than that of the surrounding paint to allow for the thickness of the inpainting...Inpainting medium should look and behave like the original medium, but must not darken with age." Cesare Brandi (1906–1988) developed the teoria del restauro,
https://en.wikipedia.org/wiki/Regulated%20rewriting
Regulated rewriting is a specific area of formal languages studying grammatical systems which are able to take some kind of control over the production applied in a derivation step. For this reason, the grammatical systems studied in Regulated Rewriting theory are also called "Grammars with Controlled Derivations". Among such grammars can be noticed: Matrix Grammars Basic concepts Definition A Matrix Grammar, , is a four-tuple where 1.- is an alphabet of non-terminal symbols 2.- is an alphabet of terminal symbols disjoint with 3.- is a finite set of matrices, which are non-empty sequences , with , and , where each , is an ordered pair being these pairs are called "productions", and are denoted . In these conditions the matrices can be written down as 4.- S is the start symbol Definition Let be a matrix grammar and let the collection of all productions on matrices of . We said that is of type according to Chomsky's hierarchy with , or "increasing length" or "linear" or "without -productions" if and only if the grammar has the corresponding property. The classic example Note: taken from Abraham 1965, with change of nonterminals names The context-sensitive language is generated by the where is the non-terminal set, is the terminal set, and the set of matrices is defined as , , , . Time Variant Grammars Basic concepts Definition A Time Variant Grammar is a pair where is a grammar and is a function from the set of natural numbers to the class of subsets of the set of productions. Programmed Grammars Basic concepts Definition A Programmed Grammar is a pair where is a grammar and are the success and fail functions from the set of productions to the class of subsets of the set of productions. Grammars with regular control language Basic concepts Definition A Grammar With Regular Control Language, , is a pair where is a grammar and is a regular expression over the alphabet of the set of productions. A naive example Con
https://en.wikipedia.org/wiki/Rolanet
Rolanet (Robotron Local Area Network) was a networking standard, developed in the former German Democratic Republic (GDR) and introduced in 1987 by the computer manufacturer Robotron. It enabled computer networking over coax cable and glass fiber with a range of . Networking speed was 500 kBd, comparable to other standards of the day. A maximum of 253 computers could be connected using Rolanet. Two variants of Rolanet existed: Rolanet 1, introduced in 1987, saw limited deployment; Rolanet 2 was planned as a successor to Rolanet 1, but presumably never got beyond the prototype stage. A scaled-down version of Rolanet, BICNet, was used for educational purposes. It is no longer possible to assemble a functioning Rolanet system today, due to lack of software and working hardware. External links More information about Robotron networking technologies on Robotrontechnik.de Computer networking Science and technology in East Germany
https://en.wikipedia.org/wiki/Multiplicative%20digital%20root
In number theory, the multiplicative digital root of a natural number in a given number base is found by multiplying the digits of together, then repeating this operation until only a single-digit remains, which is called the multiplicative digital root of . The multiplicative digital root for the first few positive integers are: 0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 0, 2, 4, 6, 8, 0, 2, 4, 6, 8, 0, 3, 6, 9, 2, 5, 8, 2, 8, 4, 0. Multiplicative digital roots are the multiplicative equivalent of digital roots. Definition Let be a natural number. We define the digit product for base to be the following: where is the number of digits in the number in base , and is the value of each digit of the number. A natural number is a multiplicative digital root if it is a fixed point for , which occurs if . For example, in base , 0 is the multiplicative digital root of 9876, as All natural numbers are preperiodic points for , regardless of the base. This is because if , then and therefore If , then trivially Therefore, the only possible multiplicative digital roots are the natural numbers , and there are no cycles other than the fixed points of . Multiplicative persistence The number of iterations needed for to reach a fixed point is the multiplicative persistence of . The multiplicative persistence is undefined if it never reaches a fixed point. In base 10, it is conjectured that there is no number with a multiplicative persistence : this is known to be true for numbers . The smallest numbers with persistence 0, 1, ... are: 0, 10, 25, 39, 77, 679, 6788, 68889, 2677889, 26888999, 3778888999, 277777788888899. The search for these numbers can be sped up by using additional properties of the decimal digits of these record-breaking numbers. These digits must be sorted, and, except for the first two digits, all digits must be 7, 8, or 9. There are also additional restrictions on the first two digits. Based on these restriction
https://en.wikipedia.org/wiki/Persistent%20binding
Host-based zoning can include WWN or LUN masking, and is typically known as “persistent binding.” In storage networking, ”persistent binding” is an option of zoning. Host-based zoning is usually referred to as persistent binding or LUN, and is perhaps the least implemented form of zoning. Because it requires the host configuration to be correct in order to avoid zoning conflicts, this form of zoning creates a greater opportunity for administrative errors and conflicting access to targets. Moreover, zoning interfaces vary among different host operating systems and HBA's — increasing the possibility for administrative errors. If a host is not configured with the zoning software, it can access all devices in the fabric and create an even higher probability of data corruption. Host-based zoning is often used when clusters are implemented to control the mapping of devices to specific target IDs. However, it should never be the only form of zoning. Augmenting host-based zoning with storage- and fabric-based zoning is the only acceptable method to reliably control device access and data security. Basically, A given LUN has it SCSI id assigned by its RAID device (typically a SAN ). But for some purposes it's useful to have the SCSI id assigned by the host itself: that's persistent binding. What is Persistent binding for ? Without persistent binding, after every reboot, the SCSI id of a LUN may change. For Example, under Linux, a LUN bound on /dev/sda could migrate to /dev/sdb after a reboot. The risks augments with multipathing. Based on that, it is obvious that many software may crash without persistent binding. from http://www.storagesearch.com/datalink-art1.html Operating systems and upper-level applications (such as backup software) typically require a static or predictable SCSI target ID for their storage reliability and persistent binding affords that happening. Types of zoning A zone can include host and LUNS. The LUNS are exported by the DISK ARRAY, the hosts
https://en.wikipedia.org/wiki/Melioration%20theory
Melioration theory in behavioral psychology is a theoretical algorithm that predicts the matching law. Melioration theory is used as an explanation for why an organism makes choices based on the rewards or reinforcers it receives. The principle of melioration states that animals will invest increasing amounts of time and/or effort into whichever alternative is better. To meliorate essentially means to "make better". Melioration theory accounts for many of the choices that organisms make when presented with two variable interval schedules. Melioration is a form of matching where the subject is constantly shifting its behavior from the poorer reinforcement schedule to the richer reinforcement schedule, until it is spending most of its time at the richest variable interval schedule. By matching, the subject is equalizing the price of the reinforcer they are working for. This is also called hyperbolic discounting. In making a choice between options, living organisms need not maximize expected payoff as classical economic theory posits. Rather than being aggregated, the options compete against one another based on differences in their local reinforcement rate. The organism continuously shifts from one alternative to the other, if one is better than the other, until the other is better than the first one, regardless of the effect on overall rate of reinforcement. Melioration is capable of accounting for behavior on both concurrent ratio and concurrent interval schedules. Melioration Equation R1/B1 = R2/B2 If this ratio is not equal, the animal will shift its behavior to the alternative that currently has the higher response ratio. When the ratio is equal, the "cost" of each reinforcer is the same for both alternatives. Melioration theory grew out of an impersonal anonymous interest in how the matching law comes to hold on. Richard J. Herrnstein (1961) reported that on concurrent VIVIVI reinforcement schedules, the proportion of responses to one alternative was approxi
https://en.wikipedia.org/wiki/List%20of%20scientific%20publications%20by%20Albert%20Einstein
Albert Einstein (1879–1955) was a renowned theoretical physicist of the 20th century, best known for his theories of special relativity and general relativity. He also made important contributions to statistical mechanics, especially his treatment of Brownian motion, his resolution of the paradox of specific heats, and his connection of fluctuations and dissipation. Despite his reservations about its interpretation, Einstein also made seminal contributions to quantum mechanics and, indirectly, quantum field theory, primarily through his theoretical studies of the photon. Einstein's scientific publications are listed below in four tables: journal articles, book chapters, books and authorized translations. Each publication is indexed in the first column by its number in the Schilpp bibliography (Albert Einstein: Philosopher–Scientist, pp. 694–730) and by its article number in Einstein's Collected Papers. Complete references for these two bibliographies may be found below in the Bibliography section. The Schilpp numbers are used for cross-referencing in the Notes (the final column of each table), since they cover a greater time period of Einstein's life at present. The English translations of titles are generally taken from the published volumes of the Collected Papers. For some publications, however, such official translations are not available; unofficial translations are indicated with a § superscript. Although the tables are presented in chronological order by default, each table can be re-arranged in alphabetical order for any column by the reader clicking on the arrows at the top of that column. For illustration, to re-order a table by subject—e.g., to group together articles that pertain to "General relativity" or "Specific heats"—one need only click on the arrows in the "Classification and Notes" columns. To print out the re-sorted table, one may print it directly by using the web-browser Print option; the "Printable version" link at the left gives o
https://en.wikipedia.org/wiki/Lipomannan
Lipomannan is a mycobacterium immune agonist. In addition, it is a major constituent of the mycobacterium cell wall. This glycoconjugate is a virulence factor that plays a key role in the human immune system via interaction with various immune cells. It is also considered to be a precursor of lipoarabinomannans. It is a trigger for TLR 2. It consists of an α-linked mannan, which consists of 50–70 residues, with some branches points linked glycosidically to a diglyceride of which the fatty acids are similar to those of the whole cell lipid. In addition, succinic acid residues are present as O-acyl substituents on about one in four of the mannose residues, the terminal carboxyl group of the succinic acid providing the whole polymer with a considerable number of acidic functions. Lipomannan has functional components that resemble lipoteichoic acids; a lipophilic region and a hydrophilic portion with frequent acid groups. Lipomannan is a phosphorylated polysaccharide associated with the cell envelope and is considered to be the multimannosylated form of PIM which is primarily located in the plasma membrane. Structurally, LM is composed of two segments: a PI anchor to which is attached an α-D-mannan domain; both play key roles in inducing cytokine production by phagocytic cells. Mannose core consists of a linear α(1-6)-linked mannan backbone extending from the c-6 of the myo-inositol; the mannan chains is further substituted by α-(1-2) man-p side branches.
https://en.wikipedia.org/wiki/Lysophosphatidylserine
Lysophosphatidylserine is a lysophospholipid which triggers TLR 2. A recent study showed that it does not stimulate normal leukocytes. It also enhances glucose transport, lowering blood glucose levels while leaving secretion of insulin unaffected.
https://en.wikipedia.org/wiki/Sum%20of%20absolute%20differences
In digital image processing, the sum of absolute differences (SAD) is a measure of the similarity between image blocks. It is calculated by taking the absolute difference between each pixel in the original block and the corresponding pixel in the block being used for comparison. These differences are summed to create a simple metric of block similarity, the L1 norm of the difference image or Manhattan distance between two image blocks. The sum of absolute differences may be used for a variety of purposes, such as object recognition, the generation of disparity maps for stereo images, and motion estimation for video compression. Example This example uses the sum of absolute differences to identify which part of a search image is most similar to a template image. In this example, the template image is 3 by 3 pixels in size, while the search image is 3 by 5 pixels in size. Each pixel is represented by a single integer from 0 to 9. Template Search image 2 5 5 2 7 5 8 6 4 0 7 1 7 4 2 7 7 5 9 8 4 6 8 5 There are exactly three unique locations within the search image where the template may fit: the left side of the image, the center of the image, and the right side of the image. To calculate the SAD values, the absolute value of the difference between each corresponding pair of pixels is used: the difference between 2 and 2 is 0, 4 and 1 is 3, 7 and 8 is 1, and so forth. Calculating the values of the absolute differences for each pixel, for the three possible template locations, gives the following: Left Center Right 0 2 0 5 0 3 3 3 1 3 7 3 3 4 5 0 2 0 1 1 3 3 1 1 1 3 4 For each of these three image patches, the 9 absolute differences are added together, giving SAD values of 20, 25, and 17, respectively. From these SAD values, it could be asserted that the right side of the search image is the most similar to the template image, because it has the lowest sum of absolute differences as compared to the other two locations.
https://en.wikipedia.org/wiki/Lipophosphoglycan
Lipophosphoglycan (LPG) is a class of molecules found on the surface of some eukaryotes, in particular protozoa. Each is made up of two parts, lipid and polysaccharide (also called glycan). They are bonded by a phosphodiester, hence the name lipo-phospho-glycan. One species with extensive lipophosphoglycan coating is Leishmania, a group of single-celled protozoan parasite which cause leishmaniasis in many mammals, including humans. Their coats help modulate their hosts' immunological responses.