source
stringlengths
31
227
text
stringlengths
9
2k
https://en.wikipedia.org/wiki/Cupin%20superfamily
The cupin superfamily is a diverse superfamily of proteins named after its conserved barrel domain (cupa being the Latin term for a small barrel). The superfamily includes a wide variety of enzymes as well as non-enzymatic seed storage proteins. Members of the superfamily play a role in allergy, especially seed storage proteins like 7S and 11S globulins, also known as vicilins and legumins, respectively. These proteins can be found at high concentrations in seeds of both mono- and dicotyledonous plants and are an important component of the normal human diet. History Thomas Burr Osborne at the end of the 19th century was the first person to systematically study seed storage proteins by their solubility characteristics. He established 4 classes of proteins: water-soluble albumins; salt soluble globulins: vicilin—typically having sedimentation coefficients, S values (a measure of the protein mass determined by sedimentation equilibrium ultracentrifugation) of about 7 Svedberg units (hence the common name 7S globulin) and legumin (11S); alcohol/water-soluble—cereal—prolamines; and a fourth class, glutelins, of difficultly soluble proteins no longer recognized and now considered low solubility prolamin or globulin storage proteins . Gluten consists of a mixture of prolamins: 'glutenin' and 'gliadin'. Osborne and his Yale colleague Lafayette Mendel are considered the 'founders' of the modern science of nutrition. Earlier, the fungus Sclerotinia sclerotiorum (Lib.) deBary was the first oxalic acid (oxalate), secreting organism to be described as early as 1886 in Botan. Z. by A. de Barry. However, since oxalate secreting fungi are not a major threat to crop cereals no studies of this interaction were made for almost 100 years. In the early 1980s a protein dubbed 'germin' was identified in germinating wheat embryos; and in the early 1990s (1992) it was found to be an enzyme having oxalate oxidase (OXO) activity converting an oxalate substrate into carbon dioxide and hydr
https://en.wikipedia.org/wiki/Feng%27s%20classification
Tse-yun Feng suggested the use of degree of parallelism to classify various computer architecture. It is based on sequential and parallel operations at a bit and word level. About degree of parallelism Maximum degree of parallelism The maximum number of binary digits that can be processed within a unit time by a computer system is called the maximum parallelism degree P. If a processor is processing P bits in unit time, then P is called the maximum degree of parallelism. Average degree of parallelism Let i = 1, 2, 3, ..., T be the different timing instants and P1, P2, ..., PT be the corresponding bits processed. Then, Processor utilization Processor utilization is defined as The maximum degree of parallelism depends on the structure of the arithmetic and logic unit. Higher degree of parallelism indicates a highly parallel ALU or processing element. Average parallelism depends on both the hardware and the software. Higher average parallelism can be achieved through concurrent programs. Types of classification According to Feng's classification, computer architecture can be classified into four. The classification is based on the way contents stored in memory are processed. The contents can be either data or instructions. Word serial bit serial (WSBS) Word serial bit parallel (WSBP) Word parallel bit serial (WPBS) Word parallel bit parallel (WPBP) Word serial bit serial (WSBS) One bit of one selected word is processed at a time. This represents serial processing and needs maximum processing time. Word serial bit parallel (WSBP) It is found in most existing computers and has been called "word slice" processing because one word of one bit is processed at a time. All bits of a selected word are processed at a time. Bit parallel means all bits of a word. Word parallel bit serial (WPBS) It has been called bit slice processing because m-bit slice is processed at a time. Word parallel signifies selection of all words. It can be considered as one bit
https://en.wikipedia.org/wiki/Carprofen
Carprofen is a nonsteroidal anti-inflammatory drug (NSAID) of the carbazole and propionic acid class that was previously for use in humans and animals but is now only available to veterinarians for prescribing as a supportive treatment for various conditions in animals. Carprofen reduces inflammation by inhibition of COX-1 and COX-2; its specificity for COX-2 varies from species to species. Marketed under many brand names worldwide, carprofen is used as a treatment for inflammation and pain, including joint pain and postoperative pain. Human use Carprofen was used in humans for almost ten years, starting in 1988, for the same conditions as in dogs; namely, joint pain and inflammation. Side effects tended to be mild, usually consisting of nausea or gastrointestinal pain and diarrhoea. It was available by prescription in 150 mg to 600 mg doses. Dosages over 250 mg were reserved for pain caused by severe trauma, such as postoperative inflammation; 150 mg doses were commonly used to relieve arthritis pain, while 200 mg doses were commonly prescribed for severe arthritis or inflammatory pain. The drug was taken orally. Pfizer voluntarily removed the medication from the market for human use on commercial grounds. Veterinary medicine Canine use Carprofen is one of eleven nonsteroidal anti-inflammatory drugs approved for use in dogs. It aids in the relief of inflammation, pain, and fever. Carprofen can be administered in pill, chewable tablet, or injection form. Carprofen can be used for long-term pain management of such conditions as osteoarthritis, which is common in canine patients, or after surgical procedures for relief of acute pain and inflammation. In patients suffering from pain, carprofen has been shown to improve energy, activity level, comfort, and general well-being. Adverse effects Most dogs respond well to carprofen use, but like all NSAIDs, it can cause gastrointestinal, liver, and kidney problemss. In 1999, the Food and Drug Administration (FDA)
https://en.wikipedia.org/wiki/Os%20%28Fringe%29
"Os" is the 16th episode of the third season of the American science fiction drama television series Fringe, and the 59th episode overall. The episode centered on the Fringe team's investigation into a series of robberies of the element osmium, which they connect to a scientist (Alan Ruck) who is able to defy the laws of physics. "Os" was written by Josh Singer and Graham Roland, while Brad Anderson served as director. Along with Ruck, the episode also guest-starred Jorge Garcia in a brief cameo as a security guard. On its first broadcast in the United States on March 11, 2011, an estimated 3.76 million viewers tuned in, giving it a 1.5/5 ratings share for those between the ages 18–49. Critical reception to the episode was generally positive, as multiple critics noted that Ruck was well-cast as a scientist just trying to help his son. Plot The Fringe team is called to the scene of a robbery of a metal depository; the body of one of the thieves, shot by a security guard, inexplicably floats off the ground, while a second culprit has gotten away. They find that the thief has taken off with a supply of osmium, one of the densest elements, and an autopsy later reveals that the dead thief's body is filled with the metal. Tracking a security card on the body, they enter a warehouse where they find the second culprit, dead, along with the bodies of several more people, all of them paraplegic. Walter (John Noble) notes that the physical properties of osmium (both gravitational and thermal) have been reversed, and by melting the osmium collected from the first victim using liquid nitrogen, they detect the presence of the rarer element lutetium, typically only present in meteorites. Recognizing that the local science museum has a display of meteorites, the Fringe team is able to secure Dr. Crick (Alan Ruck), the man behind the osmium injections, and another paraplegic who has taken Dr. Crick's injections as they attempt to steal the display. With Dr. Crick in custody, Wal
https://en.wikipedia.org/wiki/Argyle%20%28pattern%29
An argyle (, occasionally spelled argyll) pattern is made of diamonds or lozenges. The word is sometimes used to refer to an individual diamond in the design, but more commonly refers to the overall pattern. Most argyle contains layers of overlapping motifs, adding a sense of three-dimensionality, movement, and texture. Typically, there is an overlay of intercrossing diagonal lines on solid diamonds. History The argyle pattern derives loosely from the tartan of Clan Campbell of Argyll in western Scotland, used for kilts and plaids, and from the patterned socks worn by Scottish Highlanders since at least the 17th century (these were generally known as "tartan hose"). Modern argyle patterns, however, are usually not true tartans, as they have two solid colours side-by-side, which is not possible in a tartan weave (solid colours in tartan are next to blended colours and only touch other solid colours at their corners). 20th century Argyle knitwear became fashionable in Great Britain and then in the United States after the First World War of 1914–1918. Pringle of Scotland popularised the design, helped by its identification with the Duke of Windsor. Pringle's website says that "the iconic Pringle argyle design was developed" in the 1920s. The Duke, like others, used this pattern for golf clothing: both for jerseys and for the long socks needed for the plus-fours trouser fashion of the day. Contemporary use Payne Stewart (1957–1999), who won the U.S. Open in 1991 and 1999 and the PGA championship in 1989, was known for his flashy tams, knickerbockers, and argyle socks. Some sports teams use bright, contemporary interpretations of the argyle pattern. For example, the Norwegian men's curling team at the 2010 Winter Olympics and the Garmin–Slipstream professional cycling team, nicknamed the "Argyle Armada". On 27 April 2013 the professional soccer team Sporting Kansas City of Major League Soccer (MLS) in the United States announced their third kit of the 2013 seaso
https://en.wikipedia.org/wiki/Strongly%20interacting%20massive%20particle
Strongly interacting massive particles (SIMPs) are hypothetical particles that interact strongly between themselves and weakly with ordinary matter, but could form the inferred dark matter despite this. Strongly interacting massive particles have been proposed as a solution for the ultra-high-energy cosmic-ray problem and the absence of cooling flows in galactic clusters. Various experiments and observations have set constraints on SIMP dark matter from 1990 onward. SIMP annihilations would produce significant heat. DAMA set limits with NaI(Tl) crystals. Measurements of Uranus's heat excess exclude SIMPs from 150 MeV to 104 GeV. Earth's heat flow significantly constrains any cross section. See also
https://en.wikipedia.org/wiki/Synchronizing%20word
In computer science, more precisely, in the theory of deterministic finite automata (DFA), a synchronizing word or reset sequence is a word in the input alphabet of the DFA that sends any state of the DFA to one and the same state. That is, if an ensemble of copies of the DFA are each started in different states, and all of the copies process the synchronizing word, they will all end up in the same state. Not every DFA has a synchronizing word; for instance, a DFA with two states, one for words of even length and one for words of odd length, can never be synchronized. Existence Given a DFA, the problem of determining if it has a synchronizing word can be solved in polynomial time using a theorem due to Ján Černý. A simple approach considers the power set of states of the DFA, and builds a directed graph where nodes belong to the power set, and a directed edge describes the action of the transition function. A path from the node of all states to a singleton state shows the existence of a synchronizing word. This algorithm is exponential in the number of states. A polynomial algorithm results however, due to a theorem of Černý that exploits the substructure of the problem, and shows that a synchronizing word exists if and only if every pair of states has a synchronizing word. Length The problem of estimating the length of synchronizing words has a long history and was posed independently by several authors, but it is commonly known as the Černý conjecture. In 1969, Ján Černý conjectured that (n − 1)2 is the upper bound for the length of the shortest synchronizing word for any n-state complete DFA (a DFA with complete state transition graph). If this is true, it would be tight: in his 1964 paper, Černý exhibited a class of automata (indexed by the number n of states) for which the shortest reset words have this length. The best upper bound known is 0.1654n3, far from the lower bound. For n-state DFAs over a k-letter input alphabet, an algorithm by David Eppstein fi
https://en.wikipedia.org/wiki/Multiplicative%20sequence
In mathematics, a multiplicative sequence or m-sequence is a sequence of polynomials associated with a formal group structure. They have application in the cobordism ring in algebraic topology. Definition Let Kn be polynomials over a ring A in indeterminates p1, ... weighted so that pi has weight i (with p0 = 1) and all the terms in Kn have weight n (in particular Kn is a polynomial in p1, ..., pn). The sequence Kn is multiplicative if the map is an endomorphism of the multiplicative monoid , where . The power series is the characteristic power series of the Kn. A multiplicative sequence is determined by its characteristic power series Q(z), and every power series with constant term 1 gives rise to a multiplicative sequence. To recover a multiplicative sequence from a characteristic power series Q(z) we consider the coefficient of z j in the product for any m > j. This is symmetric in the βi and homogeneous of weight j: so can be expressed as a polynomial Kj(p1, ..., pj) in the elementary symmetric functions p of the β. Then Kj defines a multiplicative sequence. Examples As an example, the sequence Kn = pn is multiplicative and has characteristic power series 1 + z. Consider the power series where Bk is the k-th Bernoulli number. The multiplicative sequence with Q as characteristic power series is denoted Lj(p1, ..., pj). The multiplicative sequence with characteristic power series is denoted Aj(p1,...,pj). The multiplicative sequence with characteristic power series is denoted Tj(p1,...,pj): these are the Todd polynomials. Genus The genus of a multiplicative sequence is a ring homomorphism, from the cobordism ring of smooth oriented compact manifolds to another ring, usually the ring of rational numbers. For example, the Todd genus is associated to the Todd polynomials with characteristic power series .
https://en.wikipedia.org/wiki/Common%20fibular%20nerve
The common fibular nerve (also known as the common peroneal nerve, external popliteal nerve, or lateral popliteal nerve) is a nerve in the lower leg that provides sensation over the posterolateral part of the leg and the knee joint. It divides at the knee into two terminal branches: the superficial fibular nerve and deep fibular nerve, which innervate the muscles of the lateral and anterior compartments of the leg respectively. When the common fibular nerve is damaged or compressed, foot drop can ensue. Structure The common fibular nerve is the smaller terminal branch of the sciatic nerve. The common fibular nerve has root values of L4, L5, S1, and S2. It arises from the superior angle of the popliteal fossa and extends to the lateral angle of the popliteal fossa, along the medial border of the biceps femoris. It then winds around the neck of the fibula to pierce the fibularis longus and divides into terminal branches of the superficial fibular nerve and the deep fibular nerve. Before its division, the common fibular nerve gives off several branches in the popliteal fossa. Cutaneous branches Lateral sural cutaneous nerve (lateral cutaneous nerve of calf) - supplies the skin of the upper two-thirds of the lateral side of leg. sural communicating nerve - it runs on the posterolateral aspect of the calf and joins the sural nerve. Articular branches Superior lateral genicular nerve - accompanies artery of the same name and lies above the lateral femoral condyle. Inferior lateral genicular nerve - accompanies artery of the same name and lies just above the head of the fibula. Recurrent genicular nerve - It arises from the point of division of the common fibular nerve; then ascends anterior to the knee joint together with the anterior recurrent tibial artery to supply the knee joint and the tibialis anterior muscle. Motor branches There is only one motor branch that arises directly from the common fibular nerve, the nerve to the short head of the biceps femoris m
https://en.wikipedia.org/wiki/Loop%20management%20system
In local telephone networks, a loop management system (LMS) is a kind or a part of network management system intended to maximize local loop control. Sometimes it is referred to as local loop management (LLM) or copper loop management (CLM). Although local loop unbundling is a standard process for an incumbent (ILEC), issues remain to be solved in the local loop management process. For a CLEC which borrows lines from ILEC for DSL services provisioning process, a local loop is the most critical (and the most weak) point because of reduced management of this vital part of the network. During the provisioning process, a CLEC can request from its serving ILEC, a new cross-connect. By agreement ILEC must fulfill this request but it's almost impossible to immediately learn a local loop's quality or monitor its activity in real time. Such steps are sometimes mandatory for loop pretesting and qualification in order to validate its good (or bad) condition. Furthermore, testing and validation steps are vital in troubleshooting process. LMS, sometimes implemented as a part of a major network management system, views the local loop as an active network element. It speeds pre-qualification and reduces fault correction time. A complete LMS has software and hardware components. The latter is usually a switch fabric connected both to all required lines and line equipment at either side of a line. The functionality is similar to one performed by a main distribution frame or any other distribution frame, so sometimes is referred to as automated main distribution frame. A software component implies implementing management operations. Network management Local loop
https://en.wikipedia.org/wiki/Red%20flag%20%28idiom%29
A red flag could either be a literal red flag used for signaling or, as a metaphor, a sign of some particular problem requiring attention. Background The term and the expression "to raise the red flag" come from various usages of real flags throughout history. A red flag is frequently flown by armed forces to warn the public of live fire exercises in progress, and is sometimes flown by ships carrying munitions (in this context it is actually the flag for the letter B in the international maritime signal flag alphabet, a red swallow-tailed flag). In many countries a red flag is flown to signify that an outdoor shooting range is in use. The United States Air Force refers to its largest annual exercise as Red Flag operation. Red flags are used for various signals in team sailing races (see Racing Rules of Sailing). A red flag warning is a signal of high wildfire danger, and a red flag on the beach warns of dangerous water conditions (double red flags indicate beach closure). Red flags of various designs indicate dangerous wind and wave conditions for mariners. In auto racing, a red flag indicates that a race has been stopped. A signal of danger or a problem can be referred to as a red flag, a usage that originated in the 18th century. An infamous example of use of a red flag in warfare is Mexican General Antonio López de Santa Anna's use of the symbol to let his Texian opposition in the Alamo know that he intended to spare none of the defenders (on which he followed through). The term "red flag" is used, e.g., during screening of communications, and refers to specific words or phrases encountered that might indicate relevance to the case. For example, email spam filters make use of such "red flags". See also Red flag (racing) Red Flags Rule
https://en.wikipedia.org/wiki/Arakawa%E2%80%93Kaneko%20zeta%20function
In mathematics, the Arakawa–Kaneko zeta function is a generalisation of the Riemann zeta function which generates special values of the polylogarithm function. Definition The zeta function is defined by where Lik is the k-th polylogarithm Properties The integral converges for and has analytic continuation to the whole complex plane as an entire function. The special case k = 1 gives where is the Riemann zeta-function. The special case s = 1 remarkably also gives where is the Riemann zeta-function. The values at integers are related to multiple zeta function values by where
https://en.wikipedia.org/wiki/Jay%20Last
Jay Taylor Last (October 18, 1929 – November 11, 2021) was an American physicist, silicon pioneer, and member of the so-called "traitorous eight" that founded Silicon Valley. Early life and education Last was born in Butler, Pennsylvania, on October 18, 1929, at the beginning of the Stock Market Crash of 1929, and grew up during the Great Depression. Both his parents were teachers, but his father left teaching to work in a steel mill in hopes of earning a better living. During the depression, there was no work in the steel mills, but the family managed by growing and preserving its own food. During World War II, his father worked six to seven days a week, 12 hours a day, under demanding and dangerous physical conditions. Jay Last enjoyed hiking, walking, and exploring while growing up. Between his junior and senior years of school, at age 16, he and a friend hitch-hiked to San Jose, California, and worked for the summer picking fruit. A voracious reader, he tended to complete his schoolwork well in advance of the rest of the class. He was encouraged by his chemistry teacher, Lucille Critchlow, who recommended him to work with Frank W. Preston, a local industrial chemist whose laboratory studied glass and glass fracture. Last began working at Preston's lab as a high-school student and continued to work for him as a university student, whenever he had a break. Last graduated from Butler Senior High School in 1947 and applied for a scholarship to study Optics at the University of Rochester. Last had heard about the program from his father and did not apply anywhere else. It was a rigorous program, and three-quarters of the entering class had dropped out by the time the program was finished. The program had close ties to Eastman Kodak and to Bausch & Lomb: Last's class in optical design was taught by Rudolph Kingslake of Kodak. Last worked for a summer at the trouble-shooting department of Kodak's optical instrumentation plant, before his senior year of university
https://en.wikipedia.org/wiki/Sales%20and%20operations%20planning
Sales and operations planning (S&OP) is an integrated business management process through which the executive/leadership team continually achieves focus, alignment, and synchronization among all organization functions. The S&OP process includes an updated forecast that leads to a sales plan, production plan, inventory plan, customer lead time (backlog) plan, new product development plan, strategic initiative plan, and resulting financial plan. Plan frequency and planning horizon depend on the specifics of the context. Short product life cycles and high demand volatility require a tighter S&OP than steadily consumed products. Done well, the S&OP process also enables effective supply chain management. The Sales and Operations planning process has a twofold scope. The first scope is the horizontal alignment to balance the supply and demand through integration between the company departments and with suppliers and customers. The second aim is the vertical alignment amid strategic plan and the operational plan of a company. A properly implemented S&OP process routinely reviews customer demand and supply resources and "re-plans" quantitatively across an agreed 'rolling' horizon. The re-planning process focuses on changes from the previously agreed sales and operations plan; while it helps the management team to understand how the company achieved its current level of performance, its focus is on future actions and anticipated results. Definitions S&OP was developed with the concept of aggregated production planning (APP) in the first part of 1950, then switched to manufacturing resource planning (MRP 2) around 1985, till the current definition of business process for the alignment of supply and demand. The term S&OP and its modern meaning were conceived of in the 1980s and are generally attributed to Richard Ling, then a consultant with the management consulting firm Oliver Wight. APICS defines S&OP as the "function of setting the overall level of manufacturing outp
https://en.wikipedia.org/wiki/Honeygar
Honeygar, also Honegar, is a mix of honey and apple cider vinegar, similar to switchel. Honey and vinegar mixtures such as oxymel have been used for purported health benefits since ancient times. The name "Honegar" was used by D. C. Jarvis in his book Folk Medicine: A Vermont Doctor's Guide to Good Health (1958). Following the success of Jarvis's book in the US, honegar also enjoyed some popularity in Japan. Many in natural health circles claim that honeygar has wide ranging benefits on health. One particularly repeated claim is its benefits for the symptoms of arthritis, although there is no clinical evidence to support this. Criticism In 1960, copies of Jarvis' book Folk Medicine were seized by the Food and Drug Administration in connection with sales of honegar. Physician Louis Lasagna noted that: In Albany, New York, FDA agents seized $60,000 worth of "Honegar," a mixture of honey and apple cider vinegar, because its labeling failed to bear adequate directions for treating nearly fifty diseases and conditions for which "Honegar" was intended to be used. Seized with the mixture were reprints and quotations from Jarvis' book. (Jarvis was apparently uninvolved in the commercial manufacture of the product.)
https://en.wikipedia.org/wiki/Video%20Acceleration%20API
Video Acceleration API (VA-API) is an open source application programming interface that allows applications such as VLC media player or GStreamer to use hardware video acceleration capabilities, usually provided by the graphics processing unit (GPU). It is implemented by the free and open-source library , combined with a hardware-specific driver, usually provided together with the GPU driver. VA-API video decode/encode interface is platform and window system independent but is primarily targeted at Direct Rendering Infrastructure (DRI) in X Window System on Unix-like operating systems (including Linux, FreeBSD, Solaris), and Android, however it can potentially also be used with direct framebuffer and graphics sub-systems for video output. Accelerated processing includes support for video decoding, video encoding, subpicture blending, and rendering. The VA-API specification was originally designed by Intel for its GMA (Graphics Media Accelerator) series of GPU hardware with the specific purpose of eventually replacing the XvMC standard as the default Unix multi-platform equivalent of Microsoft Windows DirectX Video Acceleration (DxVA) API, but today the API is no longer limited to Intel-specific hardware or GPUs. Other hardware and manufacturers can freely use this open standard API for hardware accelerated video processing with their own hardware without paying a royalty fee. Overview The main motivation for VA-API is to enable hardware-accelerated video decode at various entry-points (VLD, IDCT, motion compensation, deblocking) for the prevailing coding standards today (MPEG-2, MPEG-4 ASP/H.263, MPEG-4 AVC/H.264, H.265/HEVC, and VC-1/WMV3). Extending XvMC was considered, but due to its original design for MPEG-2 MotionComp only, it made more sense to design an interface from scratch that can fully expose the video decode capabilities in today's GPUs. Supported hardware and drivers As of 2022, VA-API is natively supported by: Intel Quick Sync open-source dr
https://en.wikipedia.org/wiki/List%20of%20Australian%20herbs%20and%20spices
Australian herbs and spices were used by Aboriginal peoples to flavour food in ground ovens. The term "spice" is applied generally to the non-leafy range of strongly flavoured dried Australian bushfoods. They mainly consist of aromatic fruits and seed products, although Australian wild peppers also have spicy leaves. There are also a few aromatic leaves but unlike culinary herbs from other cultures which often come from small soft-stemmed forbs, the Australian herb species are generally trees from rainforests, open forests and woodlands. Australian herbs and spices are generally dried and ground to produce a powdered or flaked spice, either used as a single ingredient or in blends. They were used to a limited extent by colonists in the 18th and 19th centuries. Some extracts were used as flavouring during the 20th century. Australian native spices have become more widely recognized and used by non-indigenous people since the early 1980s as part of the bushfood industry, with increasing gourmet use and export. They can also be used as a fresh product. Leaves can be used whole, like a bay-leaf in cooking, or spicy fruits are added to various dishes for flavour. The distilled essential oils from leaves and twigs are also used as flavouring products. Fruit Acronychia acidula, Lemon Aspen Acronychia oblongifolia, White Aspen, Yellow Wood Austromyrtus dulcis, Midgen Berry, Silky Myrtle Citrus australasica, Finger Lime, Caviar Lime Citrus australis, Round Lime, Australian Lime Citrus glauca, Desert Lime Eupomatia laurina, Bolwarra, Native Guava, Copper Laurel Kunzea pomifera, Muntries, Emu Apples, Native Cranberries Solanum centrale, Akudjura, Australian Desert Raisin Solanum chippendalei, Chippendale's Tomato, Bush Tomato Solanum cleistogamum, Potato Bush, Bush Tomato Syzygium luehmannii, Riberry, Cherry Alder, Small Leaf Lilly Pilly Herbs Apium insulare, Flinders Island Celery Apium prostratum, Sea Celery Atherosperma Moschatum, safrole, Southern S
https://en.wikipedia.org/wiki/The%20Secret%20Island%20of%20Dr.%20Quandary
The Secret Island of Dr. Quandary is an educational computer puzzle game developed by MECC, which pits the player against a variety of mathematical and logical puzzles. It was released in 1992 for MS-DOS and Macintosh. Story The player starts as a human playing in a shooting gallery in Dr. Quandary's carnival, and is given a live-action figure when the shooting game is defeated. However, it is a ruse for Dr. Quandary to put the player in the doll and transport them to his secret island, where the player must gather and brew the Fixer Elixir in order to escape. Puzzles There are a variety of puzzles in the game, most requiring some mathematical or logic skills, with some memory challenges thrown in as well. There are also varieties of traditional puzzles, such as Tangrams, Bulls and Cows, Taxman, the Tower of Hanoi, and Nim. Beating each puzzle nets the player an ingredient for the Fixer Elixir, the recipe of which can be a puzzle in itself for the harder difficulty levels.
https://en.wikipedia.org/wiki/Nun%20Study
The Nun Study of Aging and Alzheimer's Disease is a continuing longitudinal study, begun in 1986, to examine the onset of Alzheimer's disease. David Snowdon, an Epidemiologist and the founding Nun Study investigator, started the Nun Study at the University of Minnesota, later transferring the study to the University of Kentucky in 1990. In 2008, with Snowdon's retirement, the study returned to the University of Minnesota. The Nun Study was very briefly moved from the University of Minnesota to Northwestern University in 2021 under the directorship of Dr. Margaret Flanagan. The Nun Study is currently housed at the University of Texas Health San Antonio in the Bigg's Institute for Alzheimer's and Neurodegenerative diseases under the continued directorship of Neuropathologist, Dr. Margaret Flanagan. The Sisters' autobiographies written just before they took their vows (ages 19–21) revealed that positivity was closely related to longevity and idea density, which is related to conversation and writing. This research found that higher idea density scores correlated with a higher chance of having sufficient mental capacity in late-life, despite neurological evidence that showed the onset of Alzheimer's disease. In 1992, researchers at Rush University Medical Center Rush Alzheimer's Disease Center (RADC), building on the success of the Nun Study, proposed the Rush Religious Orders Study. The Religious Orders Study was funded by the National Institute on Aging in 1993, and was ongoing, as of 2012. Origin and procedure The Nun Study began in 1986 with funding by the National Institute on Aging. This study was focused on a group of 678 American Roman Catholic Sisters who were members of the School Sisters of Notre Dame. The purpose of the study was to conclude if activities, academics, past experiences, and disposition are correlated to continued cognitive, neurological, and physical ability as individuals got older, as well as overall longevity. The Nun Study participants
https://en.wikipedia.org/wiki/Benzoic%20acid
Benzoic acid is a white (or colorless) solid organic compound with the formula , whose structure consists of a benzene ring () with a carboxyl () substituent. The benzoyl group is often abbreviated "Bz" (not to be confused with "Bn" which is used for benzyl), thus benzoic acid is also denoted as BzOH, since the benzoyl group has the formula –. It is the simplest aromatic carboxylic acid. The name is derived from gum benzoin, which was for a long time its only source. Benzoic acid occurs naturally in many plants and serves as an intermediate in the biosynthesis of many secondary metabolites. Salts of benzoic acid are used as food preservatives. Benzoic acid is an important precursor for the industrial synthesis of many other organic substances. The salts and esters of benzoic acid are known as benzoates . History Benzoic acid was discovered in the sixteenth century. The dry distillation of gum benzoin was first described by Nostradamus (1556), and then by Alexius Pedemontanus (1560) and Blaise de Vigenère (1596). Justus von Liebig and Friedrich Wöhler determined the composition of benzoic acid. These latter also investigated how hippuric acid is related to benzoic acid. In 1875 Salkowski discovered the antifungal properties of benzoic acid, which was used for a long time in the preservation of benzoate-containing cloudberry fruits. Production Industrial preparations Benzoic acid is produced commercially by partial oxidation of toluene with oxygen. The process is catalyzed by cobalt or manganese naphthenates. The process uses abundant materials, and proceeds in high yield. The first industrial process involved the reaction of benzotrichloride (trichloromethyl benzene) with calcium hydroxide in water, using iron or iron salts as catalyst. The resulting calcium benzoate is converted to benzoic acid with hydrochloric acid. The product contains significant amounts of chlorinated benzoic acid derivatives. For this reason, benzoic acid for human consumption was
https://en.wikipedia.org/wiki/Sonian%20Forest
The Sonian Forest or Sonian Wood (, , ) is a forest at the south-eastern edge of Brussels, Belgium. It is connected to the Bois de la Cambre/Ter Kamerenbos, an urban public park which enters the city up to from the city centre. The forest lies in the Flemish municipalities of Sint-Genesius-Rode, Hoeilaart, Overijse, and Tervuren, in the Brussels-Capital Region municipalities of Uccle, Watermael-Boitsfort, Auderghem, and Woluwe-Saint-Pierre, and in the Walloon towns of La Hulpe and Waterloo. Thus, it stretches out over the three Belgian Regions. It is maintained by Flanders (56%), Brussels (38%), and Wallonia (6%). There are some contiguous tracts of privately held forest and the Kapucijnenbos, the "Capuchin Wood", which belongs to the Royal Trust. As of 2017, parts of the Sonian Forest have been inscribed as a UNESCO World Heritage Site, the only Belgian component to the multinational inscription 'Primeval Beech Forests of the Carpathians and Other Regions of Europe', because of their undisturbed nature and testimony to the ecological processes governing forests in Europe since the Last Glacial Period. History The forest is part of the scattered remains of the ancient Silva Carbonaria or Charcoal Forest. Originally, it was part of the Forest of Ardennes, the Romans' Arduenna Silva. The first mention of the Sonian Forest (Soniaca Silva) dates from the early Middle Ages. Then the forest south of Brussels was crossed by the river Senne/Zenne and extended as far as Hainaut, covering most of the high ground between the Senne and the Dyle/Dijl. The 9th-century vita of Saint Foillan mentions "the forest, next to the abbey of Saint Gertrude, called the Sonesian". In the Middle Ages, the forest extended over the southern part of Brabant up to the walls of Brussels and is mentioned, under the name of Ardennes, in Byron's Childe Harold. In the 16th century, it was still seven leagues in circumference, and even at the time of the French Revolution, it was very extensive.
https://en.wikipedia.org/wiki/Nervous%20system%20network%20models
Network of human nervous system comprises nodes (for example, neurons) that are connected by links (for example, synapses). The connectivity may be viewed anatomically, functionally, or electrophysiologically. These are presented in several Wikipedia articles that include Connectionism (a.k.a. Parallel Distributed Processing (PDP)), Biological neural network, Artificial neural network (a.k.a. Neural network), Computational neuroscience, as well as in several books by Ascoli, G. A. (2002), Sterratt, D., Graham, B., Gillies, A., & Willshaw, D. (2011), Gerstner, W., & Kistler, W. (2002), and Rumelhart, J. L., McClelland, J. L., and PDP Research Group (1986) among others. The focus of this article is a comprehensive view of modeling a neural network (technically neuronal network based on neuron model). Once an approach based on the perspective and connectivity is chosen, the models are developed at microscopic (ion and neuron), mesoscopic (functional or population), or macroscopic (system) levels. Computational modeling refers to models that are developed using computing tools. Introduction The nervous system consists networks made up of neurons and synapses connected to and controlling tissues as well as impacting human thoughts and behavior. In modeling neural networks of the nervous system one has to consider many factors. The brain and the neural network should be considered as an integrated and self-contained firmware system that includes hardware (organs), software (programs), memory (short term and long term), database (centralized and distributed), and a complex network of active elements (such as neurons, synapses, and tissues) and passive elements (such as parts of visual and auditory system) that carry information within and in-and-out of the body. Why does one want to model the brain and neural network? Although highly sophisticated computer systems have been developed and used in all walks of life, they are nowhere close to the human system in hardware an
https://en.wikipedia.org/wiki/W-box
The W box is a deoxyribonucleic acid (DNA) cis-regulatory element sequence, (T)TGAC(C/T), which is recognized by the family of WRKY transcription factors. Functionality and conservation of the W-box element across plant species has been shown by gel shift experiments, random binding site selection, yeast one-hybrid screens and co-transfection assays performed with many different WRKY proteins. In silico-based studies together with functional studies of plant promoters have identified clusters of W-boxes in stress-inducible promoters. The binding of WRKY proteins to W-boxes is a feature of both biotic and abiotic stress responses, together with other plant processes such as germination. It has also been shown that multiple W-boxes have a synergistic effect on transcription. Almost all WRKY transcription factors bind preferentially to W-boxes, and since their discovery, this has raised the question as to how they show specificity for the promoters of their target genes. Ciolkowski et al. (2008) showed that although the W-box core is required, adjacent sequences also play a role in determining binding-site preference. Recent evidence suggests that the TGAC core is more degenerate, composed of a guanine adenine cytosine (GAC) core, and the upstream thymine and downstream pyrimidine flanking sequences help dictate recognition by specific WRKY factors. Basic residues of the WRKY protein domain also are believed to recognize the phosphate backbone of the cis-element. Recently, Yamasaki et al. have determined the solution structure of the C-terminal WRKY domain of Arabidopsis WRKY4 in complex with the W-box DNA by NMR. They found that a four-stranded β-sheet enters the major groove of DNA in a structure they called the β-wedge, where the sheet is nearly perpendicular to the DNA helical ais. As predicted amino acids in the conserved WRKYGQK signature motif contact the W-box DNA. External Links and Useful Resources WRKY Transcription Factor Family at The Arabidopsis
https://en.wikipedia.org/wiki/WorldBench
WorldBench is a Windows benchmark tool offered by PC World Labs since 2000. The PC World Test Center uses this same tool every month to test contenders for PC World's Top Desktop PCs and Top Notebook PCs charts. WorldBench is also used in conjunction with other tests to evaluate Top Graphics Boards, Top Hard Drives, and other product categories. The PC World Test Center is currently on WorldBench 6, which specializes in tests for the 32 and 64-bit platforms of the Windows Operating System from Windows XP on. Ongoing beta testing for a power-consumption analysis tool that will be incorporated into the next version of WorldBench, along with several other new additions, is currently taking place in the PCWorld Lab. WorldBench is no longer available to the public, and the test suite is now exclusively used by the PCWorld Lab in their benchmarks. How It Works WorldBench uses popular desktop applications to perform realistic tasks, making it easier to accurately gauge how fast a computer runs applications routinely used, rather than just listing abstract measures of individual components. The application-based tests tell exactly what is needed—how fast the computer runs real applications. WorldBench installs special versions of each application tested, enabling the automated test scripts to perform every test. It also ensures that the software configuration and version are identical for every computer evaluated. The setup program installs the user interface, testing scripts, and all documents necessary to complete the tasks. It also unpacks and installs each application as needed during testing. Once a test is completed, the program removes the unpacked application from the hard drive. For security reasons, none of the test applications can be installed from the WorldBench CD. When a test is started from WorldBench, the benchmark manager begins running the test script. This script is a series of commands—menu requests, keystrokes, and mouse clicks—that cause the app
https://en.wikipedia.org/wiki/De%20Correspondent
De Correspondent is a Dutch news website based in Amsterdam, Netherlands. It was launched on 30 September 2013 after raising more than in a crowdfunding campaign in eight days. The website distinguishes itself by rejecting the daily news cycle and focusing on in-depth and chronological coverage on a topical basis, led by individual correspondents who each focus on specific topics. Sometimes it publishes English versions of its articles. The concept and initial success of De Correspondent has inspired other projects elsewhere. A German website Krautreporter was founded in 2014 and adopted the same concept. An English-language news site, titled The Correspondent, launched on September 30, 2019. The site raised through a crowdfunding campaign in late 2018, boosted by prominent backers including Jay Rosen and Trevor Noah. However, it endured substantial criticism after it was announced that it would not open an office in the United States, as many backers had anticipated. On 10 December 2020, NiemanLab broke the news that The Correspondent would be closing down on 31 December 2020. History The project was co-founded by Dutch journalist Rob Wijnberg, creative director Harald Dunnink, CTO Sebastian Kersten, and publisher Ernst-Jan Pfauth. Wijnberg, former editor-in-chief of the Dutch newspaper NRC Next, proposed the crowdfunding idea for an ad-free news media platform on national television in March 2013. Eight days later, he and his team reached their goal of 15,000 subscribers all paying €60 for a one-year membership. Wijnberg worked with digital creative agency Momkai and its owners, Harald Dunnink and Sebastian Kersten, served as creative director and CTO respectively. Ernst-Jan Pfauth, who had been the founding editor of The Next Web and head of digital at Dutch newspaper NRC Handelsblad, joined as a publisher. The website went live in September 2013. By January 2015 the website had more than 45,000 paying subscribers. In January 2016 the number of paying subsc
https://en.wikipedia.org/wiki/Object%20co-segmentation
In computer vision, object co-segmentation is a special case of image segmentation, which is defined as jointly segmenting semantically similar objects in multiple images or video frames. Challenges It is often challenging to extract segmentation masks of a target/object from a noisy collection of images or video frames, which involves object discovery coupled with segmentation. A noisy collection implies that the object/target is present sporadically in a set of images or the object/target disappears intermittently throughout the video of interest. Early methods typically involve mid-level representations such as object proposals. Dynamic Markov networks-based methods A joint object discover and co-segmentation method based on coupled dynamic Markov networks has been proposed recently, which claims significant improvements in robustness against irrelevant/noisy video frames. Unlike previous efforts which conveniently assumes the consistent presence of the target objects throughout the input video, this coupled dual dynamic Markov network based algorithm simultaneously carries out both the detection and segmentation tasks with two respective Markov networks jointly updated via belief propagation. Specifically, the Markov network responsible for segmentation is initialized with superpixels and provides information for its Markov counterpart responsible for the object detection task. Conversely, the Markov network responsible for detection builds the object proposal graph with inputs including the spatio-temporal segmentation tubes. Graph cut-based methods Graph cut optimization is a popular tool in computer vision, especially in earlier image segmentation applications. As an extension of regular graph cuts, multi-level hypergraph cut is proposed to account for more complex high order correspondences among video groups beyond typical pairwise correlations. With such hypergraph extension, multiple modalities of correspondences, including low-level appea
https://en.wikipedia.org/wiki/65537-gon
In geometry, a 65537-gon is a polygon with 65,537 (216 + 1) sides. The sum of the interior angles of any non–self-intersecting is 11796300°. Regular 65537-gon The area of a regular is (with ) A whole regular is not visually discernible from a circle, and its perimeter differs from that of the circumscribed circle by about 15 parts per billion. Construction The regular 65537-gon (one with all sides equal and all angles equal) is of interest for being a constructible polygon: that is, it can be constructed using a compass and an unmarked straightedge. This is because 65,537 is a Fermat prime, being of the form 22n + 1 (in this case n = 4). Thus, the values and are 32768-degree algebraic numbers, and like any constructible numbers, they can be written in terms of square roots and no higher-order roots. Although it was known to Gauss by 1801 that the regular 65537-gon was constructible, the first explicit construction of a regular 65537-gon was given by Johann Gustav Hermes (1894). The construction is very complex; Hermes spent 10 years completing the 200-page manuscript. Another method involves the use of at most 1332 Carlyle circles, and the first stages of this method are pictured below. This method faces practical problems, as one of these Carlyle circles solves the quadratic equation x2 + x − 16384 = 0 (16384 being 214). Symmetry The regular 65537-gon has Dih65537 symmetry, order 131074. Since 65,537 is a prime number there is one subgroup with dihedral symmetry: Dih1, and 2 cyclic group symmetries: Z65537, and Z1. 65537-gram A 65537-gram is a 65,537-sided star polygon. As 65,537 is prime, there are 32,767 regular forms generated by Schläfli symbols {65537/n} for all integers 2 ≤ n ≤ 32768 as . See also Circle Equilateral triangle Pentagon Heptadecagon (17-sides) 257-gon
https://en.wikipedia.org/wiki/Normal%20cone%20%28functional%20analysis%29
In mathematics, specifically in order theory and functional analysis, if is a cone at the origin in a topological vector space such that and if is the neighborhood filter at the origin, then is called normal if where and where for any subset is the -saturatation of Normal cones play an important role in the theory of ordered topological vector spaces and topological vector lattices. Characterizations If is a cone in a TVS then for any subset let be the -saturated hull of and for any collection of subsets of let If is a cone in a TVS then is normal if where is the neighborhood filter at the origin. If is a collection of subsets of and if is a subset of then is a fundamental subfamily of if every is contained as a subset of some element of If is a family of subsets of a TVS then a cone in is called a -cone if is a fundamental subfamily of and is a strict -cone if is a fundamental subfamily of Let denote the family of all bounded subsets of If is a cone in a TVS (over the real or complex numbers), then the following are equivalent: is a normal cone. For every filter in if then <li> There exists a neighborhood base in such that implies and if is a vector space over the reals then we may add to this list: There exists a neighborhood base at the origin consisting of convex, balanced, -saturated sets. There exists a generating family of semi-norms on such that for all and and if is a locally convex space and if the dual cone of is denoted by then we may add to this list: For any equicontinuous subset there exists an equicontiuous such that The topology of is the topology of uniform convergence on the equicontinuous subsets of and if is an infrabarreled locally convex space and if is the family of all strongly bounded subsets of then we may add to this list: The topology of is the topology of uniform convergence on strongly bounded subsets of is a -cone in this means that the family
https://en.wikipedia.org/wiki/Graphs%20with%20few%20cliques
In graph theory, a class of graphs is said to have few cliques if every member of the class has a polynomial number of maximal cliques. Certain generally NP-hard computational problems are solvable in polynomial time on such classes of graphs, making graphs with few cliques of interest in computational graph theory, network analysis, and other branches of applied mathematics. Informally, a family of graphs has few cliques if the graphs do not have a large number of large clusters. Definition A clique of a graph is a complete subgraph, while a maximal clique is a clique that is not properly contained in another clique. One can regard a clique as a cluster of vertices, since they are by definition all connected to each other by an edge. The concept of clusters is ubiquitous in data analysis, such as on the analysis of social networks. For that reason, limiting the number of possible maximal cliques has computational ramifications for algorithms on graphs or networks. Formally, let be a class of graphs. If for every -vertex graph in , there exists a polynomial such that has maximal cliques, then is said to be a class of graphs with few cliques. Examples The Turán graph has an exponential number of maximal cliques. In particular, this graph has exactly maximal cliques when , which is asymptotically greater than any polynomial function. This graph is sometimes called the Moon-Moser graph, after Moon & Moser showed in 1965 that this graph has the largest number of maximal cliques among all graphs on vertices. So the class of Turán graphs does not have few cliques. A tree on vertices has as many maximal cliques as edges, since it contains no triangles by definition. Any tree has exactly edges, and therefore that number of maximal cliques. So the class of trees has few cliques. A chordal graph on vertices has at most maximal cliques, so chordal graphs have few cliques. Any planar graph on vertices has at most maximal cliques, so the class of planar
https://en.wikipedia.org/wiki/Immune%20reconstitution%20inflammatory%20syndrome
Immune reconstitution inflammatory syndrome (IRIS) is a condition seen in some cases of HIV/AIDS or immunosuppression, in which the immune system begins to recover, but then responds to a previously acquired opportunistic infection with an overwhelming inflammatory response that paradoxically makes the symptoms of infection worse. IRIS may also be referred to as immune reconstitution syndrome, immune reconstitution disease, immune recovery disease, and immune restoration disease. Systemic or local inflammatory responses may occur with improvement in immune function. While this inflammatory reaction is usually self-limited, there is risk of long-term symptoms and death, particularly when the central nervous system is involved. Management generally involves symptom control and treatment of the underlying infection. In severe cases of IRIS, corticosteroids are commonly used. Important exceptions to using corticosteroids include Cryptococcal meningitis and Kaposi’s sarcoma, as they have been associated with poorer outcomes. Mechanism There are two common IRIS scenarios. The first is the “unmasking” of an occult opportunistic infection. The second is the “paradoxical” symptomatic relapse of a prior infection despite microbiologic treatment success. Often in paradoxical IRIS, microbiologic cultures are sterile. In either scenario, there is hypothesized reconstitution of antigen-specific T cell-mediated immunity with activation of the immune system against persisting antigen, whether present as intact organisms, dead organisms, or debris. In HIV infection and immunosuppression The suppression of CD4 T cells by HIV (or by immunosuppressive drugs) causes a decrease in the body's normal response to certain infections. Not only does this make it more difficult to fight the infection, it may mean that a level of infection that would normally produce symptoms is instead undetected (subclinical infection). If the CD4 count rapidly increases (due to effective treatment of
https://en.wikipedia.org/wiki/Heart%20rate%20monitor
A heart rate monitor (HRM) is a personal monitoring device that allows one to measure/display heart rate in real time or record the heart rate for later study. It is largely used to gather heart rate data while performing various types of physical exercise. Measuring electrical heart information is referred to as electrocardiography (ECG or EKG). Medical heart rate monitoring used in hospitals is usually wired and usually multiple sensors are used. Portable medical units are referred to as a Holter monitor. Consumer heart rate monitors are designed for everyday use and do not use wires to connect. History Early models consisted of a monitoring box with a set of electrode leads which attached to the chest. The first wireless EKG heart rate monitor was invented in 1977 by Polar Electro as a training aid for the Finnish National Cross Country Ski team. As "intensity training" became a popular concept in athletic circles in the mid-80s, retail sales of wireless personal heart monitors started in 1983. Technologies Modern heart rate monitors commonly use one of two different methods to record heart signals (electrical and optical). Both types of signals can provide the same basic heart rate data, using fully automated algorithms to measure heart rate, such as the Pan-Tompkins algorithm. ECG (Electrocardiography) sensors measure the bio-potential generated by electrical signals that control the expansion and contraction of heart chambers, typically implemented in medical devices. PPG (Photoplethysmography) sensors use a light-based technology to measure the blood volume controlled by the heart's pumping action. Electrical The electrical monitors consist of two elements: a monitor/transmitter, which is worn on a chest strap, and a receiver. When a heartbeat is detected, a radio signal is transmitted, which the receiver uses to display/determine the current heart rate. This signal can be a simple radio pulse or a unique coded signal from the chest strap (such as Blu
https://en.wikipedia.org/wiki/Petrous%20part%20of%20the%20temporal%20bone
The petrous part of the temporal bone is pyramid-shaped and is wedged in at the base of the skull between the sphenoid and occipital bones. Directed medially, forward, and a little upward, it presents a base, an apex, three surfaces, and three angles, and houses in its interior, the components of the inner ear. The petrous portion is among the most basal elements of the skull and forms part of the endocranium. Petrous comes from the Latin word petrosus, meaning "stone-like, hard". It is one of the densest bones in the body. In other mammals, it is a separate bone, the petrosal bone. The petrous bone is important for studies of ancient DNA from skeletal remains, as it tends to contain extremely well-preserved DNA. Base The base is fused with the internal surfaces of the squamous and mastoid parts. Apex The apex, which is rough and uneven, is received into the angular interval between the posterior border of the great wing of the sphenoid bone and the basilar part of the occipital bone; it presents the anterior or internal opening of the carotid canal, and forms the postero-lateral boundary of the foramen lacerum. Surfaces Anterior surface The anterior surface forms the posterior part of the middle cranial fossa of the base of the skull, and is continuous with the inner surface of the squamous portion, to which it is united by the petrosquamous suture, remains of which are distinct even at a late period of life. It is marked by depressions for the convolutions of the brain, and presents six notable points: near the center, the arcuate eminence (eminentia arcuata), which indicates the location of the superior semicircular canal. in front of and a little lateral to this eminence, a depression indicating the position of the tympanic cavity: Here, the layer of bone that separates the tympanic from the cranial cavity is extremely thin, and is known as the tegmen tympani a shallow groove, sometimes double, leading lateralward and backward to an oblique opening, th
https://en.wikipedia.org/wiki/PCI%20configuration%20space
PCI configuration space is the underlying way that the Conventional PCI, PCI-X and PCI Express perform auto configuration of the cards inserted into their bus. Overview PCI devices have a set of registers referred to as configuration space and PCI Express introduces extended configuration space for devices. Configuration space registers are mapped to memory locations. Device drivers and diagnostic software must have access to the configuration space, and operating systems typically use APIs to allow access to device configuration space. When the operating system does not have access methods defined or APIs for memory mapped configuration space requests, the driver or diagnostic software has the burden to access the configuration space in a manner that is compatible with the operating system's underlying access rules. In all systems, device drivers are encouraged to use APIs provided by the operating system to access the configuration space of the device. Technical information One of the major improvements the PCI Local Bus had over other I/O architectures was its configuration mechanism. In addition to the normal memory-mapped and I/O port spaces, each device function on the bus has a configuration space, which is 256 bytes long, addressable by knowing the eight-bit PCI bus, five-bit device, and three-bit function numbers for the device (commonly referred to as the BDF or B/D/F, as abbreviated from bus/device/function). This allows up to 256 buses, each with up to 32 devices, each supporting eight functions. A single PCI expansion card can respond as a device and must implement at least function number zero. The first 64 bytes of configuration space are standardized; the remainder are available for vendor-defined purposes. Some high-end computers support more than one PCI domain (or PCI segment); each PCI domain supports up to 256 buses. In order to allow more parts of configuration space to be standardized without conflicting with existing uses, there can be
https://en.wikipedia.org/wiki/Cambridge%20Analytica
Cambridge Analytica Ltd (CA), previously known as SCL USA, was a British political consulting firm that came to prominence through the Facebook–Cambridge Analytica data scandal. It was started in 2013, as a subsidiary of the private intelligence company and self-described "global election management agency" SCL Group by long-time SCL executives Nigel Oakes, Alexander Nix and Alexander Oakes, with Nix as CEO. The well-connected founders had contact with, among others, the British Conservative Party, royal family, and military. The firm maintained offices in London, New York City, and Washington, D.C. The company closed operations in 2018 in the course of the Facebook–Cambridge Analytica data scandal, although firms related to both Cambridge Analytica and its parent firm SCL still exist. History Cambridge Analytica was established as a subsidiary of the private intelligence company SCL Group that was active in military and political arenas. The men who ran Cambridge Analytica and its parent SCL were described as having close ties to the Conservative Party, royalty, and the British military. Cambridge Analytica (SCL USA) was incorporated in January 2013 with its registered office in Westferry Circus, London and just one staff member, its director and CEO Alexander James Ashburner Nix (also appointed in January 2015). Nix was also the director of nine similar companies sharing the same registered offices in London, including Firecrest technologies, Emerdata and six SCL Group companies including "SCL elections limited". Nigel Oakes, known as the former boyfriend of Lady Helen Windsor, had founded the predecessor SCL Group in the 1990s, and in 2005 Oakes established SCL Group together with his brother Alexander Oakes and Alexander Nix; SCL Group was the parent company of Cambridge Analytica. former Conservative minister and MP Sir Geoffrey Pattie was the founding chairman of SCL; Lord Ivar Mountbatten also joined Oakes as a director of the company. As a result of the Fac
https://en.wikipedia.org/wiki/Physics%20World
Physics World is the membership magazine of the Institute of Physics, one of the largest physical societies in the world. It is an international monthly magazine covering all areas of physics, pure and applied, and is aimed at physicists in research, industry, physics outreach, and education worldwide. Overview The magazine was launched in 1988 by IOP Publishing Ltd, under the founding editorship of Philip Campbell. The magazine is sent free to members of the Institute of Physics, who can access a digital edition of the magazine; selected articles can be read by anyone for free online. It was redesigned in September 2005 and has an audited circulation of just under 35000. The current editor is Matin Durrani. Others on the team are Michael Banks (news editor) and Tushna Commissariat and Sarah Teah (features editors). Hamish Johnston, Margaret Harris and Tami Freeman are online editors. Alongside the print and online magazine, Physics World produces films and two podcasts. The Physics World Stories podcast is hosted by Andrew Glester and is produced monthly. The Physics World Weekly podcast is hosted by James Dacey. Breakthrough of the Year The magazine makes two awards each year. These are the Physics World Breakthrough of the Year and the Physics World Book of the Year, which have both been awarded annually since 2009. Top 10 works and winners of the Breakthrough of the Year 2009: "to August Jonathan Home and colleagues at NIST for unveiled the first small-scale device that could be described as a complete "quantum computer" Top results from Tevatron Spins spotted in room-temperature silicon Graphane makes its debut Magnetic monopoles spotted in spin ices Water on the Moon Atoms teleport information over long distance Black-hole analogue traps sound Dark matter spotted in Minnesota A 2.36 TeV big bang at the LHC 2010: "to ALPHA and the ASACUSA group at CERN for have created new ways of controlling antiatoms of hydrogen" Exoplanet atmosphere laid bar
https://en.wikipedia.org/wiki/Disease%20Models%20%26%20Mechanisms
Disease Models & Mechanisms (DMM) is a monthly peer-reviewed Open Access biomedical journal published by The Company of Biologists that launched in 2008. DMM is partnered with Publons, is part of the Review Commons initiative and has two-way integration with bioRxiv. Scope and content DMM publishes original research, resources and reviews that focus on the use of model systems to better understand, diagnose and treat human disease. Model systems of interest include: Vertebrates such as mice, zebrafish, frogs, rats and other mammals Invertebrates such as Drosophila melanogaster and Caenorhabditis elegans Unique in vitro or ex vivo models, such as stem-cell-based models, organoids and systems based on patient material Microorganisms such as yeast and Dictyostelium Other biological systems with relevance to human disease research Disease areas of interest include: Cancer Neurodegenerative and neurological diseases Psychiatric disorders Metabolic disorders, including diabetes and obesity Cardiovascular diseases, stroke and hypertension Gastrointestinal diseases Infectious diseases Autoimmunity and inflammation Developmental diseases Musculoskeletal disorders Renal or liver disease Eye disorders Drug and biomarker discovery/screening Stem cell therapies in regenerative medicine The journal operates on a continuous publication model. The final version of record is immediately released online as soon as it is ready. All papers are published as Open Access articles under the CC-BY licence. Abstracting and indexing The journal is abstracted and/or indexed by: BIOBASE CAB abstracts Cambridge Scientific Abstracts Clarivate Analytics Web of Science EMBASE Medline Scopus It is a member of OASPA (Open Access Scholarly Publishers Association) and is indexed in the DOAJ (Directory of Open Access Journals). Disease Models & Mechanisms is a signatory of the San Francisco Declaration on Research Assessment (DORA). Management The founding editor
https://en.wikipedia.org/wiki/Deletion%20mapping
In genetics and especially genetic engineering, deletion mapping is a technique used to find out the mutation sites within a gene. The principle of deletion mapping involves crossing a strain which has a point mutation in a gene, with multiple strains who each carry a deletion in a different region of the same gene. Wherever recombination occurs between the two strains to produce a wild-type (+) gene (regardless of frequency), the point mutation cannot lie within the region of the deletion. If recombination cannot produce any wild-type genes, then it is reasonable to conclude that the point mutation and deletion are found within the same stretch of DNA. This example should demonstrate how the principle works: Suppose you have a gene X, which in wild-type (+) form can be shown linearly like so: 5'-------------------------------------------------------------------------------------------------------------------3' gene X, + Suppose a strain of organisms has a point mutation in the gene (now called gene X, – to denote that it is no longer wild-type): 5'----------------------------------------------------X--------------------------------------------------------------3' gene X, – Now suppose you have two strains of organisms, each with deletions in gene X at different sites, called del-1 and del-2, respectively (the dotted line indicates the site of deletion): 5'-------------------(..............................................)------------------------------------------------3' del-1 5'------------------------------------------------------------(............................................)---------3' del-2 Because the point mutation lies within the deletion of del-1, there will be no wild-type (+) recombinants between the point mutant and the del-1 mutant. However, in a cross between the point mutant and the del-2 mutant, there could be a successful wild-type (+) recombinant produced. In genetic recombination, if a mutant allele in the donor is within the seq
https://en.wikipedia.org/wiki/39th%20meridian%20east
The meridian 39° east of Greenwich is a line of longitude that extends from the North Pole across the Arctic Ocean, Europe, Asia, Africa, the Indian Ocean, the Southern Ocean, and Antarctica to the South Pole. The 39th meridian east forms a great circle with the 141st meridian west. From Pole to Pole Starting at the North Pole and heading south to the South Pole, the 39th meridian east passes through: {| class="wikitable plainrowheaders" ! scope="col" width="125" | Co-ordinates ! scope="col" | Country, territory or sea ! scope="col" | Notes |- | style="background:#b0e0e6;" | ! scope="row" style="background:#b0e0e6;" | Arctic Ocean | style="background:#b0e0e6;" | |- | style="background:#b0e0e6;" | ! scope="row" style="background:#b0e0e6;" | Barents Sea | style="background:#b0e0e6;" | |- | ! scope="row" | | Kola Peninsula |- | style="background:#b0e0e6;" | ! scope="row" style="background:#b0e0e6;" | White Sea | style="background:#b0e0e6;" | |- | ! scope="row" | | |- | ! scope="row" | | |- | ! scope="row" | | |- | style="background:#b0e0e6;" | ! scope="row" style="background:#b0e0e6;" | Sea of Azov | style="background:#b0e0e6;" | Taganrog Bay |- | ! scope="row" | | Passing just east of Krasnodar (at ) |- | style="background:#b0e0e6;" | ! scope="row" style="background:#b0e0e6;" | Black Sea | style="background:#b0e0e6;" | |- | ! scope="row" | | |- | ! scope="row" | | |- | ! scope="row" | | |- | ! scope="row" | | |- | ! scope="row" | | |- | style="background:#b0e0e6;" | ! scope="row" style="background:#b0e0e6;" | Red Sea | style="background:#b0e0e6;" | |- | ! scope="row" | | |- | style="background:#b0e0e6;" | ! scope="row" style="background:#b0e0e6;" | Red Sea | style="background:#b0e0e6;" | |- | ! scope="row" | | Passing just east of Asmara |- | ! scope="row" | | |- | ! scope="row" | | |- | ! scope="row" | | |-valign="top" | style="background:#b0e0e6;" | ! scope="row" style="background:#b0e0e6;" | Zanzibar Channel | style="backgro
https://en.wikipedia.org/wiki/Skewed%20X-inactivation
Skewed X-chromosome inactivation (skewed X-inactivation) occurs when the X-inactivation of one X chromosome is favored over the other, leading to an uneven number of cells with each chromosome inactivated. It is usually defined as one allele being found on the active X chromosome in over 75% of cells, and extreme skewing is when over 90% of cells have inactivated the same X chromosome. It can be caused by primary nonrandom inactivation, either by chance due to a small cell pool or directed by genes, or by secondary nonrandom inactivation, which occurs by selection. X-chromosome inactivation occurs in females to provide dosage compensation between the sexes. If females kept both X chromosomes active, they would have twice the number of active X genes than males, who only have one copy of the X chromosome. At approximately the time of embryonic implantation, one of the two X chromosomes in each cell of the female embryo is randomly selected for inactivation. Cells then undergo transcriptional and epigenetic changes to ensure this inactivation is permanent (such as methylation and being modified into Barr bodies). All progeny from these initial cells will maintain the inactivation of the same chromosome, resulting in a phenotypic mosaic pattern of cells in females although not a genotypic mosaic. Most females will have some levels of skewing. It is relatively common in adult females; around 35% of women have a skewed ratio over 70:30, and 7% of women have an extreme skewed ratio of over 90:10. This is of medical significance, due to the potential for the expression of disease genes present on the X chromosome that are normally not expressed due to random X-inactivation. Causes Primary nonrandom inactivation Nonrandom X-inactivation leads to skewed X-inactivation. Nonrandom X-inactivation can be caused by chance or directed by genes. If the initial pool of cells in which X-inactivation occurs is small, chance can cause skewing to occur in some individuals by causin
https://en.wikipedia.org/wiki/Bailey%20peptide%20synthesis
The Bailey peptide synthesis is a name reaction in organic chemistry developed 1949 by J. L. Bailey. It is a method for the synthesis of a peptide from α-amino acid-N-carboxylic acid anhydrides (NCAs) and amino acids or peptide esters. The reaction is characterized by short reaction times and a high yield of the target peptide. The reaction can be carried out at low temperatures in organic solvents. The residues R1 and R2 can be organic groups or hydrogen atoms, R3 is the used amino acid or peptide ester: Reaction mechanism The reaction mechanism is not known in detail. Supposedly, the reaction begins with a nucleophilic attack of the amino group on the carbonyl carbon of the anhydride group of the N-carboxylic acid anhydride (1). After an intramolecular proton migration, a 1,4-proton shift and the cleavage of carbon dioxide follows, resulting in the peptide bond in the final product (2): Atom economy The advantage in atom economy of using NCAs for peptide formation is that there is no need for a protecting group on the functional group reacted with the amino acid. For example, the Merrifield synthesis depends on the use of Boc and Bzl protecting groups, which need be removed after the reaction. In the case of Bailey peptide synthesis, the free peptide is directly obtained after the reaction. However, unwanted and difficult to remove by-products may be formed. An N-substitution of the NCA (for example, by an o-nitrophenylsulfenyl group) can simplify the subsequent purification process, but on the other hand deteriorates the atom economy of the reaction. The synthesis of NCAs can be carried out by the Leuchs reaction or by the reaction of N-(benzyloxycarbonyl)-amino acids with oxalyl chloride. In the latter case, again the procedure is less efficient in the sense of atom economy. Synthesized peptides The following peptides were synthesized using this method by 1949: DL-Ala-Gly L-Tyr-Gly DL-Tyr-Tyr DL-Ala-DL-Ala-Gly DL-di-Ala-L-cystinyl-di-Gly DL-Ala-L-
https://en.wikipedia.org/wiki/Holmes%20%28computer%29
Holmes is a cognitive computing system developed by the Indian technology corporation Wipro and announced in 2016. Its name is a reference to IBM's Watson, and is a backronym for "Heuristics and Ontology-based Learning Machines and Experiential Systems". Its uses include development of digital virtual agents, predictive systems, cognitive process automation, visual computing applications, knowledge virtualization, robotics and drones. The HOLMES platform Vision was created by Ramprasad K.R. (Rampi), he was the chief technologist for AI at Wipro.
https://en.wikipedia.org/wiki/Oxalis%20tuberosa
Oxalis tuberosa is a perennial herbaceous plant that overwinters as underground stem tubers. These tubers are known as uqa in Quechua, oca in Spanish, yams in New Zealand and a number of other alternative names. The plant was brought into cultivation in the central and southern Andes for its tubers, which are used as a root vegetable. The plant is not known in the wild, but populations of wild Oxalis species that bear smaller tubers are known from four areas of the central Andean region. Oca was introduced to Europe in 1830 as a competitor to the potato, and to New Zealand as early as 1860. In New Zealand, oca has become a popular table vegetable and are simply called yams (although not a true yam). It is available in a range of colours, including yellow, orange, pink, apricot, and the traditional red. Cultural significance Grown primarily by Quechua and Aymara farmers, oca has been a staple of rural Andean diets for centuries. Of all Andean root and tuber crops, oca is currently second only to potato in area planted within the Central Andean region. Oca is important to local food security because of its role in crop rotations and its high nutritional content. Diversity Andean farmers, including the indigenous Quechua and Aymara people, cultivate numerous varieties of oca. Oca diversity may be described with respect to morphological characters, local cultivar names, or molecular markers. Morphological characters Oca morphotypes are distinguished by foliar, floral, fruit, stem, and tuber characteristics, as described in the International Plant Genetic Resources Institute's document on oca descriptors. The morphological diversity of oca tubers, in particular, is astounding. Tubers range from 25 to 150 mm in length by 25 mm in width; skin and flesh color may be white, cream, yellow, orange, pink, red, and/or purple and distributed in range of patterns. Local cultivar names Oca-growing communities often name varieties based primarily on tuber morphology and second
https://en.wikipedia.org/wiki/Time%20displacement
Time displacement in sociology refers to the idea that new forms of activities may replace older ones. New activities that cause time displacement are usually technology-based, most common are the information and communication technologies such as Internet and television. Those technologies are seen as responsible for declines of previously more common activities such as in- and out-of-home socializing, work, and even personal care and sleep. For example, Internet users may spend time online using it as a substitute of other activities that served similar function(s) (watching television, reading printed media, face to face interaction, etc.). Internet is not the first technology to result in time displacement. Earlier, television had a similar impact, as it shifted people's time from activities such as listening to radio, going to movie theaters or, talking in house, or spending time outside it. See also Parkinson's law
https://en.wikipedia.org/wiki/Joint%20spectral%20radius
In mathematics, the joint spectral radius is a generalization of the classical notion of spectral radius of a matrix, to sets of matrices. In recent years this notion has found applications in a large number of engineering fields and is still a topic of active research. General description The joint spectral radius of a set of matrices is the maximal asymptotic growth rate of products of matrices taken in that set. For a finite (or more generally compact) set of matrices the joint spectral radius is defined as follows: It can be proved that the limit exists and that the quantity actually does not depend on the chosen matrix norm (this is true for any norm but particularly easy to see if the norm is sub-multiplicative). The joint spectral radius was introduced in 1960 by Gian-Carlo Rota and Gilbert Strang, two mathematicians from MIT, but started attracting attention with the work of Ingrid Daubechies and Jeffrey Lagarias. They showed that the joint spectral radius can be used to describe smoothness properties of certain wavelet functions. A wide number of applications have been proposed since then. It is known that the joint spectral radius quantity is NP-hard to compute or to approximate, even when the set consists of only two matrices with all nonzero entries of the two matrices which are constrained to be equal. Moreover, the question "" is an undecidable problem. Nevertheless, in recent years much progress has been done on its understanding, and it appears that in practice the joint spectral radius can often be computed to satisfactory precision, and that it moreover can bring interesting insight in engineering and mathematical problems. Computation Approximation algorithms In spite of the negative theoretical results on the joint spectral radius computability, methods have been proposed that perform well in practice. Algorithms are even known, which can reach an arbitrary accuracy in an a priori computable amount of time. These algorithms c
https://en.wikipedia.org/wiki/Q-function
In statistics, the Q-function is the tail distribution function of the standard normal distribution. In other words, is the probability that a normal (Gaussian) random variable will obtain a value larger than standard deviations. Equivalently, is the probability that a standard normal random variable takes a value larger than . If is a Gaussian random variable with mean and variance , then is standard normal and where . Other definitions of the Q-function, all of which are simple transformations of the normal cumulative distribution function, are also used occasionally. Because of its relation to the cumulative distribution function of the normal distribution, the Q-function can also be expressed in terms of the error function, which is an important function in applied mathematics and physics. Definition and basic properties Formally, the Q-function is defined as Thus, where is the cumulative distribution function of the standard normal Gaussian distribution. The Q-function can be expressed in terms of the error function, or the complementary error function, as An alternative form of the Q-function known as Craig's formula, after its discoverer, is expressed as: This expression is valid only for positive values of x, but it can be used in conjunction with Q(x) = 1 − Q(−x) to obtain Q(x) for negative values. This form is advantageous in that the range of integration is fixed and finite. Craig's formula was later extended by Behnad (2020) for the Q-function of the sum of two non-negative variables, as follows: Bounds and approximations The Q-function is not an elementary function. However, it can be upper and lower bounded as, where is the density function of the standard normal distribution, and the bounds become increasingly tight for large x. Using the substitution v =u2/2, the upper bound is derived as follows: Similarly, using and the quotient rule, Solving for Q(x) provides the lower bound. The geometric mean of the upper and lower b
https://en.wikipedia.org/wiki/Jacobson%27s%20conjecture
In abstract algebra, Jacobson's conjecture is an open problem in ring theory concerning the intersection of powers of the Jacobson radical of a Noetherian ring. It has only been proven for special types of Noetherian rings, so far. Examples exist to show that the conjecture can fail when the ring is not Noetherian on a side, so it is absolutely necessary for the ring to be two-sided Noetherian. The conjecture is named for the algebraist Nathan Jacobson who posed the first version of the conjecture. Statement For a ring R with Jacobson radical J, the nonnegative powers are defined by using the product of ideals. Jacobson's conjecture: In a right-and-left Noetherian ring, In other words: "The only element of a Noetherian ring in all powers of J is 0." The original conjecture posed by Jacobson in 1956 asked about noncommutative one-sided Noetherian rings, however Israel Nathan Herstein produced a counterexample in 1965, and soon afterwards, Arun Vinayak Jategaonkar produced a different example which was a left principal ideal domain. From that point on, the conjecture was reformulated to require two-sided Noetherian rings. Partial results Jacobson's conjecture has been verified for particular types of Noetherian rings: Commutative Noetherian rings all satisfy Jacobson's conjecture. This is a consequence of the Krull intersection theorem. Fully bounded Noetherian rings Noetherian rings with Krull dimension 1 Noetherian rings satisfying the second layer condition
https://en.wikipedia.org/wiki/B%26M%20Baked%20Beans%20factory
The B&M Baked Beans factory is an historic cannery building in Portland, Maine, USA. The building was constructed in 1913 in the East Deering neighborhood by the Burnham & Morrill Company. Baked beans were produced in the building until 2021. The building is a prominent landmark highly visible from Interstate 295. It is located at 1 Beanpot Circle. History The Burnham and Morrill Co. began in 1867 on Franklin Street in Portland as a canning factory for vegetables, meat, and fish. The company constructed the B&M Baked Beans factory in 1913. The factory first produced canned beans in 1927. It used a traditional open-pot baking process with iron cauldrons that weighed 200 pounds attached to a ceiling-mounted rail system. The cauldrons were filled with beans, molasses, and other ingredients on the top floor of the factory and baked for hours. The cooked beans were dumped into a chute and went down to the canning line. In 2021, owner B&G Foods Inc. announced that the B&M Baked Beans factory would close. In August of that year the company removed the factory's brick smoke stack. In 2023, the new owner Roux Institute began demolition of ancillary factory buildings to make way for the construction of the Roux Institute campus. The closure of the plant moved the production of the baked beans to the midwest. Baked beans & other products Products manufactured at the factory included multiple flavors of baked beans. The flavors were original, vegetarian, home style, country style, Boston style, bacon and onion, and maple flavor. The factory also made B&M brown bread, which comes in a can in plain and raisin flavors. B&M Baked Beans are made with pea beans. By late 2022, after production moved to the midwest, B&M customers were reporting that the baked beans were undercooked, crunchy, and tasted different. Customers speculated that the beans were no longer being baked. Some customers were hoarding B&M bean cans produced at the Portland plant. Roux Institute After its cl
https://en.wikipedia.org/wiki/Continuum%20limit
In mathematical physics and mathematics, the continuum limit or scaling limit of a lattice model refers to its behaviour in the limit as the lattice spacing goes to zero. It is often useful to use lattice models to approximate real-world processes, such as Brownian motion. Indeed, according to Donsker's theorem, the discrete random walk would, in the scaling limit, approach the true Brownian motion. Terminology The term continuum limit mostly finds use in the physical sciences, often in reference to models of aspects of quantum physics, while the term scaling limit is more common in mathematical use. Application in quantum field theory A lattice model that approximates a continuum quantum field theory in the limit as the lattice spacing goes to zero may correspond to finding a second order phase transition of the model. This is the scaling limit of the model. See also Universality classes
https://en.wikipedia.org/wiki/Virtuoso%20Universal%20Server
Virtuoso Universal Server is a middleware and database engine hybrid that combines the functionality of a traditional relational database management system (RDBMS), object–relational database (ORDBMS), virtual database, RDF, XML, free-text, web application server and file server functionality in a single system. Rather than have dedicated servers for each of the aforementioned functionality realms, Virtuoso is a "universal server"; it enables a single multithreaded server process that implements multiple protocols. The free and open source edition of Virtuoso Universal Server is also known as OpenLink Virtuoso. The software has been developed by OpenLink Software with Kingsley Uyi Idehen and Orri Erling as the chief software architects. Database structure Core database engine Virtuoso provides an extended object–relational model, which combines the flexibility of relational access with inheritance, run time data typing, late binding, and identity-based access. Virtuoso Universal Server database includes physical file and in memory storage and operating system processes that interact with the storage. There is one main process, which has listeners on a specified port for HTTP, SOAP, and other protocols. Architecture Virtuoso is designed to take advantage of operating system threading support and multiple CPUs. It consists of a single process with an adjustable pool of threads shared between clients. Multiple threads may work on a single index tree with minimal interference with each other. One cache of database pages is shared among all threads and old dirty pages are written back to disk as a background process. The database has at all times a clean checkpoint state and a delta of committed or uncommitted changes to this checkpointed state. This makes it possible to do a clean backup of the checkpoint state while transactions proceed on the commit state. A transaction log file records all transactions since the last checkpoint. Transaction log files may be pr
https://en.wikipedia.org/wiki/FCGR3B
FCGR3B (Fc fragment of IgG, low affinity IIIb, receptor), also known as CD16b (Cluster of Differentiation 16b), is a human gene. Clinical relevance Mutations and copy-number variations in this gene have been associated to clinical cases of glomerulonephritis. See also CD16 Cluster of differentiation
https://en.wikipedia.org/wiki/Quartic%20graph
In the mathematical field of graph theory, a quartic graph is a graph where all vertices have degree 4. In other words, a quartic graph is a 4-regular graph. Examples Several well-known graphs are quartic. They include: The complete graph K5, a quartic graph with 5 vertices, the smallest possible quartic graph. The Chvátal graph, another quartic graph with 12 vertices, the smallest quartic graph that both has no triangles and cannot be colored with three colors. The Folkman graph, a quartic graph with 20 vertices, the smallest semi-symmetric graph. The Meredith graph, a quartic graph with 70 vertices that is 4-connected but has no Hamiltonian cycle, disproving a conjecture of Crispin Nash-Williams. Every medial graph is a quartic plane graph, and every quartic plane graph is the medial graph of a pair of dual plane graphs or multigraphs. Knot diagrams and link diagrams are also quartic plane multigraphs, in which the vertices represent the crossings of the diagram and are marked with additional information concerning which of the two branches of the knot crosses the other branch at that point. Properties Because the degree of every vertex in a quartic graph is even, every connected quartic graph has an Euler tour. And as with regular bipartite graphs more generally, every bipartite quartic graph has a perfect matching. In this case, a much simpler and faster algorithm for finding such a matching is possible than for irregular graphs: by selecting every other edge of an Euler tour, one may find a 2-factor, which in this case must be a collection of cycles, each of even length, with each vertex of the graph appearing in exactly one cycle. By selecting every other edge again in these cycles, one obtains a perfect matching in linear time. The same method can also be used to color the edges of the graph with four colors in linear time. Quartic graphs have an even number of Hamiltonian decompositions. Open problems It is an open conjecture whether all quartic Hamil
https://en.wikipedia.org/wiki/Classical%20field%20theory
A classical field theory is a physical theory that predicts how one or more physical fields interact with matter through field equations, without considering effects of quantization; theories that incorporate quantum mechanics are called quantum field theories. In most contexts, 'classical field theory' is specifically intended to describe electromagnetism and gravitation, two of the fundamental forces of nature. A physical field can be thought of as the assignment of a physical quantity at each point of space and time. For example, in a weather forecast, the wind velocity during a day over a country is described by assigning a vector to each point in space. Each vector represents the direction of the movement of air at that point, so the set of all wind vectors in an area at a given point in time constitutes a vector field. As the day progresses, the directions in which the vectors point change as the directions of the wind change. The first field theories, Newtonian gravitation and Maxwell's equations of electromagnetic fields were developed in classical physics before the advent of relativity theory in 1905, and had to be revised to be consistent with that theory. Consequently, classical field theories are usually categorized as non-relativistic and relativistic. Modern field theories are usually expressed using the mathematics of tensor calculus. A more recent alternative mathematical formalism describes classical fields as sections of mathematical objects called fiber bundles. Non-relativistic field theories Some of the simplest physical fields are vector force fields. Historically, the first time that fields were taken seriously was with Faraday's lines of force when describing the electric field. The gravitational field was then similarly described. Newtonian gravitation The first field theory of gravity was Newton's theory of gravitation in which the mutual interaction between two masses obeys an inverse square law. This was very useful for predicting t
https://en.wikipedia.org/wiki/Photo%20response%20non-uniformity
Photo response non-uniformity, pixel response non-uniformity, or PRNU, is a form of fixed-pattern noise related to digital image sensors, as used in cameras and optical instruments. Both CCD and CMOS sensors are two-dimensional arrays of photosensitive cells, each broadly corresponding to an image pixel. Due to the non-uniformity of image sensors, each cell responds with a different voltage level when illuminated with a uniform light source, and this leads to luminance inaccuracy at the pixel level. High-end and metrology camera vendors tend to characterise this non-uniformity during instrument manufacture. The sensor is illuminated with a standardized light source and a two-dimensional table of correction factors is generated. This table is either carried in camera non-volatile memory and dynamically applied to the image on each capture, or ships with the camera to be applied by an external image processing and correcting pipeline. See also Color balance Color correction Flat-field correction Image sensor Digital photography Image sensors Optics
https://en.wikipedia.org/wiki/Nature%20printing
Nature printing is a printing process, developed in the 18th century, that uses the plants, animals, rocks and other natural subjects to produce an image. The subject undergoes several stages to give a direct impression onto materials such as lead, gum, and photographic plates, which are then used in the printing process. While some sources state that Benjamin Franklin invented nature printing from leaf casts, using a copper plate press, in 1737 to thwart counterfeiters of paper money bills, other sources also report Franklin's friend, Philadelphia naturalist Joseph Breintnall, to have made contact nature prints from leaves in about 1730. Together they sent nature prints which were printed directly from inked leaves to English naturalists. Another person attributed with the invention of the process, Naturselbstdruck, is Alois Auer; the first publication, of instructions for the process, was by this Austrian printer in The Discovery of the Natural Printing Process: an Invention ... Vienna, 1853. This was written in four languages by the author. He shows the use of plants, a fossil fish, and lace impressed by roller onto a lead plate, this is hand coloured and transferred to the final print. Many others botanical and natural history illustrations had attempted to use techniques that were a 'shorthand', or for a type of accuracy, in the representation of subjects. Another printer, the Englishman Henry Bradbury, immediately used Auer's 'nature printing' process to publish work of his own. These included two major botanical works; The Ferns of Great Britain and Ireland, Moore, Thomas.(1857) and the rendition of these species was readily adapted to the process; the two dimensional print would reveal form and detail for the identification of species. The Nature-Printed British Sea-Weeds, W.G. Johnstone & A. Croall. 1859-60. Sherman Denton in his book As Nature Shows Them: Moths and Butterflies ... used the wings of the species he was describing by pressing them i
https://en.wikipedia.org/wiki/Cipher%20security%20summary
This article summarizes publicly known attacks against block ciphers and stream ciphers. Note that there are perhaps attacks that are not publicly known, and not all entries may be up to date. Table color key Best attack This column lists the complexity of the attack: If the attack doesn't break the full cipher, "rounds" refers to how many rounds were broken "time" — time complexity, number of cipher evaluations for the attacker "data" — required known plaintext-ciphertext pairs (if applicable) "memory" — how many blocks worth of data needs to be stored (if applicable) "related keys" — for related-key attacks, how many related key queries are needed Common ciphers Key or plaintext recovery attacks Attacks that lead to disclosure of the key or plaintext. Distinguishing attacks Attacks that allow distinguishing ciphertext from random data. Less common ciphers Key recovery attacks Attacks that lead to disclosure of the key. Distinguishing attacks Attacks that allow distinguishing ciphertext from random data. See also Block cipher Hash function security summary Time/memory/data tradeoff attack Transport Layer Security Bullrun (decryption program) — a secret anti-encryption program run by the U.S. National Security Agency
https://en.wikipedia.org/wiki/Klawdziy%20Duzh-Dushewski
Kłaŭdzij Sciapanavič Duž-Dušeŭski (, , ; 27 March 1891 – 25 February 1959) was a Belarusian civil engineer, architect, diplomat and journalist. He is believed to be the creator of the Flag of Belarus in 1917. Life Duzh-Dushewski was born into an impoverished family of the Lithuanian nobility. In 1912 he graduated the Vilnius non-classical secondary school where he had joined the Belarusian national movement. In 1912–1918 Duzh-Dushewski studied in the Saint Petersburg Mining Institute. Being in St. Petersburg, he also took part in Belarusian social life. For instance, he was editor of the Belarusian magazine Ranica. In 1917 he joined the Belarusian Socialist Assembly. In 1918 he worked at the Refugee Assistance Department of the Belarusian National Committee in St. Petersburg. At that time, according to his own memories, he created the draft of the white-red-white flag of Belarus that was very quickly adopted by the Belarusian national movement throughout the European part of the former Russian Empire and later adopted as the state flag of the Belarusian Democratic Republic. In 1919 he moved to Vilnius. In autumn 1919 he was appointed diplomatic representative of the Belarusian Democratic Republic in the Baltic states. In early 1921 he was arrested by Polish authorities. After that he emigrated to Lithuania and settled in Kaunas. In the 1920s–1930s he worked for several ministries there. In 1927 he graduated from the University of Lithuania in Kaunas as Civil Engineer. He also edited several Belarusian newspapers in Lithuania. In 1940 he was arrested by the Soviets who had occupied Lithuania. In August 1943 he was arrested by the Nazis for hiding Jews. He was put into the Pravieniškės labor camp. On April 13, 2004, Duzh-Dushewski was posthumously awarded the Cross of Rescue of the Perishing by the President of Lithuania. This happened at the request of the Jewish community of Lithuania for his service to the community during the war. In 1944–1946 he worked at
https://en.wikipedia.org/wiki/Transient%20receptor%20potential%20calcium%20channel%20family
The transient receptor potential Ca2+ channel (TRP-CC) family (TC# 1.A.4) is a member of the voltage-gated ion channel (VIC) superfamily and consists of cation channels conserved from worms to humans. The TRP-CC family also consists of seven subfamilies (TRPC, TRPV, TRPM, TRPN, TRPA, TRPP, and TRPML) based on their amino acid sequence homology: the canonical or classic TRPs, the vanilloid receptor TRPs, the melastatin or long TRPs, ankyrin (whose only member is the transmembrane protein 1 [TRPA1]) TRPN after the nonmechanoreceptor potential C (nonpC), and the more distant cousins, the polycystins and mucolipins. A representative list of members belonging to the TRP-CC family can be found in the Transporter Classification Database. Function Members of the TRP-CC family are characterized as cellular sensors with polymodal activation and gating properties. Many TRP channels are activated by a variety of different stimuli and function as signal integrators. These mammalian proteins have been tabulated revealing their accepted designations, activators and inhibitors, putative interacting proteins and proposed functions. The founding members of the TRP superfamily are the TRPC (TRP canonical) channels, which can be activated following the stimulation of phospholipase C and/or depletion of internal calcium stores. However, the precise mechanisms leading to TRPC activation remain unclear. TRPC channels regulate nicotine-dependent behavior. One member of the TRP-CC family, TRP-PLIK (1862 aas; AF346629), has been implicated in the regulation of cell division. It has an N-terminal TRP-CC-like sequence and a C-terminal protein kinase-like sequence. It was shown to autophosphorylate and exhibits an ATP phosphorylation-dependent, non-selective, Ca2+-permeable, outward rectifying conductance. Another long homologue, Melastatin, is associated with melanocytic tumor progression whereas another homologue, MTR1, is associated with Beckwith-Wiedemann syndrome and a pred
https://en.wikipedia.org/wiki/Clay%20Shirky
Clay Shirky (born 1964) is an American writer, consultant and teacher on the social and economic effects of Internet technologies and journalism. In 2017 he was appointed Vice Provost of Educational Technologies of New York University (NYU), after serving as Chief Information Officer at NYU Shanghai from 2014 to 2017. He also is an associate professor at the Arthur L. Carter Journalism Institute and Associate Arts Professor at the Tisch School of the Arts' Interactive Telecommunications Program. His courses address, among other things, the interrelated effects of the topology of social networks and technological networks, how our networks shape culture and vice versa. He has written and been interviewed about the Internet since 1996. His columns and writings have appeared in Business 2.0, The New York Times, The Wall Street Journal, the Harvard Business Review and Wired. Shirky divides his time between consulting, teaching, and writing on the social and economic effects of Internet technologies. His consulting practice is focused on the rise of decentralized technologies such as peer-to-peer, web services, and wireless networks that provide alternatives to the wired client–server infrastructure that characterizes the World Wide Web. He is a member of the Wikimedia Foundation's advisory board. In The Long Tail, Chris Anderson calls Shirky "a prominent thinker on the social and economic effects of Internet technologies." Education and career After graduating from Yale University with a Bachelor of Arts degree in fine art in 1986, he moved to New York. In the 1990s he founded the Hard Place Theater, a theatre company that produced non-fiction theater using only found materials such as government documents, transcripts and cultural records and also worked as a lighting designer for other theater and dance companies, including the Wooster Group, Elevator Repair Service and Dana Reitz. During this time, Shirky was vice-president of the New York chapter of the Electron
https://en.wikipedia.org/wiki/Characteristic%20property
A characteristic property is a chemical or physical property that helps identify and classify substances. The characteristic properties of a substance are always the same whether the sample being observed is large or small. Thus, conversely, if the property of a substance changes as the sample size changes, that property is not a characteristic property. Examples of physical properties that are not characteristic properties are mass and volume. Examples of characteristic properties include melting points, boiling points, density, viscosity, solubility, crystal shape, and color. Substances with characteristic properties can be separated. For example, in fractional distillation, liquids are separated using the boiling point. The water Boiling point is 212 degrees Fahrenheit. Identifying a substance Every characteristic property is unique to one given substance. Scientists use characteristic properties to identify unknown substances. However, characteristic properties are most useful for distinguishing between two or more substances, not identifying a single substance. For example, isopropanol and water can be distinguished by the characteristic property of odor. Characteristic properties are used because the sample size and the shape of the substance does not matter. For example, 1 gram of lead is the same color as 100 tons of lead. See also Intensive and extensive properties
https://en.wikipedia.org/wiki/Ferranti%20Argus
Ferranti's Argus computers were a line of industrial control computers offered from the 1960s into the 1980s. Originally designed for a military role, a re-packaged Argus was the first digital computer to be used to directly control an entire factory. They were widely used in a variety of roles in Europe, particularly in the UK, where a small number continue to serve as monitoring and control systems for nuclear reactors. Original series Blue Envoy, hearing aid computer The original concept for the computer was developed as part of the Blue Envoy missile project. This was a very long-range surface-to-air missile system with a range on the order of . To reach these ranges, the missile was "lofted" in a nearly vertical trajectory at launch, flying as quickly as possible to high altitude where it suffered less drag during the subsequent long cruise toward the target. During the vertical climb, the missile's radar would not be able to see the target, so during this period it was command guided from the ground. Argus began as a system to read the radar data, compute the required trajectory, and send that to the missile in-flight. The system not only had to develop the trajectory, but also directly controlled the control surfaces of the missile and thus had a complete control feedback system. Development was carried out by Maurice Gribble at Ferranti's Automation Division in Wythenshawe starting in 1956. The system used OC71 transistors from Mullard, originally designed for use in hearing aids. They could only be run at the low speed of 25 kHz, but this was enough for the task. Blue Envoy was cancelled in 1957 as part of the sweeping 1957 Defence White Paper. Ferranti decided to continue the development of the computer for other uses. During a visit by Prince Philip, Duke of Edinburgh in November 1957, they set up a system with an automotive headlamp connected to a handle that could be moved by hand to shine at any point on a wall, while the computer attempted to move
https://en.wikipedia.org/wiki/Shepp%E2%80%93Logan%20phantom
The Shepp–Logan phantom is a standard test image created by Larry Shepp and Benjamin F. Logan for their 1974 paper The Fourier Reconstruction of a Head Section. It serves as the model of a human head in the development and testing of image reconstruction algorithms. Definition The function describing the phantom is defined as the sum of 10 ellipses inside a 2×2 square:
https://en.wikipedia.org/wiki/Elaeagnus%20%C3%97%20submacrophylla
Elaeagnus × submacrophylla, formerly known as Elaeagnus × ebbingei, is a hybrid between Elaeagnus macrophylla and Elaeagnus pungens. Several cultivars, including 'Gilt Edge', are grown in gardens as ornamental plants. Both the hybrid and 'Gilt Edge' have gained the Royal Horticultural Society's Award of Garden Merit. Description Elaeagnus × submacrophylla is an evergreen shrub, ultimately growing to about . The upper surfaces of the leaves are dark green, sometimes appearing metallic; the lower surfaces are silvery and scaly. Small fragrant tubular white flowers appear in autumn. Taxonomy The hybrid was first discovered in 1929 by Simon Doorenbos, a Dutch horticulturalist, while he was director of the Parks Department in The Hague. He sowed seed from several Elaeagnus species growing next to one another. These included E. macrophylla and E. pungens. Doorenbos gave the hybrid the epithet ebbingei, honouring J.W.E. Ebbinge, and the name E. × ebbingei was widely used until it was realized that the name with priority was E. × submacrophylla, published by Camille Servettaz in 1909. E. × ebbingei is now regarded as an illegitimate name. Cultivars Cultivars include: 'Albert Doorenbos' – large green leaves 'Compacta' – dwarf plant 'Costal Gold' – broad leaves, pale yellow in the centre when mature 'Gilt Edge' AGM – leaves with dark green centres, golden yellow margins 'Limelight' – leaves with yellow and pale green central areas when mature 'The Hague' – small green leaves
https://en.wikipedia.org/wiki/Chimerin%201
Chimerin 1 (CHN1), also known as alpha-1-chimerin, n-chimerin, is a protein which in humans is encoded by the CHN1 gene. Chimerin 1 is a GTPase activating protein specific for RAC GTP-binding proteins. It is expressed primarily in the brain and may be involved in signal transduction. This gene encodes GTPase-activating protein for p21-rac and a phorbol ester receptor. It plays an important role in ocular motor axon pathfinding. Function CHN1 is a three-domain protein with the N-terminal SH2 domain, the C-terminal RhoGAP domain and the central C1 domain similar to protein kinase C. When lipid diacylglycerol (DAG) binds to the C1 domain, CHN1 is transferred to the plasma membrane and negatively regulates Rho-family small GTPases RAC1 and CDC42, thus causing the morphological change of axons by pruning the ends of axon dendrites. Mutational analysis suggests that un-overlapping residues of the RhoGAP domain are involved in RAC1-binding and the RAC1-GAP activity. Regulation of the RhoGAP activity of CHN1 by phorbol esters, natural compounds mimic of the lipid second messenger DAG, presents a possible way of designing agents for therapeutics. Clinical significance Heterozygous missense mutations in this gene cause Duane's retraction syndrome 2 (DURS2).
https://en.wikipedia.org/wiki/Organic%20horticulture
Organic horticulture is the science and art of growing fruits, vegetables, flowers, or ornamental plants by following the essential principles of organic agriculture in soil building and conservation, pest management, and heirloom variety preservation. The Latin words hortus (garden plant) and cultura (culture) together form horticulture, classically defined as the culture or growing of garden plants. Horticulture is also sometimes defined simply as "agriculture minus the plough". Instead of the plough, horticulture makes use of human labour and gardener's hand tools, although some small machine tools like rotary tillers are commonly employed now. General Mulches, cover crops, compost, manures, vermicompost, and mineral supplements are soil-building mainstays that distinguish this type of farming from its conventional counterpart. Through attention to good healthy soil condition, it is expected that insect, fungal, or other problems that sometimes plague plants can be minimized. However, pheromone traps, insecticidal soap sprays, and other pest-control methods available to organic farmers are also utilized by organic horticulturists. Horticulture involves five areas of study: floriculture (includes production and marketing of floral crops), landscape horticulture (includes production, marketing and maintenance of landscape plants), olericulture (includes production and marketing of vegetables), pomology (includes production and marketing of fruits), and postharvest physiology (involves maintaining quality and preventing spoilage of horticultural crops). All of these can be, and sometimes are, pursued according to the principles of organic cultivation. Organic horticulture (or organic gardening) is based on knowledge and techniques gathered over thousands of years. In general terms, organic horticulture involves natural processes, often taking place over extended periods of time, and a sustainable, holistic approach – while chemical-based horticulture focuses on
https://en.wikipedia.org/wiki/Gomashio
Gomashio (hiragana: ごま塩; also spelled gomasio) is a dry condiment, similar to furikake, made from unhulled and . It is often used in Japanese cuisine, such as a topping for sekihan. It is also sometimes sprinkled over plain rice or onigiri. Some commercially sold gomashio also has sugar mixed in with the salt. Composition and use The sesame seeds used to make gomashio may be either tan or black in color. They are toasted before being mixed with the salt. Occasionally the salt is also toasted. The ratio of sesame seeds to salt varies according to taste and diet, generally ranging between 5:1 (5 parts sesame seeds to 1 part salt) and 15:1. Gomashio is often homemade, though it is also commercially available in glass or plastic containers. Gomashio is also a part of the macrobiotic diet, where it is used as a healthier alternative to ordinary salt. Generally, the gomashio used in macrobiotic cuisine will contain less salt than traditional Japanese gomashio (a ratio of 18 parts sesame seeds to 1 part salt is recommended for some individuals with a particularly restricted diet) and made by hand grinding in a suribachi. The word gomashio is also used in the Japanese language to describe a head of hair containing both white and black hair strands that intermingle, similar to the English idiom "salt and pepper". See also List of sesame seed dishes External links Gomashio page at northof49naturals.com Blog post discussing Gomashio (includes recipe) Japanese condiments Salted foods Sesame dishes
https://en.wikipedia.org/wiki/Circular%20dichroism
Circular dichroism (CD) is dichroism involving circularly polarized light, i.e., the differential absorption of left- and right-handed light. Left-hand circular (LHC) and right-hand circular (RHC) polarized light represent two possible spin angular momentum states for a photon, and so circular dichroism is also referred to as dichroism for spin angular momentum. This phenomenon was discovered by Jean-Baptiste Biot, Augustin Fresnel, and Aimé Cotton in the first half of the 19th century. Circular dichroism and circular birefringence are manifestations of optical activity. It is exhibited in the absorption bands of optically active chiral molecules. CD spectroscopy has a wide range of applications in many different fields. Most notably, UV CD is used to investigate the secondary structure of proteins. UV/Vis CD is used to investigate charge-transfer transitions. Near-infrared CD is used to investigate geometric and electronic structure by probing metal d→d transitions. Vibrational circular dichroism, which uses light from the infrared energy region, is used for structural studies of small organic molecules, and most recently proteins and DNA. Physical principles Circular polarization of light Electromagnetic radiation consists of an electric and magnetic field that oscillate perpendicular to one another and to the propagating direction, a transverse wave. While linearly polarized light occurs when the electric field vector oscillates only in one plane, circularly polarized light occurs when the direction of the electric field vector rotates about its propagation direction while the vector retains constant magnitude. At a single point in space, the circularly polarized-vector will trace out a circle over one period of the wave frequency, hence the name. The two diagrams below show the electric field vectors of linearly and circularly polarized light, at one moment of time, for a range of positions; the plot of the circularly polarized electric vector forms a heli
https://en.wikipedia.org/wiki/Solar%20Terrestrial%20Probes%20program
NASA's Solar Terrestrial Probes program (STP) is a series of missions focused on study the Sun-Earth system. It is part of NASA's Heliophysics Science Division within the Science Mission Directorate. Objectives Understand the fundamental physical processes of the complex space environment throughout the Solar System, which includes the flow of energy and charged material, known as plasma, as well as a dynamic system of magnetic and electric fields. Understand how human society, technological systems, and the habitability of planets are affected by solar variability and planetary magnetic fields. Develop the capability to predict the extreme and dynamic conditions in space in order to maximize the safety and productivity of human and robotic explorers. Missions TIMED The TIMED (Thermosphere Ionosphere Mesosphere Energetics and Dynamics) is an orbiter mission dedicated to study the dynamics of the Mesosphere and Lower Thermosphere (MLT) portion of the Earth's atmosphere. The mission was launched from Vandenberg Air Force Base in California on December 7, 2001 aboard a Delta II rocket launch vehicle. Hinode Hinode, an ongoing collaboration with JAXA, is a mission to explore the magnetic fields of the Sun. It was launched on the final flight of the M-V-7 rocket from Uchinoura Space Center, Japan on September 22, 2006. STEREO STEREO (Solar Terrestrial Relations Observatory) is a solar observation mission. It consists in two nearly identical spacecraft, launched on October 26, 2006. MMS The Magnetospheric Multiscale Mission (MMS) is a mission to study the Earth's magnetosphere, using four identical spacecraft flying in a tetrahedral formation. The spacecraft were launched on March 13, 2015. IMAP IMAP (Interstellar Mapping and Acceleration Probe) is a heliosphere observation mission. Planned for launch in 2025, it will sample, analyze, and map particles streaming to Earth from the edges of interstellar space.
https://en.wikipedia.org/wiki/Universal%20stress%20protein
The universal stress protein (USP) domain is a superfamily of conserved genes which can be found in bacteria, archaea, fungi, protozoa and plants. Proteins containing the domain are induced by many environmental stressors such as nutrient starvation, drought, extreme temperatures, high salinity, and the presence of uncouplers, antibiotics and metals. In the presence of these stressors, Usp genes are upregulated resulting in large quantities of Usp proteins being produced by the cell. The over production of USP genes allows the organisms to better cope with stresses by largely unknown mechanisms. However, the USPs will alter the expression of a variety of genes that help to cope with stress. Function The primary function of this superfamily is to protect the organism from environmental stress such as exposure to UV light, which may induce genes containing the USP domain in order to protect the DNA and more generally the cell from further damage. During bacterial starvation the USP genes upregulated will often arrest cell growth and promote its metabolism to adapt to sparse nutrients. Recent research also suggests proteins containing this domain have functions beyond the realms of dealing with environmental stresses. Nachin et al. demonstrated in Escherichia coli that USPs are involved in actions such as adhesion and motility. The researchers, through means of "knocking out" USP genes known as UspE and UspC, saw results suggesting an inability to swim and completely lack of motility, respectively. Conversely, mutants for genes UspF and UspG were shown to have enhanced swimming abilities. Therefore, mobility is affected both positively and negatively USPs within E. coli. This demonstrates USPs influence throughout the cell could be widespread for a number of reasons. Additionally, in Halmonas elongate, there is a USP called TeaD has been described as a key regulator in the transport of Ectoine across the cell membrane. This demonstrates how versatile USPs can be.
https://en.wikipedia.org/wiki/Axial%20current
The axial current, also denoted the pseudo-vector or chiral current, is the conserved current associated to the chiral symmetry or axial symmetry of a system. Origin According to Noether's theorem, each symmetry of a system is associated a conserved quantity. For example, the rotational invariance of a system implies the conservation of its angular momentum, or spacetime invariance implies the conservation of energy–momentum. In quantum field theory, internal symmetries also result in conserved quantities. For example, the U(1) gauge transformation of QED implies the conservation of the electric charge. Likewise, if a theory possesses an internal chiral or axial symmetry, there will be a conserved quantity, which is called the axial charge. Further, just as the motion of an electrically charged particle produces an electric current, a moving axial charge constitutes an axial current. Definition The axial current resulting from the motion of an axially charged moving particle is formally defined as , where is the particle field represented by Dirac spinor (since the particle is typically a spin-1/2 fermion) and and are the Dirac gamma matrices. For comparison, the electromagnetic current produced by an electrically charged moving particle is . Meaning As explained above, the axial current is simply the equivalent of the electromagnetic current for the axial symmetry instead of the U(1) symmetry. Another perspective is given by recalling that the chiral symmetry is the invariance of the theory under the field rotation   and   (or alternatively   and   ), where denotes a left-handed field and a right-handed one. From this as well as the fact that and the definition of above, one sees that the axial current is the difference between the current due to left-handed fermions and that from right-handed ones, whilst the electromagnetic current is the sum. Chiral symmetry is exhibited by vector gauge theories with massless fermions. Since there is no know
https://en.wikipedia.org/wiki/Label%20Information%20Base
Label Information Base (LIB) is the software table maintained by IP/MPLS capable routers to store the details of port and the corresponding MPLS router label to be popped/pushed on incoming/outgoing MPLS packets. Entries are populated from label-distribution protocols. LIB functions in the control plane of router's MPLS layer. It is used by the label distribution protocol for mapping the next hop labels. MPLS networking
https://en.wikipedia.org/wiki/Evx1
Evx1 is a mammalian gene located downstream of the HoxA cluster, which encodes for a homeobox transcription factor. Evx1 is a homolog of even-skipped (eve), which is a pair-rule gene that regulates body segmentation in Drosophila. The expression of Evx1 is developmentally regulated, displaying a biphasic expression pattern with peak expression in the primitive streak during gastrulation and in interneurons during neural development. Evx1 has been shown to regulate anterior-posterior patterning during gastrulation by acting as a downstream effector of the Wnt and BMP signalling pathways. It is also a critical regulator of interneuron identity. Despite a regulatory role in both of these processes, the Evx1 KO mouse is viable and displays no overt phenotype after birth. A divergent long non-coding RNA (Evx1as) is also expressed from the locus, which is temporally and spatially co-expressed with Evx1. Evx1as does not have a function beyond that of Evx1 (in trans); however, it has been shown to regulate Evx1 at the locus (in cis).
https://en.wikipedia.org/wiki/Real%20computation
In computability theory, the theory of real computation deals with hypothetical computing machines using infinite-precision real numbers. They are given this name because they operate on the set of real numbers. Within this theory, it is possible to prove interesting statements such as "The complement of the Mandelbrot set is only partially decidable." These hypothetical computing machines can be viewed as idealised analog computers which operate on real numbers, whereas digital computers are limited to computable numbers. They may be further subdivided into differential and algebraic models (digital computers, in this context, should be thought of as topological, at least insofar as their operation on computable reals is concerned). Depending on the model chosen, this may enable real computers to solve problems that are inextricable on digital computers (For example, Hava Siegelmann's neural nets can have noncomputable real weights, making them able to compute nonrecursive languages.) or vice versa. (Claude Shannon's idealized analog computer can only solve algebraic differential equations, while a digital computer can solve some transcendental equations as well. However this comparison is not entirely fair since in Claude Shannon's idealized analog computer computations are immediately done; i.e. computation is done in real time. Shannon's model can be adapted to cope with this problem.) A canonical model of computation over the reals is Blum–Shub–Smale machine (BSS). If real computation were physically realizable, one could use it to solve NP-complete problems, and even #P-complete problems, in polynomial time. Unlimited precision real numbers in the physical universe are prohibited by the holographic principle and the Bekenstein bound. See also Hypercomputation, for other such powerful machines.
https://en.wikipedia.org/wiki/Stool%20guaiac%20test
The stool guaiac test or guaiac fecal occult blood test (gFOBT) is one of several methods that detects the presence of fecal occult blood (blood invisible in the feces). The test involves placing a fecal sample on guaiac paper (containing a phenolic compound, alpha-guaiaconic acid, extracted from the wood resin of Guaiacum trees) and applying hydrogen peroxide which, in the presence of blood, yields a blue reaction product within seconds. The American College of Gastroenterology has recommended the abandoning of gFOBT testing as a colorectal cancer screening tool, in favor of the fecal immunochemical test (FIT). Though the FIT is preferred, even the guaiac FOB testing of average risk populations may have been sufficient to reduce the mortality associated with colon cancer by about 25%. With this lower efficacy, it was not always cost effective to screen a large population with gFOBT. Methodology The stool guaiac test involves fasting from iron supplements, red meat (the blood it contains can turn the test positive), certain vegetables (which contain a chemical with peroxidase properties that can turn the test positive), and vitamin C and citrus fruits (which can turn the test falsely negative) for a period of time before the test. It has been suggested that cucumber, cauliflower and horseradish, and often other vegetables, should be avoided for three days before the test. In testing, feces are applied to a thick piece of paper attached to a thin film coated with guaiac. Either the patient or medical professional smears a small fecal sample on to the film. The fecal sample is obtained by catching the stool and transferring a sample with an applicator. Digital rectal examination specimens are also used but this method is discouraged for colorectal cancer screening due to very poor performance characteristics. Both sides of the test card can be peeled open, to access the inner guaiac paper. One side of the card is marked for application of the stool and the o
https://en.wikipedia.org/wiki/Macrolepiota%20procera
Macrolepiota procera, the parasol mushroom, is a basidiomycete fungus with a large, prominent fruiting body resembling a parasol. It is a fairly common species on well-drained soils. It is found solitary or in groups and fairy rings in pastures and occasionally in woodland. Globally, it is widespread in temperate regions. Taxonomy The fungus was first described in 1772 by Italian naturalist Giovanni Antonio Scopoli, who named it Agaricus procerus. Rolf Singer transferred it to the genus Macrolepiota in 1948. Description The height and cap diameter of a mature specimen may both reach 30 to 40 (50) cm. The stipe is relatively thin and reaches full height before the cap has expanded. The stipe is very fibrous in texture which renders it inedible (unless dried and ground). The surface is characteristically wrapped in a snakeskin-like pattern of scaly growths (therefore, known in some parts of Europe as the "snake's hat" or "snake's sponge"). The immature cap is compact and egg-shaped, with the cap margin around the stipe, sealing a chamber inside the cap. As it matures, the margin breaks off, leaving a fleshy, movable ring around the stipe. At full maturity, the cap is more or less flat, with a chocolate-brown umbo in the centre that is leathery to touch. Dark and cap-coloured flakes remain on the upper surface of the cap and can be removed easily. The gills are crowded, free, and white with a pale pink tinge sometimes present. The spore print is white. It has a pleasant nutty smell. When sliced, the white flesh may turn a pale pink. Uses Macrolepiota procera is a choice edible mushroom. It is very sought-after and popular in Europe, due in part to its large size, seasonal frequency, and versatility in the kitchen. In the UK, it may be found from July through to November. The parasol mushroom is difficult to mistake for any other, especially in regions like Europe where the poisonous look-alike Chlorophyllum molybdites is rare. The spores and lamellae of C. molybdit
https://en.wikipedia.org/wiki/Hydrobiology
Hydrobiology is the science of life and life processes in water. Much of modern hydrobiology can be viewed as a sub-discipline of ecology but the sphere of hydrobiology includes taxonomy, economic and industrial biology, morphology, and physiology. The one distinguishing aspect is that all fields relate to aquatic organisms. Most work is related to limnology and can be divided into lotic system ecology (flowing waters) and lentic system ecology (still waters). One of the significant areas of current research is eutrophication. Special attention is paid to biotic interactions in plankton assemblage including the microbial loop, the mechanism of influencing algal blooms, phosphorus load, and lake turnover. Another subject of research is the acidification of mountain lakes. Long-term studies are carried out on changes in the ionic composition of the water of rivers, lakes and reservoirs in connection with acid rain and fertilization. One goal of current research is elucidation of the basic environmental functions of the ecosystem in reservoirs, which are important for water quality management and water supply. Much of the early work of hydrobiologists concentrated on the biological processes utilized in sewage treatment and water purification especially slow sand filters. Other historically important work sought to provide biotic indices for classifying waters according to the biotic communities that they supported. This work continues to this day in Europe in the development of classification tools for assessing water bodies for the EU water framework directive. A hydrobiologist technician conducts field analysis for hydrobiology. They identify plants and living species, locate their habitat, and count them. They also identify pollutants and nuisances that can affect the aquatic fauna and flora. They take the samples and write reports of their observations for publications. A hydrobiologist engineer intervenes more in the process of the study. They define the inte
https://en.wikipedia.org/wiki/Email%20forwarding
Email forwarding generically refers to the operation of re-sending a previously delivered email to an email address to one or more different email addresses. The term forwarding, used for mail since long before electronic communications, has no specific technical meaning, but it implies that the email has been moved "forward" to a new destination. Email forwarding can also redirect mail going to a certain address and send it to one or more other addresses. Vice versa, email items going to several different addresses can converge via forwarding to end up in a single address in-box. Email users and administrators of email systems use the same term when speaking of both server-based and client-based forwarding. Server-based forwarding The domain name (the part appearing to the right of @ in an email address) defines the target server(s) for the corresponding class of addresses. A domain may also define backup servers; they have no mailboxes and forward messages without changing any part of their envelopes. By contrast, primary servers can deliver a message to a user's mailbox and/or forward it by changing some envelope addresses. ~/.forward files (see below) provide a typical example of server-based forwarding to different recipients. Email administrators sometimes use the term redirection as a synonym for server-based email-forwarding to different recipients. Protocol engineers sometimes use the term Mediator to refer to a forwarding server. Because of spam, it is becoming increasingly difficult to reliably forward mail across different domains, and some recommend avoiding it if at all possible. Uses of server-based forwarding to different recipients Role-addresses info, sales, postmaster, and similar names can appear to the left of @ in email addresses. An organization may forward messages intended for a given role to the address of the person(s) currently functioning in that role or office. Pseudonym-addresses Most domain name hosting facilities provide
https://en.wikipedia.org/wiki/Clamshell%20design
A clamshell design is a kind of form factor for electronic devices in the shape of a clamshell. Mobile phones, handheld game consoles, and especially laptops, are often designed like clamshells. Clamshell devices are usually made of two sections connected by a hinge, each section containing either a flat panel display or an alphanumeric keyboard/keypad, which can fold into contact together like a bivalve shell. A clamshell mobile phone is sometimes also called a flip phone, especially if the hinge is on the short edge. If the hinge is on a long edge (e.g., Nokia Communicators), the device is more likely to be called just a "clamshell" rather than a flip phone. Generally speaking, the interface components such as keys and display are kept inside the closed clamshell, protecting them from damage and unintentional use while also making the device shorter or narrower so it is easier to carry around. In many cases, opening the clamshell offers more surface area than when the device is closed, allowing interface components to be larger and easier to use than on devices which do not flip open. A disadvantage of the clamshell design is the connecting hinge, which is prone to fatigue or failure. Etymology The clamshell form factor is most closely associated with the cell phone market, as Motorola used to have a trademark on the term "flip phone", but the term "flip phone" has become genericized to be used more frequently than "clamshell" in colloquial speech. History A "flip phone" like communication device appears in chapter 3 of Armageddon 2419 A.D., a science fiction novella by Philip Francis Nowlan, which was first published in the August 1928 issue of the pulp magazine Amazing Stories: "Alan took a compact packet about six inches square from a holster attached to her belt and handed it to Wilma. So far as I could see, it had no special receiver for the ear. Wilma merely threw back a lid, as though she was opening a book, and began to talk. The voice that came bac
https://en.wikipedia.org/wiki/Nanonetwork
A nanonetwork or nanoscale network is a set of interconnected nanomachines (devices a few hundred nanometers or a few micrometers at most in size), which are able to perform only very simple tasks such as computing, data storing, sensing and actuation. Nanonetworks are expected to expand the capabilities of single nanomachines both in terms of complexity and range of operation by allowing them to coordinate, share and fuse information. Nanonetworks enable new applications of nanotechnology in the biomedical field, environmental research, military technology and industrial and consumer goods applications. Nanoscale communication is defined in IEEE P1906.1. Communication approaches Classical communication paradigms need to be revised for the nanoscale. The two main alternatives for communication in the nanoscale are based either on electromagnetic communication or on molecular communication. Electromagnetic This is defined as the transmission and reception of electromagnetic radiation from components based on novel nanomaterials. Recent advancements in carbon and molecular electronics have opened the door to a new generation of electronic nanoscale components such as nanobatteries, nanoscale energy harvesting systems, nano-memories, logical circuitry in the nanoscale and even nano-antennas. From a communication perspective, the unique properties observed in nanomaterials will decide on the specific bandwidths for emission of electromagnetic radiation, the time lag of the emission, or the magnitude of the emitted power for a given input energy, amongst others. For the time being, two main alternatives for electromagnetic communication in the nanoscale have been envisioned. First, it has been experimentally demonstrated that is possible to receive and demodulate an electromagnetic wave by means of a nanoradio, i.e., an electromechanically resonating carbon nanotube which is able to decode an amplitude or frequency modulated wave. Second, graphene-based nano-antennas
https://en.wikipedia.org/wiki/Skyronnes
Skyronnes are Icelandic dipping sauces made out of skyr instead of mayonnaise. They are usually served with fish and chips.
https://en.wikipedia.org/wiki/Pilgrim%20badge
Pilgrim badges are decorations worn by some of those who undertake a Christian pilgrimage to a place considered holy by the Church. They became very popular among Catholics in the later medieval period. Typically made of lead alloy, they were sold as souvenirs at sites of Christian pilgrimage and bear imagery relating to the saint venerated there. The production of pilgrim badges flourished in the Middle Ages in Europe, particularly in the 14th and 15th centuries, but declined after the Protestant Reformation of the mid-16th century. Tens of thousands have been found since the mid-19th century, predominantly in rivers. Together they form the largest corpus of medieval art objects to survive to us today. Pilgrimage sites housed a saint's relics: sometimes the whole body, sometimes a body part or significant object owned or touched by the saint. For example, St Thomas Becket was martyred at Canterbury Cathedral in England in 1170 and his body remained there, becoming the epicentre of an enormously popular cult. In 1220 it was translated into a costly shrine. The pilgrim souvenirs associated with his cult have a particularly diverse array of imagery, including that of his shrine, his head reliquary and scenes from his life. Other major sites that produced badges were Santiago de Compostela, Cologne, Our Lady of Rocamadour and Jerusalem. Their badges bore images that were iconic and easily recognisable, such as the scallop shell, the Adoration of the Magi, the St Peter or the Jerusalem Cross. Shrines to the Virgin were common all over Christendom, as are badges associable with her. They often show her holding the Infant Christ, or represent the first letter of her name. The practice is continued by some today. For example, knights and dames of the Equestrian Order of the Holy Sepulchre receive a pilgrim badge when they travel to the Holy Land. Origins Various cultural practices converged to bring about the pilgrim badge. Pilgrims had long sought natural souvenirs f
https://en.wikipedia.org/wiki/Glossary%20of%20quantum%20philosophy
This is a glossary for the terminology applied in the foundations of quantum mechanics and quantum metaphysics, collectively called quantum philosophy, a subfield of philosophy of physics. Note that this is a highly debated field, hence different researchers may have different definitions on the terms. Physics Non-classical properties of quantum mechanics nonseparability See also: entangled Nonlocality Superposition of states See also: Schrödinger's cat Quantum phenomena decoherence uncertainty principle See also: Einstein and the quantum entanglement See also: Bell's theorem, EPR paradox and CHSH inequality quantum teleportation superselection rule quantum erasure delayed choice experiment Quantum Zeno effect premeasurement ideal measurement Suggested physical entities hidden variables ensemble Terms used in the formalism of quantum mechanics Born's rule collapse postulate measurement relative state decoherent histories Metaphysics objective and subjective ontic and epistemic intrinsic and extrinsic agnostic Philosophical realism determinism causality empiricism rationalism scientific realism psychophysical parallelism Interpretations of quantum mechanics List of interpretations: Bohmian Mechanics de Broglie–Bohm theory consistent histories Copenhagen interpretation conventional interpretation Usually refer to the Copenhagen interpretation. Ensemble Interpretation Everett interpretation See relative-state interpretation. hydrodynamic interpretation Ghirardi–Rimini–Weber theory (GRW theory / GRW effect) many-worlds interpretation many-minds interpretation many-measurements interpretation modal interpretations objective collapse theory orthodox interpretation Usually refer to the Copenhagen interpretation. Penrose interpretation Pilot wave Quantum logic relative-state interpretation relational quantum mechanics stochastic interpretation transactional interpretation Uncategorized items qua
https://en.wikipedia.org/wiki/Obstructed%20defecation
Obstructed defecation syndrome (abbreviated as ODS, with many synonymous terms) is a major cause of functional constipation (primary constipation), of which it is considered a subtype. It is characterized by difficult and/or incomplete emptying of the rectum with or without an actual reduction in the number of bowel movements per week. Normal definitions of functional constipation include infrequent bowel movements and hard stools. In contrast, ODS may occur with frequent bowel movements and even with soft stools, and the colonic transit time may be normal (unlike slow transit constipation), but delayed in the rectum and sigmoid colon. Definitions and terminology Definition and classification of constipation Constipation is usually divided into two groups: primary and secondary. Primary constipation is caused by disrupted regulation of neuromuscular function of in the colon and the rectum, and also disruption of brain–gut neuroenteric function. Secondary constipation is caused by many other different factors such as diet, drugs, behavioral, endocrine, metabolic, neurological, and other disorders. There are main subtypes of primary constipation which are recognized, although overlap exists (see: Co-existence of different constipation subtypes): dyssynergic defecation, slow transit constipation (colonic dysmotility) and irritable bowel syndrome with constipation. Definition and terminology of obstructed defecation syndrome Obstructed defecation is one of the causes of chronic constipation. ODS is a loose term, consisting of a constellation of possible symptoms, caused by multiple, complex and poorly understood disorders which may include both functional and organic disorders. The topic of defecation disorders is very complicated, and there is a lot of confusion regarding terminology and classification in published literature. Occasionally some sources inappropriately treat ODS as a synonym of anismus. Although anismus is a major cause of ODS, there are other possib
https://en.wikipedia.org/wiki/Acorn%20MOS
The Machine Operating System (MOS) or OS is a discontinued computer operating system (OS) used in Acorn Computers' BBC computer range. It included support for four-channel sound, graphics, file system abstraction, and digital and analogue input/output (I/O) including a daisy-chained expansion bus. The system was single-tasking, monolithic and non-reentrant. Versions 0.10 to 1.20 were used on the BBC Micro, version 1.00 on the Electron, version 2 was used on the B+, and versions 3 to 5 were used in the BBC Master series. The final BBC computer, the BBC A3000, was 32-bit and ran RISC OS, which kept on portions of the Acorn MOS architecture and shared a number of characteristics (e.g. "star commands" CLI, "VDU" video control codes and screen modes) with the earlier 8-bit MOS. Versions 0 to 2 of the MOS were 16 KiB in size, written in 6502 machine code, and held in read-only memory (ROM) on the motherboard. The upper quarter of the 16-bit address space (0xC000 to 0xFFFF) is reserved for its ROM code and I/O space. Versions 3 to 5 were still restricted to a 16 KiB address space, but managed to hold more code and hence more complex routines, partly because of the alternative 65C102 central processing unit (CPU) with its denser instruction set plus the careful use of paging. User interface The original MOS versions, from 0 to 2, did not have a user interface per se: applications were expected to forward operating system command lines to the OS on its behalf, and the programming language BBC BASIC ROM, with 6502 assembler built in, supplied with the BBC Micro is the default application used for this purpose. The BBC Micro would halt with a Language? error if no ROM is present that advertises to the OS an ability to provide a user interface (called language ROMs). MOS version 3 onwards did feature a simple command-line interface, normally only seen when the CMOS memory did not contain a setting for the default language ROM. Application programs on ROM, and some casset
https://en.wikipedia.org/wiki/Quasi-identifier
Quasi-identifiers are pieces of information that are not of themselves unique identifiers, but are sufficiently well correlated with an entity that they can be combined with other quasi-identifiers to create a unique identifier. Quasi-identifiers can thus, when combined, become personally identifying information. This process is called re-identification. As an example, Latanya Sweeney has shown that even though neither gender, birth dates nor postal codes uniquely identify an individual, the combination of all three is sufficient to identify 87% of individuals in the United States. The term was introduced by Tore Dalenius in 1986. Since then, quasi-identifiers have been the basis of several attacks on released data. For instance, Sweeney linked health records to publicly available information to locate the then-governor of Massachusetts' hospital records using uniquely identifying quasi-identifiers, and Sweeney, Abu and Winn used public voter records to re-identify participants in the Personal Genome Project. Additionally, Arvind Narayanan and Vitaly Shmatikov discussed on quasi-identifiers to indicate statistical conditions for de-anonymizing data released by Netflix. Motwani and Ying warn about potential privacy breaches being enabled by publication of large volumes of government and business data containing quasi-identifiers. See also De-identification Differential privacy Personally identifying information
https://en.wikipedia.org/wiki/Kiki%20Kaikai
is a shoot 'em up video game developed and published by Taito for arcades in 1986. Set in Feudal Japan, the player assumes the role of a Shinto shrine maiden who must use her o-fuda scrolls and gohei wand to defeat renegade spirits and monsters from Japanese mythology. The game is noteworthy for using a traditional fantasy setting in a genre otherwise filled with science fiction motifs. The game received a number of home ports, both as a stand-alone title and as part of compilations. The original arcade game was only ever released in Japan, but a bootleg version called Knight Boy was released outside Japan. Kiki Kaikai was followed by a sequel for the Super NES in 1992 known as Pocky & Rocky outside Japan. The series, known as Kiki Kaikai in Japan and Pocky & Rocky outside Japan, has continued since then and includes several games. Plot The game follows the adventures of "Sayo-chan", a young Shinto shrine maiden living in Feudal Japan. One night, while Sayo-chan is fanning a ceremonial fire, she is visited by the Seven Lucky Gods, who warn her of a great, impending danger. Suddenly, a band of mischievous youkai appear and kidnap the gods, quickly retreating to a faraway mountain range. Sayo-chan, determined to help the gods, sets off on a journey across the countryside, where she confronts a number of strange creatures from Japanese mythology, including obake, and yurei. Gameplay Kiki Kaikai is an overhead multi-directional shooter game that requires the player to move in four directions through various levels while attacking harmful enemies as they approach from off screen. As Sayo-chan, the player can attack by either throwing her special o-fuda scrolls in eight separate directions, or by swinging her purification rod directly in front of her. These techniques can be upgraded by finding special paper slips left by defeated enemies that will either enhance their power or improve their range. Sayo can be damaged by coming in contact with an enemy, and can only
https://en.wikipedia.org/wiki/Rubik%27s%20Clock
Rubik's Clock is a mechanical puzzle invented and patented by Christopher C. Wiggs and Christopher J. Taylor. The Hungarian sculptor and professor of architecture Ernő Rubik bought the patent from them to market the product under his name. It was first marketed in 1988. Rubik's Clock is a two-sided puzzle, each side presenting nine clocks to the puzzler. There are four wheels, one at each corner of the puzzle, each allowing the corresponding corner clock to be rotated directly. (The corner clocks, unlike the other clocks, rotate on both sides of the puzzle simultaneously and can never be operated independently. Thus the puzzle contains only 14 independent clocks.) There are also four buttons which span both sides of the puzzle; each button arranged such that if it is "in" on one side it is "out" on the other. The state of each button (in or out) determines whether the adjacent corner clock is mechanically connected to the three other adjacent clocks on the front side or on the back side: thus the configuration of the buttons determines which sets of clocks can be turned simultaneously by rotating a suitable wheel. The aim of the puzzle is to set all nine clocks to 12 o'clock (straight up) on both sides of the puzzle simultaneously. The method to do so is to start by constructing a cross on both sides (at 12 o’clock) and then solving the corner clocks. The Rubik’s clock is listed as one of the 17 WCA events, with records for fastest time to solve one puzzle, and the fastest average time to solve 5 puzzles (discarding the slowest and fastest times). Combinations Since there are 14 independent clocks, with 12 settings each, there are a total of =1,283,918,464,548,864 possible combinations for the clock faces. This does not count for the number of pin positions. Notation The puzzle is oriented with 12 o'clock on top, and either side in front. The following moves can be made: Pin movements: UR (top-right): Move the top-right pin up. DR (bottom-right): Move
https://en.wikipedia.org/wiki/Rashba%E2%80%93Edelstein%20effect
The Rashba–Edelstein effect (REE) is a spintronics-related effect, consisting in the conversion of a bidimensional charge current into a spin accumulation. This effect is an intrinsic charge-to-spin conversion mechanism and it was predicted in 1990 by the scientist V.M. Edelstein. It has been demonstrated in 2013 and confirmed by several experimental evidences in the following years. Its origin can be ascribed to the presence of spin-polarized surface or interface states. Indeed, a structural inversion symmetry breaking (i.e., a structural inversion asymmetry (SIA)) causes the Rashba effect to occur: this effect breaks the spin degeneracy of the energy bands and it causes the spin polarization being locked to the momentum in each branch of the dispersion relation. If a charge current flows in these spin-polarized surface states, it generates a spin accumulation. In the case of a bidimensional Rashba gas, where this band splitting occurs, this effect is called Rashba–Edelstein effect. For what concerns a class of peculiar materials, called topological insulators (TI), spin-splitted surface states exist due to the surface topology, independently from the Rashba effect. Topological insulators, indeed, display a spin-splitted linear dispersion relation on their surfaces (i.e., spin-polarized Dirac cones), while having a band gap in the bulk (this is why these materials are called insulators). Also in this case, spin and momentum are locked and, when a charge current flows in these spin-polarized surface states, a spin accumulation is produced and this effect is called Edelstein effect. In both cases, a 2D charge-to-spin conversion mechanism occurs. The reverse process is called inverse Rashba–Edelstein effect and it converts a spin accumulation into a bidimensional charge current, resulting in a 2D spin-to-charge conversion. The Rashba–Edelstein effect and its inverse effect are classified as a spin-charge interconversion (SCI) mechanisms, as the direct and inverse
https://en.wikipedia.org/wiki/Load%20cell
A load cell converts a force such as tension, compression, pressure, or torque into a signal (electrical, pneumatic or hydraulic pressure, or mechanical displacement indicator) that can be measured and standardized. It is a force transducer. As the force applied to the load cell increases, the signal changes proportionally. The most common types of load cells are pneumatic, hydraulic, and strain gauge types for industrial applications. Typical non-electronic bathroom scales are a widespread example of a mechanical displacement indicator where the applied weight (force) is indicated by measuring the deflection of springs supporting the load platform, technically a "load cell". Strain gauge load cell Strain gauge load cells are the kind most often found in industrial settings. It is ideal as it is highly accurate, versatile, and cost-effective. Structurally, a load cell has a metal body to which strain gauges have been secured.  The body is usually made of aluminum, alloy steel, or stainless steel which makes it very sturdy but also minimally elastic. This elasticity gives rise to the term "spring element", referring to the body of the load cell.  When force is exerted on the load cell, the spring element is slightly deformed, and unless overloaded, always returns to its original shape. As the spring element deforms, the strain gauges also change shape. The resulting alteration to the resistance in the strain gauges can be measured as voltage. The change in voltage is proportional to the amount of force applied to the cell, thus the amount of force can be calculated from the load cell's output. Strain Gauges A strain gauge is constructed of very fine wire, or foil, set up in a grid pattern and attached to a flexible backing. When the shape of the strain gauge is altered, a change in its electrical resistance occurs. The wire or foil in the strain gauge is arranged in a way that, when force is applied in one direction, a linear change in resistance results. Tension
https://en.wikipedia.org/wiki/List%20of%20herbicides
This is a list of herbicides. These are chemical compounds which have been registered as herbicides. The names on the list are the ISO common name for the active ingredient which is formulated into the branded product sold to end-users. The University of Hertfordshire maintains a database of the chemical and biological properties of these materials, including their brand names and the countries and dates where and when they have been introduced. The industry-sponsored Herbicide Resistance Action Committee (HRAC) advises on the use of herbicides in crop protection and classifies the available compounds according to their chemical structures and mechanism of action so as to manage the risks of pesticide resistance developing. The 2022 HRAC poster of herbicide modes of action includes the majority of chemicals listed below. The Weed Science Society of America also classifies herbicides by their mechanism of action. 0-9 A B C D E F G H I K L M N O P Q R S T V X Z See also List of fungicides List of insecticides
https://en.wikipedia.org/wiki/LtrA
LtrA is an open reading frame found in the Lactococcus lactis group II introns LtrB. It is an intron-encoded protein, which consists of three subdomains: a reverse-transcriptase/maturase, DNA endonuclease, and DNA/RNA binding domain. LtrA helps to capture and stabilize the catalytically active conformation of the LtrB group II intron RNA. It also functions in group II intron retrohoming.
https://en.wikipedia.org/wiki/Biological%20engineering
Biological engineering or bioengineering is the application of principles of biology and the tools of engineering to create usable, tangible, economically viable products. Biological engineering employs knowledge and expertise from a number of pure and applied sciences, such as mass and heat transfer, kinetics, biocatalysts, biomechanics, bioinformatics, separation and purification processes, bioreactor design, surface science, fluid mechanics, thermodynamics, and polymer science. It is used in the design of medical devices, diagnostic equipment, biocompatible materials, renewable energy, ecological engineering, agricultural engineering, process engineering and catalysis, and other areas that improve the living standards of societies. Examples of bioengineering research include bacteria engineered to produce chemicals, new medical imaging technology, portable and rapid disease diagnostic devices, prosthetics, biopharmaceuticals, and tissue-engineered organs. Bioengineering overlaps substantially with biotechnology and the biomedical sciences in a way analogous to how various other forms of engineering and technology relate to various other sciences (such as aerospace engineering and other space technology to kinetics and astrophysics). In general, biological engineers attempt to either mimic biological systems to create products, or to modify and control biological systems. Working with doctors, clinicians, and researchers, bioengineers use traditional engineering principles and techniques to address biological processes, including ways to replace, augment, sustain, or predict chemical and mechanical processes. History Biological engineering is a science-based discipline founded upon the biological sciences in the same way that chemical engineering, electrical engineering, and mechanical engineering can be based upon chemistry, electricity and magnetism, and classical mechanics, respectively. Before WWII, biological engineering had begun being recognized as a
https://en.wikipedia.org/wiki/Quad%20in-line%20package
In microelectronics, a quad in-line package (QIP or QIL), is an electronic component package with a rectangular housing and four parallel rows of electrical connecting pins. The package may be through-hole mounted to a printed circuit board (PCB) or inserted in a socket. Rockwell used a QIP with 42 leads formed into staggered rows for their PPS-4 microprocessor family introduced in 1973, and other microprocessors and microcontrollers, some with higher lead counts, through the early 1990s. The QIP has the same dimensions as a Dual in-line package (DIP), but the leads on each side are bent into an alternating zigzag configuration so as to fit four lines of solder pads (instead of two with a DIP but similar to Zig-zag in-line package). The QIP design increased the spacing between solder pads without increasing package size, for two reasons: First it allowed more reliable soldering. This may seem odd today, given the far closer solder pad spacing in use now, but in the 1970s, the heyday of the QIL, bridging of neighbouring solder pads on DIP ICs was an issue at times, QIP also increased the possibility of running a copper track between two solder pads. This was very handy on the then standard single sided single layer PCBs. Some QIP packaged ICs had added heatsinking tabs, such as the HA1306W. Intel and 3M developed the ceramic leadless quad in-line package (QUIP), introduced in 1979, to boost microprocessor density and economy. The ceramic leadless QUIP is not designed for surface-mount use, and requires a socket. It was used by Intel for the iAPX 432 microprocessor chip set, and by Zilog for the Z8-02 external-ROM prototyping version of the Z8 microcontroller.
https://en.wikipedia.org/wiki/Cut%20%28graph%20theory%29
In graph theory, a cut is a partition of the vertices of a graph into two disjoint subsets. Any cut determines a cut-set, the set of edges that have one endpoint in each subset of the partition. These edges are said to cross the cut. In a connected graph, each cut-set determines a unique cut, and in some cases cuts are identified with their cut-sets rather than with their vertex partitions. In a flow network, an s–t cut is a cut that requires the source and the sink to be in different subsets, and its cut-set only consists of edges going from the source's side to the sink's side. The capacity of an s–t cut is defined as the sum of the capacity of each edge in the cut-set. Definition A cut is a partition of of a graph into two subsets and . The cut-set of a cut is the set of edges that have one endpoint in and the other endpoint in . If and are specified vertices of the graph , then an cut is a cut in which belongs to the set and belongs to the set . In an unweighted undirected graph, the size or weight of a cut is the number of edges crossing the cut. In a weighted graph, the value or weight is defined by the sum of the weights of the edges crossing the cut. A bond is a cut-set that does not have any other cut-set as a proper subset. Minimum cut A cut is minimum if the size or weight of the cut is not larger than the size of any other cut. The illustration on the right shows a minimum cut: the size of this cut is 2, and there is no cut of size 1 because the graph is bridgeless. The max-flow min-cut theorem proves that the maximum network flow and the sum of the cut-edge weights of any minimum cut that separates the source and the sink are equal. There are polynomial-time methods to solve the min-cut problem, notably the Edmonds–Karp algorithm. Maximum cut A cut is maximum if the size of the cut is not smaller than the size of any other cut. The illustration on the right shows a maximum cut: the size of the cut is equal to 5, and there is no cu
https://en.wikipedia.org/wiki/Phenol%20extraction
Phenol extraction is a laboratory technique that purifies nucleic acid samples using a phenol solution. Phenol is common reagent in extraction because its properties allow for effective nucleic acid extraction, particularly as it strongly denatures proteins, it is a nucleic acid preservative, and it is immiscible in water. It may also refer to the process of extracting and isolating phenols from raw materials such as coal tar. These purified phenols are used in many industrial and medical compounds and are used as precursors in some synthesis reactions. Phenol extraction of nucleic acids Phenol extraction is a widely used technique for purifying nucleic acid samples from cell lysates. To obtain nucleic acids, the cell must be lysed, and the nucleic acids separated from other cell components. Phenol is a polar substance with a higher density than water (1.07 g/cm3 compared to water's 1.00 g/cm3). When suspended in a water-phenol solution, denatured proteins and unwanted cell components dissolve in the phenol, while polar nucleic acids dissolve in the water phase. The solution may then be centrifuged to separate the phenol and water into distinct organic and aqueous phases. Purified nucleic acids can be precipitated from the aqueous phase of the solution. Phenol is often used in combination with chloroform. Adding an equal volume of chloroform and phenol ensures a distinct separation between the aqueous and organic phases. Chloroform and phenol are miscible and create a denser solution than phenol alone, aiding the separation of the organic and aqueous layers. This addition of chloroform is useful when removing the aqueous phase to obtain a purified nucleic acid sample. The pH of the solution must be adjusted specifically for each type of extraction. For DNA extraction, the pH is adjusted to 7.0–8.0. For RNA-specific extraction, the pH is adjusted to 4.5. At pH 4.5, hydrogen ions neutralize the negative charges on the phosphate groups, causing DNA to dissolve
https://en.wikipedia.org/wiki/Surface%20gravity
The surface gravity, g, of an astronomical object is the gravitational acceleration experienced at its surface at the equator, including the effects of rotation. The surface gravity may be thought of as the acceleration due to gravity experienced by a hypothetical test particle which is very close to the object's surface and which, in order not to disturb the system, has negligible mass. For objects where the surface is deep in the atmosphere and the radius not known, the surface gravity is given at the 1 bar pressure level in the atmosphere. Surface gravity is measured in units of acceleration, which, in the SI system, are meters per second squared. It may also be expressed as a multiple of the Earth's standard surface gravity, which is equal to In astrophysics, the surface gravity may be expressed as , which is obtained by first expressing the gravity in cgs units, where the unit of acceleration and surface gravity is centimeters per second squared (cm/s2), and then taking the base-10 logarithm of the cgs value of the surface gravity. Therefore, the surface gravity of Earth could be expressed in cgs units as , and then taking the base-10 logarithm ("log g") of 980.665, and we get 2.992 as "log g". The surface gravity of a white dwarf is very high, and of a neutron star even higher. A white dwarf's surface gravity is around 100,000 g () whilst the neutron star's compactness gives it a surface gravity of up to with typical values of order (that is more than 1011 times that of Earth). One measure of such immense gravity is that neutron stars have an escape velocity of around 100,000 km/s, about a third of the speed of light. For black holes, the surface gravity must be calculated relativistically. Relationship of surface gravity to mass and radius In the Newtonian theory of gravity, the gravitational force exerted by an object is proportional to its mass: an object with twice the mass produces twice as much force. Newtonian gravity also follows an inverse squ
https://en.wikipedia.org/wiki/DSTN
DSTN (double super twisted nematic), also known as dual-scan super twisted nematic or simply dual-scan, is an LCD technology in which a screen is divided in half, which are simultaneously refreshed giving faster refresh rate than traditional passive matrix screens. It is an improved form of supertwist nematic display that offers low power consumption but inferior sharpness and brightness compared to TFT screens. History For several years (early 1990s to early 2000s), TFT screens were only found in high-end laptops due to them being more expensive and lower-end laptops offering DSTN screens only. This was at a time when the screen was often the most expensive component of laptops. The price difference between a laptop with DSTN and one with TFT could easily be $400 or more. However, TFT gradually became cheaper and has essentially captured the entire market. DSTN display quality is poor compared to TFT, with visible noise, smearing, much lower contrast and slow response. Such screens are unsuitable for viewing movies or playing video games of any kind.
https://en.wikipedia.org/wiki/Zero%20one%20infinity%20rule
The Zero one infinity (ZOI) rule is a rule of thumb in software design proposed by early computing pioneer Willem van der Poel. It argues that arbitrary limits on the number of instances of a particular type of data or structure should not be allowed. Instead, an entity should either be forbidden entirely, only one should be allowed, or any number of them should be allowed. Although various factors outside that particular software could limit this number in practice, it should not be the software itself that puts a hard limit on the number of instances of the entity. Examples of this rule may be found in the structure of many file systems' directories (also known as folders): 0 – The topmost directory has zero parent directories; that is, there is no directory that contains the topmost directory. 1 – Each subdirectory has exactly one parent directory (not including shortcuts to the directory's location; while such files may have similar icons to the icons of the destination directories, they are not directories at all). Infinity – Each directory, whether the topmost directory or any of its subdirectories, according to the file system's rules, may contain any number of files or subdirectories. Practical limits to this number are caused by other factors, such as space available on storage media and how well the computer's operating system is maintained. In real-world software design, violations of this rule of thumb are common. For example, the FAT16 file system imposes a limit of 65,536 files to a directory. Authorship Van der Poel confirmed that he was the originator of the rule, but Bruce MacLennan has also claimed authorship (in the form "The only reasonable numbers are zero, one and infinity."), writing in 2015 that:
https://en.wikipedia.org/wiki/Signaling%20compression
For data compression, signaling compression, or SigComp, is a compression method designed especially for compression of text-based communication data as SIP or RTSP. SigComp had originally been defined in RFC 3320 and was later updated with RFC 4896. A Negative Acknowledgement Mechanism for Signaling Compression is defined in RFC 4077. The SigComp work is performed in the ROHC working group in the transport area of the IETF. Overview SigComp specifications describe a compression schema that is located in between the application layer and the transport layer (e.g. between SIP and UDP). It is implemented upon a virtual machine configuration which executes a specific set of commands that are optimized for decompression purposes (namely UDVM, Universal Decompressor Virtual Machine). One strong point for SigComp is that the bytecode to decode messages can be sent over SigComp itself, so this allows to use any kind of compression schema given that it is expressed as bytecode for the UDVM. Thus any SigComp compatible device may use compression mechanisms that did not exist when it was released without any firmware change. Additionally, some decoders may be already been standardised, so SigComp may recall that code so it is not needed to be sent over the connection. To assure that a message is decodable the only requirement is that the UDVM code is available, so the compression of messages is executed off the virtual machine, and native code can be used. As an independent system a mechanism to signal the application conversation (e.g. a given SIP session), a compartment mechanism is used, so a given application may have any given number of different, independent conversations, while persisting all the session status (as needed/specified per compression schema and UDVM code). General architecture
https://en.wikipedia.org/wiki/Substructural%20type%20system
Substructural type systems are a family of type systems analogous to substructural logics where one or more of the structural rules are absent or only allowed under controlled circumstances. Such systems are useful for constraining access to system resources such as files, locks, and memory by keeping track of changes of state that occur and preventing invalid states. Different substructural type systems Several type systems have emerged by discarding some of the structural rules of exchange, weakening, and contraction: Ordered type systems (discard exchange, weakening and contraction): Every variable is used exactly once in the order it was introduced. Linear type systems (allow exchange, but neither weakening nor contraction): Every variable is used exactly once. Affine type systems (allow exchange and weakening, but not contraction): Every variable is used at most once. Relevant type systems (allow exchange and contraction, but not weakening): Every variable is used at least once. Normal type systems (allow exchange, weakening and contraction): Every variable may be used arbitrarily. The explanation for affine type systems is best understood if rephrased as “every occurrence of a variable is used at most once”. Ordered type system Ordered types correspond to noncommutative logic where exchange, contraction and weakening are discarded. This can be used to model stack-based memory allocation (contrast with linear types which can be used to model heap-based memory allocation). Without the exchange property, an object may only be used when at the top of the modelled stack, after which it is popped off, resulting in every variable being used exactly once in the order it was introduced. Linear type systems Linear types corresponds to linear logic and ensures that objects are used exactly once. This allows the system to safely deallocate an object after its use, or to design software interfaces that guarantee a resource cannot be used once it has been closed or
https://en.wikipedia.org/wiki/Clinical%20and%20Translational%20Science
Clinical and Translational Science is a bimonthly peer-reviewed open-access medical journal covering translational medicine. It is published by Wiley-Blackwell and is an official journal of the American Society for Clinical Pharmacology and Therapeutics. The journal was established in 2008 and the editor-in-chief is John A. Wagner (Cygnal Therapeutics). Abstracting and indexing The journal is abstracted and indexed in: According to the Journal Citation Reports, its 2020 impact factor is 4.689.