source stringlengths 31 227 | text stringlengths 9 2k |
|---|---|
https://en.wikipedia.org/wiki/OpenGALEN | OpenGALEN is a not-for-profit organisation that provides an open source medical terminology. This terminology is written in a formal language called GRAIL (GALEN Representation And Integration Language) and also distributed in OWL.
Background
The GALEN technologies were developed with research funding provided by the European Community Framework III (GALEN Project) and Framework IV (GALEN-In-Use Project) programmes.
Early phases of the GALEN Programme developed the GRAIL concept modelling language, experimented with different structures for the GALEN Common Reference Model, and, in parallel, tested the usefulness of the approach with a series of clinical demonstrator projects.
Later phases of the GALEN Programme, during the late 1990s, have concentrated on robust implementations of GRAIL and the Terminology Server, development of the GALEN Common Reference Model in both scope and detail, and development of tools and techniques to enable the further development, scaling-up and maintenance of the model. An important additional focus has been in developing tools and techniques with which we can map the information found in existing coding and Medical classification schemes to the GALEN Common Reference Model.
OpenGALEN has been set up as a not-for-profit Dutch Foundation by the universities of Manchester and Nijmegen to make the results of the GALEN projects available to the world.
GALEN Common Reference Model
The GALEN Common Reference Model is the model of medical concepts (or clinical terminology) being built in GRAIL. This model forms the underlying structural foundation for the services provided by a GALEN Terminology Server.
The GALEN Common Reference Model is written in the formal language GRAIL (see below). The GRAIL statements in the model are equivalent with sentences like these:
Ulcer is a kind of inflammatory lesion
The process whose outcome is an ulcer is called ulceration
The stomach is a part of the GI tract
It is sensible to talk about u |
https://en.wikipedia.org/wiki/Ionophore | In chemistry, an ionophore () is a chemical species that reversibly binds ions. Many ionophores are lipid-soluble entities that transport ions across the cell membrane. Ionophores catalyze ion transport across hydrophobic membranes, such as liquid polymeric membranes (carrier-based ion selective electrodes) or lipid bilayers found in the living cells or synthetic vesicles (liposomes). Structurally, an ionophore contains a hydrophilic center and a hydrophobic portion that interacts with the membrane.
Some ionophores are synthesized by microorganisms to import ions into their cells. Synthetic ion carriers have also been prepared. Ionophores selective for cations and anions have found many applications in analysis. These compounds have also shown to have various biological effects and a synergistic effect when combined with the ion they bind.
Classification
Biological activities of metal ion-binding compounds can be changed in response to the increment of the metal concentration, and based on the latter compounds can be classified as "metal ionophores", "metal chelators" or "metal shuttles". If the biological effect is augmented by increasing the metal concentration, it is classified as a "metal ionophore". If the biological effect is decreased or reversed by increasing the metal concentration, it is classified as a "metal chelator". If the biological effect is not affected by increasing the metal concentration, and the compound-metal complex enters the cell, it is classified as a "metal shuttle". The term ionophore (from Greek ion carrier or ion bearer) was proposed by Berton Pressman in 1967 when he and his colleagues were investigating the antibiotic mechanisms of valinomycin and nigericin.
Many ionophores are produced naturally by a variety of microbes, fungi and plants, and act as a defense against competing or pathogenic species. Multiple synthetic membrane-spanning ionophores have also been synthesized.
The two broad classifications of ionophores synthesiz |
https://en.wikipedia.org/wiki/Lexicographically%20minimal%20string%20rotation | In computer science, the lexicographically minimal string rotation or lexicographically least circular substring is the problem of finding the rotation of a string possessing the lowest lexicographical order of all such rotations. For example, the lexicographically minimal rotation of "bbaaccaadd" would be "aaccaaddbb". It is possible for a string to have multiple lexicographically minimal rotations, but for most applications this does not matter as the rotations must be equivalent. Finding the lexicographically minimal rotation is useful as a way of normalizing strings. If the strings represent potentially isomorphic structures such as graphs, normalizing in this way allows for simple equality checking.
A common implementation trick when dealing with circular strings is to concatenate the string to itself instead of having to perform modular arithmetic on the string indices.
Algorithms
The Naive Algorithm
The naive algorithm for finding the lexicographically minimal rotation of a string is to iterate through successive rotations while keeping track of the most lexicographically minimal rotation encountered. If the string is of length , this algorithm runs in time in the worst case.
Booth's Algorithm
An efficient algorithm was proposed by Booth (1980).
The algorithm uses a modified preprocessing function from the Knuth–Morris–Pratt string search algorithm. The failure function for the string is computed as normal, but the string is rotated during the computation so some indices must be computed more than once as they wrap around. Once all indices of the failure function have been successfully computed without the string rotating again, the minimal lexicographical rotation is known to be found and its starting index is returned. The correctness of the algorithm is somewhat difficult to understand, but it is easy to implement.
def least_rotation(s: str) -> int:
"""Booth's lexicographically minimal string rotation algorithm."""
n = len(s)
f = [-1] * (2 |
https://en.wikipedia.org/wiki/Q%20code | The Q-code is a standardised collection of three-letter codes that each start with the letter "Q". It is an operating signal initially developed for commercial radiotelegraph communication and later adopted by other radio services, especially amateur radio. To distinguish the use of a Q-code transmitted as a question from the same Q-code transmitted as a statement, operators either prefixed it with the military network question marker "" (dit dit dah dit dah) or suffixed it with the standard Morse question mark (dit dit dah dah dit dit).
Although Q-codes were created when radio used Morse code exclusively, they continued to be employed after the introduction of voice transmissions. To avoid confusion, transmitter call signs are restricted; no country is ever issued an ITU prefix starting with "Q".
Codes in the range QAA–QNZ are reserved for aeronautical use; QOA–QQZ for maritime use and QRA–QUZ for all services.
"Q" has no official meaning, but it is sometimes assigned a word with mnemonic value, such as "Queen" for example in QFE: Queen's field elevation, or "Query", "Question", "reQuest".
Early development
The original Q-codes were created, circa 1909, by the British government as a "List of abbreviations ... prepared for the use of British ships and coast stations licensed by the Postmaster General". The Q-codes facilitated communication between maritime radio operators speaking different languages, so they were soon adopted internationally. A total of forty-five Q-codes appeared in the "List of Abbreviations to be used in Radio Communications", which was included in the Service Regulations affixed to the Second International Radiotelegraph Convention in London (The Convention was signed on July 5, 1912, and became effective July 1, 1913.)
The following table reviews a sample of the all-services Q-codes adopted by the 1912 convention:
Later use
Over the years the original Q-codes were modified to reflect changes in radio practice. For example, QSW / QSX or |
https://en.wikipedia.org/wiki/Feasible%20region | In mathematical optimization and computer science, a feasible region, feasible set, or solution space is the set of all possible points (sets of values of the choice variables) of an optimization problem that satisfy the problem's constraints, potentially including inequalities, equalities, and integer constraints. This is the initial set of candidate solutions to the problem, before the set of candidates has been narrowed down.
For example, consider the problem of minimizing the function with respect to the variables and subject to and Here the feasible set is the set of pairs (x, y) in which the value of x is at least 1 and at most 10 and the value of y is at least 5 and at most 12. The feasible set of the problem is separate from the objective function, which states the criterion to be optimized and which in the above example is
In many problems, the feasible set reflects a constraint that one or more variables must be non-negative. In pure integer programming problems, the feasible set is the set of integers (or some subset thereof). In linear programming problems, the feasible set is a convex polytope: a region in multidimensional space whose boundaries are formed by hyperplanes and whose corners are vertices.
Constraint satisfaction is the process of finding a point in the feasible region.
Convex feasible set
A convex feasible set is one in which a line segment connecting any two feasible points goes through only other feasible points, and not through any points outside the feasible set. Convex feasible sets arise in many types of problems, including linear programming problems, and they are of particular interest because, if the problem has a convex objective function that is to be maximized, it will generally be easier to solve in the presence of a convex feasible set and any local optimum will also be a global optimum.
No feasible set
If the constraints of an optimization problem are mutually contradictory, there are no points that satisfy al |
https://en.wikipedia.org/wiki/Conidial%20anastomosis%20tubes | Conidial anastomosis tubes (CATs) are cells formed from the conidia (a type of fungal asexual spores) of many filamentous fungi. These cells have a tubular shape and form an anastomosis (bridge) that allows fusion between conidia.
CATs and germ tubes (germination tubes) are some of the specialized hyphae (long cells formed by filamentous fungal species) that are formed by fungal conidia. CATs are morphologically and physiologically distinct from germ tubes and are under separate genetic control.
Germ tubes, produced during conidial germination, are different from CATs because: CATs are thinner, shorter, lack branches, exhibit determinate growth, and home toward each other.
CAT biology is not completely understood. Initially, conidia are induced to form CATs. Once they are formed, they grow homing toward each other, and eventually they fuse. Once fusion occurs, the respective nuclei can pass through the fused CATs from one conidium to the other. These are events of fungal vegetative growth (asexual reproduction) and not sexual reproduction. Part of the CAT fusion (cell fusion) have been shown to be a coordinated behaviour.
The filamentous fungus Neurospora crassa (a bread mould and fungal model organism) produces CATs from conidia and conidial germ tubes. In contrast, the fungal plant pathogen, Colletotrichum lindemuthianum, only produces CATs from conidia and not from germ tubes.
Fusion between these cells seems to be important for some fungi during early stages of colony establishment. The production of these cells has been suggested to occur in 73 different species of fungi. |
https://en.wikipedia.org/wiki/Valeriu%20Rudic | Valeriu Rudic (born 18 February 1947) is a Moldovan microbiologist, chemist, biochemist and pharmacist who was selected member of Academy of Sciences of Moldova.
External links
http://www.asm.md/?go=detalii-membri&n=62&new_language=0
1947 births
Living people
Moldovan pharmacists
Titular members of the Academy of Sciences of Moldova
Recipients of the Order of Honour (Moldova)
Moldovan chemists
Biochemists
Microbiologists |
https://en.wikipedia.org/wiki/Uncharted%20Waters%20%28video%20game%29 | Uncharted Waters is a 1990 video game published by Koei.
Gameplay
Uncharted Waters is a game in which player captains seek great trade wind routes and the major currents. The story is set in the early 16th century. It was released on the PC-88 in May, 1990.
Reception
Dave Arneson reviewed the game for Computer Gaming World, and stated that "in the final analysis, Uncharted Waters is mildly interesting as it stands. When compared to what it could have been, it is more than mildly disappointing. Someone may yet design the "ultimate" game of 16th century exploration, but it isn't here yet." In a 1993 survey of pre 20th-century strategy games the magazine gave the game two stars out of five, stating that while its geography was inaccurate and user interface "could bear improvement", "Game play can be interesting".
Reviews
All Game Guide - 1998
Computer Gaming World - Jun, 1993
GamePro (Dec, 1994)
ASM (Aktueller Software Markt) (Feb, 1993) |
https://en.wikipedia.org/wiki/Synthetic%20setae | Synthetic setae emulate the setae found on the toes of a gecko and scientific research in this area is driven towards the development of dry adhesives. Geckos have no difficulty mastering vertical walls and are apparently capable of adhering themselves to just about any surface. The five-toed feet of a gecko are covered with elastic hairs called setae and the ends of these hairs are split into nanoscale structures called spatulae (because of their resemblance to actual spatulas). The sheer abundance and proximity to the surface of these spatulae make it sufficient for van der Waals forces alone to provide the required adhesive strength. Following the discovery of the gecko's adhesion mechanism in 2002, which is based on van der Waals forces, biomimetic adhesives have become the topic of a major research effort. These developments are poised to yield families of novel adhesive materials with superior properties which are likely to find uses in industries ranging from defense and nanotechnology to healthcare and sport.
Basic principles
Geckos are renowned for their exceptional ability to stick and run on any vertical and inverted surface (excluding Teflon). However gecko toes are not sticky in the usual way like chemical adhesives. Instead, they can detach from the surface quickly and remain quite clean around everyday contaminants even without grooming.
Extraordinary adhesion
The two front feet of a tokay gecko can withstand 20.1 N of force parallel to the surface with 227 mm2 of pad area, a force as much as 40 times the gecko's weight. Scientists have been investigating the secret of this extraordinary adhesion ever since the 19th century, and at least seven possible mechanisms for gecko adhesion have been discussed over the past 175 years. There have been hypotheses of glue, friction, suction, electrostatics, micro-interlocking and intermolecular forces. Sticky secretions were ruled out first early in the study of gecko adhesion since geckos lack glandular tiss |
https://en.wikipedia.org/wiki/Porting | In software engineering, porting is the process of adapting software for the purpose of achieving some form of execution in a computing environment that is different from the one that a given program (meant for such execution) was originally designed for (e.g., different CPU, operating system, or third party library). The term is also used when software/hardware is changed to make them usable in different environments.
Software is portable when the cost of porting it to a new platform is significantly less than the cost of writing it from scratch. The lower the cost of porting software relative to its implementation cost, the more portable it is said to be.
Etymology
The term "port" is derived from the Latin portāre, meaning "to carry". When code is not compatible with a particular operating system or architecture, the code must be "carried" to the new system.
The term is not generally applied to the process of adapting software to run with less memory on the same CPU and operating system.
Software developers often claim that the software they write is portable, meaning that little effort is needed to adapt it to a new environment. The amount of effort actually needed depends on several factors, including the extent to which the original environment (the source platform) differs from the new environment (the target platform), the experience of the original authors in knowing which programming language constructs and third party library calls are unlikely to be portable, and the amount of effort invested by the original authors in only using portable constructs (platform specific constructs often provide a cheaper solution).
History
The number of significantly different CPUs and operating systems used on the desktop today is much smaller than in the past. The dominance of the x86 architecture means that most desktop software is never ported to a different CPU. In that same market, the choice of operating systems has effectively been reduced to three: Micro |
https://en.wikipedia.org/wiki/HUMARA%20assay | HUMARA Assay is one of the most widely used methods to determine the clonal origin of a tumor. The method is based on X chromosome inactivation and it takes the advantage of having different methylation status of a gene called HUMARA (short for human androgen receptor) that is located on X chromosome. Considering the fact that once one X chromosome is inactivated in a cell, all other cells derived from it will have the same X chromosome inactivated, this approach becomes a great tool to differentiate a monoclonal population from a polyclonal one in a female tissue. HUMARA gene, in particular, has three important features that make it highly convenient for the purpose.
1-) The gene is located on X chromosome and it goes through inactivation by methylation in normal embryogenesis of a female infant. The fact that most-but not all-genes on X chromosome undergo inactivation, this feature becomes an important one.
2-) Human Androgen Receptor gene alleles have varying numbers of CAG repeats. Thus, when DNA from a healthy female tissue is amplified by PCR for a specific region of the gene, two separated bands can be seen on the gel.
3-) The region that is amplified by PCR also has certain base orders that make it susceptible to be digested by HpaII (or HhaI) enzyme when it is not methylated. This detail gives the opportunity to researchers to differentiate a methylated allele from the unmethylated one.
Thanks to these qualities of HUMARA gene, clonal origin of any tissue from a female mammalian organism can be determined.
The basic process is performed as the following :
1-) DNA from the tissue is isolated.
2-) The isolated DNA is treated with the suitable enzyme (such as HpaII) in optimal conditions for a suggested amount of time (i.e. overnight).
3-) DNA is cleaned and the certain region of HUMARA gene is amplified by PCR using "suitable" primers (as an example, please see:Ref.2)
4-) After running PCR products through a gel, the gel is visualized and the result |
https://en.wikipedia.org/wiki/Culm%20%28botany%29 | A culm is the aerial (above-ground) stem of a grass or sedge. It is derived from Latin 'stalk', and it originally referred to the stem of any type of plant.
In horticulture or agriculture, it is especially used to describe the stalk or woody stems of bamboo, cane or grain grasses.
Malting
In the production of malted grains, the culms refer to the rootlets of the germinated grains. The culms are normally removed in a process known as "deculming" after kilning when producing barley malt, but form an important part of the product when making sorghum or millet malt. These culms are very nutritious and are sold off as animal feed. |
https://en.wikipedia.org/wiki/Electrocution | Electrocution is death or severe injury caused by electric shock from electric current passing through the body. The word is derived from "electro" and "execution", but it is also used for accidental death.
The term "electrocution" was coined in 1889 in the US just before the first use of the electric chair and originally referred to only electrical execution and not other electrical deaths. However, since no English word was available for non-judicial deaths due to electric shock, the word "electrocution" eventually took over as a description of all circumstances of electrical death from the new commercial electricity.
Origins
In the Netherlands in 1746 Pieter van Musschenbroek's lab assistant, Andreas Cuneus, received an extreme shock while working with a leyden jar, the first recorded injury from human-made electricity. By the mid-19th century high-voltage electrical systems came into use to power arc lighting for theatrical stage lighting and lighthouses leading to the first recorded accidental death in 1879 when a stage carpenter in Lyon, France, touched a 250-volt wire.
The spread of arc light–based street lighting systems (which at the time ran at a voltage above 3,000 volts) after 1880 led to many people dying from coming in contact with these high-voltage lines, a strange new phenomenon which seemed to kill instantaneously without leaving a mark on the victim. This would lead to execution by electricity in the electric chair in the early 1890s as an official method of capital punishment in the U.S. state of New York, thought to be a more humane alternative to hanging. After an 1881 death in Buffalo, New York, caused by a high-voltage arc lighting system, Alfred P. Southwick sought to develop this phenomenon into a way to execute condemned criminals. Southwick, a dentist, based his device on the dental chair.
The next nine years saw a promotion by Southwick, the New York state Gerry commission (which included Southwick) recommending execution by elect |
https://en.wikipedia.org/wiki/Synchronism | Synchronism may refer to:
Synchronism (Davidovsky), compositions by Argentine-American composer Mario Davidovsky incorporating acoustic instruments and electroacoustic sounds
Chronological synchronism, an event that links two chronologies such as historical and datable astronomical events
Synchronization, the coordination of events to operate a system in unison
Film
Synchronized sound, film sound technologically coupled to image
Post-synchronization, the process of re-recording dialogue after the filming process
See also
Synchromism an early 20th-century art movement, commonly misspelled as "synchronism"
Synchronicity (disambiguation)
Synchronizer (disambiguation)
Synchrony (disambiguation)
Synchronization |
https://en.wikipedia.org/wiki/SAO%20206462 | SAO 206462 is a young star, surrounded by a circumstellar disc of gas and clearly defined spiral arms. It is situated about 440 light-years away from Earth in the constellation Lupus. The presence of these spiral arms seems to be related to the existence of planets inside the disk of gas surrounding the star. The disk's diameter is about twice the size of the orbit of Pluto.
Discovery
The discovery of this object was presented in October 2011 by Carol Grady, astronomer of Eureka Scientific, headquartered in the Goddard Space Flight Center at NASA. It was the first of this class that exhibited a high degree of clarity and was made using several space telescopes (Hubble, FUSE, Spitzer) and earth telescopes (Gemini Observatory and Subaru Telescope, situated in Hawaii), through an international research program of young stars and of stars with planets. A number of astronomers of different observatories collaborated. |
https://en.wikipedia.org/wiki/Prospects%20Course%20Exchange | Prospects Course Exchange is a system that manages XCRI-CAP feeds, enabling course data from higher education providers to be visible through Prospects' postgraduate course search.
It is run and operated by Graduate Prospects.
Prospects Course Check is a free course validation checker also provided by the service.
See also
XCRI |
https://en.wikipedia.org/wiki/Deimatic%20behaviour | Deimatic behaviour or startle display means any pattern of bluffing behaviour in an animal that lacks strong defences, such as suddenly displaying conspicuous eyespots, to scare off or momentarily distract a predator, thus giving the prey animal an opportunity to escape. The term deimatic or dymantic originates from the Greek δειματόω (deimatóo), meaning "to frighten".
Deimatic display occurs in widely separated groups of animals, including moths, butterflies, mantises and phasmids among the insects. In the cephalopods, different species of octopuses, squids, cuttlefish and the paper nautilus are deimatic.
Displays are classified as deimatic or aposematic by the responses of the animals that see them. Where predators are initially startled but learn to eat the displaying prey, the display is classed as deimatic, and the prey is bluffing; where they continue to avoid the prey after tasting it, the display is taken as aposematic, meaning the prey is genuinely distasteful. However, these categories are not necessarily mutually exclusive. It is possible for a behaviour to be both deimatic and aposematic, if it both startles a predator and indicates the presence of anti-predator adaptations.
Vertebrates including several species of frog put on warning displays; some of these species have poison glands. Among the mammals, such displays are often found in species with strong defences, such as in foul-smelling skunks and spiny porcupines. Thus these displays in both frogs and mammals are at least in part aposematic.
In insects
Deimatic displays are made by insects including the praying mantises (Mantodea) and stick insects (Phasmatodea). While undisturbed, these insects are usually well camouflaged. When disturbed by a potential predator, they suddenly reveal their hind wings, which are brightly coloured. In mantises, the wing display is sometimes reinforced by showing brightly coloured front legs, and accompanied by a loud hissing sound created by stridulation. For ex |
https://en.wikipedia.org/wiki/Nucleic%20acid%20test | A nucleic acid test (NAT) is a technique used to detect a particular nucleic acid sequence and thus usually to detect and identify a particular species or subspecies of organism, often a virus or bacterium that acts as a pathogen in blood, tissue, urine, etc. NATs differ from other tests in that they detect genetic materials (RNA or DNA) rather than antigens or antibodies. Detection of genetic materials allows an early diagnosis of a disease because the detection of antigens and/or antibodies requires time for them to start appearing in the bloodstream. Since the amount of a certain genetic material is usually very small, many NATs include a step that amplifies the genetic material—that is, makes many copies of it. Such NATs are called nucleic acid amplification tests (NAATs). There are several ways of amplification, including polymerase chain reaction (PCR), strand displacement assay (SDA), or transcription mediated assay (TMA).
Virtually all nucleic acid amplification methods and detection technologies use the specificity of Watson-Crick base pairing; single-stranded probe or primer molecules capture DNA or RNA target molecules of complementary strands. Therefore, the design of probe strands is highly significant to raise the sensitivity and specificity of the detection. However, the mutants which form the genetic basis for a variety of human diseases are usually slightly different from the normal nucleic acids. Often, they are only different in a single base, e.g., insertions, deletions, and single-nucleotide polymorphisms (SNPs). In this case, imperfect probe-target binding can easily occur, resulting in false-positive outcomes such as mistaking a strain that is commensal for one that is pathogenic. Much research has been dedicated to achieving single-base specificity.
Advances
Nucleic acid (DNA and RNA) strands with corresponding sequences stick together in pairwise chains, zipping up like Velcro tumbled in a clothes dryer. But each node of the chain is not v |
https://en.wikipedia.org/wiki/Baiting%20crowd | Baiting crowd is a form of collective aggression; this is a situation where a group of individuals unify their aggression, while they may not even know each other, towards another individual or group.
Deindividuation
It is a typical situation where a person is about to jump from a high building while there are people standing below, whereas some people begin to shout that the person has to jump off.
Leon Mann (born 12 December 1937) is an Australian psychologist. He is currently Director of the Research Leadership Program and coordinator of the University of Melbourne's Mentoring Program for Research Leaders.
In the late 20th century, Leon Mann became interested in the idea of deindividuation. More specifically, he wanted to investigate how aggression of a group of people towards another individual or group, could influence one’s own behaviour. In the attempt to investigate this matter, Mann analysed 166 cases of suicide or suicide attempt, in which in 21 cases a crowd was involved. He wanted to find out when collective aggression towards an individual, who was about to commit suicide (e.g. encouraging to jump of a large building), would cause an outsider to join the baiting crowd in this process of deindividuation. In ten of these 21 cases several factors may have led to deindividuation where baiting occurred.
He found out that when the crowd was small and during daytime, people would usually not shout that the person should jump. However, when the crowd was large and it was late at night it caused people to be anonymous, therefore they would more quickly shout that the person had to jump from the building.
This is caused through deindividuation which is a process whereby people lose their sense of socialized individual identity and often show unsocialized and even antisocial behavior. In this situation people often blur their normal behaviors, leading to an increase in impulsive and deviant behavior. One becomes less an individual and more part of the mass.
|
https://en.wikipedia.org/wiki/Yota | Yota () is a Russian mobile phone brand and mobile broadband manufacturer. Yota is a trademark of Skartel LLC.
On May 9, 2012, Yota's WiMAX was replaced by its LTE network. In September 2012, 4G networks were launched in the Russian cities of Novosibirsk, Krasnodar, Moscow, Sochi, Samara, Vladivostok, Ufa, Kazan and St. Petersburg.
Garsdale Services Investment Ltd. owns 100% of Yota's shares and 50% of MegaFon's shares. In turn, Garsdale is controlled by AF Telecom (82%), Telconet Capital (13.5%), and the Russian Technologies State Corporation (4.5%).
History
In 2006, the co-owner of the St. Petersburg company Korus, Denis Sverdlov and Bulgarian businessman Sergey Adonev, established the first provider of WiMAX, a new data transfer technology. In 2006, WiMAX was used in China, India, Indonesia, Taiwan and the United States. In 2008, Skartel was the first company in Russia to deploy WiMAX standard network in Moscow and St Petersburg in a range of 2.5-2.7 GHz. In 2010 Yota announced its plans to launch LTE on its network. The first test of the new standard network took place in Kazan, on 30 August 2010. The subscribers gained access to the Internet at a rate of 20-30 Mbit/s. About 150 base stations have been installed in Kazan. The investments into the LTE network deployment constituted $20 million. The fourth generation LTE network, which was tested by the Yota provider in Kazan, was switched off the next day. At that point, Yota had no interest to use a 4G standard network in a commercial or test mode.
In April 2019, Yota filed for bankruptcy. The bankruptcy stemmed from a lawsuit filed against the company by its contracted manufacturer, Hi-P Singapore.
Long Term Evolution (LTE)
Officially, Novosibirsk was the first Russian city where the LTE network was deployed, commercially launched on December 22, 2011. Then this new format of communications was adopted in Krasnodar (29 April 2012), Moscow (10 May 2012), and Sochi (11 May 2012). Samara was connected to L |
https://en.wikipedia.org/wiki/Two-way%20satellite%20time%20and%20frequency%20transfer | Two-way satellite time and frequency transfer (TWSTFT) is a high-precision long distance time and frequency transfer mechanism used between time bureaux to determine and distribute time and frequency standards.
TWSTFT is being evaluated as an alternative to be used by the Bureau International des Poids et Mesures in the determination of International Atomic Time (TAI), as a complement to the current standard method of simultaneous observations of GPS transmissions.
External links
TWSTFT page at the National Physical Laboratory
TWSTFT page at the Physikalisch-Technische Bundesanstalt
NIST TWSTFT page
TWSTFT page at the US Naval Observatory
Time
Telecommunications techniques
Synchronization |
https://en.wikipedia.org/wiki/Delta%20update | A delta update is a software update that requires the user to download only those parts of the software's code that are new, or have been changed from their previous state, in contrast to having to download the entire program. The use of delta updates can save significant amounts of time and computing bandwidth. The name "delta" derives from the mathematical science use of the Greek letter delta, Δ or δ to denote change.
Uses
Linux
Fedora Linux has supported binary delta updates by default using the yum presto plugin since June 2009. This is based on RPM Package Manager's deltarpm system (2004), which was in turn based on bsdiff. This functionality has been inherited by Fedora-derived operating systems, including RedHat Enterprise Linux and its variant, CentOS. OpenSUSE also uses deltarpms with its zypper manager. A more primitive system, the SUSE patchrpm, worked by replacing changed files.
A similar system for the dpkg-APT package manager system of Debian is debdelta (2006); despite an apparent halt on the homepage, its package repository as well as the source code remains actively maintained. Debdelta is not installed by default and not many mirrors have been set up for it. A member of the developer team has proposed yet another format that integrates directly into the currently mirrored main repositories called patch debs in 2018. It is intended to have more integrity checks.
A descendant of Debian, Ubuntu developers have tried many times to implement delta updates for their system. During 2006 they tried to create one, but were confronted with too many options and dropped the efforts. In 2011 they tried to just set up debdelta, but once again dropped the efforts in May of the same year.
The Arch Linux package manager pacman used to support a form of delta updating using VCDIFF (xdelta). It was scrapped due to an arbitrary command execution vulnerability () due to a lack of string escaping.
Windows
Windows Update has supported delta updates since Win |
https://en.wikipedia.org/wiki/Operative%20temperature | Operative temperature () is defined as a uniform temperature of an imaginary black enclosure in which an occupant would exchange the same amount of heat by radiation plus convection as in the actual nonuniform environment. Some references also use the terms 'equivalent temperature" or 'effective temperature' to describe combined effects of convective and radiant heat transfer. In design, operative temperature can be defined as the average of the mean radiant and ambient air temperatures, weighted by their respective heat transfer coefficients. The instrument used for assessing environmental thermal comfort in terms of operative temperature is called a eupatheoscope and was invented by A. F. Dufton in 1929. Mathematically, operative temperature can be shown as;
where,
= convective heat transfer coefficient
= linear radiative heat transfer coefficient
= air temperature
= mean radiant temperature
Or
where,
= air velocity
and have the same meaning as above.
It is also acceptable to approximate this relationship for occupants engaged in near sedentary physical activity (with metabolic rates between 1.0 met and 1.3 met), not in direct sunlight, and not exposed to air velocities greater than 0.10 m/s (20 fpm).
where and have the same meaning as above.
Application
Operative temperature is used in heat transfer and thermal comfort analysis in transportation and buildings. Most psychrometric charts used in HVAC design only show the dry bulb temperature on the x-axis(abscissa), however, it is the operative temperature which is specified on the x-axis of the psychrometric chart illustrated in ANSI/ASHRAE Standard 55 – Thermal Environmental Conditions for Human occupancy.
See also
HVAC
Psychrometrics
Underfloor heating
ASHRAE |
https://en.wikipedia.org/wiki/Miniature%20Inverted-repeat%20Transposable%20Elements | Miniature Inverted-repeat Transposable Elements (MITEs) are a group of non-autonomous Class II transposable elements (DNA sequences). Being non-autonomous, MITEs cannot code for their own transposase. They exist within the genomes of animals, plants, fungi, bacteria and even viruses. MITEs are generally short (50 to 500 bp) elements with terminal inverted repeats (TIRs; 10–15 bp) and two flanking target site duplications (TSDs). Like other transposons, MITEs are inserted predominantly in gene-rich regions and this can be a reason that they affect gene expression and play important roles in accelerating eukaryotic evolution. Their high copy number in spite of small sizes has been a topic of interest.
Origin of MITEs
A detailed study of MITEs reveals that MITE subfamilies have arisen from related autonomous elements from a single genome and these subfamilies constitute the MITE families. One type of autonomous element can give rise to one or more MITE families.
Classification
Based on their relations in sequences of TIRs with known TE superfamilies, MITEs have been classified into certain families. For example, wTourist, Acrobat, Hearthealer are MITE families in some plant species are under the TE superfamily PIF/Harbinger. Stowaway is a MITE family in Pisum sativum L. with TSD TA in relation to Tc1/mariner TE superfamily. A group of MITEs known as CMITES related to Piggybac superfamily were found in certain coral species.
While most of the MITEs are grouped, some of them are yet to be allotted their TE superfamilies. Such families include AtATE in Arabidopsis thaliana and ATon family found in Aedes aegypti. Besides this, many more MITE families are likely to be discovered.
MITEs in Plant Genomes
MITEs were first discovered in plants. Elements belonging to the CACTA, hAT, Mutator, PIF, and Tc1/Mariner superfamilies have been described. Depending upon the similarity of their terminal inverted repeats and target site duplications, most of the MITEs in plant geno |
https://en.wikipedia.org/wiki/SystemC%20AMS | SystemC AMS is an extension to SystemC for analog, mixed-signal and RF functionality. The SystemC AMS 2.0 standard was released on April 6, 2016 as IEEE Std 1666.1-2016.
Language specification
ToDo: description
Language features
ToDo: description
MoC - Model of Computation
A model of computation (MoC) is a set of rules defining the behavior and interaction between SystemC AMS primitive
modules. SystemC AMS defines the following models of computation: timed data flow (TDF), linear signal flow
(LSF) and electrical linear networks (ELN).
TDF - Timed Data Flow
In the timed data flow (TDF) model, components exchange analogue values with each other
on a periodic basis at a chosen sampling rate, such as every 10 microseconds.
By the sampling theorem, this would be sufficient to convey
signals of up to 50 MHz bandwidth without aliasing artefacts.
A TDF model defines a method called `processing()' that is invoked
at the appropriate rate as simulation time advances.
A so-called cluster of models share a static schedule of when they should communicate.
This sets the relative ordering of the calls to the processing() methods
of each TDF instance in the cluster.
The periodic behaviour of TDF allows it to operate independently of the main
SystemC event-driven kernel used for digital logic.
ELN - Electrical Linear Networks
The SystemC electrical linear networks (ELN) library provides a set of
standard electrical components that enable SPICE-like
simulations to be run. The three basic components, resistors, capacitors and
inductors are, of course, available. Further voltage-controlled variants, such as a transconductance
amplifier (voltage-controlled current generator) enable most FET and other
semiconductor models to be readily created.
Current flowing in ELN networks of resistors can be solved with a suitable simultaneous equation solver.
These are called the nodal equations.
Where time-varying components, such as capacitors and inductors are included, Euler |
https://en.wikipedia.org/wiki/Richmann%27s%20law | Richmann's law, sometimes referred to as Richmann's rule, Richmann's mixing rule, Richmann's rule of mixture or Richmann's law of mixture, is a physical law for calculating the mixing temperature when pooling multiple bodies. It is named after the Baltic German physicist Georg Wilhelm Richmann, who published the relationship in 1750, establishing the first general equation for calorimetric calculations.
Origin
Through experimental measurements, Wilhelm Richmann determined that the following relationship holds when water of different temperatures is mixed:
It follows:
Here and are the masses of the two mixture components, and are their respective initial temperatures, and is the mixture temperature.
This observation is called Richmann's law in the narrower sense and applies in principle to all substances of the same state of aggregation. According to this, the mixing temperature is the weighted arithmetic mean of the temperatures of the two initial components.
Richmann's rule of mixing can also be applied in reverse, for example, to the question of the ratio in which quantities of water of given temperatures must be mixed to obtain water of a desired temperature. Determining the quantities and required for this purpose, given a total quantity , is accomplished with the mixing cross. The corresponding formula, obtained from the above equation by rearrangement, is:
or .
For the mixing ratio, this gives:
.
The physical background of the mixing rule is the fact that the heat energy of a substance is directly proportional to its mass and its absolute temperature. The proportionality factor is the specific heat capacity, which depends on the nature of the substance, but which was not described until some time after Richmann's discovery by Joseph Black. Thus, the validity of the formula is limited to mixtures of the same substance, since it assumes a uniform specific heat capacity. Another condition is that both components be uniformly warm everywhere a |
https://en.wikipedia.org/wiki/T-box%20transcription%20factor%20T | T-box transcription factor T, also known as Brachyury protein, is encoded for in humans by the TBXT gene. Brachyury functions as a transcription factor within the T-box family of genes. Brachyury homologs have been found in all bilaterian animals that have been screened, as well as the freshwater cnidarian Hydra.
History
The brachyury mutation was first described in mice by Nadezhda Alexandrovna Dobrovolskaya-Zavadskaya in 1927 as a mutation that affected tail length and sacral vertebrae in heterozygous animals. In homozygous animals the brachyury mutation is lethal at around embryonic day 10 due to defects in mesoderm formation, notochord differentiation and the absence of structures posterior to the forelimb bud (Dobrovolskaïa-Zavadskaïa, 1927). The name brachyury comes from the Greek brakhus meaning short and oura meaning tail.
In 2018 HGNC updated the human gene name from T to TBXT, presumably to overcome difficulties associated with searching for a single letter gene symbol. The mouse gene has been changed to Tbxt.
Tbxt was cloned by Bernhard Herrmann and colleagues and proved to encode a 436 amino acid embryonic nuclear transcription factor. Tbxt binds to a specific DNA element, a near palindromic sequence TCACACCT through a region in its N-terminus, called the T-box. Tbxt is the founding member of the T-box family which in mammals currently consists of 18 T-box genes.
The crystal structure of the human brachyury protein was solved in 2017 by Opher Gileadi and colleagues at the Structural Genomics Consortium in Oxford.
Role in development
The gene brachyury appears to have a conserved role in defining the midline of a bilaterian organism, and thus the establishment of the anterior-posterior axis; this function is apparent in chordates and molluscs.
Its ancestral role, or at least the role it plays in the Cnidaria, appears to be in defining the blastopore. It also defines the mesoderm during gastrulation. Tissue-culture based techniques have demonstra |
https://en.wikipedia.org/wiki/Baited%20remote%20underwater%20video | Baited remote underwater video (BRUV) is a system used in marine biology research. By attracting fish into the field of view of a remotely controlled camera, the technique records fish diversity, abundance and behaviour of species. Sites are sampled by video recording the region surrounding a baited canister which is lowered to the bottom from a surface vessel or less commonly by a submersible or remotely operated underwater vehicle. The video can be transmitted directly to the surface by cable, or recorded for later analysis.
Baited cameras are highly effective at attracting scavengers and subsequent predators, and are a non-invasive method of generating relative abundance indices for a number of marine species.
As a non-extractive technique, it offers a low environmental impact way of understanding changes in fish numbers and diversity over time. BRUV surveys were developed in Australia, and are now used around the world for a variety of projects. This is a low budget monitoring system that is less reliant on the availability of skilled labour and may make sustainable monitoring more practical, over the long term.
There are two main types of remote video technique which have been used to record reef fish populations. They can both be left free standing without the need of an operator. The first system uses one downward looking camera (D-BRUV), and the other uses either one (mono) or two (stereo) horizontally facing cameras (H-BRUV), and may use underwater lighting to illuminate the target area. Stereo BRUV recordings can use software analysis to determine the size of specimens.
The colour of the lighting used for video may influence behaviour of the target species. |
https://en.wikipedia.org/wiki/Bastiaan%20Quast | Bastiaan Quast is a Dutch Machine learning researcher. He is the author and lead maintainer of the open-source rnn and transformer deep-learning frameworks in the R programming language, and the datasets.load GUI package, as well as R packages on Global Value Chain decomposition & WIOD and on Regression Discontinuity Design. Quast is a great-great-grandson of the Nobel Peace Prize laureate Tobias Asser.
Early life and education
Bastiaan Quast graduated from University of Groningen with a bachelor's degree in Economics and bachelor's degree in Theoretical philosophy. He holds a master's degree in Econometrics from the University of St. Gallen He obtained his Ph.D from the Graduate Institute of International and Development Studies with advisors Richard Baldwin and Jean-Louis Arcand, his work on local languages and internet usage was discussed at the 2017 G20 meeting in Germany.
Career
Quast is an functionary of the United Nations at the International Telecommunication Union, as Secretary of the ITU-WHO Focus Group on Artificial Intelligence for Health and AI for Good.
Bastiaan Quast created the popular machine learning framework rnn in R, which allows native implementations of recurrent neural network architectures, such as LSTM and GRU (>100,000 downloads). While working at UNCTAD, Quast developed the popular package datasets.load, which is part of the top 10% of most downloaded R packages (>100,000). The R packages decompr and wiod have been downloaded >20,000 times.
Bibliography
Kummritz, Victor; Quast, Bastiaan (2017). Global value chains in developing economies. London, United Kingdom: VoxEU. |
https://en.wikipedia.org/wiki/Blom%27s%20scheme | Blom's scheme is a symmetric threshold key exchange protocol in cryptography. The scheme was proposed by the Swedish cryptographer Rolf Blom in a series of articles in the early 1980s.
A trusted party gives each participant a secret key and a public identifier, which enables any two participants to independently create a shared key for communicating. However, if an attacker can compromise the keys of at least k users, they can break the scheme and reconstruct every shared key. Blom's scheme is a form of threshold secret sharing.
Blom's scheme is currently used by the HDCP (Version 1.x only) copy protection scheme to generate shared keys for high-definition content sources and receivers, such as HD DVD players and high-definition televisions.
The protocol
The key exchange protocol involves a trusted party (Trent) and a group of users. Let Alice and Bob be two users of the group.
Protocol setup
Trent chooses a random and secret symmetric matrix over the finite field , where p is a prime number. is required when a new user is to be added to the key sharing group.
For example:
Inserting a new participant
New users Alice and Bob want to join the key exchanging group. Trent chooses public identifiers for each of them; i.e., k-element vectors:
.
For example:
Trent then computes their private keys:
Using as described above:
Each will use their private key to compute shared keys with other participants of the group.
Computing a shared key between Alice and Bob
Now Alice and Bob wish to communicate with one another. Alice has Bob's identifier and her private key .
She computes the shared key , where denotes matrix transpose. Bob does the same, using his private key and her identifier, giving the same result:
They will each generate their shared key as follows:
Attack resistance
In order to ensure at least k keys must be compromised before every shared key can be computed by an attacker, identifiers must be k-linearly independent: all sets of k randomly |
https://en.wikipedia.org/wiki/Mycoplasma%20haemocanis | Mycoplasma haemocanis (formerly Haemobartonella canis) is a species of bacteria in the genus Mycoplasma. It rarely causes anemia in dogs with normal spleens and normal immune systems. Clinical anemia can develop when a carrier dog is splenectomized, or when a splenectomized dog is transfused with blood from a carrier donor. It affects many species of canids like dogs and foxes. |
https://en.wikipedia.org/wiki/Curve%20of%20growth | In astronomy, the curve of growth describes the equivalent width of a spectral line as a function of the column density of the material from which the spectral line is observed. |
https://en.wikipedia.org/wiki/Cofinal%20%28mathematics%29 | In mathematics, a subset of a preordered set is said to be cofinal or frequent in if for every it is possible to find an element in that is "larger than " (explicitly, "larger than " means ).
Cofinal subsets are very important in the theory of directed sets and nets, where “cofinal subnet” is the appropriate generalization of "subsequence". They are also important in order theory, including the theory of cardinal numbers, where the minimum possible cardinality of a cofinal subset of is referred to as the cofinality of
Definitions
Let be a homogeneous binary relation on a set
A subset is said to be or with respect to if it satisfies the following condition:
For every there exists some that
A subset that is not frequent is called .
This definition is most commonly applied when is a directed set, which is a preordered set with additional properties.
Final functions
A map between two directed sets is said to be if the image of is a cofinal subset of
Coinitial subsets
A subset is said to be (or in the sense of forcing) if it satisfies the following condition:
For every there exists some such that
This is the order-theoretic dual to the notion of cofinal subset.
Cofinal (respectively coinitial) subsets are precisely the dense sets with respect to the right (respectively left) order topology.
Properties
The cofinal relation over partially ordered sets ("posets") is reflexive: every poset is cofinal in itself. It is also transitive: if is a cofinal subset of a poset and is a cofinal subset of (with the partial ordering of applied to ), then is also a cofinal subset of
For a partially ordered set with maximal elements, every cofinal subset must contain all maximal elements, otherwise a maximal element that is not in the subset would fail to be any element of the subset, violating the definition of cofinal. For a partially ordered set with a greatest element, a subset is cofinal if and only if it contains that grea |
https://en.wikipedia.org/wiki/Pearcey%20integral | In mathematics, the Pearcey integral is defined as
The Pearcey integral is a class of canonical diffraction integrals, often used in wave propagation and optical diffraction problems The first numerical evaluation of this integral was performed by Trevor Pearcey using the quadrature formula.
In optics, the Pearcey integral can be used to model diffraction effects at a cusp caustic, which corresponds to the boundary between two regions of geometric optics: on one side, each point is contained in three light rays; on the other side, each point is contained in one light ray. |
https://en.wikipedia.org/wiki/Cumulina | Cumulina (October 3, 1997 - May 5, 2000), a mouse, was the first animal cloned from adult cells that survived to adulthood. She was cloned using the Honolulu technique developed by "Team Yana", the Ryuzo Yanagimachi research group at the former campus of the John A. Burns School of Medicine located at the University of Hawai'i at Mānoa. Cumulina was a brown Mus musculus or common house mouse. She was named after the cumulus cells surrounding the developing oocyte in the ovarian follicle in mice. Nuclei from these cells were put into egg cell devoid of their original nuclei in the Honolulu cloning technique. All other mice produced by the Yanagimachi lab are just known by a number.
Cumulina was able to produce two healthy litters. She was retired after the second.
Cumulina's preserved remains can be visited at the Institute for Biogenesis Research, a part of the John A. Burns School of Medicine laboratory, in Honolulu, Hawaii.
Some of her descendants have been displayed at the Bishop Museum and the Museum of Science and Industry in Chicago, Illinois. |
https://en.wikipedia.org/wiki/Emery%27s%20rule | In 1909, the entomologist Carlo Emery noted that social parasites among insects (e.g., kleptoparasites) tend to be parasites of species or genera to which they are closely related. Over time, this pattern has been recognized in many additional cases, and generalized to what is now known as Emery's rule. The pattern is best known for various taxa of Hymenoptera. For example, the social wasp Dolichovespula adulterina parasitizes other members of its genus such as Dolichovespula norwegica and Dolichovespula arenaria. Emery's rule is also applicable to members of other kingdoms such as fungi, red algae, and mistletoe. The significance and general relevance of this pattern are still a matter of some debate, as a great many exceptions exist, though a common explanation for the phenomenon when it occurs is that the parasites may have started as facultative parasites within the host species itself (such forms of intraspecific parasitism are well-known, even in some species of bees), but later became reproductively isolated and split off from the ancestral species, a form of sympatric speciation.
When a parasitic species is a sister taxon to its host in a phylogenetic sense, the relationship is considered to be in "strict" adherence to Emery's rule. When the parasite is a close relative of the host but not its sister species, the relationship is in "loose" adherence to the rule. |
https://en.wikipedia.org/wiki/Ericsson%20Radio%20Systems | Ericsson Radio Systems AB was the name of a wholly owned subsidiary in the Ericsson sphere, founded on 1 January 1983 by buying out all former owners of Svenska Radioaktiebolaget (SRA). The company was well known in Scandinavia and elsewhere in the 1980s, as it was deploying NMT systems and developing a line of mobile telephones under the brand name Hotline. In 2002 the subsidiary changed its name to simply Ericsson AB and absorbed 19 other legal entities in the Ericsson sphere, but kept its company registration number with the Swedish state, so it is still the same legal entity as Ericsson Radio Systems. The merge of the smaller legal entities was done to cut down operating costs. The main activity within Ericsson AB is infrastructure for mobile telephony.
History
1 January 1983, the company became a wholly owned subsidiary under the name Ericsson Radio Systems AB. This year the company had 3998 employees and described its business as advanced wireless communications for civilian use. The company was organized in the following areas:
Mobile telephony, which also becomes the most expansive part of the business from 1983 and forward.
Mobile radio, in 1983 roughly 30% of the billing with products like police radio systems and radiotelephones for taxis. By the mid 1980s this business goes stagnant.
Pagers, in cooperation with the Dutch subsidiary NIRA.
Microwave communications, especially dealing with a government-backed project called Tele-X. This area changed its name to Radiotransmission in 1984 and was transferred to Ericssons business unit for defense products in the end of 1985.
29 October 1983, the company created The Ericsson Mobile Telephone Laboratory in the science park Ideon in Lund, Sweden, and all research and development of mobile telephones was transferred there. The former development site for mobile phones in Gävle, tracing its roots back to Sonab mobiltelefon AB and as far back as AGA mobilradio was decommissioned, and none of the original d |
https://en.wikipedia.org/wiki/Nanolinux | NanoLinux
is an open source, free and very lightweight Linux distribution that requires only 14 MB of disk space including tiny versions of the most common desktop applications and several games. It is based on the Core version of the Tiny Core Linux
distribution and uses Busybox, Nano-X instead of X.Org, FLTK 1.3.x as the default GUI toolkit, and SLWM (super-lightweight window manager). The included applications are mainly based on FLTK.
Applications included in the distribution
Nanolinux includes several lightweight applications, including:
Dillo graphical web browser
FlWriter word processor
Sprsht spreadsheet application
FLTDJ personal information manager
AntiPaint painting application
Fluff file manager
NXterm terminal emulator
Flcalc calculator
FlView image viewer
Fleditor text editor
FlChat IRC client
FlMusic CD player
FlRadio internet radio
Webserver, mount tool, system statistics, package install utility.
The distribution also includes several games, such as Tuxchess, Checkers, NXeyes, Mastermind, Sudoku and Blocks.
Support for TrueType fonts and UTF-8 is also provided. Nanolinux is distributed as Live CD ISO images, installation on flash disk
and hard disk
is documented on its Wiki pages.
System requirements
Minimal configuration:
The Live CD version without swapfile requires 64 MB of RAM and 14 MB of disk space.
See also
Comparison of Linux live distributions
Lightweight Linux distribution
List of Linux distributions that run from RAM |
https://en.wikipedia.org/wiki/Field%20specification | A field specification or fspec defines a portion of a word in some programming language. It has the form "(L:R)" where "L" is the leftmost byte and "R" is the rightmost byte, and counting begins at zero. For example, the first three bytes of a word would have the fspec (0:2) for bytes numbered 0, 1, and 2.
Units of information |
https://en.wikipedia.org/wiki/Marine%20clay | Marine clay is a type of clay found in coastal regions around the world. In the northern, deglaciated regions, it can sometimes be quick clay, which is notorious for being involved in landslides.
Marine clay is a particle of soil that is dedicated to a particle size class, this is usually associated with USDA's classification with sand at 0.05mm, silt at 0.05-.002mm and clay being less than 0.002 mm in diameter. Paired with the fact this size of particle was deposited within a marine system involving the erosion and transportation of the clay into the ocean.
Soil particles become suspended when in a solution with water, with sand being affected by the force of gravity first with suspended silt and clay still floating in solution. This is also known as turbidity, in which floating soil particles create a murky brown color to a water solution. These clay particles are then transferred to the abyssal plain in which they are deposited in high percentages of clay.
Once the clay is deposited on the ocean floor it can change its structure through a process known as flocculation, process by which fine particulates are caused to clump together or floc. These can be either edge to edge flocculation or edge to face flocculation. Relating to individual clay particles interacting with each other. Clays can also be aggregated or shifted in their structure besides being flocculated.
Particles configurations
Clay particles can self-assemble into various configurations, each with totally different properties.
This change in structure to the clay particles is due to a swap in cations with the basic structure of a clay particle. This basic structure of the clay particle is known as a silica tetrahedral or aluminum octahedral. They are the basic structure of clay particles composing of one cation, usually silica or aluminum surrounded by hydroxide anions, these particles form in sheets forming what we know as clay particles and have very specific properties to them including m |
https://en.wikipedia.org/wiki/Peptide%20transporter%201 | Peptide transporter 1 (PepT 1) also known as solute carrier family 15 member 1 (SLC15A1) is a protein that in humans is encoded by SLC15A1 gene. PepT 1 is a solute carrier for oligopeptides. It functions in renal oligopeptide reabsorption and in the intestines in a proton dependent way, hence acting like a cotransporter.
Function
SLC15A1is localized to the brush border membrane of the intestinal epithelium and mediates the uptake of di- and tripeptides from the lumen into the enterocytes. This protein plays an important role in the uptake and digestion of dietary proteins. This protein also facilitates the absorption of numerous peptidomimetic drugs. Peptide transporter 1 functions in nutrient and drug transport have been studied using intestinal organoids.
See also
Solute carrier family |
https://en.wikipedia.org/wiki/Tf%E2%80%93idf | In information retrieval, tf–idf (also TF*IDF, TFIDF, TF–IDF, or Tf–idf), short for term frequency–inverse document frequency, is a measure of importance of a word to a document in a collection or corpus, adjusted for the fact that some words appear more frequently in general. It was often used as a weighting factor in searches of information retrieval, text mining, and user modeling. A survey conducted in 2015 showed that 83% of text-based recommender systems in digital libraries used tf–idf.
Variations of the tf–idf weighting scheme were often used by search engines as a central tool in scoring and ranking a document's relevance given a user query.
One of the simplest ranking functions is computed by summing the tf–idf for each query term; many more sophisticated ranking functions are variants of this simple model.
Motivations
Karen Spärck Jones (1972) conceived a statistical interpretation of term-specificity called Inverse Document Frequency (idf), which became a cornerstone of term weighting:
For example, the tf and idf for some words in Shakespeare's 37 plays are as follows:
We see that "Romeo", "Falstaff", and "salad" appears in very few plays, so seeing these words, one could get a good idea as to which play it might be. In contrast, "good" and "sweet" appears in every play and are completely uninformative as to which play it is.
Definition
The tf–idf is the product of two statistics, term frequency and inverse document frequency. There are various ways for determining the exact values of both statistics.
A formula that aims to define the importance of a keyword or phrase within a document or a web page.
Term frequency
Term frequency, , is the relative frequency of term within document ,
,
where is the raw count of a term in a document, i.e., the number of times that term occurs in document . Note the denominator is simply the total number of terms in document (counting each occurrence of the same term separately). There are various other |
https://en.wikipedia.org/wiki/Mantel%20test | The Mantel test, named after Nathan Mantel, is a statistical test of the correlation between two matrices. The matrices must be of the same dimension; in most applications, they are matrices of interrelations between the same vectors of objects. The test was first published by Nathan Mantel, a biostatistician at the National Institutes of Health, in 1967. Accounts of it can be found in advanced statistics books (e.g., Sokal & Rohlf 1995).
Usage
The test is commonly used in ecology, where the data are usually estimates of the "distance" between objects such as species of organisms. For example, one matrix might contain estimates of the genetic distances (i.e., the amount of difference between two different genomes) between all possible pairs of species in the study, obtained by the methods of molecular systematics; while the other might contain estimates of the geographical distance between the ranges of each species to every other species. In this case, the hypothesis being tested is whether the variation in genetics for these organisms is correlated to the variation in geographical distance.
Method
If there are n objects, and the matrix is symmetrical (so the distance from object a to object b is the same as the distance from b to a) such a matrix contains
distances. Because distances are not independent of each other – since changing the "position" of one object would change of these distances (the distance from that object to each of the others) – we can not assess the relationship between the two matrices by simply evaluating the correlation coefficient between the two sets of distances and testing its statistical significance. The Mantel test deals with this problem.
The procedure adopted is a kind of randomization or permutation test. The correlation between the two sets of distances is calculated, and this is both the measure of correlation reported and the test statistic on which the test is based. In principle, any correlation coefficient could be |
https://en.wikipedia.org/wiki/Metal%20aromaticity | Metal aromaticity or metalloaromaticity is the concept of aromaticity, found in many organic compounds, extended to metals and metal-containing compounds. The first experimental evidence for the existence of aromaticity in metals was found in aluminium cluster compounds of the type where M stands for lithium, sodium or copper. These anions can be generated in a helium gas by laser vaporization of an aluminium / lithium carbonate composite or a copper or sodium / aluminium alloy, separated and selected by mass spectrometry and analyzed by photoelectron spectroscopy. The evidence for aromaticity in these compounds is based on several considerations. Computational chemistry shows that these aluminium clusters consist of a tetranuclear plane and a counterion at the apex of a square pyramid. The unit is perfectly planar and is not perturbed the presence of the counterion or even the presence of two counterions in the neutral compound . In addition its HOMO is calculated to be a doubly occupied delocalized pi system making it obey Hückel's rule. Finally a match exists between the calculated values and the experimental photoelectron values for the energy required to remove the first 4 valence electrons. The first fully metal aromatic compound was a cyclogallane with a Ga32- core discovered by Gregory Robinson in 1995.
D-orbital aromaticity is found in trinuclear tungsten and molybdenum metal clusters generated by laser vaporization of the pure metals in the presence of oxygen in a helium stream. In these clusters the three metal centers are bridged by oxygen and each metal has two terminal oxygen atoms. The first signal in the photoelectron spectrum corresponds to the removal of the valence electron with the lowest energy in the anion to the neutral compound. This energy turns out to be comparable to that of bulk tungsten trioxide and molybdenum trioxide. The photoelectric signal is also broad which suggests a large difference in conformation between the anion and |
https://en.wikipedia.org/wiki/2018%20in%20paleobotany | This article records new taxa of plants that are scheduled to be described during the year 2018, as well as other significant discoveries and events related to paleobotany that occurred in the year 2018.
Flowering plants
{| class="wikitable sortable" align="center" width="80%"
|-
! Name
! Novelty
! Status
! Authors
! Age
! Unit
! Location
! Notes
! Images
|-
|
Alloberberis axelrodii
|
Sp. nov
|
Valid
|
Doweld
|
Miocene
|
|
()
|
A member of the family Berberidaceae; a replacement name for the previously invalidly published Mahonia sinuata Axelrod (1985), lacking holotype designation when published.
|
|-
|
Alloberberis caeruleomontana
|
Nom. nov
|
Valid
|
Doweld
|
Miocene
|
|
()
|
A member of the family Berberidaceae; a replacement name for Ilex sinuata Chaney & Axelrod (1959).
|
|-
|
Anacolosidites eosenonicus
|
Sp. nov
|
Valid
|
Arai & Dias-Brito
|
Late Cretaceous (Santonian)
|
São Carlos Formation
|
|
A pollen taxon, possibly a member of the family Loranthaceae.
|
|-
|
Aniba caucasica
|
Nom. nov
|
Valid
|
Doweld
|
Pliocene
|
|
Abkhazia
|
A species of Aniba; a replacement name for Aniba longifolia Kolakovsky & Schakryl (1958).
|
|-
|
Anisodromum upchurchii
|
Sp. nov
|
Valid
|
Wang & Dilcher
|
Early Cretaceous (Albian)
|
Dakota Formation
|
()
|
A rosid described on the basis of fossil leaves.
|
|-
|
Annona nepalensis
|
Sp. nov
|
Valid
|
Prasad et al.
|
Miocene
|
Churia Formation
|
|
A species of Annona.
|
|-
|
Araliaephyllum popovii
|
Sp. nov
|
Valid
|
Golovneva
|
Early Cretaceous (Albian)
|
|
|
A member of Laurales described on the basis of fossil leaves.
|
|-
|
Archeampelos betulifolia
|
Sp. nov
|
Valid
|
Moiseeva, Kodrul & Herman
|
Paleocene
|
Zeya–Bureya Basin
|
|
A flowering plant described on the basis of fossil leaves, similar to leaves of members of the family Betulaceae.
|
|-
|
Austrovideira
|
Gen. et sp. nov
|
Valid
|
Rozefelds & Pace
|
Early Oligocene
|
|
|
A member of Vitaceae. Genus includes new species A. dettmannae.
|
|-
|
Berberis miopannonica
| |
https://en.wikipedia.org/wiki/Freedom%20Frog | Freedom Frog is a frog mascot character of Intervention Helpline, an Alaska counseling nonprofit organization. It is used in school animations and can be seen during Iditarod races.
Intervention Helpline
Intervention Helpline is an Alaska counseling nonprofit organization helping people to quit drugs and alcohol.
The organization uses Freedom Frog for scholar and other organization animations.
Acronym
It is proposed than free could mean Family Recovery Educates Everyone and frog Family Recovery Ongoing Growth.
Iditarod
During the ceremonial start of the Iditarod Trail Sled Dog Race in Anchorage, Freedom Frog could be seen on a sledge.
The following racers hosted Freedom Frog on their sledge:
2008: Benedikt Beisch
2009: Bill Cotter
2010: Ross Adam
Book
A book, The adventures of freedom frog, explains to children the issues of addiction. |
https://en.wikipedia.org/wiki/Collisional%20excitation | Collisional excitation is a process in which the kinetic energy of a collision partner is converted into the internal energy of a reactant species.
Astronomy
In astronomy, collisional excitation gives rise to spectral lines in the spectra of astronomical objects such as planetary nebulae and H II regions.
In these objects, most atoms are ionised by photons from hot stars embedded within the nebular gas, stripping away electrons. The emitted electrons, (called photoelectrons), may collide with atoms or ions within the gas, and excite them. When these excited atoms or ions revert to their ground state, they will emit a photon. The spectral lines formed by these photons are called collisionally excited lines (often abbreviated to CELs).
CELs are only seen in gases at very low densities (typically less than a few thousand particles per cm³) for forbidden transitions. For allowed transitions, the gas density can be substantially higher. At higher densities, the reverse process of collisional de-excitation suppresses the lines. Even the hardest vacuum produced on earth is still too dense for CELs to be observed. For this reason, when CELs were first observed by William Huggins in the spectrum of the Cat's Eye Nebula, he did not know what they were, and attributed them to a hypothetical new element called nebulium. However, the lines he observed were later found to be emitted by extremely rarefied oxygen.
CELs are very important in the study of gaseous nebulae, because they can be used to determine the density and temperature of the gas.
Mass spectrometry
Collisional excitation in mass spectrometry is the process where an ion collides with an atom or molecule and leads to an increase in the internal energy of the ion. Molecular ions are accelerated to high kinetic energy and then collide with neutral gas molecules (e.g. helium, nitrogen or argon). In the collision some of the kinetic energy is converted into internal energy which results in fragmentation in a pro |
https://en.wikipedia.org/wiki/Strep-tag | The Strep-tag system is a method which allows the purification and detection of proteins by affinity chromatography. The Strep-tag II is a synthetic peptide consisting of eight amino acids (Trp-Ser-His-Pro-Gln-Phe-Glu-Lys). This peptide sequence exhibits intrinsic affinity towards Strep-Tactin, a specifically engineered streptavidin, and can be N- or C- terminally fused to recombinant proteins. By exploiting the highly specific interaction, Strep-tagged proteins can be isolated in one step from crude cell lysates. Because the Strep-tag elutes under gentle, physiological conditions, it is especially suited for generation of functional proteins.
Development and biochemistry of the Strep-tag
Streptavidin is a tetrameric protein expressed in Streptomyces avidinii. Because of Streptavidin's high affinity for vitamin H (biotin), Streptavidin is commonly used in the fields of molecular biology and biotechnology. The Strep-tag was originally selected from a genetic library to specifically bind to a proteolytically truncated "core" version of streptavidin. Over the years, the Strep-tag was systemically optimized, to permit a greater flexibility in the choice of attachment site. Further, its interaction partner, Streptavidin, was also optimized to increase peptide-binding capacity, which resulted in the development of Strep-Tactin. The binding affinity of Strep-tag to Strep-Tactin is nearly 100 times higher than from Strep-tag to Streptavidin. The so-called Strep-tag system, consisting of Strep-tag and Strep-Tactin, has proven particularly useful for the functional isolation and analysis of protein complexes in proteome research.
The Strep-tag principle
Just like other short-affinity tags (His-tag, FLAG-tag), the Strep-tag can be easily fused to recombinant proteins during subcloning of its cDNA or gene. For its expression, various vectors for various host organisms (E. coli, yeast, insect, and mammalian cells) are available. A particular benefit of the Strep-tag is |
https://en.wikipedia.org/wiki/Malapterurus%20minjiriya | Malapterurus minjiriya is a species of electric catfish native to Burkina Faso, Ethiopia, Ghana, Mali, Nigeria and Togo. This species grows to a length of SL. |
https://en.wikipedia.org/wiki/Serial%20Input/Output%20eXchange | SIOX is an asynchronous serial communication bus that uses (default 4800) datarates. Specified by in Sweden, it is widely used in factories, plants, ships and district heating systems.
The bus can use point-to-point, bus, tree, star and ring topologies and uses a level, with a short-circuit current of . It is also recommended that twisted pair is used as well as a cable area of . are used, giving .
See also
List of network buses |
https://en.wikipedia.org/wiki/Mir-504%20microRNA%20precursor%20family | In molecular biology mir-504 microRNA is a short RNA molecule. MicroRNAs function to regulate the expression levels of other genes by several mechanisms.
See also
MicroRNA |
https://en.wikipedia.org/wiki/JQuery | jQuery is a JavaScript library designed to simplify HTML DOM tree traversal and manipulation, as well as event handling, CSS animation, and Ajax. It is free, open-source software using the permissive MIT License. , jQuery is used by 77% of the 10 million most popular websites. Web analysis indicates that it is the most widely deployed JavaScript library by a large margin, having at least 3 to 4 times more usage than any other JavaScript library.
jQuery's syntax is designed to make it easier to navigate a document, select DOM elements, create animations, handle events, and develop Ajax applications. jQuery also provides capabilities for developers to create plug-ins on top of the JavaScript library. This enables developers to create abstractions for low-level interaction and animation, advanced effects and high-level, theme-able widgets. The modular approach to the jQuery library allows the creation of powerful dynamic web pages and Web applications.
The set of jQuery core features—DOM element selections, traversal, and manipulation—enabled by its selector engine (named "Sizzle" from v1.3), created a new "programming style", fusing algorithms and DOM data structures. This style influenced the architecture of other JavaScript frameworks like YUI v3 and Dojo, later stimulating the creation of the standard Selectors API.
Microsoft and Nokia bundle jQuery on their platforms. Microsoft includes it with Visual Studio for use within Microsoft's ASP.NET AJAX and ASP.NET MVC frameworks while Nokia has integrated it into the Web Run-Time widget development platform.
Overview
jQuery, at its core, is a Document Object Model (DOM) manipulation library. The DOM is a tree-structure representation of all the elements of a Web page. jQuery simplifies the syntax for finding, selecting, and manipulating these DOM elements. For example, jQuery can be used for finding an element in the document with a certain property (e.g. all elements with an h1 tag), changing one or more of its at |
https://en.wikipedia.org/wiki/Balt%20%28company%29 | Balt is a medical equipment manufacturer specializing in medical devices designed to treat stroke and other neurovascular diseases.
Presentation
Balt was founded as a small family business in 1977 by Leopold Płowiecki, in France. His son, Nicolas, grew the company until 2018. Bridgepoint Advisers took a stake in 2015. Balt began to expand internationally. Pascal Girin CEO of Balt International since 2016, was appointed as Balt’s CEO in late 2018.
Balt initially focused on plastic micro-tubes for the pharmaceutical industry. In 1987, it developed a microcatheter enabling the treatment of arteriovenous malformations. The Silk Vista Baby, one of the company’s most recent innovations released in 2018, is the world’s smallest collapsible intracranial stent.
Balt's revenues tripled between 2015 and 2020 and its workforce, quadrupled to reach 500 employees worldwide in 2020. Balt sells its production to 100 countries around the world. Between 2016 and 2019, the company acquired the American start-up Blockade Medical, which also produces medical devices, as well as several of its distributors in Europe, China, India and Brazil. Today, Balt operates in ten different geographical locations worldwide. |
https://en.wikipedia.org/wiki/Zero-knowledge%20proof | In cryptography, a zero-knowledge proof or zero-knowledge protocol is a method by which one party (the prover) can prove to another party (the verifier) that a given statement is true, while avoiding conveying to the verifier any information beyond the mere fact of the statement's truth. The intuition underlying zero-knowledge proofs is that it is trivial to prove the possession of certain information by simply revealing it; the challenge is to prove this possession without revealing the information, or any aspect of it whatsoever.
In light of the fact that one should be able to generate a proof of some statement only when in possession of certain secret information connected to the statement, the verifier, even after having become convinced of the statement's truth, should nonetheless remain unable to prove the statement to third parties.
In the plain model, nontrivial zero-knowledge proofs (i.e., those for languages outside of BPP) demand interaction between the prover and the verifier. This interaction usually entails the selection of one or more random challenges by the verifier; the random origin of these challenges, together with the prover's successful responses to them notwithstanding, jointly convince the verifier that the prover does possess the claimed knowledge. If interaction weren't present, then the verifier, having obtained the protocol's execution transcript—that is, the prover's one and only message—could replay that transcript to a third party, thereby convincing the third party that the verifier too possessed the secret information.
In the common random string and random oracle models, non-interactive zero-knowledge proofs exist, in light of the Fiat–Shamir heuristic. These proofs, in practice, rely on computational assumptions (typically the collision-resistance of a cryptographic hash function).
Abstract examples
The Ali Baba cave
There is a well-known story presenting the fundamental ideas of zero-knowledge proofs, first published in 19 |
https://en.wikipedia.org/wiki/Lacrimal%20tubercle | The lateral margin of the groove of the frontal process of the maxilla is named the anterior lacrimal crest, and is continuous below with the orbital margin; at its junction with the orbital surface is a small tubercle, the lacrimal tubercle, which serves as a guide to the position of the lacrimal sac. |
https://en.wikipedia.org/wiki/Mycobiome | The mycobiome, mycobiota, or fungal microbiome, is the fungal community in and on an organism.
The word “mycobiome” comes from the ancient Greek μύκης (mukēs), meaning "fungus" with the suffix “biome” derived from the Greek βίος (bíos), meaning “life.” The term was first coined in the 2009 paper by Gillevet et al.
Most species of fungi are decomposers with the ability to break down complex polymers. Fungi are commonly found within plant cells in an endophytic relationship or as a pathogen. Most plants also form mutualistic relationships with fungi that accelerate nutrient uptake among their root structures. The most common phyla present in the fungal communities that live alongside animals and in aquatic environments are Ascomycota and Basidiomycota. Animals will typically form a commensal relationship with fungi with the occasional occurrence of a pathogenic interaction.
Interactions with other microbes
Fungal microbes are amongst a wide variety of other microbes involved in a symbiotic relationship involving multicellular organisms. In mammals, the gut flora is usually met with vastly diverse populations of microbes from many kingdoms, where fungal populations make up less than 1% of the entire gut biome. Due to the coexistence of fungal populations with other microbes in most cases of host-symbiont associations, it's important to assess common dynamics that may occur.
Most interactions between microbes in the gut are either competitive or cooperative. This can be seen with multiple fungal microbes as well by observing populations through the treatment of antibiotics and antifungals. Research on microbial populations in animal models has resulted in noticeable fluctuations in microbe populations. Antibiotic treatment has mostly shown increases in parasitic fungal presence, suggesting competitive behaviors between microbes against fungi. Additionally, application of antifungal molecules have resulted in colitis in mice, suggesting that commensal fungi are resp |
https://en.wikipedia.org/wiki/Birkeland%20current | A Birkeland current (also known as field-aligned current) is a set of electrical currents that flow along geomagnetic field lines connecting the Earth's magnetosphere to the Earth's high latitude ionosphere. In the Earth's magnetosphere, the currents are driven by the solar wind and interplanetary magnetic field and by bulk motions of plasma through the magnetosphere (convection indirectly driven by the interplanetary environment). The strength of the Birkeland currents changes with activity in the magnetosphere (e.g. during substorms). Small scale variations in the upward current sheets (downward flowing electrons) accelerate magnetospheric electrons which, when they reach the upper atmosphere, create the Auroras Borealis and Australis. In the high latitude ionosphere (or auroral zones), the Birkeland currents close through the region of the auroral electrojet, which flows perpendicular to the local magnetic field in the ionosphere. The Birkeland currents occur in two pairs of field-aligned current sheets. One pair extends from noon through the dusk sector to the midnight sector. The other pair extends from noon through the dawn sector to the midnight sector. The sheet on the high latitude side of the auroral zone is referred to as the Region 1 current sheet and the sheet on the low latitude side is referred to as the Region 2 current sheet.
The currents were predicted in 1908 by Norwegian explorer and physicist Kristian Birkeland, who undertook expeditions north of the Arctic Circle to study the aurora. He rediscovered, using simple magnetic field measurement instruments, that when the aurora appeared the needles of magnetometers changed direction, confirming the findings of Anders Celsius and assistant Olof Hjorter more than a century before. This could only imply that currents were flowing in the atmosphere above. He theorized that somehow the Sun emitted a cathode ray, and corpuscles from what is now known as a solar wind entered the Earth's magnetic field and |
https://en.wikipedia.org/wiki/Forensic%20Architecture | Forensic Architecture is a multidisciplinary research group based at Goldsmiths, University of London that uses architectural techniques and technologies to investigate cases of state violence and violations of human rights around the world. The group is led by architect Eyal Weizman. He received a Peabody Award in 2021 for his work with Forensic Architecture.
The agency develops new evidentiary techniques and undertakes advanced architectural and media research with and on behalf of communities affected by state violence, and routinely works in partnership with international prosecutors, human rights organisations and political and environmental justice groups. It consists of an interdisciplinary team of investigators including architects, scholars, artists, filmmakers, software developers, investigative journalists, archaeologists, lawyers, and scientists. It investigates alleged human rights violations by states or corporations on behalf of civil society groups. The group uses advanced architectural and media techniques to investigate armed conflicts and environmental destruction, as well as to cross-reference a variety of evidence sources, such as new media, remote sensing, material analysis, and witness testimony.
The term forensic architecture also refers to an academic field and an emergent field of practice developed at the Centre for Research Architecture, at Goldsmiths, University of London, concerning the production and presentation of architectural evidence, relating to buildings and urban environments and their media representations.
History
Forensic Architecture was formed in 2010 as a research project within the Centre for Research Architecture at Goldsmiths, University of London. The project developed as a response to several converging phenomena, such as the urbanisation of warfare, the erosion of trust in evidence in relation to state crimes and human rights violations, the emergence and proliferation of open source media (or 'image flotsam'), |
https://en.wikipedia.org/wiki/KLJN%20Secure%20Key%20Exchange | Random-resistor-random-temperature Kirchhoff-law-Johnson-noise key exchange, also known as RRRT-KLJN or simply KLJN, is an approach for distributing cryptographic keys between two parties that claims to offer unconditional security. This claim, which has been contested, is significant, as the only other key exchange approach claiming to offer unconditional security is Quantum key distribution.
The KLJN secure key exchange scheme was proposed in 2005 by Laszlo Kish and Granqvist. It has the advantage over quantum key distribution in that it can be performed over a metallic wire with just four resistors, two noise generators, and four voltage measuring devices---equipment that is low-priced and can be readily manufactured. It has the disadvantage that several attacks against KLJN have been identified which must be defended against.
"Given that the amount of effort and funding that goes into Quantum Cryptography is substantial (some even mock it as a distraction from the ultimate prize which is quantum computing), it seems to me that the fact that classic thermodynamic resources allow for similar inherent security should give one pause," wrote Henning Dekant, the founder of the Quantum Computing Meetup, in April 2013.
The Cybersecurity Curricula 2017, a joint project of the Association for Computing Machinery, the IEEE Computer Society, the Association for Information Systems, and the International Federation for Information Processing Technical Committee on Information Security Education (IFIP WG 11.8) recommends teaching the KLJN Scheme as part of teaching "Advanced concepts" in its knowledge unit on cryptography.
See Also/Further Reading
http://www.scholarpedia.org/article/Secure_communications_using_the_KLJN_scheme
http://noise.ece.tamu.edu/research_files/research_secure.htm
Science: Simple Noise May Stymie Spies without Quantum Weirdness, Adrian Cho, September 30, 2005. http://noise.ece.tamu.edu/news_files/science_secure.pdf |
https://en.wikipedia.org/wiki/Kiyoshi%20Oka | was a Japanese mathematician who did fundamental work in the theory of several complex variables.
Biography
Oka was born in Osaka. He went to Kyoto Imperial University in 1919, turning to mathematics in 1923 and graduating in 1924.
He was in Paris for three years from 1929, returning to Hiroshima University. He published solutions to the first and second Cousin problems, and work on domains of holomorphy, in the period 1936–1940. He received his Doctor of Science degree from Kyoto Imperial University in 1940. These were later taken up by Henri Cartan and his school, playing a basic role in the development of sheaf theory.
The Oka–Weil theorem is due to a work of André Weil in 1935 and Oka's work in 1937.
Oka continued to work in the field, and proved Oka's coherence theorem in 1950. Oka's lemma is also named after him.
He was a professor at Nara Women's University from 1949 to retirement at 1964. He received many honours in Japan.
Honors
1951 Japan Academy Prize
1954 Asahi Prize
1960 Order of Culture
1973 Order of the Sacred Treasure, 1st class
Bibliography
KIYOSHI OKA COLLECTED PAPERS
- Includes bibliographical references.
Selected papers (Sur les fonctions analytiques de plusieurs variables)
PDF TeX |
https://en.wikipedia.org/wiki/Agaricus%20langei | Agaricus langei is a species of fungus in the genus Agaricus.
See also
List of Agaricus species |
https://en.wikipedia.org/wiki/Ammonium%20chloride | Ammonium chloride is an inorganic compound with the formula NH4Cl and a white crystalline salt that is highly soluble in water. Solutions of ammonium chloride are mildly acidic. In its naturally occurring mineralogic form, it is known as sal ammoniac. The mineral is commonly formed on burning coal dumps from condensation of coal-derived gases. It is also found around some types of volcanic vents. It is mainly used as fertilizer and a flavouring agent in some types of liquorice. It is the product from the reaction of hydrochloric acid and ammonia.
Production
It is a product of the Solvay process used to produce sodium carbonate:
CO2 + 2 NH3 + 2 NaCl + H2O → 2 NH4Cl + Na2CO3
Not only is that method the principal one for the manufacture of ammonium chloride, but also it is used to minimize ammonia release in some industrial operations.
Ammonium chloride is prepared commercially by combining ammonia (NH3) with either hydrogen chloride (gas) or hydrochloric acid (water solution):
NH3 + HCl → NH4Cl
Ammonium chloride occurs naturally in volcanic regions, forming on volcanic rocks near fume-releasing vents (fumaroles). The crystals deposit directly from the gaseous state and tend to be short-lived, as they dissolve easily in water.
Reactions
Ammonium chloride appears to sublime upon heating but actually reversibly decomposes into ammonia and hydrogen chloride gas:
NH4Cl NH3 + HCl
Ammonium chloride reacts with a strong base, like sodium hydroxide, to release ammonia gas:
NH4Cl + NaOH → NH3 + NaCl + H2O
Similarly, ammonium chloride also reacts with alkali-metal carbonates at elevated temperatures, giving ammonia and alkali-metal chloride:
2 NH4Cl + Na2CO3 → 2 NaCl + CO2 + H2O + 2 NH3
A solution of 5% by mass of ammonium chloride in water has a pH in the range 4.6 to 6.0.
Some reactions of ammonium chloride with other chemicals are endothermic, such as its reaction with barium hydroxide and its dissolving in water.
Applications
The dominant application of ammon |
https://en.wikipedia.org/wiki/Dedekind%20domain | In abstract algebra, a Dedekind domain or Dedekind ring, named after Richard Dedekind, is an integral domain in which every nonzero proper ideal factors into a product of prime ideals. It can be shown that such a factorization is then necessarily unique up to the order of the factors. There are at least three other characterizations of Dedekind domains that are sometimes taken as the definition: see below.
A field is a commutative ring in which there are no nontrivial proper ideals, so that any field is a Dedekind domain, however in a rather vacuous way. Some authors add the requirement that a Dedekind domain not be a field. Many more authors state theorems for Dedekind domains with the implicit proviso that they may require trivial modifications for the case of fields.
An immediate consequence of the definition is that every principal ideal domain (PID) is a Dedekind domain. In fact a Dedekind domain is a unique factorization domain (UFD) if and only if it is a PID.
The prehistory of Dedekind domains
In the 19th century it became a common technique to gain insight into integer solutions of polynomial equations using rings of algebraic numbers of higher degree. For instance, fix a positive integer . In the attempt to determine which integers are represented by the quadratic form , it is natural to factor the quadratic form into , the factorization taking place in the ring of integers of the quadratic field . Similarly, for a positive integer the polynomial (which is relevant for solving the Fermat equation ) can be factored over the ring , where is a primitive n-th root of unity.
For a few small values of and these rings of algebraic integers are PIDs, and this can be seen as an explanation of the classical successes of Fermat () and Euler (). By this time a procedure for determining whether the ring of all algebraic integers of a given quadratic field is a PID was well known to the quadratic form theorists. Especially, Gauss had looked at the c |
https://en.wikipedia.org/wiki/Jeremiah%20Day | Jeremiah Day (August 3, 1773 – August 22, 1867) was an American academic, a Congregational minister and President of Yale College (1817–1846).
Early life
Day was the son of Rev. Jeremiah and Abigail (Noble) Osborn Day, who were descendants of Robert Day, who came from Ipswich, England in 1634, settled in Newtown Cambridge, Massachusetts, and later became one of the original proprietors of Hartford, Connecticut. He was born in the parish of New Preston, Connecticut, then a part of New Milford, but since 1779, of Washington, where his father was pastor of the Congregational Church.
One of the latter's theological pupils, David Hale, brother of Nathan, first instructed Day, and later he continued his preparation for college under John Kingsbury of Waterbury, Connecticut. He entered Yale College in 1789, left because of pulmonary trouble in 1791, reentered in 1793, having taught school in the meantime, and graduated in 1795. While at Yale, he was a member of the Linonian Society.
Career
Day then succeeded Timothy Dwight IV [q.v.], as principal of the academy which the latter had established at Greenfield Hill, Connecticut, but soon left there to become tutor at Williams College. Two years later Day accepted a similar position at Yale. On June 3, 1800, he was licensed to preach by the New Haven West Association of Ministers. During all this time Day had been suffering from tuberculosis, and in July 1801 a hemorrhage brought on by the exertion of preaching caused him to go to Bermuda where he spent nearly a year. Upon his return Day went to his father's home with little expectation of recovery, but life among the Connecticut hills arrested the disease, and in the summer of 1803 he undertook the duties of the professorship of mathematics and natural philosophy at Yale to which he had been elected shortly after his departure for Bermuda. On January 14, 1805, Day married Martha, the daughter of the Hon. Roger Sherman and Rebecca Minot Prescott, they had one son child She |
https://en.wikipedia.org/wiki/Francis%20Su | Francis Edward Su is an American mathematician. He joined the Harvey Mudd College faculty in 1996, and is currently Benediktsson-Karwa Professor of Mathematics. Su served as president of the Mathematical Association of America from 2015–2017 and is serving as a Vice President of the American Mathematical Society from 2020-2023. Su has received multiple awards from the MAA, including the Henry L. Alder Award and the Haimo Award, both for distinguished teaching. He was also a Phi Beta Kappa Visiting Scholar during the 2019-2020 term.
Su received his B.S. in Mathematics from the University of Texas, graduating Phi Beta Kappa in 1989. He went on to receive his Ph.D. from Harvard University, where his advisor was Persi Diaconis. His research area is combinatorics, and he is particularly known for his work on fair division.
Su and Michael Starbird are co-authors of the book "Topology Through Inquiry". His book, "Mathematics for Human Flourishing", was released on 7 January 2020. The latter book is based on his speech of the same title, delivered Jan 6, 2017 at the Joint Math Meetings. He won the Halmos-Ford Award for Distinguished Writing in 2018 for that speech. Three of his articles have been featured on "The Princeton Anthology of the Best Writing in Mathematics" in the years 2011, 2014, and 2018. In 2021 he received the Euler Book Prize jointly with Christopher Jackson.
Selected publications |
https://en.wikipedia.org/wiki/Stotting | Stotting (also called pronking or pronging) is a behavior of quadrupeds, particularly gazelles, in which they spring into the air, lifting all four feet off the ground simultaneously. Usually, the legs are held in a relatively stiff position. Many explanations of stotting have been proposed, though for several of them there is little evidence either for or against.
The question of why prey animals stot has been investigated by evolutionary biologists including John Maynard Smith, C. D. Fitzgibbon, and Tim Caro; all of them conclude that the most likely explanation given the available evidence is that it is an honest signal to predators that the stotting animal would be difficult to catch. Such a signal is called "honest" as it is not deceptive in any way, and would benefit both predator and prey: the predator as it avoids a costly and unproductive chase, and the prey as it does not get chased.
Etymology
Stot is a common Scots and Northern English verb meaning "bounce" or "walk with a bounce". Uses in this sense include stotting a ball off a wall, and rain stotting off a pavement. Pronking comes from the Afrikaans verb pronk-, which means "show off" or "strut", and is a cognate of the English verb "prance".
Taxonomic distribution
Stotting occurs in several deer species of North America, including mule deer, pronghorn, and Columbian black-tailed deer, when a predator is particularly threatening, and in a variety of ungulate species from Africa, including Thomson's gazelle and springbok. It is also said to occur in the blackbuck, a species found in India.
Stotting occurs in domesticated livestock such as sheep and goats, typically only in young animals.
Possible explanations
Stotting makes a prey animal more visible, and uses up time and energy that could be spent on escaping from the predator. Since it is dangerous, the continued performance of stotting by prey animals must bring some benefit to the animal (or its family group) performing the behavior. Sev |
https://en.wikipedia.org/wiki/Central%20tolerance | In immunology, central tolerance (also known as negative selection) is the process of eliminating any developing T or B lymphocytes that are autoreactive, i.e. reactive to the body itself. Through elimination of autoreactive lymphocytes, tolerance ensures that the immune system does not attack self peptides. Lymphocyte maturation (and central tolerance) occurs in primary lymphoid organs such as the bone marrow and the thymus. In mammals, B cells mature in the bone marrow and T cells mature in the thymus.
Central tolerance is not perfect, so peripheral tolerance exists as a secondary mechanism to ensure that T and B cells are not self-reactive once they leave primary lymphoid organs. Peripheral tolerance is distinct from central tolerance in that it occurs once developing immune cells exit primary lymphoid organs (the thymus and bone-marrow), prior to their export into the periphery.
Function of central tolerance
Central tolerance is essential to proper immune cell functioning because it helps ensure that mature B cells and T cells do not recognize self-antigens as foreign microbes. More specifically, central tolerance is necessary because T cell receptors (TCRs) and B cell receptors (BCRs) are made by cells through random somatic rearrangement. This process, known as V(D)J recombination, is important because it increases the receptor diversity which increases the likelihood that B cells and T cells will have receptors for novel antigens. Junctional diversity occurs during recombination and serves to further increase the diversity of BCRs and TCRs. The production of random TCRs and BCRs is an important method of defense against microbes due to their high mutation rate. This process also plays an important role in promoting the survival of a species, because there will be a variety of receptor arrangements within a species – this enables a very high chance of at least one member of the species having receptors for a novel antigen.
While the process of somatic recom |
https://en.wikipedia.org/wiki/Mathematical%20Neuroscience%20Prize | The Mathematical Neuroscience Prize is a prize awarded biennially since 2013 by the nonprofit organization Israel Brain Technologies (IBT). It is endowed with $100,000 for each laureate and honors researchers who have significantly advanced the understanding of the neural mechanisms of perception, behavior and thought through the application of mathematical analysis and modeling.
Laureates
2013 Larry Abbott (Columbia University) and Haim Sompolinsky (Hebrew University Jerusalem)
2015 Nancy Kopell (Boston University) and Bard Ermentrout (University of Pittsburgh)
2017 Fred Wolf (Max Planck Institute for Dynamics and Self-Organization) and Misha Tsodyks (Weizmann Institute of Science)
2019 Naftali Tishby (Hebrew University) and John Rinzel (New York University)
See also
The Brain Prize
The Kavli Prize
The Mind & Brain Prize
List of mathematics awards
List of neuroscience awards |
https://en.wikipedia.org/wiki/Application%20binary%20interface | In computer software, an application binary interface (ABI) is an interface between two binary program modules. Often, one of these modules is a library or operating system facility, and the other is a program that is being run by a user.
An ABI defines how data structures or computational routines are accessed in machine code, which is a low-level, hardware-dependent format. In contrast, an application programming interface (API) defines this access in source code, which is a relatively high-level, hardware-independent, often human-readable format. A common aspect of an ABI is the calling convention, which determines how data is provided as input to, or read as output from, computational routines. Examples of this are the x86 calling conventions.
Adhering to an ABI (which may or may not be officially standardized) is usually the job of a compiler, operating system, or library author. However, an application programmer may have to deal with an ABI directly when writing a program in a mix of programming languages, or even compiling a program written in the same language with different compilers.
An ABI is as important as the underlying hardware architecture. The program will fail equally if it violates any constraints of these two.
Description
Details covered by an ABI include the following:
Processor instruction set, with details like register file structure, stack organization, memory access types, etc.
Sizes, layouts, and alignments of basic data types that the processor can directly access
Calling convention, which controls how the arguments of functions are passed, and return values retrieved; for example, it controls the following:
Whether all parameters are passed on the stack, or some are passed in registers
Which registers are used for which function parameters
Whether the first function parameter passed on the stack is pushed first or last
Whether the caller or callee is responsible for cleaning up the stack after the function call
How an appli |
https://en.wikipedia.org/wiki/Andr%C3%A1s%20Hajnal | András Hajnal (May 13, 1931 – July 30, 2016) was a professor of mathematics at Rutgers University and a member of the Hungarian Academy of Sciences known for his work in set theory and combinatorics.
Biography
Hajnal was born on 13 May 1931, in Budapest, Hungary.
He received his university diploma (M.Sc. degree) in 1953 from the Eötvös Loránd University, his Candidate of Mathematical Science degree (roughly equivalent to Ph.D.) in 1957, under the supervision of László Kalmár, and his Doctor of Mathematical Science degree in 1962. From 1956 to 1995 he was a faculty member at the Eötvös Loránd University; in 1994, he moved to Rutgers University to become the director of DIMACS, and he remained there as a professor until his retirement in 2004. He became a member of the Hungarian Academy of Sciences in 1982, and directed its mathematical institute from 1982 to 1992. He was general secretary of the János Bolyai Mathematical Society from 1980 to 1990, and president of the society from 1990 to 1994. Starting in 1981, he was an advisory editor of the journal Combinatorica. Hajnal was also one of the honorary presidents of the European Set Theory Society.
Hajnal was an avid chess player.
Hajnal was the father of Peter Hajnal, the co-dean of the European College of Liberal Arts.
Research and publications
Hajnal was the author of over 150 publications. Among the many co-authors of Paul Erdős, he had the second largest number of joint papers, 56.
With Peter Hamburger, he wrote a textbook, Set Theory (Cambridge University Press, 1999, ). Some of his more well-cited research papers include
A paper on circuit complexity with Maas, Pudlak, Szegedy, and György Turán, showing exponential lower bounds on the size of bounded-depth circuits with weighted majority gates that solve the problem of computing the parity of inner products.
The Hajnal–Szemerédi theorem on equitable coloring, proving a 1964 conjecture of Erdős: let Δ denote the maximum degree of a vertex in a finite graph |
https://en.wikipedia.org/wiki/Time-bin%20encoding | Time-bin encoding is a technique used in quantum information science to encode a qubit of information on a photon. Quantum information science makes use of qubits as a basic resource similar to bits in classical computing. Qubits are any two-level quantum mechanical system; there are many different physical implementations of qubits, one of which is time-bin encoding.
While the time-bin encoding technique is very robust against decoherence, it does not allow easy interaction between the different qubits. As such, it is much more useful in quantum communication (such as quantum teleportation and quantum key distribution) than in quantum computation.
Construction of a time-bin encoded qubit
Time-bin encoding is done by having a single-photon go through a Mach–Zehnder interferometer (MZ), shown in black here. The photon coming from the left is guided through one of two paths (shown in blue and red); the guiding can be made by optical fiber or simply in free space using mirrors and polarising cubes. One of the two paths is longer than the other. The difference in path length must be longer than the coherence length of the photon to make sure the path taken can be unambiguously distinguished. The interferometer has to keep a stable phase, which means that the path length difference must vary by much less than the wavelength of light during the experiment. This usually requires active temperature stabilization.
If the photon takes the short path, it is said to be in the state ; if it takes the long path, it is said to be in the state . If the photon has a non-zero probability to take either path, then it is in a coherent superposition of the two states:
These coherent superpositions of the two possible states are called qubits and are the basic ingredient of Quantum information science.
In general, it is easy to vary the phase gained by the photon between the two paths, for example by stretching the fiber, while it is much more difficult to vary the amplitudes w |
https://en.wikipedia.org/wiki/Constraint%20%28computational%20chemistry%29 | In computational chemistry, a constraint algorithm is a method for satisfying the Newtonian motion of a rigid body which consists of mass points. A restraint algorithm is used to ensure that the distance between mass points is maintained. The general steps involved are: (i) choose novel unconstrained coordinates (internal coordinates), (ii) introduce explicit constraint forces, (iii) minimize constraint forces implicitly by the technique of Lagrange multipliers or projection methods.
Constraint algorithms are often applied to molecular dynamics simulations. Although such simulations are sometimes performed using internal coordinates that automatically satisfy the bond-length, bond-angle and torsion-angle constraints, simulations may also be performed using explicit or implicit constraint forces for these three constraints. However, explicit constraint forces give rise to inefficiency; more computational power is required to get a trajectory of a given length. Therefore, internal coordinates and implicit-force constraint solvers are generally preferred.
Constraint algorithms achieve computational efficiency by neglecting motion along some degrees of freedom. For instance, in atomistic molecular dynamics, typically the length of covalent bonds to hydrogen are constrained; however, constraint algorithms should not be used if vibrations along these degrees of freedom are important for the phenomenon being studied.
Mathematical background
The motion of a set of N particles can be described by a set of second-order ordinary differential equations, Newton's second law, which can be written in matrix form
where M is a mass matrix and q is the vector of generalized coordinates that describe the particles' positions. For example, the vector q may be a 3N Cartesian coordinates of the particle positions rk, where k runs from 1 to N; in the absence of constraints, M would be the 3Nx3N diagonal square matrix of the particle masses. The vector f represents the generaliz |
https://en.wikipedia.org/wiki/Integrated%20circuit%20layout%20design%20protection | Layout designs (topographies) of integrated circuits are a field in the protection of intellectual property.
In United States intellectual property law, a "mask work" is a two or three-dimensional layout or topography of an integrated circuit (IC or "chip"), i.e. the arrangement on a chip of semiconductor devices such as transistors and passive electronic components such as resistors and interconnections. The layout is called a mask work because, in photolithographic processes, the multiple etched layers within actual ICs are each created using a mask, called the photomask, to permit or block the light at specific locations, sometimes for hundreds of chips on a wafer simultaneously.
Because of the functional nature of the mask geometry, the designs cannot be effectively protected under copyright law (except perhaps as decorative art). Similarly, because individual lithographic mask works are not clearly protectable subject matter; they also cannot be effectively protected under patent law, although any processes implemented in the work may be patentable. So since the 1990s, national governments have been granting copyright-like exclusive rights conferring time-limited exclusivity to reproduction of a particular layout. Terms of integrated circuit rights are usually shorter than copyrights applicable on pictures.
International law
A diplomatic conference was held at Washington, D.C., in 1989, which adopted a Treaty on Intellectual Property in Respect of Integrated Circuits, also called the Washington Treaty or IPIC Treaty. The Treaty, signed at Washington on May 26, 1989, is open to member states of the United Nations (UN) World Intellectual Property Organization (WIPO) and to intergovernmental organizations meeting certain criteria. The Treaty has been incorporated by reference into the TRIPS Agreement of the World Trade Organization (WTO), subject to the following modifications: the term of protection is at least 10 (rather than eight) years from the date of f |
https://en.wikipedia.org/wiki/Von%20Neumann%20paradox | In mathematics, the von Neumann paradox, named after John von Neumann, is the idea that one can break a planar figure such as the unit square into sets of points and subject each set to an area-preserving affine transformation such that the result is two planar figures of the same size as the original. This was proved in 1929 by John von Neumann, assuming the axiom of choice. It is based on the earlier Banach–Tarski paradox, which is in turn based on the Hausdorff paradox.
Banach and Tarski had proved that, using isometric transformations, the result of taking apart and reassembling a two-dimensional figure would necessarily have the same area as the original. This would make creating two unit squares out of one impossible. But von Neumann realized that the trick of such so-called paradoxical decompositions was the use of a group of transformations that include as a subgroup a free group with two generators. The group of area-preserving transformations (whether the special linear group or the special affine group) contains such subgroups, and this opens the possibility of performing paradoxical decompositions using them.
Sketch of the method
The following is an informal description of the method found by von Neumann. Assume that we have a free group H of area-preserving linear transformations generated by two transformations, σ and τ, which are not far from the identity element. Being a free group means that all its elements can be expressed uniquely in the form for some n, where the s and s are all non-zero integers, except possibly the first and the last . We can divide this group into two parts: those that start on the left with σ to some non-zero power (we call this set A) and those that start with τ to some power (that is, is zero—we call this set B, and it includes the identity).
If we operate on any point in Euclidean 2-space by the various elements of H we get what is called the orbit of that point. All the points in the plane can thus be classed into |
https://en.wikipedia.org/wiki/Auranthine | Auranthine is an antimicrobial chemical compound isolated from a nephrotoxic strain of Penicillium fungus, Penicillium aurantiogriseum.
A total synthesis of auranthine has been reported. |
https://en.wikipedia.org/wiki/Oligoflexia | The Oligoflexia are a class of the phylum Bdellovibrionota. All species of this group are all Gram-negative. |
https://en.wikipedia.org/wiki/Population%20pyramid | A population pyramid (age structure diagram) or "age-sex pyramid" is a graphical illustration of the distribution of a population (typically that of a country or region of the world) by age groups and sex; it typically takes the shape of a pyramid when the population is growing. Males are usually shown on the left and females on the right, and they may be measured in absolute numbers or as a percentage of the total population. The pyramid can be used to visualize the age of a particular population. It is also used in ecology to determine the overall age distribution of a population; an indication of the reproductive capabilities and likelihood of the continuation of a species. Number of people per unit area of land is called population density.
Structure
A population pyramid often contains continuous stacked-histogram bars, making it a horizontal bar diagram. The population size is shown on the x-axis (horizontal) while the age-groups are represented on the y-axis (vertical). The size of each bar can be displayed either as a percentage of the total population or as a raw number. Males are conventionally shown on the left and females on the right. Population pyramids are often viewed as the most effective way to graphically depict the age and distribution of a population, partly because of the very clear image these pyramids provide. A great deal of information about the population broken down by age and sex can be read from a population pyramid, and this can shed light on the extent of development and other aspects of the population.
The measures of central tendency (mean, median, and mode) should be considered when assessing a population pyramid. For example, the average age could be used to determine the type of population in a particular region. A population with an average age of 15 would be very young compared to one with an average age of 55. Population statistics are often mid-year numbers.
A series of population pyramids could give a clear picture of how |
https://en.wikipedia.org/wiki/BRCT%20domain | BRCA1 C Terminus (BRCT) domain is a family of evolutionarily related proteins. It is named after the C-terminal domain of BRCA1, a DNA-repair protein that serves as a marker of breast cancer susceptibility.
The BRCT domain is found predominantly in proteins involved in cell cycle checkpoint functions responsive to DNA damage, for example as found in the breast cancer DNA-repair protein BRCA1. The domain is an approximately 100 amino acid tandem repeat, which appears to act as a phospho-protein binding domain.
Examples
Human proteins containing this domain include:
BARD1; BRCA1
CTDP1; TDT or DNTT
ECT2
LIG4
MCPH1; MDC1
NBN
PARP1; PARP4; PAXIP1; PES1
REV1; RFC1; TOPBP1; TP53BP1; XRCC1 |
https://en.wikipedia.org/wiki/2nd%20meridian%20west | The meridian 2° west of Greenwich is a line of longitude that extends from the North Pole across the Arctic Ocean, the Atlantic Ocean, Europe, Africa, the Southern Ocean, and Antarctica to the South Pole.
The 2nd meridian west forms a great circle with the 178th meridian east.
From Pole to Pole
Starting at the North Pole and heading south to the South Pole, the 2nd meridian west passes through:
{| class="wikitable plainrowheaders"
! scope="col" width="125" | Co-ordinates
! scope="col" | Country, territory or sea
! scope="col" | Notes
|-
| style="background:#b0e0e6;" |
! scope="row" style="background:#b0e0e6;" | Arctic Ocean
| style="background:#b0e0e6;" |
|-valign="top"
| style="background:#b0e0e6;" |
! scope="row" style="background:#b0e0e6;" | Atlantic Ocean
| style="background:#b0e0e6;" | Passing just east of the island of Foula, Scotland, (at )
|-
| style="background:#b0e0e6;" |
! scope="row" style="background:#b0e0e6;" | North Sea
| style="background:#b0e0e6;" |
|-
|
! scope="row" |
| Scotland
|-valign="top"
| style="background:#b0e0e6;" |
! scope="row" style="background:#b0e0e6;" | North Sea
| style="background:#b0e0e6;" | Passing just east of Aberdeen, Scotland, (at )
|-valign="top"
|
! scope="row" |
| England — passing through Berwick-upon-Tweed (at ) and western Birmingham (at )
|-valign="top"
| style="background:#b0e0e6;" |
! scope="row" style="background:#b0e0e6;" | English Channel
| style="background:#b0e0e6;" | Passing just west of the Cotentin Peninsula, (at ) Passing just east of the island of (at )
|-
|
! scope="row" |
|
|-
| style="background:#b0e0e6;" |
! scope="row" style="background:#b0e0e6;" | Atlantic Ocean
| style="background:#b0e0e6;" | Bay of Biscay
|-
|
! scope="row" |
| Passing just west of San Sebastián (at )
|-
| style="background:#b0e0e6;" |
! scope="row" style="background:#b0e0e6;" | Mediterranean Sea
| style="background:#b0e0e6;" | Alboran Sea
|-
|
! scope="row" |
|
|-
|
! scope="row" |
|
|-
|
! scope="row" | |
https://en.wikipedia.org/wiki/SPARK%20%28programming%20language%29 | SPARK is a formally defined computer programming language based on the Ada programming language, intended for the development of high integrity software used in systems where predictable and highly reliable operation is essential. It facilitates the development of applications that demand safety, security, or business integrity.
Originally, there were three versions of the SPARK language (SPARK83, SPARK95, SPARK2005) based on Ada 83, Ada 95 and Ada 2005 respectively.
A fourth version of the SPARK language, SPARK 2014, based on Ada 2012, was released on April 30, 2014. SPARK 2014 is a complete re-design of the language and supporting verification tools.
The SPARK language consists of a well-defined subset of the Ada language that uses contracts to describe the specification of components in a form that is suitable for both static and dynamic verification.
In SPARK83/95/2005, the contracts are encoded in Ada comments and so are ignored by any standard Ada compiler, but are processed by the SPARK "Examiner" and its associated tools.
SPARK 2014, in contrast, uses Ada 2012's built-in "aspect" syntax to express contracts, bringing them into the core of the language. The main tool for SPARK 2014 (GNATprove) is based on the GNAT/GCC infrastructure, and re-uses almost the entirety of the GNAT Ada 2012 front-end.
Technical overview
SPARK utilises the strengths of Ada while trying to eliminate all its potential ambiguities and insecure constructs. SPARK programs are by design meant to be unambiguous, and their behavior is required to be unaffected by the choice of Ada compiler. These goals are achieved partly by omitting some of Ada's more problematic features (such as unrestricted parallel tasking) and partly by introducing contracts which encode the application designer's intentions and requirements for certain components of a program.
The combination of these approaches allows SPARK to meet its design objectives, which are:
logical soundness
rigorous formal definit |
https://en.wikipedia.org/wiki/Sprite%20%28lightning%29 | Sprites or red sprites are large-scale electric discharges that occur in the mesosphere, high above thunderstorm clouds, or cumulonimbus, giving rise to a varied range of visual shapes flickering in the night sky. They are usually triggered by the discharges of positive lightning between an underlying thundercloud and the ground.
Sprites appear as luminous red-orange flashes. They often occur in clusters above the troposphere at an altitude range of . Sporadic visual reports of sprites go back at least to 1886. They were first photographed on July 4, 1989, by scientists from the University of Minnesota and have subsequently been captured in video recordings thousands of times.
Sprites are sometimes inaccurately called upper-atmospheric lightning. However, they are cold plasma phenomena that lack the hot channel temperatures of tropospheric lightning, so they are more akin to fluorescent tube discharges than to lightning discharges. Sprites are associated with various other upper-atmospheric optical phenomena including blue jets and ELVES.
History
The earliest known report is by Toynbee and Mackenzie in 1886. Nobel laureate C. T. R. Wilson had suggested in 1925, on theoretical grounds, that electrical breakdown could occur in the upper atmosphere, and in 1956 he witnessed what possibly could have been a sprite. They were first documented photographically on July 6, 1989, when scientists from the University of Minnesota, using a low-light video camera, accidentally captured the first image of what would subsequently become known as a sprite.
Several years after their discovery they were named sprites (air spirits) after their elusive nature. Since the 1989 video capture, sprites have been imaged from the ground, from aircraft and from space, and have become the subject of intensive investigations. A featured high speed video that was captured by Thomas Ashcraft, Jacob L Harley, Matthew G McHarg, and Hans Nielsen in 2019 at about 100,000 frames per second is fast e |
https://en.wikipedia.org/wiki/Prolonged%20sine | The law of the prolonged sine was observed when measuring strength of the reaction of the plant stems and roots in response to turning from their usual vertical orientation. Such organisms maintained their usual vertical growth, and, if turned, start bending back toward the vertical. The prolonged sine law was observed when measuring the dependence of the bending speed from the angle of reorientation.
The observed law
It was observed that deviation from the desired growth direction by more than the 90 degrees causes further increase of the bending speed. After turning the 135 degrees the reoriented plant or fungi understands that it is placed "head down" and bends faster than turned by just 45 degrees. Poul Larsen in 1962 proposed, that the intensity of the gravitropic reaction (bending rate) is proportional to
where α is the angle of reorientation, g - gravity vector and constants a and b are determined experimentally.
Significance
Following the popular hypothesis of the mechanism of the plant spatial orientation, the bending from the horizontal position is caused by some small heavy particles that after turning put the pressure on the side cell (statocyte) wall, irritating some system and activating the bending process. The pressure of such particle to the cell was would be proportional to the sine of the reorientation angle, being maximal at 90° reorientation. It would be equal for the reorientations by both 45° and 135°. The prolonged sine law concludes that there are very significant deviations from such a predicted reaction. |
https://en.wikipedia.org/wiki/Zinc%20finger%20protein%20502 | Zinc finger protein 502 is a protein that in humans is encoded by the ZNF502 gene. |
https://en.wikipedia.org/wiki/Dental%20abscess | A dental abscess is a localized collection of pus associated with a tooth. The most common type of dental abscess is a periapical abscess, and the second most common is a periodontal abscess. In a periapical abscess, usually the origin is a bacterial infection that has accumulated in the soft, often dead, pulp of the tooth. This can be caused by tooth decay, broken teeth or extensive periodontal disease (or combinations of these factors). A failed root canal treatment may also create a similar abscess.
A dental abscess is a type of odontogenic infection, although commonly the latter term is applied to an infection which has spread outside the local region around the causative tooth.
Classification
The main types of dental abscess are:
Periapical abscess: The result of a chronic, localized infection located at the tip, or apex, of the root of a tooth.
Periodontal abscess: begins in a periodontal pocket (see: periodontal abscess)
Gingival abscess: involving only the gum tissue, without affecting either the tooth or the periodontal ligament (see: periodontal abscess)
Pericoronal abscess: involving the soft tissues surrounding the crown of a tooth (see: Pericoronitis)
Combined periodontic-endodontic abscess: a situation in which a periapical abscess and a periodontal abscess have combined (see: Combined periodontic-endodontic lesions).
Signs and symptoms
The pain is continuous and may be described as extreme, growing, sharp, shooting, or throbbing. Putting pressure or warmth on the tooth may induce extreme pain. The area may be sensitive to touch and possibly swollen as well. This swelling may be present at either the base of the tooth, the gum, and/or the cheek, and sometimes can be reduced by applying ice packs.
An acute abscess may be painless but still have a swelling present on the gum. It is important to get anything that presents like this checked by a dental professional as it may become chronic later.
In some cases, a tooth abscess may perforate |
https://en.wikipedia.org/wiki/Kerstin%20Nordstrom | Kerstin N. Nordstrom is an American physicist who is the Clare Boothe Luce Assistant Professor of Physics in the Department of Physics at Mount Holyoke College. Her research focuses on soft matter physics; her work has been featured in the LA Times and in the BBC News.
Early life and education
Nordstrom completed a bachelor's degree in physics and mathematics at Bryn Mawr College in 2004. She joined the University of Pennsylvania as a graduate student, earning a Master of Science in 2006 and a PhD in 2010. Her doctoral thesis focused on the "Jamming and flow of soft particle suspensions." In 2011, Nordstrom joined the University of Maryland, College Park as a postdoctoral researcher. At the University of Maryland, Nordstrom worked on several topics, including how beds of granular materials respond to impact and how razor clams burrow in sand.
Research and career
In 2014, Nordstrom joined Mount Holyoke College as an Assistant Professor. She is interested in complex fluid flows, including the systems of solid particles found in granular materials.
Awards and honors
2012 AAAS Mass Media Fellow
2018 Cottrell Scholar Award
2019 National Science Foundation CAREER Award
External media
In 2016, Nordstrom appeared on Jeopardy!. |
https://en.wikipedia.org/wiki/Quality%20control | Quality control (QC) is a process by which entities review the quality of all factors involved in production. ISO 9000 defines quality control as "a part of quality management focused on fulfilling quality requirements".
This approach places emphasis on three aspects (enshrined in standards such as ISO 9001):
Elements such as controls, job management, defined and well managed processes, performance and integrity criteria, and identification of records
Competence, such as knowledge, skills, experience, and qualifications
Soft elements, such as personnel, integrity, confidence, organizational culture, motivation, team spirit, and quality relationships.
Inspection is a major component of quality control, where physical product is examined visually (or the end results of a service are analyzed). Product inspectors will be provided with lists and descriptions of unacceptable product defects such as cracks or surface blemishes for example.
History and introduction
Early stone tools such as anvils had no holes and were not designed as interchangeable parts. Mass production established processes for the creation of parts and system with identical dimensions and design, but these processes are not uniform and hence some customers were unsatisfied with the result. Quality control separates the act of testing products to uncover defects from the decision to allow or deny product release, which may be determined by fiscal constraints. For contract work, particularly work awarded by government agencies, quality control issues are among the top reasons for not renewing a contract.
The simplest form of quality control was a sketch of the desired item. If the sketch did not match the item, it was rejected, in a simple Go/no go procedure. However, manufacturers soon found it was difficult and costly to make parts be exactly like their depiction; hence around 1840 tolerance limits were introduced, wherein a design would function if its parts were measured to be within the l |
https://en.wikipedia.org/wiki/Index%20of%20physics%20articles%20%28W%29 | The index of physics articles is split into multiple pages due to its size.
To navigate by individual letter use the table of contents below.
W
W' and Z' bosons
W' boson
W. G. Unruh
W. Jason Morgan
W. Lewis Hyde
W. R. Dean
W. W. Hansen
WAMIT
WASH-1400
WASH-740
WAsP
WIMP Argon Programme
WITCH experiment
WKB approximation
WMAP cold spot
WOMBAT (diffractometer)
W and Z bosons
W band
W state
Wade Allison
Wafer (electronics)
Wagner model
Wake
Wake turbulence
Waldo K. Lyon
Wall-plug efficiency
Wallace Clement Sabine
Wallace Hampton Tucker
Wallace Smith Broecker
Walter A. Rosenblith
Walter Dieminger
Walter Dornberger
Walter Dröscher
Walter Eric Spear
Walter Franz
Walter Gear
Walter Gerlach
Walter Gilbert
Walter Gordon (physicist)
Walter Grotrian
Walter Guyton Cady
Walter H. Schottky
Walter Heitler
Walter Herrmann (physicist)
Walter Hoppe
Walter Houser Brattain
Walter Kaufmann (physicist)
Walter Kistler
Walter Kohn
Walter M. Elsasser
Walter Mauderli
Walter Oelert
Walter Rogowski
Walter Rotman
Walter Schottky Prize
Walter Selke
Walter Thirring
Walter Tollmien
Walter Zinn
Walter Zürn
Walter de Heer
Walther Bothe
Walther Kossel
Walther Meissner
Walther Müller
Walther Nernst
Walther Ritz
Wander Johannes de Haas
Wang Ganchang
Wang Zhuxi
Wannier function
Ward Plummer
Wardenclyffe Tower
Ward–Takahashi identity
Warm dark matter
Warm–hot intergalactic medium
Warp drive (Star Trek)
Warped Passages
Warped geometry
Warren J. Smith
Warren Siegel
Washburn's equation
Washburn constant
Washington Large Area Time Coincidence Array
Washout (aviation)
Water content
Water cycle
Water meter
Water pipe percolator
Water potential
Water retention curve
Water thread experiment
Water vapor
Water window
Watercraft
Waterfall plot
Waterhole (radio)
Waterspout
Watson interferometer
Watt
Watt's law
Watt W. Webb
Watt steam engine
Wave
Wave-icle
Wave-making resistance
Wave-piercing
WaveRider
Wave Motion (journal)
Wave action (continuum mechanics)
Wave base
Wave drag
Wave equation
Wave field synthesis
Wa |
https://en.wikipedia.org/wiki/No-teleportation%20theorem | In quantum information theory, the no-teleportation theorem states that an arbitrary quantum state cannot be converted into a sequence of classical bits (or even an infinite number of such bits); nor can such bits be used to reconstruct the original state, thus "teleporting" it by merely moving classical bits around. Put another way, it states that the unit of quantum information, the qubit, cannot be exactly, precisely converted into classical information bits. This should not be confused with quantum teleportation, which does allow a quantum state to be destroyed in one location, and an exact replica to be created at a different location.
In crude terms, the no-teleportation theorem stems from the Heisenberg uncertainty principle and the EPR paradox: although a qubit can be imagined to be a specific direction on the Bloch sphere, that direction cannot be measured precisely, for the general case ; if it could, the results of that measurement would be describable with words, i.e. classical information.
The no-teleportation theorem is implied by the no-cloning theorem: if it were possible to convert a qubit into classical bits, then a qubit would be easy to copy (since classical bits are trivially copyable).
Formulation
The term quantum information refers to information stored in the state of a quantum system. Two quantum states ρ1 and ρ2 are identical if the measurement results of any physical observable have the same expectation value for ρ1 and ρ2. Thus measurement can be viewed as an information channel with quantum input and classical output, that is, performing measurement on a quantum system transforms quantum information into classical information. On the other hand, preparing a quantum state takes classical information to quantum information.
In general, a quantum state is described by a density matrix. Suppose one has a quantum system in some mixed state ρ. Prepare an ensemble, of the same system, as follows:
Perform a measurement on ρ.
According to |
https://en.wikipedia.org/wiki/Journal%20of%20Food%20and%20Drug%20Analysis | The Journal of Food and Drug Analysis is a quarterly peer-reviewed open-access scientific journal published by the Taiwanese Food and Drug Administration. It contains review and research articles covering food science, pharmacology, and chemical analysis. The journal was established in 1993 and the editor-in-chief is Gow-Chin Yen (National Chung Hsing University).
Editors
The following persons are or have been editor-in-chief of the journal:
Abstracting and indexing
The journal is abstracted and indexed in:
According to the Journal Citation Reports, the journal has a 2021 impact factor of 6.157. |
https://en.wikipedia.org/wiki/Arctic%20realm | The Arctic realm is one of the planet's twelve marine realms, as designated by the WWF and Nature Conservancy. It includes the coastal regions and continental shelves of the Arctic Ocean and adjacent seas, including the Arctic Archipelago, Hudson Bay, and the Labrador Sea of northern Canada, the seas surrounding Greenland, the northern and eastern coasts of Iceland, and the eastern Bering Sea.
The Arctic realm transitions to the Temperate Northern Atlantic realm in the Atlantic Basin, and the Temperate Northern Pacific realm in the Pacific Basin.
Ecoregions
The Arctic realm is further subdivided into 19 marine ecoregions:
North Greenland
North and East Iceland
East Greenland Shelf
West Greenland Shelf
Northern Grand Banks-Southern Labrador
Northern Labrador
Baffin Bay-Davis Strait
Hudson Complex
Lancaster Sound
High Arctic Archipelago
Beaufort-Amundsen-Viscount Melville-Queen Maud
Beaufort Sea-continental coast and shelf
Chukchi Sea
Eastern Bering Sea
East Siberian Sea
Laptev Sea
Kara Sea
North and East Barents Sea
White Sea |
https://en.wikipedia.org/wiki/Boxing%20barcode | Boxing is high-capacity 2D barcode. The flexible barcode format is fully customizable in terms of frame geometry, number of symbols per pixel and forward-error-correction (FEC) method. This makes it a suitable choice for storing large amounts of any kind of digital data on storage mediums such paper, photographic film or similar.
Applications
Boxing barcode is used on piqlFilm by Piql AS to store many infos in Arctic World Archive:
the Vatican Library
the photographic collection
GitHub
and other
Format
The Boxing barcode used in the piqlFilm consists 4096 rows and 2160 cols. Each frame has:
border
four corner marks
external bars (reference, calibration, structural metadata, human-readable)
data container
sync points |
https://en.wikipedia.org/wiki/Maximal%20information%20coefficient | In statistics, the maximal information coefficient (MIC) is a measure of the strength of the linear or non-linear association between two variables X and Y.
The MIC belongs to the maximal information-based nonparametric exploration (MINE) class of statistics. In a simulation study, MIC outperformed some selected low power tests, however concerns have been raised regarding reduced statistical power in detecting some associations in settings with low sample size when compared to powerful methods such as distance correlation and Heller–Heller–Gorfine (HHG). Comparisons with these methods, in which MIC was outperformed, were made in Simon and Tibshirani and in Gorfine, Heller, and Heller. It is claimed that MIC approximately satisfies a property called equitability which is illustrated by selected simulation studies. It was later proved that no non-trivial coefficient can exactly satisfy the equitability property as defined by Reshef et al., although this result has been challenged. Some criticisms of MIC are addressed by Reshef et al. in further studies published on arXiv.
Overview
The maximal information coefficient uses binning as a means to apply mutual information on continuous random variables. Binning has been used for some time as a way of applying mutual information to continuous distributions; what MIC contributes in addition is a methodology for selecting the number of bins and picking a maximum over many possible grids.
The rationale is that the bins for both variables should be chosen in such a way that the mutual information between the variables be maximal. That is achieved whenever . Thus, when the mutual information is maximal over a binning of the data, we should expect that the following two properties hold, as much as made possible by the own nature of the data. First, the bins would have roughly the same size, because the entropies and are maximized by equal-sized binning. And second, each bin of X will roughly correspond to a bin in Y.
Be |
https://en.wikipedia.org/wiki/Serial%20digital%20interface | Serial digital interface (SDI) is a family of digital video interfaces first standardized by SMPTE (The Society of Motion Picture and Television Engineers) in 1989. For example, ITU-R BT.656 and SMPTE 259M define digital video interfaces used for broadcast-grade video. A related standard, known as high-definition serial digital interface (HD-SDI), is standardized in SMPTE 292M; this provides a nominal data rate of 1.485 Gbit/s.
Additional SDI standards have been introduced to support increasing video resolutions (HD, UHD and beyond), frame rates, stereoscopic (3D) video, and color depth. Dual link HD-SDI consists of a pair of SMPTE 292M links, standardized by SMPTE 372M in 1998; this provides a nominal 2.970 Gbit/s interface used in applications (such as digital cinema or HDTV 1080P) that require greater fidelity and resolution than standard HDTV can provide. 3G-SDI (standardized in SMPTE 424M) consists of a single 2.970 Gbit/s serial link that allows replacing dual link HD-SDI. 6G-SDI and 12G-SDI standards were published on March 19, 2015.
These standards are used for transmission of uncompressed, unencrypted digital video signals (optionally including embedded audio and time code) within television facilities; they can also be used for packetized data. SDI is used to connect together different pieces of equipment such as recorders, monitors, PCs and vision mixers. Coaxial variants of the specification range in length but are typically less than . Fiber optic variants of the specification such as 297M allow for long-distance transmission limited only by maximum fiber length or repeaters. SDI and HD-SDI are usually available only in professional video equipment because various licensing agreements restrict the use of unencrypted digital interfaces, such as SDI, prohibiting their use in consumer equipment. Several professional video and HD-video capable DSLR cameras and all uncompressed video capable consumer cameras use the HDMI interface, often called clean HDMI |
https://en.wikipedia.org/wiki/Gland%20of%20Zeis | Glands of Zeis are unilobar sebaceous glands located on the margin of the eyelid. The glands of Zeis service the eyelash. These glands produce an oily substance that is issued through the excretory ducts of the sebaceous lobule into the middle portion of the hair follicle. In the same area of the eyelid, near the base of the eyelashes are apocrine glands called the "glands of Moll".
If eyelashes are not kept clean, conditions such as folliculitis may take place, and if the sebaceous gland becomes infected, it can lead to abscesses and styes. The glands of Zeis are named after German ophthalmologist Eduard Zeis (1807–68).
See also
Meibomian gland
Moll's gland
List of specialized glands within the human integumentary system |
https://en.wikipedia.org/wiki/Einstein%E2%80%93Cartan%20theory | In theoretical physics, the Einstein–Cartan theory, also known as the Einstein–Cartan–Sciama–Kibble theory, is a classical theory of gravitation similar to general relativity. The theory was first proposed by Élie Cartan in 1922. Einstein–Cartan theory is the simplest Poincaré gauge theory.
Overview
Einstein–Cartan theory differs from general relativity in two ways: (1) it is formulated within the framework of Riemann–Cartan geometry, which possesses a locally gauged Lorentz symmetry, while general relativity is formulated within the framework of Riemannian geometry, which does not; (2) an additional set of equations are posed that relate torsion to spin. This difference can be factored into
by first reformulating general relativity onto a Riemann–Cartan geometry, replacing the Einstein–Hilbert action over Riemannian geometry by the Palatini action over Riemann–Cartan geometry; and second, removing the zero torsion constraint from the Palatini action, which results in the additional set of equations for spin and torsion, as well as the addition of extra spin-related terms in the Einstein field equations themselves.
The theory of general relativity was originally formulated in the setting of Riemannian geometry by the Einstein–Hilbert action, out of which arise the Einstein field equations. At the time of its original formulation, there was no concept of Riemann–Cartan geometry. Nor was there a sufficient awareness of the concept of gauge symmetry to understand that Riemannian geometries do not possess the requisite structure to embody a locally gauged Lorentz symmetry, such as would be required to be able to express continuity equations and conservation laws for rotational and boost symmetries, or to describe spinors in curved spacetime geometries. The result of adding this infrastructure is a Riemann–Cartan geometry. In particular, to be able to describe spinors requires the inclusion of a spin structure, which suffices to produce such a geometry.
The chief di |
https://en.wikipedia.org/wiki/Gene%20targeting | Gene targeting is a biotechnological tool used to change the DNA sequence of an organism (hence it is a form of Genome Editing). It is based on the natural DNA-repair mechanism of Homology Directed Repair (HDR), including Homologous Recombination. Gene targeting can be used to make a range of sizes of DNA edits, from larger DNA edits such as inserting entire new genes into an organism, through to much smaller changes to the existing DNA such as a single base-pair change. Gene targeting relies on the presence of a repair template to introduce the user-defined edits to the DNA. The user (usually a scientist) will design the repair template to contain the desired edit, flanked by DNA sequence corresponding (homologous) to the region of DNA that the user wants to edit; hence the edit is targeted to a particular genomic region. In this way Gene Targeting is distinct from natural homology-directed repair, during which the ‘natural’ DNA repair template of the sister chromatid is used to repair broken DNA (the sister chromatid is the second copy of the gene). The alteration of DNA sequence in an organism can be useful in both a research context – for example to understand the biological role of a gene – and in biotechnology, for example to alter the traits of an organism (e.g. to improve crop plants).
Methods
To create a gene-targeted organism, DNA must be introduced into its cells. This DNA must contain all of the parts necessary to complete the gene targeting. At a minimum this is the homology repair template, containing the desired edit flanked by regions of DNA homologous (identical in sequence to) the targeted region (these homologous regions are called “homology arms” ). Often a reporter gene and/or a selectable marker is also required, to help identify and select for cells (or “events”) where GT has actually occurred. It is also common practice to increase GT rates by causing a double-strand-break (DSB) in the targeted DNA region. Hence the genes encoding for the s |
https://en.wikipedia.org/wiki/Decorative%20knot | A decorative or ornamental knot (also fancy knot) is an often complex knot exhibiting repeating patterns. A decorative knot is generally a knot that not only has practical use but is also known for its aesthetic or ornamental qualities. Often originating from maritime use, "decorative knots are not only serviceable and functional but also enhance the ship-shape appearance of any vessel." Decorative knots may be used alone or in combination, and may consist of single or multiple strands.
Coxcombing is decorative knotwork performed by sailors during the Age of Sail to dress-up, protect, or help identify specific items and parts of ships and boats.
List
This is an alphabetical list of decorative knots.
See also |
https://en.wikipedia.org/wiki/Molecular%20Biology%20%28journal%29 | Molecular Biology is a scientific journal which covers a wide scope of problems related to molecular, cell, and computational biology including genomics, proteomics, bioinformatics, molecular virology and immunology, molecular development biology, and molecular evolution. Molecular Biology publishes reviews, mini-reviews, experimental, and theoretical works, short communications and hypotheses. In addition, the journal publishes book reviews and meeting reports. The journal also publishes special issues devoted to most rapidly developing branches of physical-chemical biology and to the most outstanding scientists on the occasion of their anniversary birthdays. The journal is published in English and Russian versions by Nauka.
External links
Molecular and cellular biology journals
Multilingual journals
Publications with year of establishment missing
Nauka academic journals
English-language journals
Russian-language journals
Springer Science+Business Media academic journals |
https://en.wikipedia.org/wiki/STXBP5 | Syntaxin-binding protein 5 is a protein that in humans is encoded by the STXBP5 gene. It is also known as tomosyn, after , "friend" in Japanese, for its role as a binding protein.
Function
Syntaxin 1 is a component of the 7S and 20S SNARE complexes which are involved in docking and fusion of synaptic vesicles with the presynaptic plasma membrane. This gene encodes a syntaxin 1 binding protein. In rat, a similar protein dissociates syntaxin 1 from the Munc18/n-Sec1/rbSec1 complex to form a 10S complex, an intermediate which can be converted to the 7S SNARE complex. Thus this protein is thought to be involved in neurotransmitter release by stimulating SNARE complex formation. Alternatively spliced variants have been identified, but their biological validity has not been determined.
Positional cloning suggested that tomosyn might inhibit neurotransmitter secretion in Caenorhabditis elegans neurons.] This hypothesis was tested and confirmed, showing that tomosyn specifically inhibits synaptic vesicle priming—the biochemical step immediately preceding vesicle fusion and neurotransmitter release.
Structure
Two functional domains were originally identified, including one which binds to syntaxin, but recent crystallization of the yeast homolog Sro7 revealed that tomosyn likely has three functional domains: one WD40 domain and one syntaxin-binding domain, as previously recognized, but also another WD40 domain. The study also suggested that tomosyn's 'syntaxin binding domain' is not the reason tomosyn is inhibitory for neurotransmitter release, as originally proposed. The Sro7-based structure is currently given on SWISS-MODEL, which includes the WD40 domains but not most of the coiled coil syntaxin-binding domain seen in the infobox.
Interactions
STXBP5 has been shown to interact with STX4 and STX1A. |
https://en.wikipedia.org/wiki/Ramification%20%28mathematics%29 | In geometry, ramification is 'branching out', in the way that the square root function, for complex numbers, can be seen to have two branches differing in sign. The term is also used from the opposite perspective (branches coming together) as when a covering map degenerates at a point of a space, with some collapsing of the fibers of the mapping.
In complex analysis
In complex analysis, the basic model can be taken as the z → zn mapping in the complex plane, near z = 0. This is the standard local picture in Riemann surface theory, of ramification of order n. It occurs for example in the Riemann–Hurwitz formula for the effect of mappings on the genus.
In algebraic topology
In a covering map the Euler–Poincaré characteristic should multiply by the number of sheets; ramification can therefore be detected by some dropping from that. The z → zn mapping shows this as a local pattern: if we exclude 0, looking at 0 < |z| < 1 say, we have (from the homotopy point of view) the circle mapped to itself by the n-th power map (Euler–Poincaré characteristic 0), but with the whole disk the Euler–Poincaré characteristic is 1, n – 1 being the 'lost' points as the n sheets come together at z = 0.
In geometric terms, ramification is something that happens in codimension two (like knot theory, and monodromy); since real codimension two is complex codimension one, the local complex example sets the pattern for higher-dimensional complex manifolds. In complex analysis, sheets can't simply fold over along a line (one variable), or codimension one subspace in the general case. The ramification set (branch locus on the base, double point set above) will be two real dimensions lower than the ambient manifold, and so will not separate it into two 'sides', locally―there will be paths that trace round the branch locus, just as in the example. In algebraic geometry over any field, by analogy, it also happens in algebraic codimension one.
In algebraic number theory
In algebraic extensions |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.