source
stringlengths 31
227
| text
stringlengths 9
2k
|
|---|---|
https://en.wikipedia.org/wiki/Sucroglyceride
|
Sucroglycerides are substances used in the manufacture of food. They are known in the E number scheme as E474.
Synopsis
Sucroglycerides have been known at least since 1963.
Sucroglycerides are obtained through a reaction between sucrose and an edible oil or fat, and consist of a mixture of mono- and di-esters of sucrose and fatty acids, and mono- and diglycerides. They are immiscible with water, so some solvents may be necessary to produce them. These are limited to dimethyl formamide, cyclohexane, isobutanol, isopropanol and ethyl acetate.
Sucroglycerides are employed as an emulsifier, stabiliser and thickener, and may be used in dairy based drinks, such as chocolate milk, eggnog, drinking yoghurt, beverage whiteners, or in dairy based desserts such as ice cream, yoghurt, sorbets, fruit based desserts, cocoa mixes, chewing gum, rice pudding or tapioca pudding. Processed meat, egg based desserts like custard, soups and broths, sauces also may be treated with sucroglycerides.
Goops and supplements for "weight reduction", infants or youth, "sport" or "electrolyte" drinks, and particulated drinks like cider, fruit wine, mead or spirituous drinks may also be treated with sucroglycerides.
|
https://en.wikipedia.org/wiki/WebAssembly
|
WebAssembly (sometimes abbreviated Wasm) defines a portable binary-code format and a corresponding text format for executable programs as well as software interfaces for facilitating interactions between such programs and their host environment.
The main goal of WebAssembly is to enable high-performance applications on web pages, "but it does not make any Web-specific assumptions or provide Web-specific features, so it can be employed in other environments as well." It is an open standard and aims to support any language on any operating system, and in practice all of the most popular languages already have at least some level of support.
Announced in and first released in , WebAssembly became a World Wide Web Consortium recommendation on 5 December 2019 and it received the Programming Languages Software Award from ACM SIGPLAN in 2021. The World Wide Web Consortium (W3C) maintains the standard with contributions from Mozilla, Microsoft, Google, Apple, Fastly, Intel, and Red Hat.
History
WebAssembly is named to evoke the concept of assembly language, a term which dates to the 1950s. The name suggests bringing assembly-like programming to the Web, where it will be executed client-side by the website-user's computer via the user's web browser. To accomplish this, WebAssembly must be much more hardware-independent than a true assembly language.
WebAssembly was first announced in 2015, and the first demonstration was executing Unity's Angry Bots in Firefox, Google Chrome, and Microsoft Edge. The precursor technologies were asm.js from Mozilla and Google Native Client, and the initial implementation was based on the feature set of asm.js. The asm.js technology already provides near-native code execution speeds and can be considered a viable alternative for browsers that don't support WebAssembly or have it disabled for security reasons.
In March 2017, the design of the minimum viable product (MVP) was declared to be finished and the preview phase ended. In late Se
|
https://en.wikipedia.org/wiki/Rad%C3%B3%27s%20theorem%20%28harmonic%20functions%29
|
See also Rado's theorem (Ramsey theory)
In mathematics, Radó's theorem is a result about harmonic functions, named after Tibor Radó. Informally, it says that any "nice looking" shape without holes can be smoothly deformed into a disk.
Suppose Ω is an open, connected and convex subset of the Euclidean space R2 with smooth boundary ∂Ω and suppose that D is the unit disk. Then, given any homeomorphism
μ : ∂D → ∂Ω, there exists a unique harmonic function u : D → Ω such that u = μ on ∂D and u is a diffeomorphism.
|
https://en.wikipedia.org/wiki/American%20Society%20for%20Microbiology
|
The American Society for Microbiology (ASM), originally the Society of American Bacteriologists, is a professional organization for scientists who study viruses, bacteria, fungi, algae, and protozoa as well as other aspects of microbiology. It was founded in 1899. The Society publishes a variety of scientific journals, textbooks, and other educational materials related to microbiology and infectious diseases. ASM organizes annual meetings, as well as workshops and professional development opportunities for its members.
History
ASM was founded in 1899 under the name the "Society of American Bacteriologists." In December 1960, it was renamed the "American Society for Microbiology."
Mission
ASM's mission is "to promote and advance the microbial sciences." The society seeks to accomplish this mission through:
Publishing highly cited publications
Running multi-disciplinary meetings
Deploying resources and expertise around the world
Advocating for scientific research
Fostering a deeper public understanding of microbiology
Membership
ASM has more than 30,000 members, including researchers, educators and health professionals. Membership is open to all and is offered at a discounted rate to students, postdoctoral fellows and emeritus faculty. Members pay annual dues to support the activities of ASM. ASM's newest Clinical Lab Scientist membership category was established in 2019.
ASM provides professional development opportunities and supports microbiology professionals through 60 fellowships and 300 travel awards, webinars, conferences, workshops, networking opportunities, continuing education and honorific awards.
Meetings
ASM hosts meetings and conferences, including the annual meeting ASM Microbe, that provide forums to explore microbiology topics.
ASM Microbe, a combination of what was formerly the ASM General Meeting and the Interscience Conference on Antimicrobial Agents and Chemotherapy, includes sessions in eight specialty tracks:
Antimicrobial Agents
|
https://en.wikipedia.org/wiki/X8%20protein%20domain
|
In molecular biology, the X8 domain, is thought to play a role in targeting the plasmodesmata by providing it with structural support. The domain is able to do this since it contains signal sequences for a glycosylphosphatidylinositol (GPI) linkage to the extracellular face of the plasma membrane. This domain is involved in carbohydrate binding.
Structure
The X8 domain contains 6 conserved cysteine residues that presumably form three disulphide bridges. The domain is also found in an Olive pollen allergen as well as at the C terminus of family 17 glycosyl hydrolases.
|
https://en.wikipedia.org/wiki/Eric%20M.%20Rains
|
Eric Michael Rains (born 23 August 1973) is an American mathematician specializing in coding theory and special functions, especially applications from and to noncommutative algebraic geometry.
Biography
Eric Rains was 14 when he began classes in 1987. He left Case Western Reserve University with bachelor's degrees in computer science and physics and a master's degree in mathematics at age 17.
By means of a Churchill Scholarship he studied mathematics and physics at the University of Cambridge for the academic year 1991–1992, receiving a Certificate of Advanced Study in Mathematics. He received his PhD in 1995 from Harvard University with thesis Topics in Probability on Compact Lie Groups under the supervision of Persi Diaconis. From 1995 to 1996, Rains worked at the IDA's Center for Communications Research (CCR) in Princeton. From 1996 to 2002 he was a researcher for AT&T Labs. From 2002 to 2003 he returned to the CCR in Princeton. In 2003, Rains became a full professor at the University of California, Davis and since 2007 has been a full professor at Caltech where he currently works. He has served as the Executive Officer of the Caltech Mathematics Department from 2019 to 2022.
In the fall of 2006 he was a visiting professor at the University of Melbourne. He is the co-author with Gabriele Nebe and Neil J. A. Sloane of the 2006 book Self-Dual Codes and Invariant Theory.
In 2007, Rains was a plenary speaker at the Western Sectional meeting of the American Mathematical Society (AMS). In 2010 he was an invited speaker at the International Congress of Mathematicians in Hyderabad. He was elected a Fellow of the AMS in the class of 2018 for "contributions to coding theory, the theory of random matrices, the study of special functions, non-commutative geometry and number theory".
Selected publications
(This article has over 1200 citations.)
|
https://en.wikipedia.org/wiki/Ehrling%27s%20lemma
|
In mathematics, Ehrling's lemma, also known as Lions' lemma, is a result concerning Banach spaces. It is often used in functional analysis to demonstrate the equivalence of certain norms on Sobolev spaces. It was named after Gunnar Ehrling.
Statement of the lemma
Let (X, ||·||X), (Y, ||·||Y) and (Z, ||·||Z) be three Banach spaces. Assume that:
X is compactly embedded in Y: i.e. X ⊆ Y and every ||·||X-bounded sequence in X has a subsequence that is ||·||Y-convergent; and
Y is continuously embedded in Z: i.e. Y ⊆ Z and there is a constant k so that ||y||Z ≤ k||y||Y for every y ∈ Y.
Then, for every ε > 0, there exists a constant C(ε) such that, for all x ∈ X,
Corollary (equivalent norms for Sobolev spaces)
Let Ω ⊂ Rn be open and bounded, and let k ∈ N. Suppose that the Sobolev space Hk(Ω) is compactly embedded in Hk−1(Ω). Then the following two norms on Hk(Ω) are equivalent:
and
For the subspace of Hk(Ω) consisting of those Sobolev functions with zero trace (those that are "zero on the boundary" of Ω), the L2 norm of u can be left out to yield another equivalent norm.
|
https://en.wikipedia.org/wiki/BacDive
|
BacDive (the Bacterial Diversity Metadatabase) is a bacterial metadatabase that provides strain-linked information about bacterial and archaeal biodiversity.
Introduction
BacDive is a resource for different kind of metadata like taxonomy, morphology, physiology, environment and molecular-biology.
The majority of data is manually annotated and curated. With the release in December 2022 BacDive offers information for 93,254 strains, including 19,313 type strains.
The database is hosted by the Leibniz Institute DSMZ - German Collection of Microorganisms and Cell Cultures GmbH and is part of de.NBI the German Network for Bioinformatics Infrastructure, as well as Elixir the European Network for Bioinformatics. In December 2022 BacDive was selected by the Global Biodata Coalition as a Global Core Biodata Resource (GCBR). GCBRs are considered critical data resources for the global research endeavour in life sciences and biomedicine.
Content and Features
Database
The December release of the database encompassed over 1000 different data fields, divided into the categories "Name and taxonomic classification", "Morphology", "Culture and growth conditions,"Physiology and metabolism", "Isolation, sampling and environmental information." "Safety information", "Sequence information" and "Strain availability". The database comprised 1,922,166 entries, linked to the according strain and reference.
The data are retrieved from internal descriptions of culture collections, expert-compiled compendia and primary scientific literature like species descriptions.
Data access
Data can be accessed either via a GUI or via the RESTful web service. Using the GUI the user can choose between a simple search for searching strains by name, Culture collection number, NCBI Tax ID or INSDC sequence accession number, or the user can use the advanced search, which enables the search in 130 data fields and gives the opportunity of complex queries by combining several fields. Data can be downloaded i
|
https://en.wikipedia.org/wiki/P-y%20method
|
In geotechnical civil engineering, the p–y is a method of analyzing the ability of deep foundations to resist loads applied in the lateral direction. This method uses the finite difference method and p-y graphs to find a solution. P–y graphs are graphs which relate the force applied to soil to the lateral deflection of the soil. In essence, non-linear springs are attached to the foundation in place of the soil. The springs can be represented by the following equation:
where '' is the non-linear spring stiffness defined by the p–y curve, is the deflection of the spring, and is the force applied to the spring.
The p–y curves vary depending on soil type.
The available geotechnical engineering software programs for the p–y method include FB-MultiPier by the Bridge Software Institute, DeepFND by Deep Excavation LLC, PileLAT by Innovative Geotechnics, LPile by Ensoft, and PyPile by Yong Technology.
|
https://en.wikipedia.org/wiki/Phase-shift%20oscillator
|
A phase-shift oscillator is a linear electronic oscillator circuit that produces a sine wave output. It consists of an inverting amplifier element such as a transistor or op amp with its output fed back to its input through a phase-shift network consisting of resistors and capacitors in a ladder network. The feedback network 'shifts' the phase of the amplifier output by 180 degrees at the oscillation frequency to give positive feedback. Phase-shift oscillators are often used at audio frequency as audio oscillators.
The filter produces a phase shift that increases with frequency. It must have a maximum phase shift of more than 180 degrees at high frequencies so the phase shift at the desired oscillation frequency can be 180 degrees. The most common phase-shift network cascades three identical resistor-capacitor stages that produce a phase shift of zero at low frequencies and 270° at high frequencies.
The first integrated circuit was a phase shift oscillator invented by Jack Kilby in 1958.
Implementations
Bipolar implementation
This schematic drawing shows the oscillator using a common-emitter connected bipolar transistor as an amplifier. The two resistors R and three capacitors C form the RC phase-shift network which provides feedback from collector to base of the transistor. Resistor Rb provides base bias current. Resistor Rc is the collector load resistor for the collector current. Resistor Rs isolates the circuit from the external load.
FET implementation
This circuit implements the oscillator with a FET. R1, R2, Rs, and Cs provide bias for the transistor. Note that the topology used for positive feedback is voltage series feedback.
Op-amp implementation
The implementation of the phase-shift oscillator shown in the diagram uses an operational amplifier (op-amp), three capacitors and four resistors.
The circuit's modeling equations for the oscillation frequency and oscillation criterion are complicated because each RC stage loads the preceding ones. Assum
|
https://en.wikipedia.org/wiki/Level%20staff
|
A level staff, also called levelling rod, is a graduated wooden or aluminium rod, used with a levelling instrument to determine the difference in height between points or heights of points above a vertical datum.
When used for stadiametric rangefinding, the level staff is called a stadia rod.
Rod construction and materials
Levelling rods can be one piece, but many are sectional and can be shortened for storage and transport or lengthened for use. Aluminum rods may be shortened by telescoping sections inside each other, while wooden rod sections can be attached to each other with sliding connections or slip joints, or hinged to fold when not in use.
There are many types of rods, with names that identify the form of the graduations and other characteristics. Markings can be in imperial or metric units. Some rods are graduated on one side only while others are marked on both sides. If marked on both sides, the markings can be identical or can have imperial units on one side and metric on the other.
Reading a rod
In the photograph on the right, both a metric (left) and imperial (right) levelling rod are seen. This is a two-sided aluminum rod, coated white with markings in contrasting colours. The imperial side has a bright yellow background.
The metric rod has major numbered graduations in meters and tenths of meters (e.g. 18 is 1.8 m - there is a tiny decimal point between the numbers). Between the major marks are either a pattern of squares and spaces in different colours or an E shape (or its mirror image) with horizontal components and spaces between of equal size. In both parts of the pattern, the squares, lines or spaces are precisely one centimetre high. When viewed through an instrument's telescope, the observer can visually interpolate a 1 cm mark to a tenth of its height, yielding a reading with precision in mm. Usually readings are recorded with millimetre precision. On this side of the rod, the colours of the markings alternate between red and black wit
|
https://en.wikipedia.org/wiki/Tinkertoy
|
The Tinkertoy Construction Set is a construction set toy for children. It was designed in 1914—six years after the Frank Hornby's Meccano sets—by Charles H. Pajeau, who formed the Toy Tinker Company in Evanston, Illinois to manufacture them. Pajeau, a stonemason, designed the toy after seeing children play with sticks and empty spools of thread. Pajeau partnered with Robert Pettit and Gordon Tinker to market a toy that would allow and inspire children to use their imaginations. After an initially-slow start, over a million were sold.
The cornerstone of the set is a wooden spool roughly 1.4 inches in diameter, with holes drilled every 45 degrees around the perimeter and one through the center. Unlike the center, the perimeter holes do not go all the way through. With the differing-length sticks, the set was intended to be based on the Pythagorean progressive right triangle.
The sets were introduced to the public through displays in and around Chicago which included model Ferris wheels. Tinkertoys have been used to construct complex machines, including Danny Hillis's tic-tac-toe-playing computer (now in the collection of the Computer History Museum in Mountain View, California) and a robot at Cornell University in 1998.
One of Tinkertoy’s distinctive features is the toy’s packaging. Initially, the mailing tube design was chosen to reduce shipping costs. Early versions of the packaging included an address label on the tube with space for postage. To assist buyers in differentiating between the various offerings, sets were placed in mail tube packages of different sizes and also delineated with a number (e.g.: 116, 136) and a name (e.g.: major, prep, big boy, junior, grad). A colorful "how-to" instruction guide accompanied each set. In the 1950s, color was added and the wooden sticks appeared in red, green, blue, and peach.
The main manufacturing location was a 65,000-square-foot four-story plant at 2012 Ridge Avenue, Evanston, Illinois.
Tinkertoys were inducted in
|
https://en.wikipedia.org/wiki/Nisin
|
Nisin is a polycyclic antibacterial peptide produced by the bacterium Lactococcus lactis that is used as a food preservative. It has 34 amino acid residues, including the uncommon amino acids lanthionine (Lan), methyllanthionine (MeLan), didehydroalanine (Dha), and didehydroaminobutyric acid (Dhb). These unusual amino acids are introduced by posttranslational modification of the precursor peptide. In these reactions a ribosomally synthesized 57-mer is converted to the final peptide. The unsaturated amino acids originate from serine and threonine, and the enzyme-catalysed addition of cysteine residues to the didehydro amino acids result in the multiple (5) thioether bridges.
Subtilin and epidermin are related to nisin. All are members of a class of molecules known as lantibiotics.
In the food industry, nisin is obtained from the culturing of L. lactis on natural substrates, such as dextrose, and it is not chemically synthesized.
It was originally isolated in the late 1930s, and produced since the 1950s as Nisaplin from naturally occurring sources by Aplin and Barrett in laboratories in Beaminster in Dorset (now owned by International_Flavors_%26_Fragrances), and approved as an additive for food use in the US in the late 1960s.
Properties
While in general most bacteriocins inhibit only closely related species, nisin is a rare example of a "broad-spectrum" bacteriocin effective against many Gram-positive organisms, including lactic acid bacteria (commonly associated to avoid food spoilage), Listeria monocytogenes (a known pathogen), Staphylococcus aureus, Bacillus cereus, Clostridium botulinum, etc. It is also particularly effective against spores. Gram-negative bacteria are protected by their outer membrane but may become susceptible to nisin action after a heat shock or when this is coupled with the chelator EDTA. When used in combination with EDTA, nisin has the ability to inhibit E. coli O157:H7 and Salmonella enterica. Nisin, as a class I bacteriocin, is very
|
https://en.wikipedia.org/wiki/Accelerated%20Math
|
Accelerated Math is a daily, progress-monitoring software tool that monitors and manages mathematics skills practice, from preschool math through calculus. It is primarily used by primary and secondary schools, and it is published by Renaissance Learning, Inc. Currently, there are five versions: a desktop version and a web-based version in Renaissance Place, the company's web software for Accelerated Math and a number of other software products (e.g. Accelerated Reader). In Australia and the United Kingdom, the software is referred to as "Accelerated Maths".
Research
Below is a sample of some of the most current research on Accelerated Math.
Sadusky and Brem (2002) studied the impact of first-year implementation of Accelerated Math in a K-6 urban elementary school during the 2001–2002 school year. The researchers found that teachers were able to immediately use data to make decisions about instruction in the classroom. The students in classrooms using Accelerated Math had twice the percentile gains when tested as compared to the control classrooms that did not use Accelerated Math.
Ysseldkyke and Tardrew (2003) studied 2,202 students in 125 classrooms encompassing 24 states. The results showed that when students using Accelerated Math were compared to a control group, those students using the software made a significant gains on the STAR Math test. Students in grades 3 through 10 that were using Accelerated Math had more than twice the percentile gains on these tests than students in the control group.
Ysseldyke, Betts, Thill, and Hannigan (2004) conducted a quasi-experimental study with third- through sixth-grade Title I students. They found that Title I students who used Accelerated Math outperformed students who did not. Springer, Pugalee, and Algozzine (2005) also discovered a similar pattern. They studied students that failed to pass the AIMS test in order to graduate. Over half of the students passed the test after taking a course in which Accelerated Math
|
https://en.wikipedia.org/wiki/BIOTEC
|
The National Center for Genetic Engineering and Biotechnology (BIOTEC) is a research center in Thailand under the umbrella of the National Science and Technology Development Agency (NSTDA), Ministry of Higher Education, Science, Research and Innovation. The center was founded on 20 September 1983 and became a member of NSTDA in 1991.
History
1983–1992
BIOTEC provided basic infrastructure in biotechnology. To achieve its goals, the center built up a network of biotech experts and supported the development of laboratories in academic institutions. BIOTEC also supported the competency enhancement of personnel in the biotechnology field by granting scholarships for master's and doctoral degrees.
1993–2002
As new technology quickly evolved, BIOTEC worked in collaboration with several institutions in Thailand to achieve biotechnology excellence. In preparing the country for biotechnology, its successes included biotechnology policy development, intellectual property management, biosafety guidelines, raising public awareness, and educating people about biotech and its applications.
2003–present
Thailand's National Biotechnology Policy Framework was implemented with emphasis on healthcare services, modern biotechnology, education, training, biotechnology business and intellectual property rights. While its new technologies and discoveries have been transferred and applied for the benefit of private companies and the public, especially in rural communities, enhancement of the competency of personnel in the biotechnology field remains its focus.
Research units
BIOTEC has established 16 distinct research units for conducting research and providing technical services. Some laboratories act as stand-alone research centers in their own right, while others are collaborative ventures set up jointly with government agencies and universities.
|
https://en.wikipedia.org/wiki/Featural%20writing%20system
|
In a featural writing system, the shapes of the symbols (such as letters) are not arbitrary but encode phonological features of the phonemes that they represent. The term featural was introduced by Geoffrey Sampson to describe the Korean alphabet and Pitman shorthand.
Joe Martin introduced the term featural notation to describe writing systems that include symbols to represent individual features rather than phonemes. He asserts that "alphabets have no symbols for anything smaller than a phoneme".
A featural script represents finer detail than an alphabet. Here, symbols do not represent whole phonemes, but rather the elements (features) that make up the phonemes, such as voicing or its place of articulation. In the Korean alphabet, the featural symbols are combined into alphabetic letters, and these letters are in turn joined into syllabic blocks, so the system combines three levels of phonological representation.
Some scholars (e.g. John DeFrancis) reject this class or at least labeling the Korean alphabet as such. Others include stenographies and constructed scripts of hobbyists and fiction writers (such as Tengwar), many of which feature advanced graphic designs corresponding to phonologic properties. The basic unit of writing in these systems can map to anything from phonemes to words. It has been shown that even the Latin script has sub-character "features".
Examples of featural systems
This is a small list of examples of featural writing systems by date of creation. The languages for which each system was developed are also shown.
15th century
Hangul Korean
19th century
Canadian Aboriginal syllabics several Algonquian, Eskimo-Aleut and Athabaskan languages
Gregg shorthand many languages from different families
Duployan shorthand originally French, later English, German, Spanish, Romanian, Chinook Jargon and others
Visible Speech (a phonetic script) no specific language. Developed to aid the deaf and teach them to speak properly
20th century
Shavian a
|
https://en.wikipedia.org/wiki/Schick%20test
|
The Schick test, developed in 1913, is a skin test used to determine whether or not a person is susceptible to diphtheria. It was named after its inventor, Béla Schick (1877–1967), a Hungarian-born American pediatrician.
Procedure
The test is a simple procedure. A small amount (0.1 ml) of diluted (1/50 MLD) diphtheria toxin is injected intradermally into one arm of the person and a heat inactivated toxin on the other as a control. If a person does not have enough antibodies to fight it off, the skin around the injection will become red and swollen, indicating a positive result. This swelling disappears after a few days. If the person has an immunity, then little or no swelling and redness will occur, indicating a negative result.
Results can be interpreted as:
Positive: when the test results in a wheal of 5–10 mm diameter, reaching its peak in 4–7 days. The control arm shows no reaction. This indicates that the subject lacks antibodies against the toxin and hence is susceptible to the disease.
Pseudo-positive: when there is only a red-colored inflammation (erythema) and it disappears within 4 days. This happens on both the arms since the subject is immune but hypersensitive to the toxin.
Negative reaction: Indicates that the person is immune.
Combined reaction: Initial picture is like that of the pseudo-reaction but the erythema fades off after 4 days only in the control arm. It progresses on the test arm to a typical positive. The subject is interpreted to be both susceptible and hypersensitive.
The test was created when immunizing agents were scarce and not very safe; however, as newer and safer toxoids became available, susceptibility tests were no longer required.
|
https://en.wikipedia.org/wiki/Metal-coded%20affinity%20tag
|
Metal-coded affinity tag is a method used for quantitative proteomics by mass spectrometry that uses a metal chelate complex 1,4,7,10-tetraazacyclododecane-1,4,7,10-tetraacetate (DOTA) coupled to different lanthanide ions. The metal complexes attach to the cysteine residues of proteins in a sample.
Proteomic analysis
For bottom-up proteomics, the proteins can be separated by two-dimensional gel electrophoresis and analyzed by matrix-assisted laser desorption/ionization (MALDI) or electrospray ionization mass spectrometry for relative quantification or by inductively coupled plasma mass spectrometry for absolute quantification. For top-down proteomics, the undigested labeled proteins are analyzed.
See also
Mass cytometry
|
https://en.wikipedia.org/wiki/Rietdijk%E2%80%93Putnam%20argument
|
In philosophy, the Rietdijk–Putnam argument, named after and Hilary Putnam, uses 20th-century findings in physicsspecifically in special relativityto support the philosophical position known as four-dimensionalism.
If special relativity is true, then each observer will have their own plane of simultaneity, which contains a unique set of events that constitutes the observer's present moment. Observers moving at different relative velocities have different planes of simultaneity, and hence different sets of events that are present. Each observer considers their set of present events to be a three-dimensional universe, but even the slightest movement of the head or offset in distance between observers can cause the three-dimensional universes to have differing content. If each three-dimensional universe exists, then the existence of multiple three-dimensional universes suggests that the universe is four-dimensional. The argument is named after the discussions by Rietdijk (1966) and Putnam (1967). It is sometimes called the Rietdijk–Putnam–Penrose argument.
Andromeda paradox
Roger Penrose advanced a form of this argument that has been called the Andromeda paradox in which he points out that two people walking past each other on the street could have very different present moments. If one of the people were walking towards the Andromeda Galaxy, then events in this galaxy might be hours or even days advanced of the events on Andromeda for the person walking in the other direction. If this occurs, it would have dramatic effects on our understanding of time. Penrose highlighted the consequences by discussing a potential invasion of Earth by aliens living in the Andromeda Galaxy. As Penrose put it:
The "paradox" consists of two observers who are, from their conscious perspective, in the same place and at the same instant having different sets of events in their "present moment". Notice that neither observer can actually "see" what is happening in Andromeda, because ligh
|
https://en.wikipedia.org/wiki/Hayashi%20limit
|
The Hayashi limit is a theoretical constraint upon the maximum radius of a star for a given mass. When a star is fully within hydrostatic equilibrium—a condition where the inward force of gravity is matched by the outward pressure of the gas—the star can not exceed the radius defined by the Hayashi limit. This has important implications for the evolution of a star, both during the formulative contraction period and later when the star has consumed most of its hydrogen supply through nuclear fusion.
A Hertzsprung-Russell diagram displays a plot of a star's surface temperature against the luminosity. On this diagram, the Hayashi limit forms a nearly vertical line at about 3,500 K. The outer layers of low temperature stars are always convective, and models of stellar structure for fully convective stars do not provide a solution to the right of this line. Thus in theory, stars are constrained to remain to the left of this limit during all periods when they are in hydrostatic equilibrium, and the region to the right of the line forms a type of "forbidden zone". Note, however, that there are exceptions to the Hayashi limit. These include collapsing protostars, as well as stars with magnetic fields that interfere with the internal transport of energy through convection.
Red giants are stars that have expanded their outer envelope in order to support the nuclear fusion of helium. This moves them up and to the right on the H-R diagram. However, they are constrained by the Hayashi limit not to expand beyond a certain radius.
The Hayashi limit is named after Chūshirō Hayashi, a Japanese astrophysicist.
See also
Hayashi track
Eddington limit
|
https://en.wikipedia.org/wiki/KOBO%20%28whale%29
|
KOBO (King of the Blue Ocean) is the skeleton of a juvenile blue whale on display at the New Bedford Whaling Museum in New Bedford, Massachusetts. The whale was accidentally struck and killed by a tanker and brought ashore in Rhode Island in March 1998. It was named by New Bedford sixth-grade student Katie Hallett and put on display in 2000. It shares the gallery with three other whale skeletons: a male humpback, a female North Atlantic right whale who was pregnant at the time of her death, and the right whale's fetus.
Discovery and necropsy
KOBO was found on March 3, 1998 when the crew of a pilot boat in Narragansett Bay, Rhode Island spotted a dead whale draped across the bulbous bow of the tanker Botany Triumph. The tanker's crew appeared unaware that they had struck a whale, and the tanker reversed thrust to remove it from the bow. Several days later, scientists identified it as a juvenile male blue whale between 4 and 6 years of age. This was an extremely rare find, as there are only an estimated 500 blue whales in the North Atlantic, and there had not been a recorded stranding of a blue whale on the East Coast since 1891.
After identification, the Coast Guard towed the carcass to Second Beach near Sachuest Point National Wildlife Refuge in Middletown, Rhode Island. There, a necropsy was performed, and the whale was dissected. Parts of the whale were donated to scientific institutions around the world for research on genetics, toxicology, vision, hearing, respiration, and other subjects. The necropsy found that the whale's jaw was broken and deeply cut. The whale was also bruised around its midsection, but had been relatively healthy prior to its injuries. Researchers concluded that the whale had been first struck by a propeller, possibly belonging to a different ship, and then was killed when the Botany Triumph collided with it. The exact location and date of its death is uncertain, but it was estimated that the whale was struck off the southeast coast
|
https://en.wikipedia.org/wiki/Kline%20sphere%20characterization
|
In mathematics, a Kline sphere characterization, named after John Robert Kline, is a topological characterization of a two-dimensional sphere in terms of what sort of subset separates it. Its proof was one of the first notable accomplishments of R. H. Bing; Bing gave an alternate proof using brick partitioning in his paper Complementary domains of continuous curves
A simple closed curve in a two-dimensional sphere (for instance, its equator) separates the sphere into two pieces upon removal. If one removes a pair of points from a sphere, however, the remainder is connected. Kline's sphere characterization states that the converse is true: If a nondegenerate locally connected metric continuum is separated by any simple closed curve but by no pair of points, then it is a two-dimensional sphere.
|
https://en.wikipedia.org/wiki/Genetics%20of%20synesthesia
|
The genetic mechanism of synesthesia has long been debated, with researchers previously claiming it was a single X-linked trait due to seemingly higher prevalence in women and no evidence of male-male transmission This is where the only synesthetic parent is male and the male child has synesthesia, meaning that the trait cannot be solely linked to the X chromosome.
The Mendelian nature of the trait was further disproven when case studies showed that the Phenotype of synesthesia could be differentially expressed in monozygotic (genotypically identical) twins While both twins had the same genome with the potential for phenotypic expression of synesthesia, only one had documented synesthesia. Therefore, the condition is now thought to be oligogenic, with Locus heterogeneity and multiple forms of inheritance, and expression, implying that synesthesia is determined by more than one gene, more than one location in those genes, and a complex mode of inheritance. Several full genome linkage scans have shown particular areas of the genome whose inheritance seem to correlate with the inheritance of synesthesia.
Using the LOD score which describes the likelihood that two genes are near each other on a chromosome, and thus will be inherited together, areas of strong or suggestive linkage with inheritance of synesthesia were found. The area with the highest LOD score in the genome of an individual with auditory-visual synesthesia has been shown to be linked with autism as well, another disorder with sensory and perceptual abnormalities. Other regions of linkage include genes that are related to the development of the cerebral cortex (TBR1), dyslexia, and apoptosis (EFHC1), the last of which could be potentially related to the retention of the neonatal synesthetic pathways in the universal synesthesia/pruning hypothesis. This hypothesis posits that every person is born a synesthete and the ‘extra’ connections are pruned during normal neurodevelopment in non-synesthetes, and n
|
https://en.wikipedia.org/wiki/Force10
|
Dell Force10 (formerly nCore Networks, Force10 Networks), was a United States company that developed and marketed 10 Gigabit and 40 Gigabit Ethernet switches for computer networking to corporate, educational, and governmental customers. It had offices in North America, Europe, and the Asia Pacific region.
In August 2011, Dell completed the acquisition of Force10 and changed the name to Dell Force10.
In mid 2013, the Force10 designation was dropped from the products in favor of the data center networking line of the Dell Networking brand, and some of the other product lines were sold.
History
Founding
The company was founded by PK Dubey, Naresh Nigam and Som Sikdar. It was named by founder Som Sikdar, an avid sailor, after Beaufort Force 10 (Storm, Whole gale) on the Beaufort scale for wind speeds, indicating a storm with high speed winds, and matched their focus on 10 Gigabit Ethernet switching and routing products.
Acquisition
In January 2009, Force10 was acquired by Turin Networks (Founded by Philip Yim), which had previously purchased Carrier Access Corporation and White Rock Networks. Carrier Access Corporation itself had previously purchased Mangrove Systems and White Rock Networks had previously purchased Seranoa Networks.
On July 20, 2011 Dell announced it intended to fully acquire Force10 for an undisclosed amount. With the acquisition, Dell offered products for the data center where Dell focuses on the Ethernet switches. Dell Force10 continued to offer their non-Ethernet backhaul and metro-access platforms as well.
Telmar Network Technology of Plano, Texas, announced the acquisition of the Force10 Turin transport product lines from Dell in May, 2013, and has resumed support and development of the Traverse, TraverseEdge, TransAccess, TransNav, MasterSeries, Adit, Wide Bank, and Broadmore products. Telmar Network Technology, Inc. is a wholly owned subsidiary of Jabil Circuit, Inc. of St. Petersburg, FL.
iQor of St. Petersburg, FL, announced, in Dece
|
https://en.wikipedia.org/wiki/Totally%20positive%20matrix
|
In mathematics, a totally positive matrix is a square matrix in which all the minors are positive: that is, the determinant of every square submatrix is a positive number. A totally positive matrix has all entries positive, so it is also a positive matrix; and it has all principal minors positive (and positive eigenvalues). A symmetric totally positive matrix is therefore also positive-definite. A totally non-negative matrix is defined similarly, except that all the minors must be non-negative (positive or zero). Some authors use "totally positive" to include all totally non-negative matrices.
Definition
Let
be an n × n matrix. Consider any and any p × p submatrix of the form
where:
Then A is a totally positive matrix if:
for all submatrices that can be formed this way.
History
Topics which historically led to the development of the theory of total positivity include the study of:
the spectral properties of kernels and matrices which are totally positive,
ordinary differential equations whose Green's function is totally positive (by M. G. Krein and some colleagues in the mid-1930s),
the variation diminishing properties (started by I. J. Schoenberg in 1930),
Pólya frequency functions (by I. J. Schoenberg in the late 1940s and early 1950s).
Examples
For example, a Vandermonde matrix whose nodes are positive and increasing is a totally positive matrix.
See also
Compound matrix
|
https://en.wikipedia.org/wiki/Fulkerson%E2%80%93Chen%E2%80%93Anstee%20theorem
|
The Fulkerson–Chen–Anstee theorem is a result in graph theory, a branch of combinatorics. It provides one of two known approaches solving the digraph realization problem, i.e. it gives a necessary and sufficient condition for pairs of nonnegative integers to be the indegree-outdegree pairs of a simple directed graph; a sequence obeying these conditions is called "digraphic". D. R. Fulkerson (1960) obtained a characterization analogous to the classical Erdős–Gallai theorem for graphs, but in contrast to this solution with exponentially many inequalities. In 1966 Chen improved this result in demanding the additional constraint that the integer pairs must be sorted in non-increasing lexicographical order leading to n inequalities. Anstee (1982) observed in a different context that it is sufficient to have . Berger reinvented this result and gives a direct proof.
Statement
A sequence of nonnegative integer pairs with is digraphic if and only if and the following inequality holds for k such that :
Stronger versions
Berger proved that it suffices to consider the th inequality such that with and for .
Other notations
The theorem can also be stated in terms of zero-one matrices. The connection can be seen if one realizes that each directed graph has an adjacency matrix where the column sums and row sums correspond to and . Note that the diagonal of the matrix only contains zeros. There is a connection to the relation majorization. We define a sequence with . Sequence can also be determined by a corrected Ferrers diagram. Consider sequences , and as -dimensional vectors , and . Since by applying the principle of double counting, the theorem above states that a pair of nonnegative integer sequences with nonincreasing is digraphic if and only if vector majorizes .
Generalization
A sequence of nonnegative integer pairs with is digraphic if and only if and there exists a sequence such that the pair is digraphic and majorizes .
Characterizations for
|
https://en.wikipedia.org/wiki/Sirtris%20Pharmaceuticals
|
Sirtris Pharmaceuticals, Inc. was a biotechnology company based in Cambridge, MA that developed therapies for type 2 diabetes, cancer, and other diseases. Conceived in 2004 by Harvard University biologist David Sinclair and serial entrepreneur Andrew Perlman, and founded that year by Sinclair and Perlman, along with Christoph Westphal, Richard Aldrich, Richard Pops, and Paul Schimmel, the company was focused on developing Sinclair's research into activators of sirtuins, work that began in the laboratory of Leonard P. Guarente where Sinclair worked as a post-doc before starting his own lab.
Synopsis
The company was specifically focused on resveratrol formulations and derivatives as activators of the SIRT1 enzyme; Sinclair became known for making statements about resveratrol like: “(It's) as close to a miraculous molecule as you can find.... One hundred years from now, people will maybe be taking these molecules on a daily basis to prevent heart disease, stroke, and cancer.” Most of the anti-aging field was more cautious, especially with regard to what else resveratrol might do in the body and its lack of bioavailability. The company's initial product was called SRT501, and was a formulation of resveratrol.
Sirtris went public in 2007 and was subsequently purchased and made a subsidiary of GlaxoSmithKline in 2008 for $720 million. GSK paid $22.50/share, when Sirtris's stock was trading at $12/share, down 45% from its highest price of the previous year.
Studies published in 2009 and early 2010 by scientists from Amgen and Pfizer cast doubt on whether SIRT1 was directly activated by resveratrol and showed that the apparent activity was actually due to a fluorescent reagent used in the experiments.
In August 2010, a nonprofit called the Healthy Lifespan Institute, which had been formed the year before by Westphal and Michelle Dipp, who joined GSK from Sirtris, began selling SRT501 as a dietary supplement online; when this become public GSK required Westphal and Di
|
https://en.wikipedia.org/wiki/Silent%20mutation
|
Silent mutations are mutations in DNA that do not have an observable effect on the organism's phenotype. They are a specific type of neutral mutation. The phrase silent mutation is often used interchangeably with the phrase synonymous mutation; however, synonymous mutations are not always silent, nor vice versa. Synonymous mutations can affect transcription, splicing, mRNA transport, and translation, any of which could alter phenotype, rendering the synonymous mutation non-silent. The substrate specificity of the tRNA to the rare codon can affect the timing of translation, and in turn the co-translational folding of the protein. This is reflected in the codon usage bias that is observed in many species. Mutations that cause the altered codon to produce an amino acid with similar functionality (e.g. a mutation producing leucine instead of isoleucine) are often classified as silent; if the properties of the amino acid are conserved, this mutation does not usually significantly affect protein function.
Genetic code
The genetic code translates mRNA nucleotide sequences to amino acid sequences. Genetic information is coded using this process with groups of three nucleotides along the mRNA which are commonly known as codons. The set of three nucleotides almost always produce the same amino acid with a few exceptions like UGA which typically serves as the stop codon but can also encode tryptophan in mammalian mitochondria. Most amino acids are specified by multiple codons demonstrating that the genetic code is degenerate–different codons result in the same amino acid. Codons that code for the same amino acid are termed synonyms. Silent mutations are base substitutions that result in no change of the amino acid or amino acid functionality when the altered messenger RNA (mRNA) is translated. For example, if the codon AAA is altered to become AAG, the same amino acid – lysine – will be incorporated into the peptide chain.
Mutations are often linked to diseases or negativ
|
https://en.wikipedia.org/wiki/Paternal%20care
|
In biology, paternal care is parental investment provided by a male to his own offspring. It is a complex social behaviour in vertebrates associated with animal mating systems, life history traits, and ecology. Paternal care may be provided in concert with the mother (biparental care) or, more rarely, by the male alone (so called exclusive paternal care).
The provision of care, by either males or females, is presumed to increase growth rates, quality, and/or survival of young, and hence ultimately increase the inclusive fitness of parents. In a variety of vertebrate species (e.g., about 80% of birds and about 6% of mammals), both males and females invest heavily in their offspring. Many of these biparental species are socially monogamous, so individuals remain with their mate for at least one breeding season.
Exclusive paternal care has evolved multiple times in a variety of organisms, including invertebrates, fishes, and amphibians.
Mammals
Male mammals employ different behaviors to enhance their reproductive success (e.g. courtship displays, mate choice). However, the benefits of paternal care have rarely been studied in mammals, largely because only 5-10% of mammals exhibit such care (mostly present in primates, rodents and canids). In those species in which males provide extensive care for their offspring, indirect evidence suggests that its costs can be substantial. For example, mammalian fathers that care for their young may undergo changes in body mass and an increase in production of a number of costly hormones (e.g. androgens, glucocorticoids, leptin). Nonetheless, there is evidence that suggest that across all mammals, when males carry and groom their offspring their female partner fecundity increases, and if males provision the females, their litter size tend to be larger.
Humans
Human cultures and societies vary widely in the expression of paternal care. Some cultures recognize paternal care via celebration of Father's Day. According to CARTA ,
|
https://en.wikipedia.org/wiki/Descriptive%20interpretation
|
According to Rudolf Carnap, in logic, an interpretation is a descriptive interpretation (also called a factual interpretation) if at least one of the undefined symbols of its formal system becomes, in the interpretation, a descriptive sign (i.e., the name of single objects, or observable properties). In his Introduction to Semantics (Harvard Uni. Press, 1942) he makes a distinction between formal interpretations which are logical interpretations (also called mathematical interpretation or logico-mathematical interpretation) and descriptive interpretations: a formal interpretation is a descriptive interpretation if it is not a logical interpretation.
Attempts to axiomatize the empirical sciences, Carnap said, use a descriptive interpretation to model reality.: the aim of these attempts is to construct a formal system for which reality is the only interpretation. - the world is an interpretation (or model) of these sciences, only insofar as these sciences are true.
Any non-empty set may be chosen as the domain of a descriptive interpretation, and all n-ary relations among the elements of the domain are candidates for assignment to any predicate of degree n.
Examples
A sentence is either true or false under an interpretation which assigns values to the logical variables. We might for example make the following assignments:
Individual constants
a: Socrates
b: Plato
c: Aristotle
Predicates:
Fα: α is sleeping
Gαβ: α hates β
Hαβγ: α made β hit γ
Sentential variables:
p "It is raining."
Under this interpretation the sentences discussed above would represent the following English statements:
p: "It is raining."
F(a): "Socrates is sleeping."
H(b,a,c): "Plato made Socrates hit Aristotle."
x(F(x)): "Everybody is sleeping."
z(G(a,z)): "Socrates hates somebody."
xyz(H(x,y,z)): "Somebody made everybody hit somebody."
xz(F(x)G(a,z)): Everybody is sleeping and Socrates hates somebody.
xyz (G(a,z)H(x,y,z)): Either Socrates hates somebody or somebody made
|
https://en.wikipedia.org/wiki/Neuropixels
|
Neuropixels probes (or "Neuropixels") are electrodes developed in 2017 to record the activity of hundreds of neurons in the brain. The probes are based on CMOS technology and have 1,000 recording sites arranged in two rows on a thin, 1-cm long shank.
The probes are used in hundreds of neuroscience laboratories including the International Brain Laboratory, to record brain activity mostly in mice and rats. By revealing the activity of vast numbers of neurons, Neuropixels probes are allowing new approaches to the study of brain processes such as sensory processing, decision making, internal state, and emotions and to create brain-machine interfaces.
The probes were announced in 2017. They are designed and fabricated by imec, an electronics research center in Belgium. In 2022, Neuropixels probes were inserted in human patients.
|
https://en.wikipedia.org/wiki/Pediastrum
|
Pediastrum is a genus of green algae, in the family Hydrodictyaceae. It is a photoautotrophic, nonmotile coenobial green alga that inhabits freshwater environments.
Morphology
Pediastrum develops colonies with a fixed number of cells, termed coenobia. In this case, the coenobia are composed of between 22 and 27 cells, which are orderly arranged in a flat disk. The diameter of a single coenobium ranges from 20 to 80 μm, making them microalgae. Cells in Pediastrum are dimorphic, consisting of interior cells and peripheral cells, distinguished by their position in the colony and by their shape. Some species have inter-cellular spaces between their interior cells. The peripheral cells surround the interior cells, and they usually possess bristles, V-like cutting edges, or wavy projections.
Pediastrum shows lots of morphological diversity, both between and within species. Although most species produce flat, circular coenobia, others produce coenobia that are more oval-shaped, or curved and bowl-shaped. For species identification, the shape of the cells (particularly marginal ones), presence or absence of intercellular gaps, and ornamentation of the cell wall are all important distinguishing characteristics.
Phylogeny
The genus of Pediastrum belongs to the Hydrodictyaceae family, along with Pseudopediastrum, Tetraedron, Hydrodictyon and so on. Pediastrum spp. on the phylogenetic tree can be divided into Group I, Group II, Group III. Group I, which includes Pediastrum duplex, are mainly distributed in North America and Europe. Group II is mostly from Australia, containing among others P. angulosum and P. alternans. Group I, Group II form a monophyletic group (clade) in the phylogenetic tree together with Hydrodictyon spp. lastly, Group III forms a sister group with Monactinus spp.
For the morphological characteristics of these groups, Group I has intracellular spaces and V-like incisions in the cells along the outside, while some species in Group II lack intercellula
|
https://en.wikipedia.org/wiki/Tripod%20%28surveying%29
|
A surveyor's tripod is a device used to support any one of a number of surveying instruments, such as theodolites, total stations, levels or transits.
History
The modern sturdy, but portable, tripod stand with three leg pairs hinged to a triangular metal head was invented and first manufactured for sale by Sir Francis Ronalds in the late 1820s in Croydon. He sold 140 of the stands in the decade 1830-40 and his design was soon imitated by others.
Older surveying tripods had slightly different features compared to modern ones. For example, on some older tripods, the instrument had its own footplate and did not need to move laterally relative to the tripod head. For this reason, the head of the tripod was not a flat footplate but was simply a large diameter fitting. Threads on the outside of the head engaged threads on the instrument's footplate. No other mounting screw was used.
Fixed length legs were also seen on older instruments. Instrument height was adjusted by changing the angle of the legs. Widely spaced tripod feet resulted in a lower instrument while closely spaced legs raised the instrument. This was considerably less convenient than having variable length legs.
Materials for older tripods were predominantly wood and brass, with some steel for high wear items like the feet or foot points.
Usage
The tripod is placed in the location where it is needed. The surveyor will press down on the legs' platforms to securely anchor the legs in soil or to force the feet to a low position on uneven, pock-marked pavement. Leg lengths are adjusted to bring the tripod head to a convenient height and make it roughly level.
Once the tripod is positioned and secure, the instrument is placed on the head. The mounting screw is pushed up under the instrument to engage the instrument's base and screwed tight when the instrument is in the correct position. The flat surface of the tripod head is called the foot plate and is used to support the adjustable feet of the instrume
|
https://en.wikipedia.org/wiki/Norverapamil
|
Norverapamil is a calcium channel blocker. It is the main active metabolite of verapamil.
|
https://en.wikipedia.org/wiki/Daniel%20P.%20Sanders
|
Daniel P. Sanders is an American mathematician. He is known for his 1996 efficient proof (algorithm) of proving the Four color theorem (with Neil Robertson, Paul Seymour, and Robin Thomas). He used to be a guest professor of the department of computer science at Columbia University.
Sanders received his Ph.D. in algorithms, combinatorics, and optimization from Georgia Tech in 1993 under the guidance of professor Robin Thomas. He was the Graph Theory Resources editor of www.graphtheory.com. Sanders is a quantitative strategist at Renaissance Technologies. He has been on the faculty of the mathematics departments of Ohio State University and Princeton University.
Select work
On linear recognition of tree-width at most four, DP Sanders - SIAM Journal on Discrete Mathematics, 1996 - link.aip.org
Efficiently four-coloring planar graphs, - gatech.edu [PS], N Robertson, DP Sanders, P Seymour, R Thomas - Proceedings of the twenty-eighth annual ACM symposium on …, 1996 - portal.acm.org
|
https://en.wikipedia.org/wiki/CDR%20coding
|
In computer science CDR coding is a compressed data representation for Lisp linked lists. It was developed and patented by the MIT Artificial Intelligence Laboratory, and implemented in computer hardware in a number of Lisp machines derived from the MIT CADR.
CDR coding is in fact a fairly general idea; whenever a data object A ends in a reference to another data structure B, we can instead place the structure B itself there, overlapping and running off the end of A. By doing this we free the space required by the reference, which can add up if done many times, and also improve locality of reference, enhancing performance on modern machines. The transformation is especially effective for the cons-based lists it was created for; we free about half of the space for each node we perform this transformation on.
It is not always possible to perform this substitution, because there might not be a large enough chunk of free space beyond the end of A. Thus, some objects will end in a real reference, and some with the referenced object, and the machine must be able to tell by reading the final cell which one it is. This can be accomplished with some inefficiency in software by the use of tagged pointers, which allow a pointer in a final position to be specifically tagged as such, but is best done in hardware.
In the presence of mutable objects, CDR coding becomes more complex. If a reference is updated to point to another object, but currently has an object stored in that field, the object must be relocated, along with any other pointers to it. Not only are such moves typically expensive or impossible, but over time they cause fragmentation of the store. This problem is typically avoided by using CDR coding only on immutable data structures.
External links
Lisp (programming language)
Data compression
|
https://en.wikipedia.org/wiki/Nested%20stack%20automaton
|
In automata theory, a nested stack automaton is a finite automaton that can make use of a stack containing data which can be additional stacks.
Like a stack automaton, a nested stack automaton may step up or down in the stack, and read the current symbol; in addition, it may at any place create a new stack, operate on that one, eventually destroy it, and continue operating on the old stack. This way, stacks can be nested recursively to an arbitrary depth; however, the automaton always operates on the innermost stack only.
A nested stack automaton is capable of recognizing an indexed language, and in fact the class of indexed languages is exactly the class of languages accepted by one-way nondeterministic nested stack automata.
Nested stack automata should not be confused with embedded pushdown automata, which have less computational power.
Formal definition
Automaton
A (nondeterministic two-way) nested stack automaton is a tuple where
Q, Σ, and Γ is a nonempty finite set of states, input symbols, and stack symbols, respectively,
[, ], and ] are distinct special symbols not contained in Σ ∪ Γ,
[ is used as left endmarker for both the input string and a (sub)stack string,
] is used as right endmarker for these strings,
] is used as the final endmarker of the string denoting the whole stack.
An extended input alphabet is defined by Σ' = Σ ∪ {[,]}, an extended stack alphabet by Γ' = Γ ∪ {]}, and the set of input move directions by D = {-1,0,+1}.
δ, the finite control, is a mapping from Q × Σ' × (Γ' ∪ [Γ' ∪ {], []}) into finite subsets of Q × D × ([Γ* ∪ D), such that δ maps
Informally, the top symbol of a (sub)stack together with its preceding left endmarker "[" is viewed as a single symbol; then δ reads
the current state,
the current input symbol, and
the current stack symbol,
and outputs
the next state,
the direction in which to move on the input, and
the direction in which to move on the stack, or the string of symbols to replace the topmos
|
https://en.wikipedia.org/wiki/Comparison%20of%20HTML%20parsers
|
HTML parsers are software for automated Hypertext Markup Language (HTML) parsing. They have two main purposes:
HTML traversal: offer an interface for programmers to easily access and modify the "HTML string code". Canonical example: DOM parsers.
HTML clean: to fix invalid HTML and to improve the layout and indent style of the resulting markup. Canonical example: HTML Tidy.
* Latest release (of significant changes) date.
** sanitize (generating standard-compatible web-page, reduce spam, etc.) and clean (strip out surplus presentational tags, remove XSS code, etc.) HTML code.
*** Updates HTML4.X to XHTML or to HTML5, converting deprecated tags (ex. CENTER) to valid ones (ex. DIV with style="text-align:center;").
|
https://en.wikipedia.org/wiki/Finite%20impulse%20response
|
In signal processing, a finite impulse response (FIR) filter is a filter whose impulse response (or response to any finite length input) is of finite duration, because it settles to zero in finite time. This is in contrast to infinite impulse response (IIR) filters, which may have internal feedback and may continue to respond indefinitely (usually decaying).
The impulse response (that is, the output in response to a Kronecker delta input) of an Nth-order discrete-time FIR filter lasts exactly samples (from first nonzero element through last nonzero element) before it then settles to zero.
FIR filters can be discrete-time or continuous-time, and digital or analog.
Definition
For a causal discrete-time FIR filter of order N, each value of the output sequence is a weighted sum of the most recent input values:
where:
is the input signal,
is the output signal,
is the filter order; an th-order filter has terms on the right-hand side
is the value of the impulse response at the ith instant for of an -order FIR filter. If the filter is a direct form FIR filter then is also a coefficient of the filter.
This computation is also known as discrete convolution.
The in these terms are commonly referred to as s, based on the structure of a tapped delay line that in many implementations or block diagrams provides the delayed inputs to the multiplication operations. One may speak of a 5th order/6-tap filter, for instance.
The impulse response of the filter as defined is nonzero over a finite duration. Including zeros, the impulse response is the infinite sequence:If an FIR filter is non-causal, the range of nonzero values in its impulse response can start before , with the defining formula appropriately generalized.
Properties
An FIR filter has a number of useful properties which sometimes make it preferable to an infinite impulse response (IIR) filter. FIR filters:
Require no feedback. This means that any rounding errors are not compounded by summed iterations.
|
https://en.wikipedia.org/wiki/Scott%20Yanoff
|
Scott Yanoff (born October 20, 1969) is an IT manager and web developer who was a key person in the early days of the internet, most notably for creating and maintaining the Yanoff List, an alphabetical list of internet sites.
Career
Yanoff authored the Inter-Network Mail Guide, a text written in 1997 documenting the different methods of sending email from one network to another. He was also a co-author of The Web Site Administrator's Survival Guide with Jerry Ablan, a book that explains how to set up, administer, care for, and feed your own Web server. Most of this work was accomplished as an undergraduate student at the University of Wisconsin–Milwaukee, while working as a mainframe/UNIX consultant for the university.
He has worked for SpectraCom, Inc., and the now-defunct Strong Capital Management in Menomonee Falls, Wisconsin and at Northwestern Mutual in Milwaukee, Wisconsin from February, 2004 to June, 2023.
The Yanoff List
In the early and mid-1990s, before the use of search engines, the Yanoff List became an important tool for internet users. The list consisted of internet sites listed alphabetically and grouped by subject acting as a type of internet yellow pages containing hundreds of FTP, gopher, and web locations relevant to each subject. Users of the internet in the early 1990s would eagerly await the latest version of this list. As a minor tribute to his service, a popular Palm-based newsreader, Yanoff, was named after him.
Additional work
Yanoff created a visual basic script called "iTunesStats" in 2008 that can be run on Windows-based computers to generate a file of statistics of one's listening habits based upon the user's iTunes library. Additionally, he transposed popular music guitar tablature in the 1990s including that of The Beatles, R.E.M., Bruce Springsteen, and U2.
|
https://en.wikipedia.org/wiki/Applix
|
Applix Inc. was a computer software company founded in 1983 based in Westborough, Massachusetts that published Applix TM1, a multi-dimensional online analytical processing (MOLAP) database server, and related presentation tools, including Applix Web and Applix Executive Viewer. Together, Applix TM1, Applix Web and Applix Executive Viewer were the three core components of the Applix Business Analytics Platform. (Executive Viewer was subsequently discontinued by IBM.)
On October 25, 2007, Applix was acquired by Cognos. Cognos rebranded all Applix products under its name following the acquisition. On January 31, 2008, Cognos was itself acquired by IBM.
Prior to OLAP industry consolidation in 2007, Applix was the purest OLAP vendor among publicly traded independent business intelligence vendors, and had the greatest growth rate. TM1 is now marketed as IBM Cognos TM1; version 10.2 became publicly available on September 13, 2013.
Products and technology
Applix TM1 is enterprise planning software used for collaborative planning, budgeting and forecasting, as well as analytical and reporting applications. Data in TM1 is stored and represented as multidimensional cubes, with data being stored at the "leaf" level. Computations on the leaf data are performed in real-time (for example, to aggregate numbers up a dimensional hierarchy). IBM Cognos TM1 also includes a data orchestration environment for accessing external data and systems, as well as capabilities designed for common business planning and budgeting requirements (e.g. workflow, top-down adjustments).
See also
Applixware
|
https://en.wikipedia.org/wiki/Interchromatin%20granule
|
An interchromatin granule is a cluster in the nucleus of a mammal cell which is enriched in pre-mRNA splicing factors. Interchromatin granules are located in the interchromatin regions of the mammalian Cell nuclei. They usually appear as irregularly shaped structures that vary in size and number. They can be observed by immunofluorescence microscopy.
Interchromatin granules are structures undergoing constant change, and their components exchange continuously with the nucleoplasm, active transcription sites and other nuclear locations.
Research on dynamics of interchromatin granules has provided new insight into the functional organisation of the nucleus and gene expression.
Interchromatin granule clusters vary in size anywhere between one and several micrometers in diameter. They are composed of 20–25 nm granules that are connected in a beaded chain fashion appearance by thin fibrils.
Interchromatin granule clusters (IGCs) have been proposed to be stockpiles of fully mature snRNPs and other RNA processing components that are ready to be used in the production of mRNA.
See also
are subnuclear structures that are enriched in pre-messenger RNA splicing factors
|
https://en.wikipedia.org/wiki/Insulin%20glargine
|
Insulin glargine sold under the brand name Lantus among others is a long-acting modified form of medical insulin, used in the management of type I and type II diabetes. It is injected just under the skin. Effects generally begin an hour after use.
Common side effects include low blood sugar, problems at the site of injection, itchiness, and weight gain. Other serious side effects include low blood potassium. NPH insulin rather than insulin glargine is generally preferred in pregnancy. After injection, microcrystals slowly release insulin for about 24 hours. This insulin causes body tissues to absorb glucose from the blood and decreases glucose production by the liver.
Insulin glargine was approved for medical use in the United States in 2000. It is on the World Health Organization's List of Essential Medicines. In 2020, it was the 32nd most commonly prescribed medication in the United States with more than 17million prescriptions. In July 2021, the US Food and Drug Administration (FDA) approved an interchangeable biosimilar insulin product called Semglee (insulin glargine-yfgn) for the treatment of diabetes.
Medical uses
The long-acting insulin class, which includes insulin glargine, do not appear much better than neutral protamine Hagedorn (NPH) insulin, but do have a greater cost, making them, as of 2010, not cost effective for the treatment of type 2 diabetes. In a previous review it was unclear if there is a difference in hypoglycemia, as there was not enough data to determine any differences with respect to long term outcomes, however a more recent Cochrane systematic review did not find clinically significant difference when comparing insulin glargine to NPH insulin, insulin detemir or insulin degludec in the management of type I Diabetes in neither adults or children in periods of 6 months or longer. It is not typically the recommended long acting insulin in the United Kingdom.
Semglee is indicated to improve glycemic control in adults and children with
|
https://en.wikipedia.org/wiki/FORE%20Systems
|
FORE Systems was a computer network switching equipment company based in Pittsburgh, Pennsylvania. Founded in 1990 to supply Asynchronous Transfer Mode (ATM) cards for workstation computers, it soon branched out to become a major supplier in the ATM switch market and the extended those product lines to add Internet Protocol switching and other devices.
FORE was purchased by the British General Electric Company (GEC) in 1999, just before the dot-com bubble, which eventually drove GEC to reform the company as Marconi. Ericsson purchased most of Marconi in 2006.
History
Francois Bitz, Onat Menzilcioglu, Robert Sansom and Eric Cooper were researchers at Carnegie Mellon University who felt their 1989 project to supply a US Naval Research Laboratories Request for Development had commercial potential. They formed FORE, using the first letter of their first names, in 1990 in the Pittsburgh suburb of Warrendale, Pennsylvania. The company went public in May 1994.
FORE initially produced Asynchronous Transfer Mode (ATM) network interface cards for Sun Microsystems' SBus and DEC's TURBOchannel based computers in addition to ATM adapters for Silicon Graphics computers. FORE's first ATM switch – the ASX-100 – connected sixteen ATM ports each with a capacity of 155.52 Mbit/s. Later interface adapters for the follow-on ASX-200, ASX-1000, and ASX-4000 switches allowed connections up to 2.5 Gbit/s.
The company established a leading position in the market for ATM switching equipment. FORE created a memory-based ATM switch that captured a strong portion of the ATM market. Other technologies include Internet Protocol, Gigabit Ethernet and Firewall switching. FORE Systems also supported advanced dynamic routing protocols such as the ForeThought Private Network Network Interface (PNNI) and the ATM Forum's PNNI protocol.
FORE was acquired by London-based GEC (now Marconi Corporation plc) on 26 April 1999 to complement its Marconi Communications business and to increase its presence i
|
https://en.wikipedia.org/wiki/Baryon%20number
|
In particle physics, the baryon number is a strictly conserved additive quantum number of a system. It is defined as
where is the number of quarks, and is the number of antiquarks. Baryons (three quarks) have a baryon number of +1, mesons (one quark, one antiquark) have a baryon number of 0, and antibaryons (three antiquarks) have a baryon number of −1. Exotic hadrons like pentaquarks (four quarks, one antiquark) and tetraquarks (two quarks, two antiquarks) are also classified as baryons and mesons depending on their baryon number.
Baryon number vs. quark number
Quarks carry not only electric charge, but also charges such as color charge and weak isospin. Because of a phenomenon known as color confinement, a hadron cannot have a net color charge; that is, the total color charge of a particle has to be zero ("white"). A quark can have one of three "colors", dubbed "red", "green", and "blue"; while an antiquark may be either "anti-red", "anti-green" or "anti-blue".
For normal hadrons, a white color can thus be achieved in one of three ways:
A quark of one color with an antiquark of the corresponding anticolor, giving a meson with baryon number 0,
Three quarks of different colors, giving a baryon with baryon number +1,
Three antiquarks of different anticolors, giving an antibaryon with baryon number −1.
The baryon number was defined long before the quark model was established, so rather than changing the definitions, particle physicists simply gave quarks one third the baryon number. Nowadays it might be more accurate to speak of the conservation of quark number.
In theory, exotic hadrons can be formed by adding pairs of quarks and antiquarks, provided that each pair has a matching color/anticolor. For example, a pentaquark (four quarks, one antiquark) could have the individual quark colors: red, green, blue, blue, and antiblue. In 2015, the LHCb collaboration at CERN reported results consistent with pentaquark states in the decay of bottom Lambda baryons
|
https://en.wikipedia.org/wiki/Lateral%20cutaneous%20nerve%20of%20forearm
|
The lateral cutaneous nerve of forearm (or lateral antebrachial cutaneous nerve) is a sensory nerve representing the continuation of the musculocutaneous nerve beyond the lateral edge of the tendon of the biceps brachii muscle. The lateral cutaneous nerve provides sensory innervation to the skin of the lateral forearm. It pierces the deep fascia of forearm to enter the subcutaneous compartment before splitting into a volar branch and a dorsal branch.
Anatomy
Course and relations
It passes behind the cephalic vein and divides opposite the elbow-joint into a volar branch and a dorsal branch.
Branches
Volar branch
The volar branch (ramus volaris; anterior branch) descends along the radial border of the forearm to the wrist, and supplies the skin over the lateral half of its volar surface.
At the wrist-joint it is placed in front of the radial artery, and some filaments, piercing the deep fascia, accompany that vessel to the dorsal surface of the carpus.
The nerve then passes downward to the ball of the thumb, where it ends in cutaneous filaments.
It communicates with the superficial branch of the radial nerve, and with the palmar cutaneous branch of the median nerve.
Dorsal branch
The dorsal branch (ramus dorsalis; posterior branch) descends, along the dorsal surface of the radial side of the forearm to the wrist.
It supplies the skin of the lower two-thirds of the dorso-lateral surface of the forearm, communicating with the superficial branch of the radial nerve and the posterior cutaneous nerve of forearm of the radial nerve.
See also
Medial cutaneous nerve of forearm
Superior lateral cutaneous nerve of arm
Additional images
|
https://en.wikipedia.org/wiki/ZO%20sex-determination%20system
|
The ZO sex-determination system is a system that determines the sex of offspring in several moths. In those species, there is one sex chromosome, Z. Males have two Z chromosomes, whereas females have one Z. Males are ZZ, while females are ZO.
See also
Y-chromosomal Adam
Sex Determination in Silene
Sex-determination system
Sexual differentiation
Haplodiploid sex-determination system
XY sex-determination system
XO sex-determination system
ZW sex-determination system
Temperature-dependent sex determination
X chromosome
Y chromosome
|
https://en.wikipedia.org/wiki/Vlasov%20equation
|
The Vlasov equation is a differential equation describing time evolution of the distribution function of plasma consisting of charged particles with long-range interaction, such as the Coulomb interaction. The equation was first suggested for the description of plasma by Anatoly Vlasov in 1938 and later discussed by him in detail in a monograph.
Difficulties of the standard kinetic approach
First, Vlasov argues that the standard kinetic approach based on the Boltzmann equation has difficulties when applied to a description of the plasma with long-range Coulomb interaction. He mentions the following problems arising when applying the kinetic theory based on pair collisions to plasma dynamics:
Theory of pair collisions disagrees with the discovery by Rayleigh, Irving Langmuir and Lewi Tonks of natural vibrations in electron plasma.
Theory of pair collisions is formally not applicable to Coulomb interaction due to the divergence of the kinetic terms.
Theory of pair collisions cannot explain experiments by Harrison Merrill and Harold Webb on anomalous electron scattering in gaseous plasma.
Vlasov suggests that these difficulties originate from the long-range character of Coulomb interaction. He starts with the collisionless Boltzmann equation (sometimes called the Vlasov equation, anachronistically in this context), in generalized coordinates:
explicitly a PDE:
and adapted it to the case of a plasma, leading to the systems of equations shown below. Here is a general distribution function of particles with momentum at coordinates and given time . Note that the term is the force acting on the particle.
The Vlasov–Maxwell system of equations (Gaussian units)
Instead of collision-based kinetic description for interaction of charged particles in plasma, Vlasov utilizes a self-consistent collective field created by the charged plasma particles. Such a description uses distribution functions and for electrons and (positive) plasma ions. The distribution functi
|
https://en.wikipedia.org/wiki/Internal%20iliac%20vein
|
The internal iliac vein (hypogastric vein) begins near the upper part of the greater sciatic foramen, passes upward behind and slightly medial to the internal iliac artery and, at the brim of the pelvis, joins with the external iliac vein to form the common iliac vein.
Structure
Several veins unite above the greater sciatic foramen to form the internal iliac vein. It does not have the predictable branches of the internal iliac artery but its tributaries drain the same regions. The internal iliac vein emerges from above the level of the greater sciatic notch It runs backwards, upwards and towards the midline to join the external iliac vein in forming the common iliac vein in front of the sacroiliac joint. It usually lies lateral to the internal iliac artery. It is wide and 3 cm long.
Tributaries
Originating outside the pelvis, its tributaries are the gluteal, internal pudendal and obturator veins. Running from the anterior surface of the sacrum are the lateral sacral veins. Coming from the pelvic plexuses and appropriate to gender are the middle rectal, vesical, prostatic, uterine and vaginal veins.
Variation
On the left, the internal iliac vein lies lateral to the internal iliac artery 73% of the time. On the right, this is 93% of the time.
Function
The internal iliac veins drain the pelvic organs, sacrum, and coccyx.
Clinical significance
If thrombosis disrupts blood flow in the external iliac systems, the internal iliac tributaries offer a major route of venous return from the femoral system. Damage to internal iliac vein tributaries during surgery can seriously compromise venous drainage and cause swelling of one or both legs.
Additional images
|
https://en.wikipedia.org/wiki/Decision-theoretic%20rough%20sets
|
In the mathematical theory of decisions, decision-theoretic rough sets (DTRS) is a probabilistic extension of rough set classification. First created in 1990 by Dr. Yiyu Yao, the extension makes use of loss functions to derive and region parameters. Like rough sets, the lower and upper approximations of a set are used.
Definitions
The following contains the basic principles of decision-theoretic rough sets.
Conditional risk
Using the Bayesian decision procedure, the decision-theoretic rough set (DTRS) approach allows for minimum-risk decision making based on observed evidence. Let be a finite set of
possible actions and let be a finite set of states. is
calculated as the conditional probability of an object being in state given the object description
. denotes the loss, or cost, for performing action when the state is .
The expected loss (conditional risk) associated with taking action is given
by:
Object classification with the approximation operators can be fitted into the Bayesian decision framework. The
set of actions is given by , where , , and represent the three
actions in classifying an object into POS(), NEG(), and BND() respectively. To indicate whether an
element is in or not in , the set of states is given by . Let
denote the loss incurred by taking action when an object belongs to
, and let denote the loss incurred by take the same action when the object
belongs to .
Loss functions
Let denote the loss function for classifying an object in into the POS region, denote the loss function for classifying an object in into the BND region, and let denote the loss function for classifying an object in into the NEG region. A loss function denotes the loss of classifying an object that does not belong to into the regions specified by .
Taking individual can be associated with the expected loss actions and can be expressed as:
where , , and , , or .
Minimum-risk decision rules
If we consider the loss functions and
|
https://en.wikipedia.org/wiki/Water%20Talkies
|
Water Talkies are devices which makes talking underwater possible. They were invented by Richie Stachowski, Jr. in 1995 at age 11.
It is a colorful, cone-shaped device that amplifies voices underwater for up to . What started as one product turned into a company producing water toys "made by a kid for kids" - a line of eight pool toys.
Stachowski won the 1999 Entrepreneur of the Year Award, the youngest ever, for this invention.
See also
Megaphone
|
https://en.wikipedia.org/wiki/Davidson%20Seamount
|
Davidson Seamount is a seamount (underwater volcano) located off the coast of Central California, southwest of Monterey and west of San Simeon. At long and wide, it is one of the largest known seamounts in the world. From base to crest, the seamount is tall, yet its summit is still below the sea surface. The seamount is biologically diverse, with 237 species and 27 types of deep-sea coral having been identified.
Discovered during the mapping of California's coast in 1933, Davidson Seamount is named after geographer George Davidson of the United States Coast and Geodetic Survey. Studied only sparsely for decades, NOAA expeditions to the seamount in 2002 and 2006 cast light upon its unique deep-sea coral ecosystem. Davidson Seamount is populated by a dense population of large, ancient corals, some of which are over 100 years of age. The data gathered during the studies fueled the making of Davidson Seamount into a part of the Monterey Bay National Marine Sanctuary in 2009.
Geology
A seamount such as Davidson is an underwater volcano; this one rises above the surrounding ocean floor. Although there are over 30,000 seamounts in the Pacific Ocean alone, only about 0.1% of them have been explored. The aqueous environment of the seamount means that it behaves differently from volcanoes on land. Its surface is composed mostly of blocky lava flows, although some pillow lava, which is the typical lava type of a seamount, prevails at the deeper flank. The summit is composed of layered deposits of volcanic ash and pyroclastic material. These rocks indicate mildly explosive eruptions of gas-rich lava near the summit of the volcano. The base of Davidson is probably buried in a deep layer of muds.
At long and wide, Davidson Seamount is impressively large. If it were on land, it would dominate the landscape in a way similar to how Mount Shasta dominates the horizon of northern California. Put in perspective, the size of the seamount is enough to fill Monterey Bay from t
|
https://en.wikipedia.org/wiki/Bithionol
|
Bithionol is an antibacterial, anthelmintic, and algaecide. It is used to treat Anoplocephala perfoliata (tapeworms) in horses and Fasciola hepatica (liver flukes).
Mechanism of action
Bithionol has been shown to be a potent inhibitor of soluble adenylyl cyclase, an intracellular enzyme important in the catalysis of adenosine triphosphate (ATP) to cyclic adenosine monophosphate (cAMP). Soluble adenylyl cyclase is uniquely activated by bicarbonate. The cAMP formed by this enzyme is associated with capacitation of sperm, eye pressure regulation, acid-base regulation, and astrocyte/neuron communication.
It is related to the organochlorine hexachlorophene, which has been shown to be an isomer-specific inhibitor of soluble adenylyl cyclase. Bithionol has two aromatic rings with a sulfur atom bonded between them and multiple chlorine ions and hydroxyl groups attached to the phenyl groups. These functional groups are capable of hydrophobic, ionic, and polar interactions.
These intermolecular interactions are responsible for the binding of bithionol to the bicarbonate binding site of soluble adenylyl cyclase efficiently enough to cause competitive inhibition with the usual bicarbonate substrate. The side chain of arginine 176 within the bicarbonate binding site interacts significantly with the aromatic ring of the bithionol molecule. This allosteric, conformational change interferes with the ability of the active site of soluble adenylyl cyclase to adequately bind ATP to convert it into cAMP. Arginine 176 usually interacts with the ATP and other catalytic ions at the active site, so when it turns from its normal position to interact with the bithionol inhibitor, it no longer functions in keeping the ATP bound to the active site.
In another form of inhibition, bithionol is a much larger molecule than simple sodium bicarbonate, so it is large enough to reach through a small channel in the soluble adenylyl cyclase and interfere with binding of ATP, preventing its conve
|
https://en.wikipedia.org/wiki/James%20McKernan
|
James McKernan (born 1964) is a mathematician, and a professor of mathematics at the University of California, San Diego. He was a professor at MIT from 2007 until 2013.
Education
McKernan was educated at the Campion School, Hornchurch, and Trinity College, Cambridge, before going on to earn his Ph.D. from Harvard University in 1991. His dissertation, On the Hyperplane Sections of a Variety in Projective Space, was supervised by Joe Harris.
Recognition
McKernan was the joint winner of the Cole Prize in 2009, and joint recipient of the Clay Research Award in 2007. Both honors were received jointly with his colleague Christopher Hacon. He gave an invited talk at the International Congress of Mathematicians in 2010, on the topic of "Algebraic Geometry". He was the joint winner (with Christopher Hacon) of the 2018 Breakthrough Prize in Mathematics.
He was elected as a Fellow of the American Mathematical Society in the 2020 Class, for "contributions to algebraic geometry, in particular his proof of the finite generation of the canonical ring, the existence of flips and the boundedness of varieties of log general type".
|
https://en.wikipedia.org/wiki/Acetic%20acid/hydrocortisone
|
Acetic acid/hydrocortisone is a commonly used combination drug to treat infections of the outer ear and ear canal. Branded as Vosol HC and Acetasol HC, it combines the antibacterial and antifungal action of acetic acid with the anti-inflammatory functions of hydrocortisone.
|
https://en.wikipedia.org/wiki/Dynamic%20linker
|
In computing, a dynamic linker is the part of an operating system that loads and links the shared libraries needed by an executable when it is executed (at "run time"), by copying the content of libraries from persistent storage to RAM, filling jump tables and relocating pointers. The specific operating system and executable format determine how the dynamic linker functions and how it is implemented.
Linking is often referred to as a process that is performed when the executable is compiled, while a dynamic linker is a special part of an operating system that loads external shared libraries into a running process and then binds those shared libraries dynamically to the running process. This approach is also called dynamic linking or late linking.
Implementations
Microsoft Windows
Dynamic-link library, or DLL, is Microsoft's implementation of the shared library concept in the Microsoft Windows and OS/2 operating systems. These libraries usually have the file extension DLL, OCX (for libraries containing ActiveX controls), or DRV (for legacy system drivers). The file formats for DLLs are the same as for Windows EXE files that is, Portable Executable (PE) for 32-bit and 64-bit Windows, and New Executable (NE) for 16-bit Windows. As with EXEs, DLLs can contain code, data, and resources, in any combination.
Data files with the same file format as a DLL, but with different file extensions and possibly containing only resource sections, can be called resource DLLs. Examples of such DLLs include multi-language user interface libraries with extension MUI, icon libraries, sometimes having the extension ICL, and font files, having the extensions FON and FOT.
Unix-like systems using ELF, and Darwin-based systems
In most Unix-like systems, most of the machine code that makes up the dynamic linker is actually an external executable that the operating system kernel loads and executes first in a process address space newly constructed as a result of calling exec or posix_spa
|
https://en.wikipedia.org/wiki/Generative%20Design%20in%20Minecraft
|
GDMC (short for Generative Design in Minecraft) is a programming competition to create procedurally generated settlements in Minecraft. The competition is organized by academics from New York University, the University of Hertfordshire and the Queen Mary University of London.
Organisers
Michael Cerny Green (New York University)
Christoph Salge (University of Hertfordshire)
Rodrigo Canaan (New York University)
Christian Guckelsberger (Queen Mary University of London)
Julian Togelius (New York University)
|
https://en.wikipedia.org/wiki/Network%20scheduler
|
A network scheduler, also called packet scheduler, queueing discipline (qdisc) or queueing algorithm, is an arbiter on a node in a packet switching communication network. It manages the sequence of network packets in the transmit and receive queues of the protocol stack and network interface controller. There are several network schedulers available for the different operating systems, that implement many of the existing network scheduling algorithms.
The network scheduler logic decides which network packet to forward next. The network scheduler is associated with a queuing system, storing the network packets temporarily until they are transmitted. Systems may have a single or multiple queues in which case each may hold the packets of one flow, classification, or priority.
In some cases it may not be possible to schedule all transmissions within the constraints of the system. In these cases the network scheduler is responsible for deciding which traffic to forward and what gets dropped.
Terminology and responsibilities
A network scheduler may have responsibility in implementation of specific network traffic control initiatives. Network traffic control is an umbrella term for all measures aimed at reducing network congestion, latency and packet loss. Specifically, active queue management (AQM) is the selective dropping of queued network packets to achieve the larger goal of preventing excessive network congestion. The scheduler must choose which packets to drop. Traffic shaping smooths the bandwidth requirements of traffic flows by delaying transmission packets when they are queued in bursts. The scheduler decides the timing for the transmitted packets. Quality of service (QoS) is the prioritization of traffic based on service class (Differentiated services) or reserved connection (Integrated services).
Algorithms
In the course of time, many network queueing disciplines have been developed. Each of these provides specific reordering or dropping of network packe
|
https://en.wikipedia.org/wiki/Hyperbow
|
The Hyperbow is an electronic violin bow interface that was developed as a result of an in-depth research project by students at MIT. The instrument is intended for use only by accomplished players and was designed to amplify their gestures, which lead to supplementary sound or musical control possibilities. It offers the violin player a range of expressive possibilities in its form as an augmented bow controller that lends itself to the control of bowed string physical models.
Development
The creation of musical instruments that utilize technology to increase the functionality and subtlety of control has been an endeavour of Tod Machover and his research group since 1986. This collaboration has produced a series of "augmented instruments" called Hyperinstruments such as the HyperCello, an acoustic cello created in 1991 for Yo-Yo Ma. The Hyperbow itself has an older version called Hyperviolin, which was designed for Ani Kavafian.
The Hyperbow project sought to capture the most intricate aspects of violin bowing technique, the subtle elements that immediately and directly impact the sound of the instrument while playing. It uses accelerometers and electromagnetic field sensor in order to track the instrument's motion. There is also a mechanism that measures the force of the bow called single strain gauge. Both of these collect data that are then sent to the radio transmitter wirelessly. The data allows the evaluation of bowing techniques and drives a specified calibration procedure. The physical gesture data collected can also be used to control audio effects and synthesis algorithms in real-time.
The Hyperbow interface was featured in Toy Symphony performances with violinists Joshua Bell and Cora Venus Lunny, and several pieces for cello and Hyperbow have recently been developed and recorded by colleagues at the Royal Academy of Music.
Current research uses the Hyperbow as a measurement system to investigate the interaction of bowing parameters (acceleration,
|
https://en.wikipedia.org/wiki/IP%20traceback
|
IP traceback is any method for reliably determining the origin of a packet on the Internet. The IP protocol does not provide for the authentication of the source IP address of an IP packet, enabling the source address to be falsified in a strategy called IP address spoofing, and creating potential internet security and stability problems.
Use of false source IP addresses allows denial-of-service attacks (DoS) or one-way attacks (where the response from the victim host is so well known that return packets need not be received to continue the attack). IP traceback is critical for identifying sources of attacks and instituting protection measures for the Internet. Most existing approaches to this problem have been tailored toward DoS attack detection. Such solutions require high numbers of packets to converge on the attack path(s).
Probabilistic packet marking
Savage et al. suggested probabilistically marking packets as they traverse routers through the Internet. They propose that the router mark the packet with either the router’s IP address or the edges of the path that the packet traversed to reach the router.
For the first alternative, marking packets with the router's IP address, analysis shows that in order to gain the correct attack path with 95% accuracy as many as 294,000 packets are required. The second approach, edge marking, requires that the two nodes that make up an edge mark the path with their IP addresses along with the distance between them. This approach would require more state information in each packet than simple node marking but would converge much faster. They suggest three ways to reduce the state information of these approaches into something more manageable.
The first approach is to XOR each node forming an edge in the path with each other. Node a inserts its IP address into the packet and sends it to b. Upon being detected at b (by detecting a 0 in the distance), b XORs its address with the address of a. This new data entity is ca
|
https://en.wikipedia.org/wiki/Food%20dehydrator
|
A food dehydrator is a device that removes moisture from food to aid in its preservation. Food drying is a method of preserving fruit, vegetables and meats that has been practiced since antiquity.
Design
Most modern food dehydrators are low-power convection ovens that use heated air flow to reduce the water content of foods. The water content of food is usually very high, typically 80–95% for various fruits and vegetables and 50–75% for various meats. Removing moisture from food restrains various bacteria from growing and spoiling food. Further, removing moisture from food dramatically reduces the weight and often volume of the food, making it easier to storage. Thus, food dehydrators are used to preserve and extend the shelf life of various foods.
Food dehydrators require heat sources such as solar energy, electric power or biofuel, and vary in form from large-scale dehydration projects to do-it-yourself projects or commercially sold appliances for domestic use. A commercial food dehydrator's basic parts usually consist of a heating element, an electric fan, air vents which allow air to circulate, and food trays to lay food upon. As shown on the right, the trays most commonly have slits to provide more surface area between the food and the air. A dehydrator's heating element, fans and vents simultaneously work to direct hot air over the food, accelerate surface evaporation and warm the food causing moisture to be also released from its interior. This process continues until the food is dried to a substantially lower water content, usually less than 20%.
Most foods are dehydrated at , although meats being made into jerky should be dehydrated at a higher temperature of — or preheated to that temperature — to guard against pathogens that may already be in the meat. These temperatures are similar to those used in pasteurization, and achieve similar effects. The key to successful food dehydration is the application of a constant temperature and adequate a
|
https://en.wikipedia.org/wiki/Borden%20Base%20Line
|
The Borden Base Line is a historic survey line (7.42 miles, long) running north/south through Hatfield and South Deerfield, Massachusetts. It was completed in 1831. It was designated a National Historic Civil Engineering Landmark by the American Society of Civil Engineers in 1981.
The baseline measurement was the first project of its kind undertaken in America, and essential for Massachusetts' pioneering Trigonometrical Survey, performed under chief engineer Robert Treat Paine. Its careful measurement was critical since the accuracy of the whole triangulation network depended on it.
The baseline was measured with greater accuracy than previously possible by using a new measuring device invented by Simeon Borden, which employed a bi-metallic measuring instrument to provide constant readings despite temperature variations. His apparatus was long, enclosed in a tube, and employed with four compound microscopes.
Borden was a highly competent engineer whose ability was widely recognized. Indeed, the entire project became generally known as the Borden Survey. He measured the baseline with a nominal accuracy of better than one part in 5 million. As Professor A. D. Butterfield has written, "The work performed and results obtained far surpassed in magnitude and attainment of any previous work of this kind in America."
It appears that the north end of the baseline lies just south of the intersection of today's Route 116 and Route 5 in South Deerfield, Massachusetts. According to the Valley Historians, the south end is still marked by a copper plug set into a boulder, located in the back yard of the house at 30 Bridge Street, Hatfield, Massachusetts.
See also
Surveying
External links
American Society of Civil Engineers -Borden baseline 1831 landmark
|
https://en.wikipedia.org/wiki/Ziggurat%20algorithm
|
The ziggurat algorithm is an algorithm for pseudo-random number sampling. Belonging to the class of rejection sampling algorithms, it relies on an underlying source of uniformly-distributed random numbers, typically from a pseudo-random number generator, as well as precomputed tables. The algorithm is used to generate values from a monotonically decreasing probability distribution. It can also be applied to symmetric unimodal distributions, such as the normal distribution, by choosing a value from one half of the distribution and then randomly choosing which half the value is considered to have been drawn from. It was developed by George Marsaglia and others in the 1960s.
A typical value produced by the algorithm only requires the generation of one random floating-point value and one random table index, followed by one table lookup, one multiply operation and one comparison. Sometimes (2.5% of the time, in the case of a normal or exponential distribution when using typical table sizes) more computations are required. Nevertheless, the algorithm is computationally much faster than the two most commonly used methods of generating normally distributed random numbers, the Marsaglia polar method and the Box–Muller transform, which require at least one logarithm and one square root calculation for each pair of generated values. However, since the ziggurat algorithm is more complex to implement it is best used when large quantities of random numbers are required.
The term ziggurat algorithm dates from Marsaglia's paper with Wai Wan Tsang in 2000; it is so named because it is conceptually based on covering the probability distribution with rectangular segments stacked in decreasing order of size, resulting in a figure that resembles a ziggurat.
Theory of operation
The ziggurat algorithm is a rejection sampling algorithm; it randomly generates a point in a distribution slightly larger than the desired distribution, then tests whether the generated point is inside the
|
https://en.wikipedia.org/wiki/Adnation
|
Adnation in angiosperms is the fusion of two or more whorls of a flower, e.g. stamens to petals". This is in contrast to connation, the fusion among a single whorl.
|
https://en.wikipedia.org/wiki/Pickardt%20syndrome
|
Pickardt syndrome denotes a rare form of tertiary hypothyroidism that is caused by interruption of the portal veins connecting hypothalamus and pituitary.It was characterized in 1972 and 1973 by Renate Pickardt and Rudolf Fahlbusch.
Signs and symptoms
Cause
Interruption of the portal system may be caused by tumors compressing the infundibulum. Other causes for Pickardt's syndrome are inflammatory disorders and traumatic brain injury. An inborn variant of Pickardt's syndrome that is associated with certain mutations (HESX1 or LHX4) is referred to as pituitary stalk interruption syndrome (PSIS).
Endocrine consequences
Typical manifestations of Pickardt–Fahlbusch syndrome are hypothyroidism with reduced TSH values and functional hyperprolactinemia (which is caused by disinhibition of prolactin release). Other endocrine disorders that are usually associated with Pickardt syndrome are suprasellar failures like secondary hypogonadism, reduced levels of growth hormone and, in more severe cases, secondary adrenal insufficiency.
Diagnosis
Pickardt's syndrome may cause difficulties in differential diagnosis of pituitary adenomas, as both suprasellar hormone-inactive adenomas and prolactinomas may be associated with increased prolactin levels, central hypogonadism and central hypothyroidism. Usually, the prolactin levels are higher in case of a true prolactinoma, but the concentration ranges overlap.
Treatment
Treatment modality depends on the cause. Tumors may be removed surgically, but pituitary stalk interruption may persist. Usually, replacement of those hormones that are reduced due to failed feedback control systems will be necessary.
|
https://en.wikipedia.org/wiki/Rubber%20duck%20debugging
|
In software engineering, rubber duck debugging (or rubberducking) is a method of debugging code by articulating a problem in spoken or written natural language. The name is a reference to a story in the book The Pragmatic Programmer in which a programmer would carry around a rubber duck and debug their code by forcing themselves to explain it, line by line, to the duck. Many other terms exist for this technique, often involving different (usually) inanimate objects, or pets such as a dog or a cat. Teddy bears are also widely used.
Approach
Many programmers have had the experience of explaining a problem to someone else, possibly even to someone who knows nothing about programming, and then hitting upon the solution in the process of explaining the problem. In describing what the code is supposed to do and observing what it actually does, any incongruity between these two becomes apparent. More generally, teaching a subject forces its evaluation from different perspectives and can provide a deeper understanding. By using an inanimate object, the programmer can try to accomplish this without having to interrupt anyone else, and with better results than have been observed from merely thinking aloud without an audience. This approach has been taught in computer science and software engineering courses.
In popular culture
On 1 April 2018, Stack Exchange introduced a rubber duck avatar on their websites as a new "feature" called Quack Overflow as an April Fools' Day joke. The duck appeared at the bottom right corner of the browser viewport, and attempted to help visitors by listening to their problems and responding with solutions. However, the duck merely produced a quack sound after apparently thinking and typing. It referenced rubber ducking as a powerful method for solving problems.
See also
Code review
Pair programming
Socratic method
Desk checking
Software walkthrough
Think aloud protocol
Pointing and calling
Rogerian method
Worry dolls
|
https://en.wikipedia.org/wiki/P-Cymene
|
p-Cymene is a naturally occurring aromatic organic compound. It is classified as an alkylbenzene related to monocyclic monoterpenes. Its structure consists of a benzene ring para-substituted with a methyl group and an isopropyl group. p-Cymene is insoluble in water, but miscible with organic solvents.
Isomers and production
In addition to p-cymene, two less common geometric isomers are o-cymene, in which the alkyl groups are ortho-substituted, and m-cymene, in which they are meta-substituted. p-Cymene is the only natural isomer, as expected from the terpene rule. All three isomers form the group of cymenes.
Cymene is also produced by alkylation of toluene with propene.
Related compounds
It is a constituent of a number of essential oils, most commonly the oil of cumin and thyme. Significant amounts are formed in sulfite pulping process from the wood terpenes.
p-Cymene is a common ligand for ruthenium. The parent compound is [(η6-cymene)RuCl2]2. This half-sandwich compound is prepared by the reaction of ruthenium trichloride with the terpene α-phellandrene. The osmium complex is also known.
Hydrogenation gives the saturated derivative p-menthane.
|
https://en.wikipedia.org/wiki/Production%20truck
|
A television production truck or OB van is a small mobile production control room to allow filming of events and video production at locations outside a regular television studio. They are used for remote broadcasts, outside broadcasting (OB), and electronic field production (EFP). Some require a crew of as many as 30 people, with additional trucks for additional equipment as well as a satellite truck, which transmits video back to the studio by sending it up through a communications satellite using a satellite dish, which then transmits it back down to the studio. Alternatively, some production trucks include a satellite transmitter and satellite dish for this purpose in a single truck body to save space, time and cost.
Other television production trucks are smaller in size and generally require two or three people in the field to manage. For instance broadcast journalism news reporters providing live television, local news in the field electronic news gathering (ENG) outside a formal television studio. In some cases, it can be a station wagon, people carrier or even a motorbike (especially in cities with congested streets or where a rapid response is needed and a motorbike is more manoeuvrable).
History
One of the BBC's early Outside Broadcast vehicles, MCR 1 (short for Mobile Control Room), was built by the joint Marconi-EMI company and delivered to the BBC just in time to televise the Coronation of George VI and Elizabeth in May 1937. MCR 2 was identical to MCR 1 and was delivered in the summer of 1938. The MCRs could handle three cameras. Initially, they were standard Emitrons, but were later supplemented by Super Emitrons, which performed much better than the standard ones in low light. The MCRs were built on the chassis of an AEC Regal single-decker bus.
After World War II, the joint company Marconi-EMI ceased. The BBC ordered two 3-camera MCRs from EMI. The cameras were equipped with CPS tubes, had electronic viewfinders and a 3 lens turret. MCR 4 was
|
https://en.wikipedia.org/wiki/Friction%20motor
|
A friction motor is a simple mechanism to propel toy cars, trucks, trains, action figures and similar toys. The motor consists of a large flywheel which is connected to the drive wheels of the toy via a very low gear ratio, so that the flywheel revolves faster. The flywheel's axis is perpendicular to the direction in which the toy faces and moves. When the toy is pushed forward, the drive wheels engage the flywheel. Pushing the vehicle forward repeatedly spins this flywheel up to speed. When let go, the flywheel drives the vehicle forward. The flywheel stores the kinetic energy of the initial acceleration and propels the toy after it is released, by forcing the perpetual motor that revolves the kinetic energy.
As the flywheel, unlike the spring of a pullback motor, is continuously rotating and not held, the motor may be "pumped up" by pushing the car repeatedly forward. The cars also typically work in forward and reverse. Some used a zip cord pulled from the vehicle body to accelerate the flywheel directly. Another system was the Turbo Tower of Power (TTP) in which air expelled from a hand-operated pump pushed turbine blades on the flywheel's rim.
These toys were especially popular in the 1960s to 1980s though they continue to be available today.
|
https://en.wikipedia.org/wiki/Camo%20%28app%29
|
Camo is a freemium webcam app by British software company Reincubate allowing phones and other mobile devices to be used as webcams and document cameras.
The app runs on macOS and Microsoft Windows and is compatible with iOS and Android phones.
The app comes in a free and Pro version. The free version uses the mobile device's main camera, while the Pro version gives accesses to all cameras.
Using Camo requires downloading an app to the mobile device and a counterpart app, Camo Studio.
|
https://en.wikipedia.org/wiki/Arthur%20J.%20Krener
|
Arthur James Krener (born October 8, 1942) is a distinguished visiting professor in the department of applied mathematics at the Naval Postgraduate School. He has made contributions in the areas of control theory, nonlinear control, and stochastic processes.
Biography
He was born in Brooklyn, New York, on October 8, 1942. He received BS in mathematics (1964) from College of the Holy Cross and an MA (1967) and PhD (1971), both in mathematics, from the University of California, Berkeley.
He was a professor of mathematics at the University of California, Davis for 35 years. He retired from Davis as a distinguished professor in 2006 and joined the department of applied mathematics, Naval Postgraduate School at that time. His research interests are in developing methods for the control and estimation of nonlinear dynamical systems and stochastic processes.
In 1988 he founded the SIAM Activity Group on Control and Systems Theory and was its first chair. He was again chair of the SIAG CST in 2005–07.
In 2012, he received the Richard E. Bellman Control Heritage Award from the AACC. The citation reads "For contributions to the control and estimation of nonlinear systems."
In 2016, he received the IEEE Control Systems Award for "contributions to the analysis, control, and estimation of nonlinear control systems."
Work
In his PhD dissertation, Krener showed that the Lie bracket played an important role in nonlinear controllability by proving a time directed version of Chow's theorem.
Several years later with Hermann, he gave the definitive treatment of controllability and observability for nonlinear systems. This work was later cited by the IEEE Control Systems Society as one of Twenty Five Seminal Papers in Control, published in the twentieth century, which have made a major impact on the field of control.
With Isidori, Gori-Giorgi and Monaco, he gave conditions for the existence and construction of decoupling and noninteracting control laws for nonlinear systems.
|
https://en.wikipedia.org/wiki/British%20Oceanographic%20Data%20Centre
|
The British Oceanographic Data Centre (BODC) is a national facility for looking after and distributing data about the marine environment. BODC is the designated marine science data centre for the United Kingdom and part of the National Oceanography Centre (NOC) — primarily at its facility in Liverpool, with small number of its staff in Southampton. The centre provides a resource for science, education and industry, as well as the general public.
History
The origins of BODC go back to 1969 when NERC created the British Oceanographic Data Service (BODS). Located at the National Institute of Oceanography, Wormley in Surrey, its purpose was to:
Act as the UK's National Oceanographic Data Centre
Participate in the international exchange of data as part of the Intergovernmental Oceanographic Commission (IOC) network of national data centres
In 1975, BODS was transferred to Bidston Observatory on the Wirral, near Liverpool, as part of the newly formed Institute of Oceanographic Sciences. The following year BODS became the Marine Information and Advisory Service (MIAS). Its primary activity was to manage the data collected from weather ships, oil rigs and data buoys.
The data banking component of MIAS was restructured to form BODC in April 1989. BODC pioneered a start to finish approach to marine data management. This involved:
Assisting in the collection of data at sea
Quality control of data
Assembling the data for use by the scientists
The publication of data sets on CD-ROM
In December 2004, BODC moved to the purpose-built Joseph Proudman Building on the campus of the University of Liverpool. A smaller number of its staff are based in the National Oceanography Centre (NOC), Southampton.
Aims
Work alongside scientists during marine research projects
Provide quality control and archiving of oceanographic data
Maintain an online source of information and improve public access to marine data
Provide innovative marine data products
National role
BODC is o
|
https://en.wikipedia.org/wiki/Index%20of%20physics%20articles%20%28I%29
|
The index of physics articles is split into multiple pages due to its size.
To navigate by individual letter use the table of contents below.
I
I. M. Dharmadasa
IBM Zurich Research Laboratory
ICARUS (experiment)
ICRANet
IEEE Antennas & Propagation Society
IEEE Antennas and Wireless Propagation Letters
IEEE Council on Superconductivity
IEEE Electron Devices Society
IEEE Fellow
IEEE Journal of Quantum Electronics
IEEE Journal of Selected Topics in Quantum Electronics
IEEE Marie Sklodowska-Curie Award
IEEE Microwave Magazine
IEEE Microwave and Guided Wave Letters
IEEE Microwave and Wireless Components Letters
IEEE Nikola Tesla Award
IEEE Photonics Award
IEEE Transactions on Antennas and Propagation
IEEE Transactions on Applied Superconductivity
IEEE Transactions on Electron Devices
IEEE Transactions on Microwave Theory and Techniques
IEEE Transactions on Semiconductor Manufacturing
IFAE
IGNITOR
II-VI semiconductor
IKAROS
INAF
INDO
IR/UV mixing
ISABELLE
ISIS neutron source
ISKRA lasers
ISOLDE
ISOLTRAP
ISO 31
ISO 31-13
ISO 31-3
ISO 31-4
ISO 31-5
ISO 31-6
ISO 31-7
ISO 31-8
ISO 31-9
ITER
IZMIRAN
Ian Fells
Ian G. Enting
Ian Grant (physicist)
Ice
IceCube Neutrino Observatory
Ice II
Ice III
Ice IX
Ice Ic
Ice Ih
Ice VII
Ice VIII
Ice XI
Ice XII
Ice XV
Ice crystals
Ice nucleus
Ice rules
Ichiji Tasaki
Ida Noddack
Ideal gas
Ideal gas law
Ideal solution
Ideally hard superconductor
Identical particles
IEEE Microwave and Guided Wave Letters
Igal Talmi
Ignace-Gaston Pardies
Ignacij Klemenčič
Igor Dmitriyevich Novikov
Igor Klebanov
Igor Kurchatov
Igor Sutyagin
Igor Tamm
Igor Ternov
Il Cimento
Il Nuovo Cimento
Il Nuovo Cimento A
Il Nuovo Cimento B
Il Nuovo Cimento C
Il Nuovo Cimento D
Ille Gebeshuber
Illuminance
Ilya Frank
Ilya Lifshitz
Ilya Prigogine
Image-based flow visualization
Image noise
Imaginary time
Imaging spectrometer
Imbert–Fedorov effect
Immersed boundary method
Immersion lithography
Immirzi parameter
Immittance
Immunophysics
Impact (mechanics)
Impact ionization
Impact p
|
https://en.wikipedia.org/wiki/Cataclysmic%20pole%20shift%20hypothesis
|
The cataclysmic pole shift hypothesis is a pseudo-scientific claim that there have been recent, geologically rapid shifts in the axis of rotation of Earth, causing calamities such as floods and tectonic events or relatively rapid climate changes.
There is evidence of precession and changes in axial tilt, but this change is on much longer time-scales and does not involve relative motion of the spin axis with respect to the planet. However, in what is known as true polar wander, the Earth rotates with respect to a fixed spin axis. Research shows that during the last 200 million years a total true polar wander of some 30° has occurred, but that no rapid shifts in Earth's geographic axial pole were found during this period. A characteristic rate of true polar wander is 1° or less per million years. Between approximately 790 and 810 million years ago, when the supercontinent Rodinia existed, two geologically rapid phases of true polar wander may have occurred. In each of these, the magnetic poles of Earth shifted by approximately 55° due to a large shift in the crust.
Definition and clarification
The geographic poles are defined by the points on the surface of Earth that are intersected by the axis of rotation. The pole shift hypothesis describes a change in location of these poles with respect to the underlying surface – a phenomenon distinct from the changes in axial orientation with respect to the plane of the ecliptic that are caused by precession and nutation, and is an amplified event of a true polar wander. Geologically, a surface shift separate from a planetary shift, enabled by earth's molten core.
Pole shift hypotheses are not connected with plate tectonics, the well-accepted geological theory that Earth's surface consists of solid plates which shift over a viscous, or semifluid asthenosphere; nor with continental drift, the corollary to plate tectonics which maintains that locations of the continents have moved slowly over the surface of Earth, resulting
|
https://en.wikipedia.org/wiki/Whole-tone%20scale
|
In music, a whole-tone scale is a scale in which each note is separated from its neighbors by the interval of a whole tone. In twelve-tone equal temperament, there are only two complementary whole-tone scales, both six-note or hexatonic scales. A single whole-tone scale can also be thought of as a "six-tone equal temperament".
The whole-tone scale has no leading tone and because all tones are the same distance apart, "no single tone stands out, [and] the scale creates a blurred, indistinct effect". This effect is especially emphasised by the fact that triads built on such scale tones are all augmented triads. Indeed, all six tones of a whole-tone scale can be played simply with two augmented triads whose roots are a major second apart. Since they are symmetrical, whole-tone scales do not give a strong impression of the tonic or tonality.
The composer Olivier Messiaen called the whole-tone scale his first mode of limited transposition. The composer and music theorist George Perle calls the whole-tone scale interval cycle 2, or C2. Since there are only two possible whole-tone-scale positions (that is, the whole-tone scale can be transposed only once), it is either C20 or C21. For this reason, the whole-tone scale is also maximally even and may be considered a generated collection.
Due to this symmetry, the hexachord consisting of the whole-tone scale is not distinct under inversion or more than one transposition. Thus many composers have used one of the "almost whole-tone" hexachords, whose "individual structural differences can be seen to result only from a difference in the 'location', or placement, of a semitone within the otherwise whole-tone series." Alexander Scriabin's mystic chord is a primary example, being a whole-tone scale with one note raised a semitone; this alteration allows for a greater variety of resources through transposition.
Classical music
In 1662, Johann Rudolf Ahle wrote a melody to the lyrics of Franz Joachim Burmeister's "Es ist genug"
|
https://en.wikipedia.org/wiki/Averch%E2%80%93Johnson%20effect
|
The Averch–Johnson effect is the tendency of regulated companies to engage in excessive amounts of capital accumulation in order to expand the volume of their profits. If companies' profits to capital ratio is regulated at a certain percentage then there is a strong incentive for companies to over-invest in order to increase profits overall. This investment goes beyond any optimal efficiency point for capital that the company may have calculated as higher profit is almost always desired over and above efficiency.
Excessive capital accumulation under rate-of-return regulation is informally known as gold plating.
But the so-called Averch-Johnson effect of overcapitalization does not as a general case involve "gold-plating".
Mathematical derivation
Suppose that a regulated firm wishes to maximize its profit:where is the revenue function, is the firm's capital stock, is the firm's labor stock, is the wage rate, and is the cost of capital. The firm's profit is constrained such that:where is the allowable rate of return. Assume that . We may then form a functional to find the firm's optimal action:where is the Lagrange multiplier (also known as the shadow price). The derivatives of this functional are:Taken together, this implies that:The ratio of the marginal product of capital and the marginal product of labor is:Since this new cost of capital is perceived to be less than the market cost of capital, the firm will tend to overinvest in capital.
See also
Law and economics
Public utilities commission
Rate-of-return regulation
|
https://en.wikipedia.org/wiki/Neotropical%20Ornithological%20Society
|
The Neotropical Ornithological Society, or Sociedad de Ornitología Neotropical, is an ornithological non-profit organization, with its principal objective the study and conservation of Neotropical birds and their habitats, including both their breeding and non-breeding areas. It was founded in 1987 by Dr Mario A. Ramos Olmos, and is a member of the Ornithological Council.
The Society produces the journal Neotropical Ornithology, which publishes papers in Spanish, English and Portuguese. It organises the Neotropical Ornithological Congress and the Francois Vuilleumier Fund for research on Neotropical birds.
See also
Neotropical Birds Online, an online encyclopedia
|
https://en.wikipedia.org/wiki/Multi-fragment%20algorithm
|
The multi-fragment (MF) algorithm is a heuristic or approximation algorithm for the travelling salesman problem (TSP) (and related problems). This algorithm is also sometimes called the "greedy algorithm" for the TSP.
The algorithm builds a tour for the traveling salesman one edge at a time and thus maintains multiple tour fragments, each of which is a simple path in the complete graph of cities. At each stage, the algorithm selects the edge of minimal cost that either creates a new fragment, extends one of the existing paths or creates a cycle of length equal to the number of cities.
|
https://en.wikipedia.org/wiki/Holy%20trinity%20%28cooking%29
|
The "holy trinity" in Cajun cuisine and Louisiana Creole cuisine is the base for several dishes in the regional cuisines of Louisiana and consists of onions, bell peppers and celery. The preparation of Cajun/Creole dishes such as crawfish étouffée, gumbo, and jambalaya all start from this base.
Variants use garlic, parsley, or shallots in addition to the three trinity ingredients. The addition of garlic to the holy trinity is sometimes referred to as adding "the pope."
The holy trinity is the Cajun and Louisiana Creole variant of mirepoix; traditional mirepoix is two parts onions, one part carrots, and one part celery, whereas the holy trinity is typically one or two parts onions, one part green bell pepper, and one part celery. It is also an evolution of the Spanish sofrito, which contains onion, garlic, bell peppers, and tomatoes.
Origin of the name
The name is an allusion to the Christian doctrine of the Trinity.
The term is first attested in 1981 and was probably popularized by Paul Prudhomme.
See also
Mirepoix
Sofrito
Soffritto
Epis
|
https://en.wikipedia.org/wiki/Molecular%20Sciences%20Institute
|
The Molecular Sciences Institute (MSI), now located in Milpitas, California was founded in Berkeley, California by Sydney Brenner in 1996. Its mission was to operate as an independent non-profit research laboratory that combined genomic experimentation with computer modeling. Current efforts include co-curricular STEAM programming for youth and quick launch strategies for products/technologies addressing global health. In the last few years, MSI has integrated into the startup community through providing research and development facilities and mentorship.
This organization has been supported by federal grants from the National Institutes of Health (NIH), the Defense Advanced Research Projects Agency (DARPA) and other funds provided by foundations and corporations.
|
https://en.wikipedia.org/wiki/Pelargonium%20%C3%97%20fragrans
|
Pelargonium × fragrans (or 'Fragrans') is a pelargonium hybrid between Pelargonium odoratissimum and Pelargonium exstipulatum. It is in the subgenus Reniforme along with Pelargonium sidoides and Pelargonium abrotanifolium.
Etymology
Pelargonium comes from the Greek; Pelargos which means stork. Another name for pelargoniums is storksbills due to the shape of their fruit. Fragrans refers to the fragrant leaves.
Description
Pelargonium × fragrans, like its parent Pelargonium odoratissimum, is a small, spreading species which only grows up to 30 cm high and 60 cm wide. It has small white flowers and its leaves are waxy, green and ovate with slightly fringed edges. It has a sweet, slightly spicy, eucalyptus like scent.
Cultivars and hybrids
There are many cultivars and hybrids of Pelargonium × fragrans. These include:
Pelargonium × fragrans 'Old Spice' - A variety of P. × fragrans with leaves that smell like Old Spice.
Pelargonium × fragrans 'Nutmeg' - A nutmeg scented variety of P. ×fragrans.
Pelargonium 'Lillan Pottinger' - An apple scented hybrid of P. × fragrans and P.odoratissimum.
Pelargonium 'Ardwyck Cinnamon' - A cinnamon/nutmeg scented hybrid between P. × fragrans and Pelargonium abrotanifolium. A smaller, compacter shape than P. × fragrans with more fringed leaves.
Pelargonium 'Brilliantine' - An eau de cologne scented hybrid, likely between P. x fragrans and Pelargonium ionidiflorum but possibly with some addition parentage from other species in the reniforme subgroup.
Uses
As well as being a houseplant or outdoor perennial depending on climate, Pelargonium × fragrans has another use. The leaves are very aromatic and can be used in something like potpourri.
|
https://en.wikipedia.org/wiki/Differential%20entropy
|
Differential entropy (also referred to as continuous entropy) is a concept in information theory that began as an attempt by Claude Shannon to extend the idea of (Shannon) entropy, a measure of average (surprisal) of a random variable, to continuous probability distributions. Unfortunately, Shannon did not derive this formula, and rather just assumed it was the correct continuous analogue of discrete entropy, but it is not. The actual continuous version of discrete entropy is the limiting density of discrete points (LDDP). Differential entropy (described here) is commonly encountered in the literature, but it is a limiting case of the LDDP, and one that loses its fundamental association with discrete entropy.
In terms of measure theory, the differential entropy of a probability measure is the negative relative entropy from that measure to the Lebesgue measure, where the latter is treated as if it were a probability measure, despite being unnormalized.
Definition
Let be a random variable with a probability density function whose support is a set . The differential entropy or is defined as
For probability distributions which do not have an explicit density function expression, but have an explicit quantile function expression, , then can be defined in terms of the derivative of i.e. the quantile density function as
.
As with its discrete analog, the units of differential entropy depend on the base of the logarithm, which is usually 2 (i.e., the units are bits). See logarithmic units for logarithms taken in different bases. Related concepts such as joint, conditional differential entropy, and relative entropy are defined in a similar fashion. Unlike the discrete analog, the differential entropy has an offset that depends on the units used to measure . For example, the differential entropy of a quantity measured in millimeters will be more than the same quantity measured in meters; a dimensionless quantity will have differential entropy of more than the s
|
https://en.wikipedia.org/wiki/Ceph%20%28software%29
|
Ceph (pronounced ) is a free and open-source software-defined storage platform that provides object storage, block storage, and file storage built on a common distributed cluster foundation. Ceph provides completely distributed operation without a single point of failure and scalability to the exabyte level, and is freely available. Since version 12 (Luminous), Ceph does not rely on any other, conventional filesystem and directly manages HDDs and SSDs with its own storage backend BlueStore and can expose a POSIX filesystem.
Ceph replicates data with fault tolerance, using commodity hardware and Ethernet IP and requiring no specific hardware support. Ceph is highly available and ensures strong data durability through techniques including replication, erasure coding, snapshots and clones. By design, the system is both self-healing and self-managing, minimizing administration time and other costs.
Large-scale production Ceph deployments include CERN, OVH and DigitalOcean.
Design
Ceph employs five distinct kinds of daemons:
Cluster monitors () that keep track of active and failed cluster nodes, cluster configuration, and information about data placement and global cluster state.
OSDs () that manage bulk data storage devices directly via the BlueStore back end, which since the v12.x release replaces the Filestore back end, which was implemented on top of a traditional filesystem)
Metadata servers () that maintain and broker access to inodes and directories inside a CephFS filesystem
HTTP gateways () that expose the object storage layer as an interface compatible with Amazon S3 or OpenStack Swift APIs
Managers () that perform cluster monitoring, bookkeeping, and maintenance tasks, and interface to external monitoring systems and management (e.g. balancer, dashboard, Prometheus, Zabbix plugin)
All of these are fully distributed, and may be deployed on disjoint, dedicated servers or in a converged topology. Clients with different needs directly interact with appro
|
https://en.wikipedia.org/wiki/Interferon%20alfacon-1
|
Interferon alfacon-1 is a recombinant synthetic type I interferon used for the treatment of hairy cell leukemia, malignant melanoma and AIDS-related Kaposi's sarcoma.
|
https://en.wikipedia.org/wiki/Bernstein%E2%80%93von%20Mises%20theorem
|
In Bayesian inference, the Bernstein–von Mises theorem provides the basis for using Bayesian credible sets for confidence statements in parametric models. It states that under some conditions, a posterior distribution converges in the limit of infinite data to a multivariate normal distribution centered at the maximum likelihood estimator with covariance matrix given by , where is the true population parameter and is the Fisher information matrix at the true population parameter value:
The Bernstein–von Mises theorem links Bayesian inference with frequentist inference. It assumes there is some true probabilistic process that generates the observations, as in frequentism, and then studies the quality of Bayesian methods of recovering that process, and making uncertainty statements about that process. In particular, it states that Bayesian credible sets of a certain credibility level will asymptotically be confidence sets of confidence level , which allows for the interpretation of Bayesian credible sets.
Heuristic statement
In a model , under certain regularity conditions (finite-dimensional, well-specified, smooth, existence of tests), if the prior distribution on has a density with respect to the Lebesgue measure which is smooth enough (near bounded away from zero), the total variation distance between the rescaled posterior distribution (by centring and rescaling to ) and a Gaussian distribution centred on any efficient estimator and with the inverse Fisher information as variance will converge in probability to zero.
Bernstein–von Mises and maximum likelihood estimation
In case the maximum likelihood estimator is an efficient estimator, we can plug this in, and we recover a common, more specific, version of the Bernstein–von Mises theorem.
Implications
The most important implication of the Bernstein–von Mises theorem is that the Bayesian inference is asymptotically correct from a frequentist point of view. This means that for large amounts of data, one
|
https://en.wikipedia.org/wiki/Robert%20V.%20Hogg
|
Robert Vincent ("Bob") Hogg (8 November 1924 – 23 December 2014) was an American statistician and professor of statistics of the University of Iowa. Hogg is known for his widely used textbooks on statistics (with his 1963 Ph.D. student Elliot Alan Tanis) and on mathematical statistics (with his 1950 Ph.D. advisor Allen Thornton Craig). Hogg has received recognition for his research on
robust and adaptive nonparametric statistics and for his scholarship on total quality management and statistics education.
Academic career
Early life
Born on 8 November 1924 in Hannibal, Missouri, Hogg served three years in the US Navy from 1943 through 1946. In 1947, he graduated from the University of Illinois with a bachelor's degree in mathematics. With the goal of becoming an actuary, Hogg matriculated at the mathematics department of the University of Iowa (then the "State University of Iowa"). However, Hogg studied statistics under Allen Craig, who became his mentor and helped him obtain a job teaching statistics at the Mathematics Department. Hogg earned his Ph.D. 1950 under Allen Craig. After graduating, Hogg remained at the Mathematics Department, where he remained to become a long-serving professor.
Basu's theorem: Special cases
Hogg independently discovered a special case of "Basu's theorem", a few years before the publication by Deb Basu. Hogg's second paper on the topic of Basu's theorem was never published, because of a negative report by an anonymous referee in 1953. Later, Basu refers "to Hogg and Craig (1956) for several interesting uses [of Basu's theorem] in proving results in distribution theory".
Collaboration and friendship with Allen Craig
The textbook "Hogg and Craig" was innovative, particularly in emphasizing sufficient statistics: Sufficient statistics were treated not only for parametric families but also for nonparametric probability distributions: In particular, the sufficiency and completeness of the order statistics from a continuous distribution we
|
https://en.wikipedia.org/wiki/Canada%20Gairdner%20Wightman%20Award
|
The Canada Gairdner Wightman Award is annually awarded by the Gairdner Foundation to a Canadian who has demonstrated outstanding leadership in the field of medicine and medical science.
Award winners
Source: Gairdner- Past Recipients
See also
Gairdner Foundation
Gairdner Foundation Global Health Award
Gairdner Foundation International Award
List of medicine awards
External links
Canada Gairdner Wightman Award
The Gairdner Foundation
Canadian science and technology awards
Medicine awards
|
https://en.wikipedia.org/wiki/Minimum%20bactericidal%20concentration
|
The minimum bactericidal concentration (MBC) is the lowest concentration of an antibacterial agent required to kill a particular bacterium. It can be determined from broth dilution minimum inhibitory concentration (MIC) tests by subculturing to agar plates that do not contain the test agent. The MBC is identified by determining the lowest concentration of antibacterial agent that reduces the viability of the initial bacterial inoculum by ≥99.9%. The MBC is complementary to the MIC; whereas the MIC test demonstrates the lowest level of antimicrobial agent that inhibits growth, the MBC demonstrates the lowest level of antimicrobial agent that results in microbial death. This means that even if a particular MIC shows inhibition, plating the bacteria onto agar might still result in organism proliferation because the antimicrobial did not cause death. Antibacterial agents are usually regarded as bactericidal if the MBC is no more than four times the MIC. Because the MBC test uses colony-forming units as a proxy measure of bacterial viability, it can be confounded by antibacterial agents which cause aggregation of bacterial cells. Examples of antibacterial agents which do this include flavonoids and peptides.
|
https://en.wikipedia.org/wiki/MORE%20protocol
|
MORE, which stands for MAC independent Opportunistic Routing, is an opportunistic routing protocol designed for wireless mesh networks. The protocol removes the dependency that other opportunistic routing protocols, such as ExOR and SOAR have on the MAC layer. Both of these protocols make use of a scheduler, to co-ordinate transmission among the nodes. Only one node transmits at a given point of time and all the other nodes listen to this. The nodes that listen remove the packets which they have queued for retransmission. This ensures that the same packet is not redundantly retransmitted by different nodes.
MORE makes use of network encoding techniques and brings about spatial re-use by allowing all the nodes to transmit at the same time. Given a file, the source node breaks up the file into K packets. The number of packets each file is divided into varies. The uncoded packets are called "native packets". The source node then creates a linear combination of K packets and forwards them. The code vector represents the random co-efficients chosen by the node to perform encoding. The source also attaches a MORE header to each packet along with a forwarding list. The forwarders listen to the transmission of the source node. If the node that listens to this packet is in the forwarding list, it checks if the packet has any new information which are called as innovative packets. If the packet is innovative, it performs a linear recombination of the packets. This is essentially the linear recombination of the native packets again. The node ignores all non-innovative packets. The destination receives the packets and checks for innovative-ness. Upon receiving K innovative packets, it sends back ACK to the source and continues decoding the packets. The intermediate nodes hear this ACK and stop further transmission followed by the purging of packets in the buffer.
Practical Challenges
Calculating the number of pockets that the forwarder has to send. The paper suggests a distr
|
https://en.wikipedia.org/wiki/Niobium%E2%80%93germanium
|
Niobium-germanium (Nb3Ge) is an intermetallic chemical compound of niobium (Nb) and germanium (Ge). It has A15 phase structure.
It is a superconductor with a critical temperature of 23.2 K.
Sputtered films have been reported to have an upper critical field of 37 teslas at 4.2 K.
History
Nb3Ge was discovered to be a superconductor in 1973 and for 13 years (until the discovery in 1986 of the cuprate superconductors) it held the record as having the highest critical temperature.
It has not been as widely used for superconductive applications as niobium–tin or niobium–titanium.
Related alloys
Niobium-germanium-aluminium has an upper critical field of about 10 teslas.
|
https://en.wikipedia.org/wiki/A.%20Aneesh
|
Aneesh Aneesh is a sociologist of globalization, labor, and technology. He is Executive Director of the School of Global Studies and Languages at the University of Oregon and a Professor of Global Studies and Sociology. Previously, he served as a professor of sociology and director of the Institute of World Affairs and the global studies program at the University of Wisconsin, Milwaukee. In the early 2000s, he taught in the science and technology program at Stanford University and formulated a theory of algocracy, distinguishing it from bureaucratic, market, and surveillance-based governance systems, pioneering the field of algorithmic governance in the social sciences. Author of Virtual Migration: The Programming of Globalization (Duke 2006) and Neutral Accent: How Language, Labor and Life Become Global (Duke 2015), Aneesh is currently completing a manuscript on the rise of what he calls modular citizenship.
Education
Aneesh studied Physics, Economics, and Philosophy at the University of Allahabad, earning a Bachelor's degree there in 1987. After pre-doctoral study in Philosophy at Jawaharlal Nehru University he came to the University of California, Irvine for a Master's degree in social relations in 1996, and completed a Ph.D. in Sociology at Rutgers University in 2001.
Books
Aneesh has written or edited the following books:
Neutral Accent: How Language, Labor and Life Become Global (2015)
The Long 1968: Revisions and New Perspectives (co-edited, 2012)
Beyond Globalization: Making New Worlds in Media, Art, and Social Practices (co-edited, 2011)
Virtual Migration: the Programming of Globalization (2006)
|
https://en.wikipedia.org/wiki/Stubline%20transmitter
|
Stubline transmitter is a broadcasting facility for mediumwave and shortwave near Zvečka, Serbia. The mediumwave transmitter broadcast on 684 kHz and, until its destruction on May 30, 1999, it was one of the strongest in Europe having used 2000 kilowatts of power.
The station was rebuilt with a 225.5 metres tall mast, but nowadays uses less power.
See also
List of tallest structures in Serbia
|
https://en.wikipedia.org/wiki/Game%20integrated%20development%20environment
|
A Game Engine (game environment) is a specialized development environment for creating video games. The features one provides depends on the type and the granularity of control allowed by the underlying framework. Some may provide diagrams, a windowing environment and debugging facilities. Users build the game with the game IDE, which may incorporate a game engine or call it externally. Game IDEs are typically specialized and tailored to work with one specific game engine.
This is not to be confused with game environment art, which is "the setting or location in which [a] game takes place." This is also in distinction from domain-specific entertainment languages, where all is needed is a text editor. They are distinct from integrated development environments which are more general, and may provide different sets of features.
There is also a distinction from Visual programming language in that programming languages are more general than Game Engines.
Examples
Below are some game engines and frameworks which come with specialized IDEs.
3D Game Creation System
Adventure Game Studio
Blender Game Engine (discontinued)
Buildbox
Construct
Clickteam Fusion
CryEngine
FPS Creator
Game Core
Game Editor
GameMaker
Gamut from CMU (not Stanford)
Gamestudio
GDevelop
Godot
Goji Editor
GameSalad
Magic Work Station
PlayCanvas
Roblox
RPG Maker
SdlBasic
SharpLudus
Stencyl
The 3D Gamemaker
Unity
Unreal Engine
Virtual Play Table
VASSAL
|
https://en.wikipedia.org/wiki/Infinity%20%28philosophy%29
|
In philosophy and theology, infinity is explored in articles under headings such as the Absolute, God, and Zeno's paradoxes.
In Greek philosophy, for example in Anaximander, 'the Boundless' is the origin of all that is. He took the beginning or first principle to be an endless, unlimited primordial mass (ἄπειρον, apeiron). The Jain metaphysics and mathematics were the first to define and delineate different "types" of infinities. The work of the mathematician Georg Cantor first placed infinity into a coherent mathematical framework. Keenly aware of his departure from traditional wisdom, Cantor also presented a comprehensive historical and philosophical discussion of infinity. In Christian theology, for example in the work of Duns Scotus, the infinite nature of God invokes a sense of being without constraint, rather than a sense of being unlimited in quantity.
Early thinking
Egyptian
Greek
Anaximander
An early engagement with the idea of infinity was made by Anaximander who considered infinity to be a foundational and primitive basis of reality. Anaximander was the first in the Greek philosophical tradition to propose that the universe was infinite.
Anaxagoras
Anaxagoras (500–428 BCE) was of the opinion that matter of the universe had an innate capacity for infinite division.
The Atomists
A group of thinkers of ancient Greece (later identified as the Atomists) all similarly considered matter to be made of an infinite number of structures as considered by imagining dividing or separating matter from itself an infinite number of times.
Aristotle and after
Aristotle, alive for the period 384–322 BCE, is credited with being the root of a field of thought, in his influence of succeeding thinking for a period spanning more than one subsequent millennium, by his rejection of the idea of actual infinity.
In Book 3 of the work entitled Physics, written by Aristotle, Aristotle deals with the concept of infinity in terms of his notion of actuality and of potentiality.
|
https://en.wikipedia.org/wiki/Cohesin
|
Cohesin is a protein complex that mediates sister chromatid cohesion, homologous recombination, and DNA looping. Cohesin is formed of SMC3, SMC1, SCC1 and SCC3 (SA1 or SA2 in humans). Cohesin holds sister chromatids together after DNA replication until anaphase when removal of cohesin leads to separation of sister chromatids. The complex forms a ring-like structure and it is believed that sister chromatids are held together by entrapment inside the cohesin ring. Cohesin is a member of the SMC family of protein complexes which includes Condensin, MukBEF and SMC-ScpAB.
Cohesin was separately discovered in budding yeast (Saccharomyces cerevisiae) both by Douglas Koshland and Kim Nasmyth in 1997.
Structure
Cohesin is a multi-subunit protein complex, made up of SMC1, SMC3, RAD21 and SCC3 (SA1 or SA2). SMC1 and SMC3 are members of the Structural Maintenance of Chromosomes (SMC) family. SMC proteins have two main structural characteristics: an ATP-binding cassette-like 'head' domain with ATPase activity (formed by the interaction of the N- and C- terminals) and a hinge domain that allows dimerization of SMCs. The head and the hinge domains are connected to each other via long anti-parallel coiled coils. The dimer is present in a V-shaped form, connected by the hinges.
The N-terminal domain of RAD21 contains two α-helices which forms a three helix bundle with the coiled coil of SMC3. The central region of RAD21 is thought to be largely unstructured but contains several binding sites for regulators of cohesin. This includes a binding site for SA1 or SA2, recognition motifs for separase cleavage and a region that is competitively bound by PDS5A, PDS5B or NIPBL. The C-terminal domain of RAD21 forms a winged helix that binds two β-sheets in the Smc1 head domain.
Once RAD21 binds the SMC proteins, SCC3 can also associate with RAD21. When RAD21 binds on both SMC1 and SMC3, the cohesin complex forms a closed ring structure. The interfaces between the SMC subunits and RAD21 c
|
https://en.wikipedia.org/wiki/Exopher
|
Exophers are a type of membrane-bound extracellular vesicle (EV) that are released by budding out of cells into the extracellular space. Exophers can be released by neurons and muscle in the nematode Caenorhabditis elegans and also from murine cardiomyocytes. Exophers are notable for their large size, averaging approximately four microns in diameter, and they are able to expel whole organelles, such as mitochondria and lysosomes as cargo. An exopher can initially remain attached to the cell that produced it by a membranous filament that resembles a tunneling nanotube. Exophers share similarities with large oncosomes, but they differ in that they are produced by physiologically normal cells instead of aberrant cells associated with tumors.
Exopher production is thought to be a mechanism cells use to preserve homeostasis. Exophers are produced in response to numerous stressors including intracellular protein aggregation, reactive oxygen species (ROS), heat, osmotic hyertonicity, starvation, and even space flight. Mechanistically, exopher production has been found to depend on extracellular receptor signaling. Two MAPK pathways, epidermal growth factor (EGF) and fibroblast growth factor (FGF) signaling have been implicated in exopher production in nematodes. Extracellular signaling receptor MERTK, expressed by cardiac-resident macrophages, is necessary for exopher clearance by phagocytosis in mouse-derived cardiac tissue.
Exophers may be relevant to disease. In mouse heart, eliminating macrophages or blocking their ability to engulf exophers lead to inflammation and ventricular dysregulation. Exophers may also promote pathological protein spreading in neurodegenerative diseases due to their ability to carry aggregated proteins outside of neurons, including human huntingtin protein.
|
https://en.wikipedia.org/wiki/Urethral%20bulking%20injections
|
A urethral bulking injection is a gynecological procedure and medical treatment used to treat involuntary leakage of urine: urinary incontinence in women. Injectional materials are used to control stress incontinence. Bulking agents are injected into the mucosa surrounding the bladder neck and proximal urethra. This reduces the diameter of the urethra and creates resistance to urine leakage. After the procedure, the pressure forcing the urine from the bladder through the urethra is resisted by the addition of the bulking agent in the tissue surrounding the proximal urethra. Most of the time this procedure prevents urinary stress incontinence in women.
Uses
A urethral bulking injection is one type of treatment for incontinence in women. Urethral bulking injections are considered by a clinician when the woman has urinary sphincter dysfunction, urethral hypermobility, persistent stress urinary incontinence after a urethral sling or urethropexy, or stress urinary incontinence in women who cannot undergo surgery due to other illnesses or conditions. It restores the ability to retain urine during coughing, laughing and other normal occurrences that increase inter-abdominal pressure Some women choose to have urethral bulking injections because they wish to avoid surgery or the use of mesh material. Another reason other treatments may not be chosen include the desire to maintain fertility. Woman who benefit most from this treatment are those who have a stable bladder neck and an inadequate muscle strength of the sphincter muscles that close the urethra. Urethral instability is identified by the angle of the urethra. This can be evaluated when the woman lies on her back and the angle of the urethral is greater than 30° during coughing. Tests are performed to confirm the diagnosis of stress urinary incontinence such as the bladder capacity to hold urine and the strength of the bladder muscle contractions.
Diagnosis is aided before the procedure by a thorough evaluation of
|
https://en.wikipedia.org/wiki/Feferman%E2%80%93Sch%C3%BCtte%20ordinal
|
In mathematics, the Feferman–Schütte ordinal Γ0 is a large countable ordinal.
It is the proof-theoretic ordinal of several mathematical theories, such as arithmetical transfinite recursion.
It is named after Solomon Feferman and Kurt Schütte, the former of whom suggested the name Γ0.
There is no standard notation for ordinals beyond the Feferman–Schütte ordinal. There are several ways of representing the Feferman–Schütte ordinal, some of which use ordinal collapsing functions: , , , or .
Definition
The Feferman–Schütte ordinal can be defined as the smallest ordinal that cannot be obtained by starting with 0 and using the operations of ordinal addition and the Veblen functions φα(β). That is, it is the smallest α such that φα(0) = α.
Properties
This ordinal is sometimes said to be the first impredicative ordinal, though this is controversial, partly because there is no generally accepted precise definition of "predicative". Sometimes an ordinal is said to be predicative if it is less than Γ0.
Any recursive path ordering whose function symbols are well-founded with order type less than that of itself has order type .
|
https://en.wikipedia.org/wiki/Chengdu%20Research%20Base%20of%20Giant%20Panda%20Breeding
|
The Chengdu Research Base of Giant Panda Breeding (or simply Chengdu Panda Base) is a public non-profit breeding and research institute for giant pandas, red pandas, and other rare animals in Chengdu, Sichuan, China.
Chengdu Panda Base was founded in 1987 by the Chengdu Municipal People's Government. It started with 6 giant pandas that were rescued from the wild. By 2008, it had 124 panda births, and the captive panda population has grown to 83.
Its stated goal is to "be a world-class research facility, conservation education center, and international educational tourism destination."
Partnerships
Chengdu Panda Base has partnered with many organizations in improving ways to conserve giant pandas. For example, its partnership with Zoo Atlanta helped the zoo secure the loan of 2 giant pandas. To date, these 2 giant pandas, Yang Yang and Lun Lun, have produced five off-spring: Mei Lan in 2006, Xi Lan in 2008, Po on November 3, 2010, twins Mei Lun and Mei Huan on July 15, 2013 and twins Ya Lun and Xi Lun on September 3, 2016.
Other research partners include:
Adventure World in Shirahama, Wakayama, Japan
East Bay Zoological Society, Oakland, California, United States
University of Liverpool, England, UK
National Institute of Health/National Cancer Institution, United States
National Zoological Park, Washington, D.C., United States
Nihon University, Tokyo, Japan
North of England Zoological Society, England, UK
The Oakland China Wildlife Preservation Foundation, California, United States
San Diego Zoo, California, United States
University of Japan
Edinburgh Zoo, Scotland, UK
Calgary Zoo, Calgary, Alberta, Canada
Zoo/Tierpark Berlin, Germany
On April 11, 2013, Chengdu Research Base of Giant Panda Breeding and CNTV reached an agreement on the establishment of iPanda.com after an official signing ceremony, and they immediately started preparing for the test launch (which was estimated in June, 2013).
See also
Captive breeding
Wolong National Nature Reserve
Giant pandas
|
https://en.wikipedia.org/wiki/Plectasin
|
Plectasin is an antibiotic protein from the mushroom Pseudoplectania nigrella. It was initially discovered in 2005 and commercialised by Novozymes. Plectasin belongs to the antimicrobial peptide class called fungal defensins, which is also present in invertebrates such as flies and mussels.
Clinical trials
Pre-clinical tests in mice have shown promising results in that multiresistant bacteria have problems mutating resistance against plectasin, which acts by directly binding the bacterial cell-wall precursor Lipid II.
At the end of 2008, Novozymes signed a global licensing agreement with Sanofi-Aventis for the further development and marketing of NZ2114, a derivative of plectasin, as a treatment for gram-positive bacterial infections, e.g. Streptococcus and Staphylococcus which are resistant to all existing antibiotics.
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.