source
stringlengths
31
227
text
stringlengths
9
2k
https://en.wikipedia.org/wiki/Cambridge%20Neuropsychological%20Test%20Automated%20Battery
The Cambridge Neuropsychological Test Automated Battery (CANTAB), originally developed at the University of Cambridge in the 1980s but now provided in a commercial capacity by Cambridge Cognition, is a computer-based cognitive assessment system consisting of a battery of neuropsychological tests, administered to subjects using a touch screen computer. The CANTAB tests were co-invented by Professor Trevor Robbins and Professor Barbara Sahakian. The 25 tests in CANTAB examine various areas of cognitive function, including: general memory and learning, working memory and executive function, visual memory, attention and reaction time (RT), semantic/verbal memory, decision making and response control. The CANTAB combines the accuracy and rigour of computerised psychological testing whilst retaining the wide range of ability measures demanded of a neuropsychological battery. It is suitable for young and old subjects, and aims to be culture and language independent through the use of non-verbal stimuli in the majority of the tests. The CANTAB PAL touchscreen test, which assesses visual memory and new learning, was included in a REF submission at the University of Cambridge. This submission (which included research from across the University unrelated to CANTAB PAL) received a 4* grade from the Research Excellence Framework (REF) 2014. CANTAB and CANTAB PAL were highlighted in the Medical Schools Council ‘Health of the Nation’ 2015 publication. See also Cognitive test Computer-based assessment
https://en.wikipedia.org/wiki/Mathlete
A mathlete is a person who competes in mathematics competitions at any level or any age. More specifically, a Mathlete is a student who participates in any of the MATHCOUNTS programs, as Mathlete is a registered trademark of the MATHCOUNTS Foundation in the United States. The term is a portmanteau of the words mathematics and athlete. Top Mathletes from MATHCOUNTS often go on to compete in the AIME, USAMO, and ARML competitions in the United States. Those in other countries generally participate in national olympiads to qualify for the International Mathematical Olympiad. Participants in World Math Day also are commonly referred to as mathletes. Mathletic competitions The Putnam Exam: The William Lowell Putnam Competition is the preeminent undergraduate level mathletic competition in the United States. Administered by the Mathematical Association of America, students compete as individuals and as teams (as chosen by their Institution) for scholarships and team prize money. The exam is administered on the first saturday in December. Mathletic off-season training The academic off-season (traditionally referred to as "summer") can be especially difficult on mathletes, though various training regimens have been proposed to keep mathletic ability at its peak. Publications such as the MAA's The American Mathematical Monthly and the AMS's Notices of the American Mathematical Society are widely read to maintain and hone mathematical ability. Some coaches suggest seeking research internships or grants, many of which are funded by the National Science Foundation. At higher levels, mathletes can obtain funding from host institutions to work on summer research projects. For example, the University of Delaware offers the Groups Exploring the Mathematical Sciences project (GEMS project) to first year graduate students. The students act as the principal investigator and work with an undergraduate research assistant and a faculty adviser who will oversee their summer research
https://en.wikipedia.org/wiki/Negative%20testing
Negative testing is a method of testing an application or system that ensures that the plot of the application is according to the requirements and can handle the unwanted input and user behavior. Invalid data is inserted to compare the output against the given input. Negative testing is also known as failure testing or error path testing. When performing negative testing exceptions are expected. This shows that the application is able to handle improper user behavior. Users input values that do not work in the system to test its ability to handle incorrect values or system failure. Purpose The purpose of negative testing is to prevent the application from crashing and it also helps improve the quality of an application by detecting defects. Negative testing helps you to improve the testing coverage of the application. Negative testing makes the application more stable and reliable. Negative testing together with positive testing allows users to test the application with any valid (or invalid) input data. Benefits of negative testing Negative testing is done to check that the product deals properly with the circumstance for which it is not programmed. The fundamental aim of this testing is to check how bad data is taken care of by the systems, and appropriate errors are shown to the client when bad data is entered. Both positive and negative testing play an important role. Positive testing ensures that the application does what it is implied for and performs each function as expected. Negative testing is opposite of positive testing. Negative testing discovers diverse approaches to make the application crash and handle the crash effortlessly. Example If there is a text box that can only take numeric values but the user tries to type a letter, the correct behavior would be to display a message such as "(Incorrect data) Please enter a number." If the user is to fill the name field and there are ground rules that the name text is mandatory to fill, but that
https://en.wikipedia.org/wiki/Normative%20model%20of%20decision-making
Victor Vroom, a professor at Yale University and a scholar on leadership and decision-making, developed the normative model of decision-making. Drawing upon literature from the areas of leadership, group decision-making, and procedural fairness, Vroom’s model predicts the effectiveness of decision-making procedures. Specifically, Vroom’s model takes into account the situation and the importance of the decision to determine which of Vroom’s five decision-making methods will be most effective. Decision-making processes Vroom identified five types of decision-making processes, each varying on degree of participation by the leader. Decide: The leader makes the decision or solves the problem alone and announces his/her decision to the group. The leader may gather information from members of the group. Consult (Individually): The leader approaches group members individually and presents them with the problem. The leader records the group member’s suggestions and makes a decision, deciding whether or not to use the information provided by group members. Consult (Group): The leader holds a group meeting where he/she presents the problem to the group as a whole. All members are asked to contribute and make suggestions during the meeting. The leader makes his/her decision alone, choosing which information obtained from the group meeting to use or discard. Facilitate: The leader holds a group meeting where he/she presents the problem to the group as a whole. This differs from consulting approach as the leader ensures that his/her opinions are not given any more weight than those of the group. The decision is made by group consensus, and not solely by the leader. Delegate: The leader does not actively participate in the decision-making process. Instead, the leader provides resources (e.g., information about the problem) and encouragement. Situational influence of decision-making Vroom identified seven situational factors that leaders should consider when choosing a decisi
https://en.wikipedia.org/wiki/List%20of%20partial%20differential%20equation%20topics
This is a list of partial differential equation topics. General topics Partial differential equation Nonlinear partial differential equation list of nonlinear partial differential equations Boundary condition Boundary value problem Dirichlet problem, Dirichlet boundary condition Neumann boundary condition Stefan problem Wiener–Hopf problem Separation of variables Green's function Elliptic partial differential equation Singular perturbation Cauchy–Kovalevskaya theorem H-principle Atiyah–Singer index theorem Bäcklund transform Viscosity solution Weak solution Loewy decomposition of linear differential equations Specific partial differential equations Broer–Kaup equations Burgers' equation Euler equations Fokker–Planck equation Hamilton–Jacobi equation, Hamilton–Jacobi–Bellman equation Heat equation Laplace's equation Laplace operator Harmonic function Spherical harmonic Poisson integral formula Klein–Gordon equation Korteweg–de Vries equation Modified KdV–Burgers equation Maxwell's equations Navier–Stokes equations Poisson's equation Primitive equations (hydrodynamics) Schrödinger equation Wave equation Numerical methods for PDEs Finite difference Finite element method Finite volume method Boundary element method Multigrid Spectral method Computational fluid dynamics Alternating direction implicit Related areas of mathematics Calculus of variations Harmonic analysis Ordinary differential equation Sobolev space Partial differential equations
https://en.wikipedia.org/wiki/Baltic%20Way%20%28mathematical%20contest%29
The Baltic Way mathematical contest has been organized annually since 1990, usually in early November, to commemorate the Baltic Way demonstration of 1989. Unlike most international mathematical competitions, Baltic Way is a true team contest. Each team consists of five secondary-school students, who are allowed and expected to collaborate on the twenty problems during the four and a half hours of the contest. Originally, the three Baltic states participated, but the list of invitees has since grown to include all countries around the Baltic Sea; Germany sends a team representing only its northernmost parts, and Russia a team from St. Petersburg. Iceland is invited on grounds of being the first state to recognize the newfound independence of the Baltic states. Extra "guest" teams are occasionally invited at the discretion of the organizers: Israel was invited in 2001, Belarus in 2004 and 2014, Belgium in 2005, South Africa in 2011, the Netherlands in 2015 and Ireland in 2021. Responsibility for organizing the contest circulates among the regular participants. History Notes External links and references Problems, solutions, results and links (some of them broken) to web sites 1990-2010 Baltic Way contest web sites Problems European student competitions Mathematics_competitions Recurring events established in 1990
https://en.wikipedia.org/wiki/Chessboard%20paradox
The chessboard paradox or paradox of Loyd and Schlömilch is a falsidical paradox based on an optical illusion. A chessboard or a square with a side length of 8 units is cut into four pieces. Those four pieces are used to form a rectangle with side lengths of 13 and 5 units. Hence the combined area of all four pieces is 64 area units in the square but 65 area units in the rectangle, this seeming contradiction is due an optical illusion as the four pieces don't fit exactly in the rectangle, but leave a small barely visible gap around the rectangle's diagonal. The paradox is sometimes attributed to the American puzzle inventor Sam Loyd (1841–1911) and the German mathematician Oskar Schlömilch (1832–1901) Analysis Upon close inspection one can see that the four pieces don't fit quite together but leave a small barely visible gap around the diagonal of the rectangle. This gap has the shape of a parallelogram, which can be checked by showing that the opposing angles are of equal size. An exact fit of the four pieces along the rectangle's requires the parallelogram to collapse into a line segments, which means its need to have the following sizes: Since the actual angles deviate only slightly from those values, it creates the optical illusion of the parallelogram being just a line segment and the pieces fitting exactly. Alternatively one can verify the parallelism by placing the reactangle in a coordinate system and compare slopes or vector representation of the sides. The side length and diagonals of the parallelogram are: Using Heron's formula one can compute the area of half of the parallelogram (). The halved circumference is which yields the area of the whole parallelogram: So the area of the gap accounts exactly for the additional area of the rectangle. Generalization The line segments occurring in the drawing of the last chapters are of length 2, 3, 5, 8 and 13. These are all sequential Fibonacci numbers, suggesting a generalization of the dissectio
https://en.wikipedia.org/wiki/Breadfruit
Breadfruit (Artocarpus altilis) is a species of flowering tree in the mulberry and jackfruit family (Moraceae) believed to be a domesticated descendant of Artocarpus camansi originating in New Guinea, the Maluku Islands, and the Philippines. It was initially spread to Oceania via the Austronesian expansion. It was further spread to other tropical regions of the world during the Colonial Era. British and French navigators introduced a few Polynesian seedless varieties to Caribbean islands during the late 18th century. Today it is grown in some 90 countries throughout South and Southeast Asia, the Pacific Ocean, the Caribbean, Central America and Africa. Its name is derived from the texture of the moderately ripe fruit when cooked, similar to freshly baked bread and having a potato-like flavor. The trees have been widely planted in tropical regions, including lowland Central America, northern South America, and the Caribbean. In addition to the fruit serving as a staple food in many cultures, the light, sturdy timber of breadfruit has been used for outriggers, ships, and houses in the tropics. Breadfruit is closely related to Artocarpus camansi (breadnut or seeded breadfruit) of New Guinea, the Maluku Islands, and the Philippines, Artocarpus blancoi (tipolo or antipolo) of the Philippines, and Artocarpus mariannensis (dugdug) of Micronesia, all of which are sometimes also referred to as "breadfruit". It is also closely related to the jackfruit. Description Breadfruit trees grow to a height of . The large and thick leaves are deeply cut into pinnate lobes. All parts of the tree yield latex, which is useful for boat caulking. The trees are monoecious, with male and female flowers growing on the same tree. The male flowers emerge first, followed shortly afterward by the female flowers. The latter grow into capitula, which are capable of pollination just three days later. Pollination occurs mainly by fruit bats, but cultivated varieties produce fruit without pollina
https://en.wikipedia.org/wiki/Realized%20niche%20width
Realized niche width is a phrase relating to ecology, is defined by the actual space that an organism inhabits and the resources it can access as a result of limiting pressures from other species (e.g. superior competitors). An organism's ecological niche is determined by the biotic and abiotic factors that make up that specific ecosystem that allow that specific organism to survive there. The width of an organism's niche is set by the range of conditions a species is able to survive in that specific environment. Definition The fundamental niche width of an organism refers to the theoretical range of conditions that an organism could survive and reproduce in without considering interspecific interactions. The fundamental niche exclusively considers limiting biotic and abiotic factors such as appropriate food sources and a suitable climate. The fundamental niche width often differs from the realized niche width (the areas where actually inhabited by a given species). This differentiation is due to interspecific competition with other species within their ecosystem while still considering the biotic and abiotic limiting factors. A species' realized niche is usually much narrower than its fundamental niche width as it is forced to adjust its niche around the superior competing species. The physical area where a species lives, is its habitat. The set of environmental features essential to that species' survival, is its "niche." (Ecology. Begon, Harper, Townsend) Importance The difference between the realized and the fundamental niche is important in understanding how interactions with a variety of different species in one environment affects the fitness of another species. This is not only important in understanding how a species functions in an ecosystem, but it is also important in determining the potential and realized success of invasive species. Invasive species could thrive or be killed off in an environment where they would theoretically be able to exist bas
https://en.wikipedia.org/wiki/Extensible%20Forms%20Description%20Language
Extensible Forms Description Language (XFDL) is a high-level computer language that facilitates defining a form as a single, stand-alone object using elements and attributes from the Extensible Markup Language (XML). Technically, it is a class of XML originally specified in a World Wide Web Consortium (W3C) Note. See Specifications below for links to the current versions of XFDL. XFDL It offers precise control over form layout, permitting replacement of existing business/government forms with electronic documents in a human-readable, open standard. In addition to precision layout control, XFDL provides multiple page capabilities, step-by-step guided user experiences, and digital signatures. XFDL also provides a syntax for in-line mathematical and conditional expressions and data validation constraints as well as custom items, options, and external code functions. Current versions of XFDL (see Specifications below) are capable of providing these interactive features via open standard markup languages including XForms, XPath, XML Schema and XML Signatures. XFDL not only supports multiple digital signatures, but the signatures can apply to specific sections of a form and prevent changes to signed content. These advantages to XFDL led large organizations such as the United States Army and Air Force to migrate to XFDL from using forms in other formats. Later, though, the lack of portable software capable of creating XFDL led them to investigate moving away from it. The Army migrated to Adobe fillable PDFs in 2014.
https://en.wikipedia.org/wiki/Alvy%20Ray%20Smith
Alvy Ray Smith III (born September 8, 1943) is an American computer scientist who co-founded Lucasfilm's Computer Division and Pixar, participating in the 1980s and 1990s expansion of computer animation into feature film. Education In 1965 Alvy Smith received his bachelor's degree in electrical engineering from New Mexico State University (NMSU). He created his first computer graphic in 1965 at NMSU. In 1970 he received a Ph.D. in computer science from Stanford University, with a dissertation on cellular automata theory jointly supervised by Michael A. Arbib, Edward J. McCluskey, and Bernard Widrow. Career His first art show was at the Stanford Coffeehouse. From 1969 to 1973 he was an associate professor of Electrical Engineering and Computer Science at New York University, under chairman Herbert Freeman, one of the earliest computer graphics researchers. He taught briefly at the University of California, Berkeley in 1974. While at Xerox PARC in 1974, Smith worked with Richard Shoup on SuperPaint, one of the first computer raster graphics editor, or 'paint', programs. Smith's major contribution to this software was the creation of the HSV color space, also known as HSB. He created his first computer animations on the SuperPaint system. In 1975 Smith joined the new Computer Graphics Laboratory at New York Institute of Technology (NYIT), where he was given the job title "Information Quanta". There, working alongside a traditional cel animation studio, he met Ed Catmull and several core personnel of Pixar. Smith worked on a series of newer paint programs, including Paint3, the first true-color raster graphics editor. As part of this work he co-invented the concept of the alpha channel. He was also the programmer and collaborator on Ed Emshwiller's animation Sunstone, included in the collection of the Museum of Modern Art in New York. Smith worked at NYIT until 1979 and then briefly at the Jet Propulsion Laboratory with Jim Blinn on the Carl Sagan Cosmos: A Person
https://en.wikipedia.org/wiki/Bus%20encoding
Bus encoding refers to converting/encoding a piece of data to another form before launching on the bus. While bus encoding can be used to serve various purposes like reducing the number of pins, compressing the data to be transmitted, reducing cross-talk between bit lines, etc., it is one of the popular techniques used in system design to reduce dynamic power consumed by the system bus. Bus encoding aims to reduce the Hamming distance between 2 consecutive values on the bus. Since the activity is directly proportional to the Hamming distance, bus encoding proves to be effective in reducing the overall activity factor thereby reducing the dynamic power consumption in the system. In the context of this article, a system can refer to anything where data is transferred from one element to another over bus (viz. System on a Chip (SoC), a computer system, an embedded system on board, etc.). Motivation Power consumption in electronic systems is a matter of concern today for the below reasons: Battery-operated devices: Due to ubiquity of battery operated devices and the need to maximize the duration between two subsequent charging of the battery, it is necessary that the system consumes as less power (and energy) as possible. Environmental constraints: In an attempt to protect the environment, we need to conserve the usable energy. Since the energy consumed by electronic systems is increasing drastically, minimizing the energy consumption of electronic systems is critical to save the environment. Power dissipation: As per the Moore's law, semiconductor devices have been packing more and more transistors in smaller amount of area. This leads to higher power dissipation per unit area and makes packaging and thermal cooling system design complex and costly. Hence, low power electronic systems are needed to tackle this issue. The dynamic power dissipated by an electronic circuit is directly proportional to the activity factor and the load capacitance as seen by the outpu
https://en.wikipedia.org/wiki/Intersymbol%20interference
In telecommunication, intersymbol interference (ISI) is a form of distortion of a signal in which one symbol interferes with subsequent symbols. This is an unwanted phenomenon as the previous symbols have a similar effect as noise, thus making the communication less reliable. The spreading of the pulse beyond its allotted time interval causes it to interfere with neighboring pulses. ISI is usually caused by multipath propagation or the inherent linear or non-linear frequency response of a communication channel causing successive symbols to blur together. The presence of ISI in the system introduces errors in the decision device at the receiver output. Therefore, in the design of the transmitting and receiving filters, the objective is to minimize the effects of ISI, and thereby deliver the digital data to its destination with the smallest error rate possible. Ways to alleviate intersymbol interference include adaptive equalization and error correcting codes. Causes Multipath propagation One of the causes of intersymbol interference is multipath propagation in which a wireless signal from a transmitter reaches the receiver via multiple paths. The causes of this include reflection (for instance, the signal may bounce off buildings), refraction (such as through the foliage of a tree) and atmospheric effects such as atmospheric ducting and ionospheric reflection. Since the various paths can be of different lengths, this results in the different versions of the signal arriving at the receiver at different times. These delays mean that part or all of a given symbol will be spread into the subsequent symbols, thereby interfering with the correct detection of those symbols. Additionally, the various paths often distort the amplitude and/or phase of the signal, thereby causing further interference with the received signal. Bandlimited channels Another cause of intersymbol interference is the transmission of a signal through a bandlimited channel, i.e., one where the
https://en.wikipedia.org/wiki/Absorbance
Absorbance is defined as "the logarithm of the ratio of incident to transmitted radiant power through a sample (excluding the effects on cell walls)". Alternatively, for samples which scatter light, absorbance may be defined as "the negative logarithm of one minus absorptance, as measured on a uniform sample". The term is used in many technical areas to quantify the results of an experimental measurement. While the term has its origin in quantifying the absorption of light, it is often entangled with quantification of light which is “lost” to a detector system through other mechanisms. What these uses of the term tend to have in common is that they refer to a logarithm of the ratio of a quantity of light incident on a sample or material to that which is detected after the light has interacted with the sample. The term absorption refers to the physical process of absorbing light, while absorbance does not always measure only absorption; it may measure attenuation (of transmitted radiant power) caused by absorption, as well as reflection, scattering, and other physical processes. Sometimes the term "attenuance" or "experimental absorbance" is used to emphasize that radiation is lost from the beam by processes other than absorption, with the term "internal absorbance" used to emphasize that the necessary corrections have been made to eliminate the effects of phenomena other than absorption. History and uses of the term absorbance Beer-Lambert law The roots of the term absorbance are in the Beer–Lambert law. As light moves through a medium, it will become dimmer as it is being "extinguished". Bouguer recognized that this extinction (now often called attenuation) was not linear with distance traveled through the medium, but related by what we now refer to as an exponential function. If is the intensity of the light at the beginning of the travel and is the intensity of the light detected after travel of a distance the fraction transmitted, is given by where
https://en.wikipedia.org/wiki/Game%20Developers%20Session
The Game Developers Session (GDS) is an annual Czech conference on video game development. It is the oldest conference dedicated to game development, technologies, businesses, and marketing in Central Europe. It has been organized annually since 2003, most being held in Prague of the Czech Republic. It is notable for being the first event of its kind in the city with a tradition spanning around 20 years. In 2009, the conference gained popularity outside of the Czech Republic by presenting a set of lectures by developers from 2K Czech, mainly on the technical background of the game Mafia II. From there, GDS began to grow, shaping the game development community within the country. GDS has hosted 900+ game developers, producers, and journalists with over 150 game studios participating every year. GDS includes the presentation of projects and talks from speakers from art, business, design, programming, and other backgrounds. Social media & press The GDS is backed by the Czech Game Developers’ Association. It represents an interconnected platform in the context of the European game industry and brings professional communities together from Slovakia, Slovenia, and Poland. GDS has had the support of the City hall of Prague for many consecutive years. In previous years the event was attended by government officials, investors, and government agencies including the Minister of Culture of the Czech Republic and the Director of The Business and Investment Development Agency CzechInvest. Venue and format The conference is designed as an international event. It is open and accessible to foreigners. The main program line and official conference's language channels are in English. The 2022 GDS presented 40 speakers over two days at the historical venue of National House Smíchov in Prague, with more than 1000 visitors confirmed. Sponsors included big names like Beat Games, Wargaming, and Microsoft. Universities and schools In order to help the next generation of game d
https://en.wikipedia.org/wiki/Paleoanthropology
Paleoanthropology or paleo-anthropology is a branch of paleontology and anthropology which seeks to understand the early development of anatomically modern humans, a process known as hominization, through the reconstruction of evolutionary kinship lines within the family Hominidae, working from biological evidence (such as petrified skeletal remains, bone fragments, footprints) and cultural evidence (such as stone tools, artifacts, and settlement localities). The field draws from and combines primatology, paleontology, biological anthropology, and cultural anthropology. As technologies and methods advance, genetics plays an ever-increasing role, in particular to examine and compare DNA structure as a vital tool of research of the evolutionary kinship lines of related species and genera. Etymology The term paleoanthropology derives from Greek palaiós (παλαιός) "old, ancient", ánthrōpos (ἄνθρωπος) "man, human" and the suffix -logía (-λογία) "study of". Hominoid taxonomies Hominoids are a primate superfamily, the hominid family is currently considered to comprise both the great ape lineages and human lineages within the hominoid superfamily. The "Homininae" comprise both the human lineages and the African ape lineages. The term "African apes" refers only to chimpanzees and gorillas. The terminology of the immediate biological family is currently in flux. The term "hominin" refers to any genus in the human tribe (Hominini), of which Homo sapiens (modern humans) is the only living specimen. History 18th century In 1758 Carl Linnaeus introduced the name Homo sapiens as a species name in the 10th edition of his work Systema Naturae although without a scientific description of the species-specific characteristics. Since the great apes were considered the closest relatives of human beings, based on morphological similarity, in the 19th century, it was speculated that the closest living relatives to humans were chimpanzees (genus Pan) and gorilla (genus Gorilla), and bas
https://en.wikipedia.org/wiki/Daily%20build
A daily build or nightly build is the practice of completing a software build of the latest version of a program, on a daily basis. This is so it can first be compiled to ensure that all required dependencies are present, and possibly tested to show no bugs have been introduced. The daily build is also often publicly available allowing access to the latest features for feedback. In this context, a build is the result of compiling and linking all the files that make up a program. The use of such disciplined procedures as daily builds is particularly necessary in large organizations where many programmers are working on a single piece of software. Performing daily builds helps ensure that developers can work knowing with reasonable certainty that any new bugs that show up are a result of their own work done within the last day. Daily builds typically include a set of tests, sometimes called a "smoke test." These tests are included to assist in determining what may have been broken by the changes included in the latest build. The critical piece of this process is to include new and revised tests as the project progresses. Continuous integration builds Although daily builds were considered a best practice of software development in the 1990s, they have now been superseded. Continuous integration is now run on an almost continual basis, with a typical cycle time of around 20-30 minutes since the last change to the source code. Continuous integration servers continually monitor the source code control system. When these servers detect new changes, they use a build tool to rebuild the software. Good practice today is also to use continuous integration as part of continuous testing, so that unit tests are re-run for each build, and more extensive functional testing (which takes longer to perform than the build) performed as frequently as its duration permits. See also Neutral build Smoke testing in software development External links IEEE Best software practices
https://en.wikipedia.org/wiki/Deep%20petrosal%20nerve
The deep petrosal nerve is a post-ganglionic branch of the (sympathetic) internal carotid (nervous) plexus (which is in turn derived from the superior cervical ganglion, a part of the cervical sympathetic trunk) that enters the cranial cavity through the carotid canal, then passes perpendicular to the carotid canal in the cartilaginous substance which fills the foramen lacerum to unite with the (parasympathetic) greater petrosal nerve to form the nerve of pterygoid canal (Vidian nerve). Anatomy intermediate grey column (of spinal cord at around the level of T1) → white rami communicantes (of cervical part of sympathetic chain) → superior cervical ganglion (synapse) → gray rami communicantes → internal carotid plexus → deep petrosal nerve → nerve of pterygoid canal → pterygopalatine ganglion (fibres pass through without synapsing) → zygomatic nerve → zygomaticotemporal nerve → lacrimal nerve Origin The cell bodies of pre-ganglionic sympathetic axons that subsequently give synapse with neurons of the deep petrosal nerve reside in the intermediate grey column of the spinal cord at around the spinal level of T1. The pre-ganglionic axons ascend in the sympathetic trunk to synapse at the superior cervical ganglion where the cell bodies of the fibres of the deep petrosal nerve are situated. The post-ganglionic fibres do not synapse again and ultimately innervate their target tissues directly. Function The deep petrosal nerve carries post-ganglionic sympathetic axons which are ultimately distributed to the blood vessels (to mediate vasoconstriction), and exocrine glands of the lacrimal gland, nasal cavity, and oral cavity (to mediate secretomotor function). Additional images
https://en.wikipedia.org/wiki/Shop%20drawing
A shop drawing is a drawing or set of drawings produced by the contractor, supplier, manufacturer, subcontractor, consultants, or fabricator. Shop drawings are typically required for prefabricated components. Examples of these include: elevators, structural steel, trusses, pre-cast concrete, windows, appliances, cabinets, air handling units, and millwork. Also critical are the installation and coordination shop drawings of the MEP trades such as sheet metal ductwork, piping, plumbing, fire protection, and electrical. Shop drawings are produced by contractors and suppliers under their contract with the owner. The shop drawing is the manufacturer’s or the contractor’s drawn version of information shown in the construction documents. The shop drawing normally shows more detail than the construction documents. It is drawn to explain the fabrication and/or installation of the items to the manufacturer’s production crew or contractor's installation crews. The style of the shop drawing is usually very different from that of the architect’s drawing. The shop drawing’s primary emphasis is on the particular product or installation and excludes notation concerning other products and installations, unless integration with the subject product is necessary. Information required to be included in Shop Drawings Comparison information for the architect and engineer The shop drawings should include information for the architect and engineer to compare to the specifications and drawings. The shop drawing should address the appearance, performance, and prescriptive descriptions in the specifications and construction drawings. The shop drawing often is more detailed than the information shown in the construction documents to give the architect and engineer the opportunity to review the fabricator’s version of the product, prior to fabrication. References to the construction documents, drawings, and specifications assist the architect and engineer in their review of the shop drawings.
https://en.wikipedia.org/wiki/AMD%20K8
The AMD K8 Hammer, also code-named SledgeHammer, is a computer processor microarchitecture designed by AMD as the successor to the AMD K7 Athlon microarchitecture. The K8 was the first implementation of the AMD64 64-bit extension to the x86 instruction set architecture. Features Processors Processors based on the K8 core include: Athlon 64 - first 64-bit consumer desktop Athlon 64 X2 - first dual-core ('X2') desktop Athlon X2 - later model dual-core desktop with '64' omitted Athlon 64 FX - enthusiast desktop (multipliers unlocked) Sempron - low-end, low-cost desktop Opteron - server market Turion 64 - mobile computing market Turion 64 X2 - dual-core mobile processor The K8 core is very similar to the K7. The most radical change is the integration of the AMD64 instructions and an on-chip memory controller. The memory controller drastically reduces memory latency and is largely responsible for most of the performance gains from K7 to K8. Nomenclature It is perceived by the PC community that after the use of the codename K8 for the Athlon 64 processor family, AMD no longer uses K-nomenclatures (which originally stood for Kryptonite) since no K-nomenclature naming convention beyond K8 has appeared in official AMD documents and press releases after the beginning of 2005. AMD now refers to the codename K8 processors as the Family 0Fh processors. 10h and 0Fh refer to the main result of the CPUID x86 processor instruction. In hexadecimal numbering, 0F(h) (where the h represents hexadecimal numbering) equals the decimal number 15, and 10(h) equals the decimal number 16. (The "K10h" form that sometimes pops up is an improper hybrid of the "K" code and Family identifier number.) See also List of AMD Athlon 64 processors - desktop List of AMD Athlon X2 processors - desktop List of AMD Sempron processors - low end List of AMD Opteron processors - server List of AMD Turion processors - mobile AMD K9 AMD 10h Jim Keller (engineer)
https://en.wikipedia.org/wiki/Sporadic%20group
In mathematics, a sporadic group is one of the 26 exceptional groups found in the classification of finite simple groups. A simple group is a group G that does not have any normal subgroups except for the trivial group and G itself. The classification theorem states that the list of finite simple groups consists of 18 countably infinite families plus 26 exceptions that do not follow such a systematic pattern. These 26 exceptions are the sporadic groups. They are also known as the sporadic simple groups, or the sporadic finite groups. Because it is not strictly a group of Lie type, the Tits group is sometimes regarded as a sporadic group, in which case there would be 27 sporadic groups. The monster group, or friendly giant, is the largest of the sporadic groups, and all but six of the other sporadic groups are subquotients of it. Names Five of the sporadic groups were discovered by Mathieu in the 1860s and the other 21 were found between 1965 and 1975. Several of these groups were predicted to exist before they were constructed. Most of the groups are named after the mathematician(s) who first predicted their existence. The full list is: Mathieu groups M11 (M11), M12 (M12), M22 (M22), M23 (M23), M24 (M24) Janko groups J1 (J1), J2 or HJ (J2), J3 or HJM (J3), J4 (J4) Conway groups Co1 (Co1), Co2 (), Co3 (Co3) Fischer groups Fi22 (Fi22), Fi23 (Fi23), Fi24′ or F3+ (Fi24) Higman–Sims group HS McLaughlin group McL Held group He or F7+ or F7 Rudvalis group Ru Suzuki group Suz or F3− O'Nan group O'N (ON) Harada–Norton group HN or F5+ or F5 Lyons group Ly Thompson group Th or F3|3 or F3 Baby Monster group B or F2+ or F2 Fischer–Griess Monster group M or F1 Various constructions for these groups were first compiled in , including character tables, individual conjugacy classes and lists of maximal subgroup, as well as Schur multipliers and orders of their outer automorphisms. These are also listed online at , updated with their group presentations and semi-p
https://en.wikipedia.org/wiki/Bilberry
Bilberries () or blueberries are a primarily Eurasian species of low-growing shrubs in the genus Vaccinium (family Ericaceae), bearing edible, dark blue berries. The species most often referred to is Vaccinium myrtillus L., but there are several other closely related species. Etymology and common names The name "bilberry" appears to have a Scandinavian origin, possibly from as early as 1577, being similar to the Danish word bølle for whortleberry with the addition of "berry". In Scandinavian languages, terms for bilberries have names that carry the meaning "blueberry": e.g. blåbär in Swedish and blåbær in Danish and Norwegian. The bilberry (especially Vaccinium myrtillus) is generally known as blaeberry in Scottish and Northern English regional dialects, and whortleberry in southern England. Description and species Bilberries—which are native to Europe—are different from North American blueberries, although the species are closely related and belong to the same genus, Vaccinium. Bilberry are non-climacteric fruits with a smooth, circular outline at the end opposite the stalk, whereas blueberries retain persistent sepals there, leaving a rough, star-shaped pattern of five flaps. Bilberries grow singly or in pairs rather than in clusters, as blueberries do, and blueberries have more evergreen leaves. Bilberries are dark in colour, and often appear near black with a slight shade of purple. Bilberries and blueberries contain diverse anthocyanins, specifically anthocyanidins, including delphinidin and cyanidin glycosides. While blueberry fruit pulp is light green, bilberry is red or purple. The high anthocyanin content may cause staining of the fingers, lips, and tongue. Bilberries include several closely related species of the genus Vaccinium, including: Vaccinium myrtillus L. (bilberry) Vaccinium uliginosum L. (bog bilberry, bog blueberry, bog whortleberry, bog huckleberry, northern bilberry, ground hurts) Vaccinium cespitosum Michx. (dwarf bilberry) Va
https://en.wikipedia.org/wiki/Anthropization
In geography and ecology, anthropization is the conversion of open spaces, landscapes, and natural environments by human action. Anthropic erosion is the process of human action degrading terrain and soil. An area may be classified as anthropized even though it looks natural, such as grasslands that have been deforested by humans. It can be difficult to determine how much a site has been anthropized in the case of urbanization because one must be able to estimate the state of the landscape before significant human action. Origin The earliest known stages of anthropization can be found as early as the Neolithic era and the basic farmland created in that time. With the continually-growing population of humans, the land that the Earth provides has been appropriated over the years. The ecological footprint created by anthropization is continually growing despite efficiency and technique improvements made in anthropization.. Whether anthropized or not, all land seldom a few locations has been claimed. Outside of the largely inhospitable Arctic and Antarctic circles and large portions of other uninhabitable landscapes, much of the globe has been used or altered in some direct way by humans. Land has been appropriated for many different reasons, but ultimately the outcome is typically a short-term benefit for humans. An area is anthropized is some way to make land available for housing, to harvest the resources, to create space for some anthropological reason, or many other possibilities. Processes and effects Agriculture The root of many early forms of civilization, agriculture has been a primary reason for anthropization. To cultivate food or breed animals, humans must alter land—till soil or build structures—to facilitate agriculture. This can lead to soil erosion and pollution (pesticides, greenhouse gas emissions, etc.), and subsequently habitat fragmentation and overall an increased ecological footprint. Agriculture and industry often overlap, and indus
https://en.wikipedia.org/wiki/Insulantarctica
Insulantarctica is a biogeographic province of the Antarctic Realm according to the classification developed by Miklos Udvardy in 1975. It comprises scattered islands of the Southern Ocean, which show clear affinity to each other. These islands belong to different countries. Some of them constitute UNESCO's protected areas. New Zealand Subantarctic Islands protected area (New Zealand): Auckland Islands National Nature Reserve Ia Campbell Islands National Nature Reserve Ia Antipodes Islands National Nature Reserve Ia Snares Islands National Nature Reserve Ia Bounty Islands National Nature Reserve Ia Auckland Islands Marine Mammal Sanctuary - Category unassigned Territorial seas at Campbell, Antipodes, Snares and Bounty Islands - Category unassigned Heard Island and McDonald Islands (HIMI) protected area (Australia) Macquarie Island (Australia), on World Heritage List since 1997 Kerguelen Islands protected area (France) Tristan da Cunha Islands (United Kingdom), on World Heritage List since 1995 Prince Edward Islands protected area (South Africa) Gough Island Wildlife Reserve (UK)
https://en.wikipedia.org/wiki/Riesz%E2%80%93Fischer%20theorem
In mathematics, the Riesz–Fischer theorem in real analysis is any of a number of closely related results concerning the properties of the space L2 of square integrable functions. The theorem was proven independently in 1907 by Frigyes Riesz and Ernst Sigismund Fischer. For many authors, the Riesz–Fischer theorem refers to the fact that the Lp spaces from Lebesgue integration theory are complete. Modern forms of the theorem The most common form of the theorem states that a measurable function on is square integrable if and only if the corresponding Fourier series converges in the Lp space This means that if the Nth partial sum of the Fourier series corresponding to a square-integrable function f is given by where the nth Fourier coefficient, is given by then where is the -norm. Conversely, if is a two-sided sequence of complex numbers (that is, its indices range from negative infinity to positive infinity) such that then there exists a function f such that f is square-integrable and the values are the Fourier coefficients of f. This form of the Riesz–Fischer theorem is a stronger form of Bessel's inequality, and can be used to prove Parseval's identity for Fourier series. Other results are often called the Riesz–Fischer theorem . Among them is the theorem that, if A is an orthonormal set in a Hilbert space H, and then for all but countably many and Furthermore, if A is an orthonormal basis for H and x an arbitrary vector, the series converges (or ) to x. This is equivalent to saying that for every there exists a finite set in A such that for every finite set B containing B0. Moreover, the following conditions on the set A are equivalent: the set A is an orthonormal basis of H for every vector Another result, which also sometimes bears the name of Riesz and Fischer, is the theorem that (or more generally ) is complete. Example The Riesz–Fischer theorem also applies in a more general setting. Let R be an inner product space consis
https://en.wikipedia.org/wiki/224%20%28number%29
224 (two hundred [and] twenty-four) is the natural number following 223 and preceding 225. In mathematics 224 is a practical number, and a sum of two positive cubes . It is also , making it one of the smallest numbers to be the sum of distinct positive cubes in more than one way. 224 is the smallest k with λ(k) = 24, where λ(k) is the Carmichael function. The mathematician and philosopher Alex Bellos suggested in 2014 that a candidate for the lowest uninteresting number would be 224 because it was, at the time, "the lowest number not to have its own page on [the English-language version of] Wikipedia". In other areas In the SHA-2 family of six cryptographic hash functions, the weakest is SHA-224, named because it produces 224-bit hash values. It was defined in this way so that the number of bits of security it provides (half of its output length, 112 bits) would match the key length of two-key Triple DES. The ancient Phoenician shekel was a standardized measure of silver, equal to 224 grains, although other forms of the shekel employed in other ancient cultures (including the Babylonians and Hebrews) had different measures. Likely not coincidentally, as far as ancient Burma and Thailand, silver was measured in a unit called a tikal, equal to 224 grains. See also 224 (disambiguation)
https://en.wikipedia.org/wiki/Field%20metabolic%20rate
Field metabolic rate (FMR) refers to a measurement of the metabolic rate of a free-living animal. Method Measurement of the Field metabolic rate is made using the doubly labeled water method, although alternative techniques, such as monitoring heart rates, can also be used. The advantages and disadvantages of the alternative approaches have been reviewed by Butler, et al. Several summary reviews have been published.
https://en.wikipedia.org/wiki/ChIL-sequencing
ChIL sequencing (ChIL-seq), also known as Chromatin Integration Labeling sequencing, is a method used to analyze protein interactions with DNA. ChIL-sequencing combines antibody-targeted controlled cleavage by Tn5 transposase with massively parallel DNA sequencing to identify the binding sites of DNA-associated proteins. It can be used to map global DNA binding sites precisely for any protein of interest. Currently, ChIP-Seq is the most common technique utilized to study protein–DNA relations, however, it suffers from a number of practical and economical limitations that ChIL-Sequencing does not. ChIL-Seq is a precise technique that reduces sample loss could be applied to single-cells. Uses ChIL-sequencing can be used to examine gene regulation or to analyze transcription factor and other chromatin-associated protein binding. Protein-DNA interactions regulate gene expression and are responsible for many biological processes and disease states. This epigenetic information is complementary to genotype and expression analysis. ChIL-Seq is an alternative to the current standard of ChIP-seq. ChIP-Seq suffers from limitations due to the cross linking step in ChIP-Seq protocols that can promote epitope masking and generate false-positive binding sites. As well, ChIP-seq suffers from suboptimal signal-to-noise ratios and poor resolution. ChIL-sequencing has the advantage of being a simpler technique suitable for low sample input due to the high signal-to-noise ratio, requiring less depth in sequencing for higher sensitivity. Specific DNA sites in direct physical interaction with proteins such as transcription factors can be isolated by Protein-A (pA) conjugated Tn5 bound to a protein of interest. Tn5 mediated cleavage produces a library of target DNA sites bound to a protein of interest in situ. Sequencing of prepared DNA libraries and comparison to whole-genome sequence databases allows researchers to analyze the interactions between target proteins and DNA, as well as d
https://en.wikipedia.org/wiki/Nukemap
Nukemap (stylised in all caps) is an interactive map using Mapbox API and declassified nuclear weapons effects data, created by Alex Wellerstein, a historian of science at the Stevens Institute of Technology who studies the history of nuclear weapons. The initial version was created in February 2012, with major upgrades in July 2013, which enables users to model the explosion of nuclear weapons (contemporary, historical, or of any given arbitrary yield) on virtually any terrain and at virtually any altitude of their choice. A variation of the script, Nukemap3D, featured rough models of mushroom clouds in 3D, scaled to their appropriate sizes. (Nukemap3D is no longer functional as Google had deprecated the Google Earth plugin.) The computer simulation of the effects of nuclear detonations has been described both as "stomach-churning" (by Wellerstein himself) and as "the most fun I’ve had with Google Maps since… well, possibly ever" despite the admittedly abjectly grim nature of the subject. Originally intended in part as a pedagogical device to illustrate the stark difference in scale between fission and fusion bombs, Nukemap went viral in 2013, necessitating a move to new servers. The website averages five "nukes" per visitor. Wellerstein's creation has garnered some popularity amongst nuclear strategists as an open source tool for calculating the costs of nuclear exchanges. As of May 2023, more than 300 million nukes have been "dropped" on the site. The Nukemap was a finalist for the National Science Foundation's Visualization Challenge in 2014. See also Computer simulation
https://en.wikipedia.org/wiki/Markov%27s%20principle
Markov's principle, named after Andrey Markov Jr, is a conditional existence statement for which there are many equivalent formulations, as discussed below. The principle is logically valid classically, but not in intuitionistic constructive mathematics. However, many particular instances of it are nevertheless provable in a constructive context as well. History The principle was first studied and adopted by the Russian school of constructivism, together with choice principles and often with a realizability perspective on the notion of mathematical function. In computability theory In the language of computability theory, Markov's principle is a formal expression of the claim that if it is impossible that an algorithm does not terminate, then for some input it does terminate. This is equivalent to the claim that if a set and its complement are both computably enumerable, then the set is decidable. In intuitionistic logic In predicate logic, a predicate P over some domain is called decidable if for every x in the domain, either P(x) is true, or P(x) is not true. This is not trivially true constructively. For a decidable predicate P over the natural numbers, Markov's principle then reads: That is, if P cannot be false for all natural numbers n, then it is true for some n. Markov's rule Markov's rule is the formulation of Markov's principle as a rule. It states that is derivable as soon as is, for decidable. Formally, Anne Troelstra proved that it is an admissible rule in Heyting arithmetic. Later, the logician Harvey Friedman showed that Markov's rule is an admissible rule in all of intuitionistic logic, Heyting arithmetic, and various other intuitionistic theories, using the Friedman translation. In Heyting arithmetic Markov's principle is equivalent in the language of arithmetic to: for a total recursive function on the natural numbers. In the presence of Church's thesis principle, the principle is equivalent to its form for primitive recursive fu
https://en.wikipedia.org/wiki/Dip%20soldering
Dip soldering is a small-scale soldering process by which electronic components are soldered to a printed circuit board (PCB) to form an electronic assembly. The solder wets to the exposed metallic areas of the board (those not protected with solder mask), creating a reliable mechanical and electrical connection. Dip soldering is used for both through-hole printed circuit assemblies, and surface mount. It is one of the cheapest methods to solder and is extensively used in the small scale industries of developing countries . Dip soldering is the manual equivalent of automated wave soldering. The apparatus required is just a small tank containing molten solder. A PCB with mounted components is dipped manually into the tank so that the molten solder sticks to the exposed metallic areas of the board. Dip solder process Dip soldering is accomplished by submerging parts to be joined into a molten solder bath. Thus, all components surfaces are coated with filler metal. Solders have low surface tension and high wetting capability. There are many types of solders, each used for different applications: Lead–silver is used for strength at higher-than-room temperature. Tin–lead is used as a general-purpose solder Tin–zinc is used for aluminium Cadmium–silver is used for strength at high temperatures Zinc–aluminium is used for aluminium and corrosion resistance Tin–silver and tin–bismuth are used for electronics. Because of the toxicity of lead, lead-free solders are being developed and more widely used. The molten bath can be any suitable filler metal, but the selection is usually confined to the lower melting point elements. The most common dip soldering operations use zinc-aluminum and tin-lead solders. Solder pot metal: cast iron or steel, electrically heated. Bath temperature: (for binary tin-lead alloys) or (for lead-free alloys) Solder composition: 60% , 40% or eutectic alloy. Process schematic The workpieces to be joined are treated with cleaning fl
https://en.wikipedia.org/wiki/TMEM143
TMEM143 (Transmembrane protein 143) is a protein that in humans is encoded by TMEM143 gene. TMEM143, a dual-pass protein (two transmembrane domains), is predicted to reside in the mitochondria and high expression has been found in both human skeletal muscle and the heart. Interaction with other proteins indicate that TMEM143 could potentially play a role in tumor suppression/expression and cancer regulation. Gene Located on the negative strand of human DNA, TMEM143 spans 31,882 base pairs on human chromosome 19 (19q13.33), neighbored by genes Coiled-coil domain containing 114 (CCDC114) and ER lumen protein-retaining receptor 1 (KDELR1). Transcript In humans, there are five transcript variants encoded by TMEM143 gene (1-5). Variant 1 is the longest mRNA transcript, with a coding region of 2577 nucleotides (nt) and a total of 8 exons, and possibly most indicative of function. Compared to variant 1, variant 2 (2472 nt, 424 amino acid protein) and variant 3 (2382nt, 394 amino acid protein) lack an in-frame exon in the 5' coding region while variant 4 (2277 nt, 359 amino acid protein) lacks two in-frame exons in the 5’ coding region, all leading to an N-terminally truncated protein. Transcript variant 5 is a non-coding RNA, approximately 2231 nt long, resulting in a transcript candidate for nonsense-mediated mRNA. Protein There are four protein isoforms, corresponding to a matching variant. Variant 1 codes for isoform a (the longest protein), and variants 2, 3, 4 code for isoforms b, c, and d, respectively. TMEM143 isoform a is 459 amino acids in length, has a molecular weight of 51.6 kDa and an isoelectric point of 9.7 in humans. A domain of unknown function (DUF3754) is present within which two transmembrane domains reside, 24 and 16 amino acids in length, both helical in nature. The transmembrane domains encompass the uncharged region present at amino acids 278 to 302. A predicted mitochondrial target peptide resides at the N-terminus spanning 52 amino acids bef
https://en.wikipedia.org/wiki/Rank-maximal%20allocation
Rank-maximal (RM) allocation is a rule for fair division of indivisible items. Suppose we have to allocate some items among people. Each person can rank the items from best to worst. The RM rule says that we have to give as many people as possible their best (#1) item. Subject to that, we have to give as many people as possible their next-best (#2) item, and so on. In the special case in which each person should receive a single item (for example, when the "items" are tasks and each task has to be done by a single person), the problem is called rank-maximal matching or greedy matching. The idea is similar to that of utilitarian cake-cutting, where the goal is to maximize the sum of utilities of all participants. However, the utilitarian rule works with cardinal (numeric) utility functions, while the RM rule works with ordinal utilities (rankings). Definition There are several items and several agents. Each agent has a total order on the items. Agents can be indifferent between some items; for each agent, we can partition the items to equivalence classes that contain items of the same rank. For example, If Alice's preference-relation is x > y,z > w, it means that Alice's 1st choice is x, which is better for her than all other items; Alice's 2nd choice is y and z, which are equally good in her eyes but not as good as x; and Alice's 3rd choice is w, which she considers worse than all other items. For every allocation of items to the agents, we construct its rank-vector as follows. Element #1 in the vector is the total number of items that are 1st-choice for their owners; Element #2 is the total number of items that are 2nd-choice for their owners; and so on. A rank-maximal allocation is one in which the rank-vector is maximum, in lexicographic order. Example Three items, x y and z, have to be divided among three agents whose rankings are: Alice: x > y > z Bob: x > y > z Carl: y > x > z In the allocation (x, y, z), Alice gets her 1st choice (x), Bob ge
https://en.wikipedia.org/wiki/Nested%20intervals
In mathematics, a sequence of nested intervals can be intuitively understood as an ordered collection of intervals on the real number line with natural numbers as an index. In order for a sequence of intervals to be considered nested intervals, two conditions have to be met: Every interval in the sequence is contained in the previous one ( is always a subset of ). The length of the intervals get arbitrarily small (meaning the length falls below every possible threshold after a certain index ). In other words, the left bound of the interval can only increase (), and the right bound can only decrease (). Historically - long before anyone defined nested intervals in a textbook - people implicitly constructed such nestings for concrete calculation purposes. For example, the ancient Babylonians discovered a method for computing square roots of numbers. In contrast, the famed Archimedes constructed sequences of polygons, that inscribed and surcumscribed a unit circle, in order to get a lower and upper bound for the circles circumference - which is the circle number Pi (). The central question to be posed is the nature of the intersection over all the natural numbers, or, put differently, the set of numbers, that are found in every Interval (thus, for all ). In modern mathematics, nested intervals are used as a construction method for the real numbers (in order to complete the field of rational numbers). Historic motivation As stated in the introduction, historic users of mathematics discovered the nesting of intervals and closely related algorithms as methods for specific calculations. Some variations and modern interpretations of these ancient techniques will be introduced here: Computation of square roots One intuitive algorithm is so easy to understand, that it could well be found by engaged high school students. When trying to find the square root of a number , one can be certain that , which gives the first interval , in which has to be found. If one kn
https://en.wikipedia.org/wiki/Pituicyte
Pituicytes are glial cells of the posterior pituitary. Their main role is to assist in the storage and release of neurohypophysial hormones. Structure Pituicytes are located in the pars nervosa of the posterior pituitary and interspersed with unmyelinated axons and Herring bodies. They generally stain dark purple with an H&E stain and are among the easiest structures to identify in the region. Pituicytes have an irregular and branched shape which resembles that of another type of glial cell: the astrocyte. Like astrocytes, their cytoplasm presents specific intermediate filaments made up of glial fibrillary acidic protein (GFAP). Function Pituicytes are similar to astrocytes, another type of glial cell. Their main role is to assist in the storage and release of hormones of the posterior pituitary. Pituicytes surround axonal endings and regulate hormone secretion by releasing their processes from these endings. Clinical significance Pituicytomas are rare tumors that arise from pituicytes. They may be mistaken for the much more common pituitary adenoma, as well as craniopharyngioma and meningioma. Symptoms from the mass effect of the tumor usually include vision disorders, and less often headaches, hypopituitarism (decreased function of the pituitary gland), fatigue, and decreased libido. See also List of distinct cell types in the adult human body
https://en.wikipedia.org/wiki/Indentation%20size%20effect
The indentation size effect (ISE) is the observation that hardness tends to increase as the indent size decreases at small scales. When an indent (any small mark, but usually made with a special tool) is created during material testing, the hardness of the material is not constant. At the small scale, materials will actually be harder than at the macro-scale. For the conventional indentation size effect, the smaller the indentation, the larger the difference in hardness. The effect has been seen through nanoindentation and microindentation measurements at varying depths. Dislocations increase material hardness by increasing flow stress through dislocation blocking mechanisms. Materials contain statistically stored dislocations (SSD) which are created by homogeneous strain and are dependent upon the material and processing conditions. Geometrically necessary dislocations (GND) on the other hand are formed, in addition to the dislocations statistically present, to maintain continuity within the material. These additional geometrically necessary dislocations (GND) further increase the flow stress in the material and therefore the measured hardness. Theory suggests that plastic flow is impacted by both strain and the size of the strain gradient experienced in the material. Smaller indents have higher strain gradients relative to the size of the plastic zone and therefore have a higher measured hardness in some materials. For practical purposes this effect means that hardness in the low micro and nano regimes cannot be directly compared if measured using different loads. However, the benefit of this effect is that it can be used to measure the effects of strain gradients on plasticity. Several new plasticity models have been developed using data from indentation size effect studies, which can be applied to high strain gradient situations such as thin films.
https://en.wikipedia.org/wiki/Non-Darwinian%20Evolution%20%28paper%29
"Non-Darwinian Evolution" is a scientific paper written by Jack Lester King and Thomas H. Jukes and published in 1969. It is credited, along with Motoo Kimura's 1968 paper "Evolutionary Rate at the Molecular Level", with proposing what became known as the neutral theory of molecular evolution. The paper brings together a wide variety of evidence, ranging from protein sequence comparisons to studies of the Treffers mutator gene in E. coli to analysis of the genetic code to comparative immunology, to argue that most protein evolution is due to neutral mutations and genetic drift. It was published in the journal Science on May 16, 1969. The idea of evolution at the molecular level being driven by the random processes of mutation and genetic drift, largely independent from natural selection, was controversial at the time; the provocative title further emphasized the break with mainstream evolutionary thought, which was dominated by the synthetic theory of evolution, often referred to as "Neo-Darwinism". Although they argued for essentially the same conclusion as Motoo Kimura's earlier paper, King and Jukes criticized one of Kimura's central arguments, an estimate of the rate of amino acid change in proteins that according to Kimura would indicate an impossibly high genetic load if the changes were caused by natural selection. The paper was initially rejected by its reviewers (one thought it was trivial and the other thought it was totally wrong ), but was published after an appeal. This time, the reviewer was James F. Crow, Motoo Kimura's collaborator. Despite the intentionally inflammatory title and "antiauthoritarian tone"—which according to historian Michael R. Dietrich "undoubtedly struck a nerve", especially since King and Jukes worked at UC Berkeley during that period of political unrest—the paper acknowledges the significance of natural selection; it merely argues against panselectionism (as advocated at the molecular level by G. G. Simpson and Emil L. Smith
https://en.wikipedia.org/wiki/Regulated%20function
In mathematics, a regulated function, or ruled function, is a certain kind of well-behaved function of a single real variable. Regulated functions arise as a class of integrable functions, and have several equivalent characterisations. Regulated functions were introduced by Nicolas Bourbaki in 1949, in their book "Livre IV: Fonctions d'une variable réelle". Definition Let X be a Banach space with norm || - ||X. A function f : [0, T] → X is said to be a regulated function if one (and hence both) of the following two equivalent conditions holds true: for every t in the interval [0, T], both the left and right limits f(t−) and f(t+) exist in X (apart from, obviously, f(0−) and f(T+)); there exists a sequence of step functions φn : [0, T] → X converging uniformly to f (i.e. with respect to the supremum norm || - ||∞). It requires a little work to show that these two conditions are equivalent. However, it is relatively easy to see that the second condition may be re-stated in the following equivalent ways: for every δ > 0, there is some step function φδ : [0, T] → X such that f lies in the closure of the space Step([0, T]; X) of all step functions from [0, T] into X (taking closure with respect to the supremum norm in the space B([0, T]; X) of all bounded functions from [0, T] into X). Properties of regulated functions Let Reg([0, T]; X) denote the set of all regulated functions f : [0, T] → X. Sums and scalar multiples of regulated functions are again regulated functions. In other words, Reg([0, T]; X) is a vector space over the same field K as the space X; typically, K will be the real or complex numbers. If X is equipped with an operation of multiplication, then products of regulated functions are again regulated functions. In other words, if X is a K-algebra, then so is Reg([0, T]; X). The supremum norm is a norm on Reg([0, T]; X), and Reg([0, T]; X) is a topological vector space with respect to the topology induced by the supremum norm. As noted
https://en.wikipedia.org/wiki/Nunuk%20Nuraini
Nunuk Nuraini (1961 – 27 January 2021), also known as Bu Nunuk ("Mrs. Nunuk"), was an Indonesian food scientist who invented Indomie's mi goreng-flavor instant noodles. West Java governor Ridwan Kamil called her ("hero for the boarding house kids"). The mi goreng flavor is described as a "cult favorite". The noodles are popular around the world, so much so that starting in 1995 the company had built three factories in Nigeria alone. Nuraini was born in Bandung, Indonesia. She graduated from Padjadjaran University in that city with a degree in food technology. Nuraini was a follower of Islam. She worked for Indofood as the flavor department manager, developing recipes there for almost three decades. She developed the popular mi goreng flavor in 1982, which was the first "dry" flavor sold without broth. Nuraini originated the flavor with traditional ingredients. Other flavors she invented include chicken curry, green chili, rendang, salted egg, sambal matah, and soto. Despite having many fans, Nuraini rarely appeared in public. Her death was confirmed by Indofood's head of public relations, Nurlita Novi Arlaida, who wrote: ("Mrs. Hajjah Nunuk Nuraini has died today and is returning peacefully to God's side"). The cause of Nuraini's death was not specified. Following her death, numerous tributes to Nuraini were posted on social media, thanking her for her work and expressing pride for it being a product of Indonesia.
https://en.wikipedia.org/wiki/SAML%201.1
Security Assertion Markup Language (SAML) is an XML standard for exchanging authentication and authorization data between security domains. SAML is a product of the OASIS (organization) Security Services Technical Committee. SAML 1.1 was ratified as an OASIS standard in September 2003. The critical aspects of SAML 1.1 are covered in detail in the official documents SAMLCore and SAMLBind. If you are new to SAML, you should probably read the introductory SAML topic first, and then the SAMLOverview document from OASIS. Prior to SAML 1.1, SAML 1.0 was adopted as an OASIS standard in November 2002. SAML has undergone one minor (V1.1) and one major revision (V2.0) since V1.0, which itself is a relatively simple protocol. SAML 1.0 is of more than historical interest, however, since the US Federal E-Authentication Initiative has adopted SAML 1.0 as its core technology. Versions 1.0 and 1.1 of SAML are similar. See SAMLDiff for specific differences between the two standards. This article concentrates on SAML 1.1 since it is an important standard upon which many other standards and implementations depend. Warning: Implementers and deployers should note well that all code examples in this article are non-normative and for illustration purposes only. Consult the OASIS SAML specifications for normative requirements. SAML 1.1 Assertions SAML assertions contain statements that service providers use to make access control decisions. For instance, authentication statements assert to the service provider that the principal did indeed authenticate with the identity provider at a particular time using a particular method of authentication. Other information about the principal may be disclosed in an authentication statement. In the authentication statement below, for example, the e-mail address of the principal is asserted to the service provider: <saml:Assertion xmlns:saml="urn:oasis:names:tc:SAML:1.0:assertion" MajorVersion="1" MinorVersion="1" AssertionID=
https://en.wikipedia.org/wiki/Program%20test%20authority
In broadcasting, program test authority (PTA) is an authorization to conduct on-air testing of broadcast station facilities authorized to be built under a construction permit. Once this testing is successfully completed, and all measured parameters match what was authorized in the permit, the permittee can apply to the broadcasting authority for a broadcast license to cover the permit. PTA lasts until the license is issued (or, rarely, denied). "Program" refers to the permission to broadcast regular radio programming or TV programming, instead of just a test transmission such as a test card or bars and tone (TV only), broadcast callsign or other required station identification, or dead air (which may not be permissible). Otherwise, only brief tests are allowed without PTA, in order to verify proper installation and functioning of all transmission equipment, such as the transmitter components (exciter and amplifier), feedline, antenna, and any diplexers that may be used. In the United States, the Federal Communications Commission (FCC) gives most stations automatic program test authority if they are using an omnidirectional antenna. For stations with a directional antenna, PTA is only for half of regular power, and is not granted until the permittee's broadcast engineer certifies that the antenna is installed properly so that the station's radiation pattern is correct. Stations must apply for a license within 30 days of starting broadcasts under program test authority. Upon issuance of the license, directional stations then commence full-power operation. Broadcast engineering Broadcast law
https://en.wikipedia.org/wiki/Wind%20turbine
A wind turbine is a device that converts the kinetic energy of wind into electrical energy. , hundreds of thousands of large turbines, in installations known as wind farms, were generating over 650 gigawatts of power, with 60 GW added each year. Wind turbines are an increasingly important source of intermittent renewable energy, and are used in many countries to lower energy costs and reduce reliance on fossil fuels. One study claimed that, wind had the "lowest relative greenhouse gas emissions, the least water consumption demands and the most favorable social impacts" compared to photovoltaic, hydro, geothermal, coal and gas energy sources. Smaller wind turbines are used for applications such as battery charging and remote devices such as traffic warning signs. Larger turbines can contribute to a domestic power supply while selling unused power back to the utility supplier via the electrical grid. Wind turbines are manufactured in a wide range of sizes, with either horizontal or vertical axes, though horizontal is most common. History The windwheel of Hero of Alexandria (10–70 CE) marks one of the first recorded instances of wind powering a machine. However, the first known practical wind power plants were built in Sistan, an Eastern province of Persia (now Iran), from the 7th century. These "Panemone" were vertical axle windmills, which had long vertical drive shafts with rectangular blades. Made of six to twelve sails covered in reed matting or cloth material, these windmills were used to grind grain or draw up water, and were used in the gristmilling and sugarcane industries. Wind power first appeared in Europe during the Middle Ages. The first historical records of their use in England date to the 11th and 12th centuries; there are reports of German crusaders taking their windmill-making skills to Syria around 1190. By the 14th century, Dutch windmills were in use to drain areas of the Rhine delta. Advanced wind turbines were described by Croatian invent
https://en.wikipedia.org/wiki/Animal%20latrine
Animal latrines (latrine areas, animal toilets, defecation sites) are places where wildlife animals habitually defecate and urinate. Many kinds of animals are highly specific in this respect and have stereotyped routines, including approach and departure. Many of them have communal, i.e., shared, latrines. Animals with dedicated defecation sites Animals with communal latrines include raccoons, Eurasian badgers, elephants, deer, antelopes, horses, and (prehistorically) dicynodonts (a 240-million-year-old site was called the "world's oldest public toilet"). A regularly used toilet area or dunghill, created by many mammals, such as hyraxes or moles, is also called a midden. Some lizards, such as yakka skinks (Egernia rugosa) and thorny devils use dedicated defecation sites. European rabbits may deposit their pellets both randomly over the range and at communal latrine sites. Function and impact Territoriality Middens and other types of defecation sites may serve as territorial markers. Elaborate "dungpile rituals" are reported for adult stallions, and deer bucks, which are thought to serve for confrontation avoidance. In contrast, female and young animals exhibit no such behavior. Sanitation Dedicated defecation sites are thought to be the result of sanitation-driven behavior. For example, the spider mite Stigmaeopsis miscanthi constructs woven nests, and nest members defecate at only one site inside the nest. Dedicated latrine areas observed by free-roaming horses mean that grazing area is kept parasite-free. Even stabled horses seem to have vestiges of such behavior. Herbivoral livestock is at risk of parasite/pathogen exposure from feces during grazing, therefore there is an interest in research of livestock behavior in the presence of feces both of their own species, and others, including wildlife, including the dependence on defecation patterns. Ecological impact Latrines of herbivores, such as antelopes, play an important role in ecology by providi
https://en.wikipedia.org/wiki/RBFOX1
Fox-1 homolog A, also known as ataxin 2-binding protein 1 (A2BP1) or hexaribonucleotide-binding protein 1 (HRNBP1) or RNA binding protein, fox-1 homolog (Rbfox1), is a protein that in humans is encoded by the RBFOX1 gene. Function Rbfox1 has an RNA recognition motif that is highly conserved among RNA-binding proteins. Rbfox1, and the related protein Rbfox2, bind the consensus RNA sequence motif (U)GCAUG within introns to exert their functions as alternative splicing factors. Additionally, the Rbfox1/A2BP1 protein binds to the C-terminus of ataxin-2, and may contribute to the restricted pathology of spinocerebellar ataxia type 2 (SCA2). Ataxin-2 is the gene product of the SCA2 gene which causes familial neurodegenerative diseases. Several alternatively spliced transcript variants have been found for this gene. Some of these variants localize to the nucleus and some other to the cytoplasm. Nuclear variants have a well-established role in tissue specific alternative splicing. Rbfox1 cytoplasmic variants modulate mRNA stability and translation. In stressed cells, Rbfox1 has been demonstrated to localize to cytoplasmic stress granules. See also Alternative splicing RNA-binding protein
https://en.wikipedia.org/wiki/Remote%20control%20animal
Remote control animals are animals that are controlled remotely by humans. Some applications require electrodes to be implanted in the animal's nervous system connected to a receiver which is usually carried on the animal's back. The animals are controlled by the use of radio signals. The electrodes do not move the animal directly, as if controlling a robot; rather, they signal a direction or action desired by the human operator and then stimulate the animal's reward centres if the animal complies. These are sometimes called bio-robots or robo-animals. They can be considered to be cyborgs as they combine electronic devices with an organic life form and hence are sometimes also called cyborg-animals or cyborg-insects. Because of the surgery required, and the moral and ethical issues involved, there has been criticism aimed at the use of remote control animals, especially regarding animal welfare and animal rights, especially when relatively intelligent complex animals are used. Non-invasive applications may include stimulation of the brain with ultrasound to control the animal. Some applications (used primarily for dogs) use vibrations or sound to control the movements of the animals. Several species of animals have been successfully controlled remotely. These include moths, beetles, cockroaches, rats, dogfish sharks, mice and pigeons. Remote control animals can be directed and used as working animals for search and rescue operations, covert reconnaissance, data-gathering in hazardous areas, or various other uses. Mammals Rats Several studies have examined the remote control of rats using micro-electrodes implanted into their brains and rely on stimulating the reward centre of the rat. Three electrodes are implanted; two in the ventral posterolateral nucleus of the thalamus which conveys facial sensory information from the left and right whiskers, and a third in the medial forebrain bundle which is involved in the reward process of the rat. This third electro
https://en.wikipedia.org/wiki/So%C5%82tan%20argument
The Sołtan argument is an astrophysical theory outlined in 1982 by Polish astronomer . It maintains that if quasars were powered by accretion onto a supermassive black hole, then such supermassive black holes must exist in our local universe as "dead" quasars. History As early as 1969, Donald Lynden-Bell wrote a paper suggesting that "dead quasars" were found at the center of the Milky Way and nearby galaxies by arguing that given the quasar-number counts, luminosities, distances, and the efficiency of accretion into a "Schwarzschild throat" through the last stable circular orbit (note that the term black hole had been coined only two years earlier and was still gaining popular usage), roughly 1010 quasars existed in the observable universe. This number density of "dead quasars" was attributed by Lynden-Bell to high mass-to-light ratio objects found at the center of galaxies. This is essentially the Sołtan argument, though the direct connection between black hole masses and quasar luminosity functions is missing. In the paper, Lynden-Bell also suggests some radical ideas that are now fully integrated into modern understanding of astrophysics including the model that accretion disks are supported by magnetic fields, that extragalactic cosmic rays are accelerated in them, and he estimates to within an order of magnitude the masses of several of the closest supermassive black holes including the ones in the Milky Way, M31, M32, M81, M82, M87, and NGC 4151. Thirteen years later, Sołtan explicitly showed that the luminosity () of quasars was due to the accretion rate of mass onto black holes given by: where is the efficiency factor is the time rate of mass falling into the black hole is the speed of light Given the number of observed quasars at various redshifts, he was able to derive an integrated energy density due to quasar output. Since observers on Earth are flux limited, there are always more quasars that exist than are observed and thus the energy density
https://en.wikipedia.org/wiki/Magic%20angle%20%28EELS%29
The magic angle is a particular value of the collection angle of an electron microscope at which the measured energy-loss spectrum "magically" becomes independent of the tilt angle of the sample with respect to the beam direction. The magic angle is not uniquely defined for isotropic samples, but the definition is unique in the (typical) case of small angle scattering on materials with a "c-axis", such as graphite. The "magic" angle depends on both the incoming electron energy (which is typically fixed) and the energy loss suffered by the electron. The ratio of the magic angle to the characteristic angle is roughly independent of the energy loss and roughly independent of the particular type of sample considered. Mathematical definition For the case of a relativistic incident electron, the "magic" angle is defined by the equality of two different functions (denoted below by and ) of the collection angle : and where is the speed of the incoming electron divided by the speed of light (N.B., the symbol is also often used in the older literature to denote the collection angle instead of ). Of course, the above integrals may easily be evaluated in terms of elementary functions, but they are presented as above because in the above form it is easier to see that the former integral is due to momentum transfers which are perpendicular to the beam direction, whereas the latter is due to momentum transfers parallel to the beam direction. Using the above definition, it is then found that
https://en.wikipedia.org/wiki/Monus
In mathematics, monus is an operator on certain commutative monoids that are not groups. A commutative monoid on which a monus operator is defined is called a commutative monoid with monus, or CMM. The monus operator may be denoted with the − symbol because the natural numbers are a CMM under subtraction; it is also denoted with the symbol to distinguish it from the standard subtraction operator. Notation Definition Let be a commutative monoid. Define a binary relation on this monoid as follows: for any two elements and , define if there exists an element such that . It is easy to check that is reflexive and that it is transitive. is called naturally ordered if the relation is additionally antisymmetric and hence a partial order. Further, if for each pair of elements and , a unique smallest element exists such that , then is called a commutative monoid with monus and the monus of any two elements and can be defined as this unique smallest element such that . An example of a commutative monoid that is not naturally ordered is , the commutative monoid of the integers with usual addition, as for any there exists such that , so holds for any , so is not a partial order. There are also examples of monoids that are naturally ordered but are not semirings with monus. Other structures Beyond monoids, the notion of monus can be applied to other structures. For instance, a naturally ordered semiring (sometimes called a dioid) is a semiring where the commutative monoid induced by the addition operator is naturally ordered. When this monoid is a commutative monoid with monus, the semiring is called a semiring with monus, or m-semiring. Examples If is an ideal in a Boolean algebra, then is a commutative monoid with monus under and . Natural numbers The natural numbers including 0 form a commutative monoid with monus, with their ordering being the usual order of natural numbers and the monus operator being a saturating variant of standard subtrac
https://en.wikipedia.org/wiki/Snap%20freezing
Snap freezing (or cook-chill or blast freezing) is the process of rapid cooling of a substance for the purpose of preservation. It is widely used in the culinary and scientific industries. Culinary uses Cooked meals can be preserved by rapid freezing after cooking is complete. The main target group for these products are those with little time for cooking such as schools, prisons, and hospitals. The process involves the cooking of meals at a central factory then rapidly chilling them for storage until they are needed. Snap frozen foods need to be packed in shallow trays to make the process more efficient. The food is cooled to a temperature under 3 degrees Celsius within 90 minutes of cooking and stored at a temperature of 0 to 3 degrees Celsius. The meals can then be transported in refrigerated transport to where the food is to be reheated and consumed when needed. The length of storage depends on the method used but is usually five days. For longer storage the food may be subjected to pasteurization after cooking. These processes have the advantage that preparation and cooking of meals is not tied to the times when the food is to be served, enabling staff and equipment to be used more efficiently. A properly managed operation is capable of supplying high-quality meals economically despite high initial equipment costs. There are potential problems; careful attention has to be paid to hygiene as there are a number of points in the process where food pathogens can gain access. This requires careful attention to both the control of the process and to staff training. Scientific use Snap-freeze is a term often used in scientific papers to describe a process by which a sample is very quickly lowered to temperatures below -70 °C. This is often accomplished by submerging a sample in liquid nitrogen. This prevents water from crystallising when it forms ice, and so better preserves the structure of the sample (e.g. RNA, protein, or live cells) See also Flash freezi
https://en.wikipedia.org/wiki/Francis%20F.%20Lee
Francis Fan Lee (李凡, born January 28, 1927) is an inventor, entrepreneur, and professor emeritus of Electrical Engineering and Computer Science at the Massachusetts Institute of Technology (MIT). Lee is the founder of Lexicon (company) (originally American Data Sciences). He is best known for three inventions: the Digital Cardiac Monitor (1969), the Digital Audio Signal Processor (1971), and the Digital Time Compression System (1972). In 1984, Lexicon won an Emmy Award for Engineering Excellence for the Model 1200 Audio Time Compressor and Expander, widely used in the television industry. Education Lee was born January 28, 1927, in Nanjing, China. In September 1948, during the Chinese Civil War, Lee left Shanghai aboard the USS General W.H. Gordon to complete his undergraduate education at the Massachusetts Institute of Technology. He studied Electrical Engineering and earned his Bachelor of Science in 1950 and Master of Science in 1951. In Fall 1952, Lee entered the PhD program at MIT. He withdrew in 1954 to pursue his career, becoming a naturalized US citizen on November 15, 1954. He returned to academia in 1964, and completed his PhD in Electrical Engineering in 1965. Early career In 1954, Lee took a full-time position as Research Engineer with the Servomechanism Laboratory. He was part of a team working on the first Digitally Controlled Milling Machine In 1955, Lee joined the Bizmac Computer Division of RCA. He left a year later to join the UNIVAC super-computer division of Remington Rand. Project MAC In 1963, Lee accepted a one-year appointment to work on Project MAC, a time-sharing Multiple Access Computer being developed at MIT under the direction of Robert Fano. Fano launched Project MAC with a 6-week summer session that drew 57 people (including Lee) from universities, industry, and government for brainstorming and collaboration. At the end of the session, Lee described his work on speeding up computer memory in "Lookaside Memory Implementation" (1963).
https://en.wikipedia.org/wiki/Susan%20Goldstine
Susan Goldstine is an American mathematician active in mathematics and fiber arts. She is a professor of mathematics at St. Mary's College of Maryland, and (for 2019–2022) the Steven Muller Distinguished Professor in the Sciences at St. Mary's College. Education and career Goldstine graduated summa cum laude from Amherst College in 1993. She completed a Ph.D. in mathematics at Harvard University in 1998. Her dissertation, Spin Representations and Lattices, was supervised by Benedict Gross. After postdoctoral and visiting assistant professorships at McMaster University, Ohio State University, and Amherst College, she joined the St. Mary's College faculty in 2004. Contributions Goldstine has made and exhibited many pieces of mathematical art, often involving textiles. A set of bead crochet jewelry pieces by her visualizing the map coloring problem on three different manifolds won the prize for "best textile, sculpture, or other medium" in the art show of the 2015 Joint Mathematics Meetings. She is the coauthor of the book Crafting Conundrums: Puzzles and Patterns for the Bead Crochet Artist (with Ellie Baker, A K Peters / CRC Press, 2014). Combining her interests in mathematics and fiber arts she is one of 24 mathematicians and artists who make up the Mathemalchemy Team. Personal life Goldstine is the granddaughter of teacher and author Bel Kaufman and the great-great-granddaughter of Sholem Aleichem.
https://en.wikipedia.org/wiki/Xinetd
In computer networking, xinetd (Extended Internet Service Daemon) is an open-source super-server daemon which runs on many Unix-like systems, and manages Internet-based connectivity. It offers a more secure alternative to the older inetd ("the Internet daemon"), which most modern Linux distributions have deprecated. Description xinetd listens for incoming requests over a network and launches the appropriate service for that request. Requests are made using port numbers as identifiers and xinetd usually launches another daemon to handle the request. It can be used to start services with both privileged and non-privileged port numbers. xinetd features access control mechanisms such as TCP Wrapper ACLs, extensive logging capabilities, and the ability to make services available based on time. It can place limits on the number of servers that the system can start, and has deployable defense mechanisms to protect against port scanners, among other things. On some implementations of Mac OS X, this daemon starts and maintains various Internet-related services, including FTP and telnet. As an extended form of inetd, it offers enhanced security. It replaced inetd in Mac OS X v10.3, and subsequently launchd replaced it in Mac OS X v10.4. However, Apple has retained inetd for compatibility purposes. Configuration Configuration of xinetd resides in the default configuration file /etc/xinetd.conf, and configuration of the services it supports resides in configuration files stored in the /etc/xinetd.d directory. The configuration for each service usually includes a switch to control whether xinetd should enable or disable the service. An example configuration file for the RFC 868 time server: # default: off # description: An RFC 868 time server. This protocol provides a # site-independent, machine readable date and time. The Time service sends back # to the originating source the time in seconds since midnight on January first # 1900. # This is the tcp version. service
https://en.wikipedia.org/wiki/National%20Sleep%20Foundation
The National Sleep Foundation (NSF) is an American non-profit, charitable organization. Founded in 1990, its stated goal is to provide expert information on health-related issues concerning sleep. It is largely funded by pharmaceutical and medical device companies. Research NSF Sleep Duration Recommendations In 2015 NSF released the results of a research study on sleep duration recommendations. The paper titled "National Sleep Foundation's sleep time duration recommendations: methodology and results summary" was published in the peer-reviewed Sleep Health Journal. NSF convened an expert panel of 18 leading scientists and researchers tasked with updating the official sleep duration recommendations. The panelists included sleep specialists and representatives from leading organizations including the American Academy of Pediatrics, American Association of Anatomy, American College of Chest Physicians, American Congress of Obstetricians and Gynecologists, American Geriatrics Society, American Neurological Association, American Physiological Society, American Psychiatric Association, American Thoracic Society, Gerontological Society of America, Human Anatomy and Physiology Society, and Society for Research in Human Development. The panelists participated in a rigorous scientific process that included reviewing over 300 current scientific publications and voting on how much sleep is appropriate throughout the lifespan. Sleep Health Index NSF developed Sleep Health Index to measure sleep health at a global group or at an individual level. It was created with the help of sleep experts and public opinion research experts. It is composed of three sub-component scales: sleep duration, sleep quality, and sleep disorders. The Index is fielded quarterly and results are publicly available. Sleep in America Poll NSF has conducted a national poll called Sleep in America Poll to catalog the state of sleep in America since 1991.  This poll provides valuable information to the
https://en.wikipedia.org/wiki/T.%20William%20Olle
T. William (Bill) Olle (born 1933 and died March 2019) was a British computer scientist and consultant and President of T. William Olle Associates, England. Biography Bill Olle was educated at Boston Grammar School (1943-1950). He received an M.Sc. degree in 1954 and a Ph.D. degree in 1957, both in Astrophysics at the University of Manchester, which involved extensive programming work on the Manchester University Electronic Computer. In 1957, he moved to the Netherlands, where he worked in computing for a NATO organization. In 1964, he moved to the United States, where he was employed by Control Data Corporation in Palo Alto, California until 1966. From 1967 to 1971, he was employed by the RCA Corporation in Cherry Hill, New Jersey. In 1972, after a year in Norway, he returned to the UK to establish his own consultancy firm, T. William Olle Associates, specializing in database management applications and information systems methodologies. He consulted clients in Europe, Australia, and Canada, and presented lectures on database topics around the world. He retired in 1993. Beginning in the 1970s, Olle became active in the CODASYL organization as Chairman of its Systems Committee and spearheaded the preparation of two early analytical reports on "Generalized Database Management Systems". He represented the British Computer Society on IFIP TC8 from its inception in 1977. He was also active in database standards work in ISO and was chairman of the BSI standards committee for many years. Bill Olle was awarded an honorary doctorate by Middlesex University in 2001. Work Olle's research interest in the field of computing started in 1953 at the University of Manchester. In the 1960s, he became interested in database applications, and after his retirement in the 1990s, he focused on the history of computing and on "professionalism in the computer field". Publications Olle published numerous books and articles. The following is a selection: 1971. Feature Analysis of Gen
https://en.wikipedia.org/wiki/Scalar%20potential
In mathematical physics, scalar potential, simply stated, describes the situation where the difference in the potential energies of an object in two different positions depends only on the positions, not upon the path taken by the object in traveling from one position to the other. It is a scalar field in three-space: a directionless value (scalar) that depends only on its location. A familiar example is potential energy due to gravity. A scalar potential is a fundamental concept in vector analysis and physics (the adjective scalar is frequently omitted if there is no danger of confusion with vector potential). The scalar potential is an example of a scalar field. Given a vector field , the scalar potential is defined such that: where is the gradient of and the second part of the equation is minus the gradient for a function of the Cartesian coordinates . In some cases, mathematicians may use a positive sign in front of the gradient to define the potential. Because of this definition of in terms of the gradient, the direction of at any point is the direction of the steepest decrease of at that point, its magnitude is the rate of that decrease per unit length. In order for to be described in terms of a scalar potential only, any of the following equivalent statements have to be true: where the integration is over a Jordan arc passing from location to location and is evaluated at location . where the integral is over any simple closed path, otherwise known as a Jordan curve. The first of these conditions represents the fundamental theorem of the gradient and is true for any vector field that is a gradient of a differentiable single valued scalar field . The second condition is a requirement of so that it can be expressed as the gradient of a scalar function. The third condition re-expresses the second condition in terms of the curl of using the fundamental theorem of the curl. A vector field that satisfies these conditions is said to be irro
https://en.wikipedia.org/wiki/MTR%20%28software%29
My traceroute, originally named Matt's traceroute (MTR), is a computer program that combines the functions of the traceroute and ping programs in one network diagnostic tool. MTR probes routers on the route path by limiting the number of hops individual packets may traverse, and listening to responses of their expiry. It will regularly repeat this process, usually once per second, and keep track of the response times of the hops along the path. History The original Matt's traceroute program was written by Matt Kimball in 1997. Roger Wolff took over maintaining MTR (renamed My traceroute) in October 1998. Fundamentals MTR is licensed under the terms of the GNU General Public License (GPL) and works under modern Unix-like operating systems. It normally works under the text console, but it also has an optional GTK+-based graphical user interface (GUI). MTR relies on Internet Control Message Protocol (ICMP) Time Exceeded (type 11, code 0) packets coming back from routers, or ICMP Echo Reply packets when the packets have hit their destination host. MTR also has a User Datagram Protocol (UDP) mode (invoked with "-u" on the command line or pressing the "u" key in the curses interface) that sends UDP packets, with the time to live (TTL) field in the IP header increasing by one for each probe sent, toward the destination host. When the UDP mode is used, MTR relies on ICMP port unreachable packets (type 3, code 3) when the destination is reached. MTR also supports IPv6 and works in a similar manner but instead relies on ICMPv6 messages. The tool is often used for network troubleshooting. By showing a list of routers traversed, and the average round-trip time as well as packet loss to each router, it allows users to identify links between two given routers responsible for certain fractions of the overall latency or packet loss through the network. This can help identify network overuse problems. Examples This example shows MTR running on Linux tracing a route from the
https://en.wikipedia.org/wiki/Curvilinear%20motion
The motion of an object moving in a curved path is called curvilinear motion. Example: A stone thrown into the air at an angle'''. Curvilinear motion describes the motion of a moving particle that conforms to a known or fixed curve. The study of such motion involves the use of two co-ordinate systems, the first being planar motion and the latter being cylindrical motion. Planar motion In planar motion, the velocity and acceleration components of the particle are always tangential and normal to the fixed curve. The velocity is always tangential to the curve and the acceleration can be broken up into both a tangential and normal component. Cylindrical components With cylindrical co-ordinates which are described as î and j, the motion is best described in polar form with components that resemble polar vectors. As with planar motion, the velocity is always tangential to the curve, but in this form acceleration consist of different intermediate components that can now run along the radius and its normal vector. This type of co-ordinate system is best used when the motion is restricted to the plane upon which it travels.
https://en.wikipedia.org/wiki/Caldicellulosiruptor
Caldicellulosiruptor is a genus of thermophilic, anaerobic, Gram-positive, non-spore forming bacteria. Originally placed within the highly polyphyletic class Clostridia, order Thermoanaerobacterales and family Thermoanaerobacterales Family III according to the NCBI and LPSN, it is now thought to lie outside of the Bacillota. Caldicellulosiruptor is known to degrade and ferment complex carbohydrates from plant matter, such as cellulose and hemicellulose (hence its name), and certain species in the genus have been identified as potential candidates for biofuel production. Phylogeny The currently accepted taxonomy is based on the List of Prokaryotic names with Standing in Nomenclature (LPSN) and National Center for Biotechnology Information (NCBI) See also List of bacterial orders List of bacteria genera
https://en.wikipedia.org/wiki/Parrot%20OS
Parrot OS is a Linux distribution based on Debian with a focus on security, privacy, and development. Core Parrot is based on Debian's "testing" branch, with a Linux 6.1 kernel. It follows a rolling release development model. The desktop environment is MATE, and the default display manager is LightDM. The system is certified to run on devices which have a minimum of 256MB of RAM, and it is suitable for both 32-bit (i386) and 64-bit (amd64) processor architectures. Moreover, the project is available for ARMv7 (armhf) architectures. In June 2017, the Parrot Team announced they were considering to change from Debian to Devuan, mainly because of problems with systemd. As of January 21st, 2019, the Parrot team has begun to phase out the development of their 32-bit (i386) ISO. In August 2020, the Parrot OS officially supports Lightweight Xfce Desktop. Editions Parrot has multiple editions that are based upon Debian, with various desktop environments available. Home Edition Parrot OS Home Edition is the base edition of Parrot designed for daily use, and it targets regular users who need a "lightweight" system on their laptops or workstations. The distribution is useful for daily work. Parrot Home also includes programs to chat privately with people, encrypt documents, or browse the internet anonymously. The system can also be used as a starting point to build a system with a custom set of security tools. Security Edition Parrot OS Security Edition is designed for penetration testing, vulnerability assessment and mitigation, computer forensics, and anonymous web browsing. Parrot ARM Parrot ARM is a lightweight Parrot release for embedded systems. It is currently available for Raspberry Pi devices. Parrot Architect & IoT ParrotOS with nothing pre-installed. Install any software and DE with this edition. Parrot OS tools There are multiple tools in Parrot OS which are specially designed for Security Researchers and are related to penetration testing. A few
https://en.wikipedia.org/wiki/Differential%20display
Differential display (also referred to as DDRT-PCR or DD-PCR) is a laboratory technique that allows a researcher to compare and identify changes in gene expression at the mRNA level between two or more eukaryotic cell samples. It was the most commonly used method to compare expression profiles of two eukaryotic cell samples in the 1990s. By 2000, differential display was superseded by DNA microarray approaches. In differential display, first all the RNA in each sample is reverse transcribed using a set of 3 "anchored primers" (having a short sequence of deoxy-thymidine nucleotides at the end) to create a cDNA library for each sample, followed by PCR amplification using arbitrary 3 primers for cDNA strand amplification together with anchored 3 primers for RNA strand amplification, identical to those used to create the library; about forty arbitrary primers is the optimal number to transcribe almost all of the mRNA. The resulting transcripts are then separated by electrophoresis and visualized, so that they can be compared. The method was prone to error due to different mRNAs migrated into single bands, differences in less abundant mRNAs getting drowned by more abundant mRNAs, sensitivity to small changes in cell culture conditions, and a tendency to amplify 3 fragments rather than full mRNAs, and the necessity to use about 300 primers to catch all the mRNA. The method was first published in Science in 1992.
https://en.wikipedia.org/wiki/List%20of%20web%20browsers%20for%20Unix%20and%20Unix-like%20operating%20systems
The following is a list of web browsers for various Unix and Unix-like operating systems. Not all of these browsers are specific to these operating systems; some are available on non-Unix systems as well. Some, but not most, have a mobile version. Graphical Colored items in this table are discontinued. Text-based Links ELinks Line-mode browser Lynx w3m See also List of web browsers Comparison of web browsers Comparison of lightweight web browsers
https://en.wikipedia.org/wiki/Flavorist
A flavorist, also known as flavor chemist, is someone who uses chemistry to engineer artificial and natural flavors. The tools and materials used by flavorists are almost the same as that used by perfumers with the exception that flavorists seek to mimic or modify both the olfactory and gustatory properties of various food products rather than creating just abstract smells. Additionally, the materials and chemicals that a flavorist utilizes for flavor creation must be safe for human consumption. The profession of flavorists came about when affordable refrigeration for the home spurred of food processing technology, which could affect the quality of the flavor of the food. In some cases, these technologies can remove naturally occurring flavors. To remedy the flavor loss, the food processing industry created the flavor industry. The chemists that resolved the demand of the food processing industry became known as flavorists. Education Educational requirements for the profession known as flavorist are varied. Flavorists are often graduated either in Chemistry, Biology or Food Science up to PhDs obtained in subjects such as Biochemistry and Chemistry. Because, however, the training of a flavorist is mostly done on-the-job and specifically at a flavor company known as a flavor house, this training is similar to the apprentice system. Located in Versailles (France), ISIPCA French School offers two years of high-standard education in food flavoring including 12 months traineeship in a flavor company. This education program provides students with solid background in Flavoring formulation, flavor application, and flavor chemistry (analysis and sensory). The British Society of Flavourists together with Reading University provide, every year, a three-week flavorist training course for flavorists from all around the world. Flavorist societies In the United States, a certified flavorist must be a member of the Society of Flavor Chemists, which meets in New Jersey, Cincin
https://en.wikipedia.org/wiki/Minimal%20coupling
In analytical mechanics and quantum field theory, minimal coupling refers to a coupling between fields which involves only the charge distribution and not higher multipole moments of the charge distribution. This minimal coupling is in contrast to, for example, Pauli coupling, which includes the magnetic moment of an electron directly in the Lagrangian. Electrodynamics In electrodynamics, minimal coupling is adequate to account for all electromagnetic interactions. Higher moments of particles are consequences of minimal coupling and non-zero spin. Non-relativistic charged particle in an electromagnetic field In Cartesian coordinates, the Lagrangian of a non-relativistic classical particle in an electromagnetic field is (in SI Units): where is the electric charge of the particle, is the electric scalar potential, and the are the components of the magnetic vector potential that may all explicitly depend on and . This Lagrangian, combined with Euler–Lagrange equation, produces the Lorentz force law and is called minimal coupling. Note that the values of scalar potential and vector potential would change during a gauge transformation, and the Lagrangian itself will pick up extra terms as well, but the extra terms in the Lagrangian add up to a total time derivative of a scalar function, and therefore still produce the same Euler–Lagrange equation. The canonical momenta are given by Note that canonical momenta are not gauge invariant, and are not physically measurable. However, the kinetic momenta are gauge invariant and physically measurable. The Hamiltonian, as the Legendre transformation of the Lagrangian, is therefore This equation is used frequently in quantum mechanics. Under a gauge transformation, where f(r,t) is any scalar function of space and time, the aforementioned Lagrangian, canonical momenta and Hamiltonian transform like which still produces the same Hamilton's equation: In quantum mechanics, the wave function will a
https://en.wikipedia.org/wiki/Clearing%20the%20neighbourhood
"Clearing the neighbourhood" (or dynamical dominance) around a celestial body's orbit describes the body becoming gravitationally dominant such that there are no other bodies of comparable size other than its natural satellites or those otherwise under its gravitational influence. "Clearing the neighbourhood" is one of three necessary criteria for a celestial body to be considered a planet in the Solar System, according to the definition adopted in 2006 by the International Astronomical Union (IAU). In 2015, a proposal was made to extend the definition to exoplanets. In the end stages of planet formation, a planet, as so defined, will have "cleared the neighbourhood" of its own orbital zone, i.e. removed other bodies of comparable size. A large body that meets the other criteria for a planet but has not cleared its neighbourhood is classified as a dwarf planet. This includes Pluto, whose orbit intersects with Neptune's orbit and shares its orbital neighbourhood with many Kuiper belt objects. The IAU's definition does not attach specific numbers or equations to this term, but all IAU-recognised planets have cleared their neighbourhoods to a much greater extent (by orders of magnitude) than any dwarf planet or candidate for dwarf planet. The phrase stems from a paper presented to the 2000 IAU general assembly by the planetary scientists Alan Stern and Harold F. Levison. The authors used several similar phrases as they developed a theoretical basis for determining if an object orbiting a star is likely to "clear its neighboring region" of planetesimals based on the object's mass and its orbital period. Steven Soter prefers to use the term "dynamical dominance", and Jean-Luc Margot notes that such language "seems less prone to misinterpretation". Prior to 2006, the IAU had no specific rules for naming planets, as no new planets had been discovered for decades, whereas there were well-established rules for naming an abundance of newly discovered small bodies such as
https://en.wikipedia.org/wiki/Automodello
Automodello is a manufacturer of resin-cast hand-built models in a variety of scales. The company is headquartered in Buffalo Grove, Illinois. Inaugural introduction Automodello was started by James Cowen and Raffi Minasian, released the inaugural model, a 1964 Griffith Series 200, in 2010 at the prestigious Amelia Island Concours d'Elegance where it was designated the Official Model for the event. Models were signed by Andrew "Jack" Griffith, founder of Griffith sports car that later became the TVR. Marques offered Automodello tends to focus on British and American vehicles that have been overlooked by the modeling community. Through 2014, Automodello has released or will release thirty-six separate models. Though the company started out with 1:43 scale reproductions to exacting standards, larger scales have increasingly been offered, including a very large 1:8 (a traditionally popular scale for clay models made by designers in the large auto companies). The following are the models currently offered or planned: 1:43 scale: 1934 Duesenberg J Graber 1937 Delahaye 135MS Figoni & Falaschi Coupe 1938 Lincoln K V12 Judkins Coupe 1938 Packard Twelve Convertible Victoria was the first 1/43rd model awarded Model of the Year in the 20-year history of the Diecast Zone. 1939-1940 Checker Model A Taxi in both New York City and Chicago livery 1948 Timbs Streamliner 1952-1954 Cunningham C-3 Cabriolet 1956-1958 Dual Ghia 1962 Ford Mustang I Concept 1963 Studebaker Avanti R2 supercharged 1964 Marcos 1800 1964 Griffith Series 200 1964 Sunbeam Tiger Mark I 1964-1968 DeTomaso Vallelunga 1965 Ford GT40 Roadster Prototype Pricing 1966 Griffith Series 600 1966 Fitch (Corvair) Sprint 1966 Fitch Phoenix 1967 to 1972 Intermeccanica Italia 1969-1973 Gilbern Genie Invader 1972-1979 TVR M series 1974 to 1976 Bricklin SV1 2011-2014 Noble M600 1:24 scale: 1930 Cord L-29 Brooks Stevens Speedster 1934 Duesenburg J Graber 1935 Duesenberg SJ Speedster Mormon Meteor
https://en.wikipedia.org/wiki/Antimicrobials%20in%20aquaculture
Antimicrobials destroy bacteria, viruses, fungi, algae, and other microbes. The cells of bacteria (prokaryotes), such as salmonella, differ from those of higher-level organisms (eukaryotes), such as fish. Antibiotics are chemicals designed to either kill or inhibit the growth of pathogenic bacteria while exploiting the differences between prokaryotes and eukaryotes in order to make them relatively harmless in higher-level organisms. Antibiotics are constructed to act in one of three ways: by disrupting cell membranes of bacteria (rendering them unable to regulate themselves), by impeding DNA or protein synthesis, or by hampering the activity of certain enzymes unique to bacteria. Antibiotics are used in aquaculture to treat diseases caused by bacteria. Sometimes the antibiotics are used to treat diseases, but more commonly antibiotics are used to prevent diseases by treating the water or fish before disease occurs. While this prophylactic method of preventing disease is profitable because it prevents loss and allows fish to grow more quickly, there are several downsides. The overuse of antibiotics can create antibiotic-resistant bacteria. Antibiotic-resistant bacteria can spontaneously arise when selective pressure to survive results in changes to the DNA sequence of a bacterium allowing that bacterium to survive antibiotic treatments. Because some of the same antibiotics are used to treat fish that are used to treat human disease, pathogenic bacteria causing human disease can also become resistant to antibiotics as a result of treatment of fish with antibiotics. For this reason, the overuse of antibiotics in treatment of fish aquaculture (among other agricultural uses) could create public health issues. Overview The issue has two sides. In some countries, clean water supplies for aquaculture are extremely limited. Untreated animal manure and human waste are used as feed in shrimp farms and tilapia farms in China and Thailand, in addition to the collection
https://en.wikipedia.org/wiki/N-Methyl-2-pyrrolidone
N-Methyl-2-pyrrolidone (NMP) is an organic compound consisting of a 5-membered lactam. It is a colorless liquid, although impure samples can appear yellow. It is miscible with water and with most common organic solvents. It also belongs to the class of dipolar aprotic solvents such as dimethylformamide and dimethyl sulfoxide. It is used in the petrochemical, polymer and battery industries as a solvent, exploiting its nonvolatility and ability to dissolve diverse materials (including polyvinylidene difluoride, PVDF). Preparation NMP is produced industrially by a typical ester-to-amide conversion, by treating butyrolactone with methylamine. Alternative routes include the partial hydrogenation of N-methylsuccinimide and the reaction of acrylonitrile with methylamine followed by hydrolysis. About 200,000 to 250,000 tons are produced annually. Applications NMP is used to recover certain hydrocarbons generated in the processing of petrochemicals, such as the recovery of 1,3-butadiene and acetylene. It is used to absorb hydrogen sulfide from sour gas and hydrodesulfurization facilities. Its good solvency properties have led to NMP's use to dissolve a wide range of polymers. Specifically, it is used as a solvent for surface treatment of textiles, resins, and metal coated plastics or as a paint stripper. It is also used as a solvent in the commercial preparation of polyphenylene sulfide. In the pharmaceutical industry, N-methyl-2-pyrrolidone is used in the formulation for drugs by both oral and transdermal delivery routes. It is also used heavily in lithium ion battery fabrication, as a solvent for electrode preparation, because NMP has a unique ability to dissolve polyvinylidene fluoride binder. Due to NMP's toxicity and high boiling point, there is much effort to replace it in battery manufacturing with other solvent(s), like water. Health hazards N-Methyl-2-pyrrolidone is an agent that causes the production of physical defects in the developing embryo. It also
https://en.wikipedia.org/wiki/Link%20Labs
Link Labs is an American company based in Annapolis, Maryland, that develops computer network technology for business and industrial customers. Link Labs technologies are marketed for Internet of things (IoT) applications and devices. History The company was founded in 2014 by Brian Ray and 3 engineers from the Johns Hopkins Applied Physics Laboratory. In August 2015, it announced a venture capital investment of $5.7 million. The investment round was led by TCP and joined by the Maryland Venture Fund, Blu Venture, Inflection Point Partners, and others. Products Symphony Link is a low power, wide-area wireless network (LPWAN) that allows for monitoring and two-way communication with sensor devices. According to Link Labs, Symphony Link can support up to 250,000 endpoints on each gateway, and ranges up to 7 miles. Additionally, Symphony Link supports upgrading firmware over the air and allows for sending and receiving compressed bidirectional message acknowledgements. AirFinder is a product division of Link Labs and a real-time location system (RTLS). It utilizes open-source iBeacon and Bluetooth Low Energy (BLE) technology to track assets and individuals. According to Link Labs, AirFinder is used to improve efficiencies through location tracking in healthcare organizations, manufacturing plants, and transport hubs. Marketing In August 2015, Link Labs partnered with Stream Technologies, allowing customers a more robust subscription, billing, and data management platform. In June 2016, the Stanley Mechanical Solutions unit division of Stanley Black & Decker unveiled its Shelter system, an IoT-enabled school safety system. Link Labs worked in partnership with Stanley Mechanical to create a technological solution with extended battery life, quick response time, and long-range signal capability. In October 2016, M2M Spectrum Networks announced plans to purchase up to 10,000 base stations. In November 2016, Link Labs announced via press release that it is devel
https://en.wikipedia.org/wiki/Resistor%20ladder
A resistor ladder is an electrical circuit made from repeating units of resistors, in specific configurations. An R–2R ladder configuration is a simple and inexpensive way to perform digital-to-analog conversion (DAC), using repetitive arrangements of precise resistor networks in a ladder-like configuration. A string resistor ladder configuration implements the non-repetitive reference network. History A 1953 paper "Coding by Feedback Methods" describes "decoding networks" that convert numbers (in any base) represented by voltage sources or current sources connected to resistor networks in a "shunt resistor decoding network" (which in base 2 corresponds to the binary-weighted configuration) or in a "ladder resistor decoding network" (which in base 2 corresponds to R–2R configuration) into a single voltage output. The paper gives an advantage of R–2R that impedances seen by the sources are more equal. Another historic description is in US Patent 3108266, filed in 1955, "Signal Conversion Apparatus". String resistor ladder network A string of many resistors connected between two reference voltages is called a "resistor string". The resistors act as voltage dividers between the referenced voltages. A Kelvin divider or string DAC is a string of equal valued resistors. Analog-to-digital conversion Each tap of the string generates a different voltage, which can be compared with another voltage: this is the basic principle of a flash ADC (analog-to-digital converter). Often a voltage is converted to a current, enabling the possibility to use an R–2R ladder network. Disadvantage: for an n-bit ADC, the number of resistors grows exponentially, as resistors are required, while the R–2R resistor ladder only increases linearly with the number of bits, as it needs only resistors. Advantage: higher impedance values can be reached using the same number of components. Digital-to-analog conversion A string resistor can function as a DAC by having the bits of the binary
https://en.wikipedia.org/wiki/Memory%20organisation
There are several ways to organise memories with respect to the way they are connected to the cache: one-word-wide memory organisation wide memory organisation interleaved memory organisation independent memory organisation One-Word-Wide The memory is one word wide and connected via a one word wide bus to the cache. Wide The memory is more than one word wide (usually four words wide) and connected by an equally wide bus to the low level cache (which is also wide). From the cache multiple busses of one word wide go to a MUX which selects the correct bus to connect to the high level cache. Interleaved There are several memory banks which are one word wide, and one word wide bus. There is some logic in the memory that selects the correct bank to use when the memory gets accessed by the cache. Memory interleaving is a way to distribute individual addresses over memory modules. Its aim is to keep the most of modules busy as computations proceed. With memory interleaving, the low-order k bits of the memory address generally specify the module on several buses. Computer memory See also Cache hierarchy Memory hierarchy Memory geometry
https://en.wikipedia.org/wiki/Uniformization%20theorem
In mathematics, the uniformization theorem says that every simply connected Riemann surface is conformally equivalent to one of three Riemann surfaces: the open unit disk, the complex plane, or the Riemann sphere. The theorem is a generalization of the Riemann mapping theorem from simply connected open subsets of the plane to arbitrary simply connected Riemann surfaces. Since every Riemann surface has a universal cover which is a simply connected Riemann surface, the uniformization theorem leads to a classification of Riemann surfaces into three types: those that have the Riemann sphere as universal cover ("elliptic"), those with the plane as universal cover ("parabolic") and those with the unit disk as universal cover ("hyperbolic"). It further follows that every Riemann surface admits a Riemannian metric of constant curvature, where the curvature can be taken to be 1 in the elliptic, 0 in the parabolic and -1 in the hyperbolic case. The uniformization theorem also yields a similar classification of closed orientable Riemannian 2-manifolds into elliptic/parabolic/hyperbolic cases. Each such manifold has a conformally equivalent Riemannian metric with constant curvature, where the curvature can be taken to be 1 in the elliptic, 0 in the parabolic and -1 in the hyperbolic case. History Felix and Henri conjectured the uniformization theorem for (the Riemann surfaces of) algebraic curves. extended this to arbitrary multivalued analytic functions and gave informal arguments in its favor. The first rigorous proofs of the general uniformization theorem were given by and . Paul Koebe later gave several more proofs and generalizations. The history is described in ; a complete account of uniformization up to the 1907 papers of Koebe and Poincaré is given with detailed proofs in (the Bourbaki-type pseudonym of the group of fifteen mathematicians who jointly produced this publication). Classification of connected Riemann surfaces Every Riemann surface is the quot
https://en.wikipedia.org/wiki/Switching%20circuit%20theory
Switching circuit theory is the mathematical study of the properties of networks of idealized switches. Such networks may be strictly combinational logic, in which their output state is only a function of the present state of their inputs; or may also contain sequential elements, where the present state depends on the present state and past states; in that sense, sequential circuits are said to include "memory" of past states. An important class of sequential circuits are state machines. Switching circuit theory is applicable to the design of telephone systems, computers, and similar systems. Switching circuit theory provided the mathematical foundations and tools for digital system design in almost all areas of modern technology. In an 1886 letter, Charles Sanders Peirce described how logical operations could be carried out by electrical switching circuits. During 1880–1881 he showed that NOR gates alone (or alternatively NAND gates alone) can be used to reproduce the functions of all the other logic gates, but this work remained unpublished until 1933. The first published proof was by Henry M. Sheffer in 1913, so the NAND logical operation is sometimes called Sheffer stroke; the logical NOR is sometimes called Peirce's arrow. Consequently, these gates are sometimes called universal logic gates. In 1898, Martin Boda described a switching theory for signalling block systems. Eventually, vacuum tubes replaced relays for logic operations. Lee De Forest's modification, in 1907, of the Fleming valve can be used as a logic gate. Ludwig Wittgenstein introduced a version of the 16-row truth table as proposition 5.101 of Tractatus Logico-Philosophicus (1921). Walther Bothe, inventor of the coincidence circuit, got part of the 1954 Nobel Prize in physics, for the first modern electronic AND gate in 1924. Konrad Zuse designed and built electromechanical logic gates for his computer Z1 (from 1935 to 1938). From 1934 to 1936, NEC engineer Akira Nakashima, Claude Shannon and
https://en.wikipedia.org/wiki/Improved%20Layer%202%20Protocol
IL2P (Improved Layer 2 Protocol) is a data link layer protocol originally derived from layer 2 of the X.25 protocol suite and designed for use by amateur radio operators. It is used exclusively on amateur packet radio networks. IL2P occupies the data link layer, the second layer of the OSI model. It is responsible for establishing link-layer connections, transferring data encapsulated in frames between nodes, and detecting errors introduced by the communications channel. The Improved Layer 2 Protocol (IL2P) was created by Nino Carrillo, KK4HEJ, based on AX.25 version 2.0 and implements Reed Solomon Forward Error Correction for greater accuracy and throughput than either AX.25 or FX.25. Specifically, in order to achieve greater stability on link speeds higher than 1200 baud. IL2P can be used with a variety of modulation methods including AFSK and GFSK. The direwolf software TNC contains the first open source implementation of the protocol. IL2P Specification The IL2P draft specification v0.5 was published via the Terrestrial Amateur Radio Packet Network (TARPN) on June 10, 2022. As of version 0.5, the Weak-Signal-Extensions were added which adds several features to the protocols, intended primarily for SSB links. The automatic ID transmission for the FM/1200-baud and faster modes is still in place, but is not enabled if these lower speed weak-signal SSB modes are selected. Implementations IL2P was first implemented in the closed source and proprietary ninoTNC to solve for lossy network links due to low Signal-to-noise ratio or weak signal strength. The specification itself outlines several design goals including: Forward error correction Eliminating bit-stuffing Streamlining the AX.25 header format Improved packet detection in the absence of Decode (DCD) and for open-squelch receive Produce a bitstream suitable for modulation on various physical layers Avoid bit-error-amplifying methods (differential encoding and free-running LFSRs) Increase efficie
https://en.wikipedia.org/wiki/Anton%20Kapustin
Anton Nikolayevich Kapustin (born November 10, 1971, Moscow) is a Russian-American theoretical physicist and the Earle C. Anthony Professor of Theoretical Physics at the California Institute of Technology. His interests lie in quantum field theory and string theory, and their applications to particle physics and condensed matter theory. He is the son of the pianist-composer Nikolai Kapustin. Education Kapustin obtained a B.S. in physics from Moscow State University in 1993. He received a Ph.D. in physics from the California Institute of Technology in 1997 with John Preskill as his advisor. Research He has made several contributions to dualities and other aspects of quantum field theories, in particular topological field theories and supersymmetric gauge theories. With Edward Witten he discovered deep connections between the S-duality of supersymmetric gauge theories and the geometric Langlands correspondence. In recent years, he has focused on mathematical structures in and classification schemes of topological field theories and symmetry-protected topological phases.
https://en.wikipedia.org/wiki/Rhabdosphaeraceae
Rhabdosphaeraceae is a family of algae belonging to the order Syracosphaerales. Genera The following is a list of Rhabdosphaeraceae genera: Acanthoica Lohmann, 1903 Algirosphaera Schlauder, 1945 Amitha Shafik, 1989 Anacanthoica G. Deflandre, 1952 Anthosphaera Blackites W.W. Hay & K.M. Towe, 1962 Cepekiella P.H. Roth, 1970 Cruxia O. Varol, 1989 Cyrtosphaera A. Kleijne, 1992 Discosphaera Haeckel, 1894 Discoturbella P.H. Roth, 1970 Naninfula K. Perch-Nielsen, 1968 Notiocyrtolithus Shafik, 1989 Ommatolithus Shafik, 1989 Palusphaera Lecal, 1966 Pseudorhabdosphaera Lecel & Bernheim, 1960 Pseudotriquetrorhabdulus S.W. Wise, 1976 Rhabdolithes O. Schmidt, 1870 Solisphaera J. Bollmann, M.Y. Cortés, A. Kleijne, J.B. Østerogaard, & J.R. Young, 2006
https://en.wikipedia.org/wiki/CLIP%20%28protein%29
CLIP or Class II-associated invariant chain peptide is the part of the invariant chain (Ii) that binds to the peptide binding groove of MHC class II and remains there until the MHC receptor is fully assembled. CLIP is one of the most prevalent self peptides found in the thymic cortex of most antigen-presenting cells. The purpose of CLIP is to prevent the degradation of MHC II dimers before antigenic peptides bind, and to prevent autoimmunity. During MHC II assembly in the endoplasmic reticulum, the invariant chain polypeptide complexes with MHC II heterodimers. In a late endosome/early lysosome, cathepsin S cleaves the invariant chain, leaving CLIP bound to the MHC II complex. In the presence of antigenic peptide fragments, HLA-DM partially binds to the MHC II peptide binding groove and acts as a catalyst, releasing CLIP and allowing peptides to bind. Antigenic peptides have a high affinity for the MHC II groove, and are readily exchanged for CLIP. This occurs in most cells expressing MHC II–however, in B cells, HLA-DO functions as the accessory protein. Both HLA-DM and HLA-DO interact with each other to act as chaperone proteins and prevent the denaturing of MHC II. MHC II with bound antigen is then transported to the plasma membrane for presentation. CLIP also can affect the differentiation of T cells. MHC II + CLIP complexes are upregulated on maturing dendritic cells, which activate and differentiate T cells into Thelper (Th) and Tcytotoxic (Tc) cells. Th cells can polarize into Th1 or Th2 effector cells depending on the presence of cytokines. High expression of CLIP favors the release of IL-4 and Th2 cell polarization. CLIP plays an important role in preventing autoimmunity. Since MHC is a polymorphic molecule, mutations that prevent CLIP from binding to MHC II leave the peptide binding groove empty. This could lead to the binding of other self-peptides and destruction of healthy cells. Autoimmune diseases such as rheumatoid arthritis, juvenile dermatomyos
https://en.wikipedia.org/wiki/Histone%20code
The histone code is a hypothesis that the transcription of genetic information encoded in DNA is in part regulated by chemical modifications (known as histone marks) to histone proteins, primarily on their unstructured ends. Together with similar modifications such as DNA methylation it is part of the epigenetic code. Histones associate with DNA to form nucleosomes, which themselves bundle to form chromatin fibers, which in turn make up the more familiar chromosome. Histones are globular proteins with a flexible N-terminus (taken to be the tail) that protrudes from the nucleosome. Many of the histone tail modifications correlate very well to chromatin structure and both histone modification state and chromatin structure correlate well to gene expression levels. The critical concept of the histone code hypothesis is that the histone modifications serve to recruit other proteins by specific recognition of the modified histone via protein domains specialized for such purposes, rather than through simply stabilizing or destabilizing the interaction between histone and the underlying DNA. These recruited proteins then act to alter chromatin structure actively or to promote transcription. For details of gene expression regulation by histone modifications see table below. The hypothesis The hypothesis is that chromatin-DNA interactions are guided by combinations of histone modifications. While it is accepted that modifications (such as methylation, acetylation, ADP-ribosylation, ubiquitination, citrullination, SUMO-ylation and phosphorylation) to histone tails alter chromatin structure, a complete understanding of the precise mechanisms by which these alterations to histone tails influence DNA-histone interactions remains elusive. However, some specific examples have been worked out in detail. For example, phosphorylation of serine residues 10 and 28 on histone H3 is a marker for chromosomal condensation. Similarly, the combination of phosphorylation of serine residu
https://en.wikipedia.org/wiki/Brotli
Brotli is a lossless data compression algorithm developed by Google. It uses a combination of the general-purpose LZ77 lossless compression algorithm, Huffman coding and 2nd-order context modelling. Brotli is primarily used by web servers and content delivery networks to compress HTTP content, making internet websites load faster. A successor to gzip, it is supported by all major web browsers and has become increasingly popular, as it provides better compression than gzip. History Google employees Jyrki Alakuijala and Zoltán Szabadka initially developed Brotli in 2013 to decrease the size of transmissions of WOFF web font. Alakuijala and Szabadka completed the Brotli specification during 20132016. The specification was accompanied with a reference implementation developed by two additional authors, Evgenii Kliuchnikov and Lode Vandevenne, who had previously developed Google's zopfli implementation of deflate and gzip compatible compression in 2013. Unlike zopfli, which was a reimplementation of an existing data format specification, Brotli was a new data format and allowed the authors to improve compression ratios even further. The Brotli specification was generalized in September 2015 for HTTP stream compression (content-encoding type "br"). This generalized iteration also improved the compression ratio by using a predefined dictionary of frequently used words and phrases. The version of Brotli released in September 2015 by the Google software engineers contained enhancements in generic lossless data compression, with particular emphasis on use for HTTP compression. The encoder was partly rewritten, with the result that the compression ratio improved, both the encoder and the decoder have been sped up, the streaming API was improved, and more compression quality levels have been added. Additionally, the new release shows performance improvements across platforms, with decoding memory reduction. The Internet Engineering Task Force approved the Brotli compresse
https://en.wikipedia.org/wiki/Alim%20Louis%20Benabid
Alim Louis Benabid is a French-Algerian emeritus professor, neurosurgeon and member of the French Academy of Sciences, who has had a global impact in the development of deep brain stimulation (DBS) for Parkinson's disease and other movement disorders. He became emeritus professor of biophysics at the Joseph Fourier University in Grenoble in September 2007, and chairman of the board of the Edmond J. Safra Biomedical Research Center in 2009 at Clinatec, a multidisciplinary institute he co-founded in Grenoble that applies nanotechnologies to neurosciences. Biography Alim Louis Benabid was born May 2, 1942, in Grenoble, France. The son of a doctor from Algeria and of a French nurse, Benabid was quoted as saying he could not easily decide between studying physics or medicine. He received his medical degree in 1970 and a doctorate in physics in 1978, both from Joseph Fourier University (now part of the Université Grenoble Alpes) in Grenoble. He became a staff neurosurgeon at Joseph Fourier University in 1972, professor of experimental medicine in 1978, and professor of biophysics from 1983 to 2007. Benabid also had a fellowship in 1979 – 1980 in preclinical neuropharmacology in the laboratory of Floyd Bloom at the Salk Institute in La Jolla, California. From 1988 to 2007, he directed the preclinical neurosciences unit at the French biomedical and public health research institution INSERM, and from 1989 to 2007, served as head of the neurosurgery department at the University Hospital of Grenoble. In other roles, Benabid coordinated the Claudio Munari Center for Surgery of Epilepsy and Movement Disorders at Ospedale (Hospital) Niguarda in Milan, Italy from 1998 to 2007, and was a staff consultant at the Cleveland Clinic Foundation in Ohio from 2000 to 2003. In 2007, Benabid joined the French Commissariat d'Energie Atomique as a scientific adviser during the time a campus for public-private innovation was being created, the Grenoble Innovation for Advanced New Technologie
https://en.wikipedia.org/wiki/Christos%20Papadimitriou
Christos Charilaos Papadimitriou (; born August 16, 1949) is a Greek theoretical computer scientist and the Donovan Family Professor of Computer Science at Columbia University. Education Papadimitriou studied at the National Technical University of Athens, where in 1972 he received his Bachelor of Arts degree in electrical engineering. He then pursued graduate studies at Princeton University, where he received his Ph.D. in electrical engineering and computer science in 1976 after completing a doctoral dissertation titled "The complexity of combinatorial optimization problems." Career Papadimitriou has taught at Harvard, MIT, the National Technical University of Athens, Stanford, UCSD, University of California, Berkeley and is currently the Donovan Family Professor of Computer Science at Columbia University. Papadimitriou co-authored a paper on pancake sorting with Bill Gates, then a Harvard undergraduate. Papadimitriou recalled "Two years later, I called to tell him our paper had been accepted to a fine math journal. He sounded eminently disinterested. He had moved to Albuquerque, New Mexico to run a small company writing code for microprocessors, of all things. I remember thinking: 'Such a brilliant kid. What a waste.'" The company was Microsoft. Papadimitriou co-authored "The Complexity of Computing a Nash Equilibrium" with his students Constantinos Daskalakis and Paul W. Goldberg, for which they received the 2008 Kalai Game Theory and Computer Science Prize from the Game Theory Society for "the best paper at the interface of game theory and computer science", in particular "for its key conceptual and technical contributions"; and the Outstanding Paper Prize from the Society for Industrial and Applied Mathematics. In 2001, Papadimitriou was inducted as a Fellow of the Association for Computing Machinery and in 2002 he was awarded the Knuth Prize. Also in 2002, he became a member of the U.S. National Academy of Engineering for contributions to complexity theor
https://en.wikipedia.org/wiki/Minecraft
{{Infobox video game | title = Minecraft | image = Minecraft cover.png | alt = The default player skin, Steve, running across a grassy plain while carrying an Iron pickaxe. Alongside him is a tame wolf. In the background, there is a pig, a chicken, a cow, a skeleton, a zombie, and a creeper. Mountains and cliffs fill the background, and the sky is blue, filled with clouds. Hovering over the scene is the Minecraft logo. | developer = Mojang Studios | publisher = | designer = | artist = | composer = | series = Minecraft | platforms = | released = {{Collapsible list | title = | Windows, macOS, Linux | | Android | | iOS | | Xbox 360 | | Raspberry Pi | | PlayStation 3 | | Fire OS | | PlayStation 4 | | Xbox One | | PlayStation Vita | | Windows Phone | | Windows 10 | | Wii U | | tvOS | | Nintendo Switch | | New Nintendo 3DS | | ChromeOS | }} | genre = | modes = | caption = Minecraft: Bedrock Edition box art | engine = Lightweight Java Game Library (Java)Render Dragon (Bedrock) }}Minecraft' is a sandbox game developed by Mojang Studios and originally released in 2009. The game was created by Markus "Notch" Persson in the Java programming language. Following several early private testing versions, it was first made public in May 2009 before being fully released in November 2011, with Notch stepping down and Jens "Jeb" Bergensten taking over development. Minecraft has become the best-selling video game in history, with over 300 million copies sold and nearly 140 million monthly active players . It has been ported to several platforms. In Minecraft, players explore a blocky, procedurally generated, three-dimensional world with virtually infinite terrain. Players can discover and extract raw materials, craft tools and items, and build structures, earthworks, and machines. Depending on their chosen game mode, players can fight hostile mobs, as well as cooperate with or compete against
https://en.wikipedia.org/wiki/Thermization
Thermization, also spelled thermisation, is a method of sanitizing raw milk with low heat. "Thermization is a generic description of a range of subpasteurization heat treatments (57 to 68°C × 10 to 20 s) that markedly reduce the number of spoilage bacteria in milk with minimal heat damage." The process is not used on other food products, and is similar to pasteurization but uses lower temperatures, allowing the milk product to retain more of its original taste. In Europe, there is a distinction between cheeses made of thermized milk and raw-milk cheeses. However, the United States' Food and Drug Administration (FDA) places the same regulations on all unpasteurized cheeses. As a result, cheeses from thermized milk must be aged for 60 days or more before being sold in the United States, the same restriction placed on raw-milk cheeses by the FDA. Thermization involves heating milk at temperatures of around for 15 seconds, while pasteurization involves heating milk at for 15 seconds or at for 30 minutes. Thermization is used to extend the keeping quality of raw milk (the length of time that milk is suitable for consumption) when it cannot be immediately used in other products, such as cheese. Thermization can also be used to extend the storage life of fermented milk products by inactivating microorganisms in the product. Thermization inactivates psychrotrophic bacteria in milk and allows the milk to be stored below for three days, or stored at for seven days. Later, the milk may be given stronger heat treatment to be preserved longer. Cooling thermized milk before reheating is necessary to delay/prevent the outgrowth of bacterial spores. When the milk is first heated, spores can begin to germinate, but their growth can be halted or delayed when the milk is refrigerated, depending on the microorganisms' growth requirements. Germinated spores are sensitive to subsequent heating, however since germination is not a homogeneous process, not all spores will germinate
https://en.wikipedia.org/wiki/Debra%20Roberts
Debra C. Roberts is a South African scientist and one of the six co-chairs of the Intergovernmental Panel on Climate Change. She was elected co-chair of Working Group II for the sixth assessment in 2015. She is head of the Sustainable and Resilient City Initiatives Unit in eThekwini Municipality (Durban, South Africa). Education and career Roberts has a PhD in Urban Biogeography at the (then) University of Natal, South Africa (1991). After working as a post-doctoral researcher, she joined local government in 1994. She established the Environmental Planning and Climate Protection Department of eThekwini Municipality (Durban, South Africa) which she led from 1994 to 2016. In 2016 she was appointed to establish the Sustainable and Resilient City Initiatives Unit in Durban and is the city’s first Chief Resilience Officer. She is a part time professor at the University of Twente. IPCC work She was a lead author of Chapter 8 (Urban Areas) of Working Group II of the IPCC Fifth Assessment Report and was elected as Co-Chair of Working Group II for the IPCC’s sixth assessment cycle in 2015. She was also a lead author of the IPCC Special Report on Global Warming of 1.5 °C. Recognition In 2014, the AfriCAN Climate Consortium gave Roberts their AfriCAN Climate Research Award. She was the 2016 Barbara Ward Lecturer of the International Institute for Environment and Development. She is an honorary professor of the University of KwaZulu-Natal, and was listed in 2019 by public servant organization Apolitical as one of the 100 most influential people in climate change in the world. She received honorary doctorates from the University of Twente in 2022 and from the University of Cape Town in 2023.
https://en.wikipedia.org/wiki/Programming%20Perl
Programming Perl, best known as the Camel Book among programmers, is a book about writing programs using the Perl programming language, revised as several editions (1991-2012) to reflect major language changes since Perl version 4. Editions have been co-written by the creator of Perl, Larry Wall, along with Randal L. Schwartz, then Tom Christiansen and then Jon Orwant. Published by O'Reilly Media, the book is considered the canonical reference work for Perl programmers. With over 1,000 pages, the various editions contain complete descriptions of each Perl language version and its interpreter. Examples range from trivial code snippets to the highly complex expressions for which Perl is widely known. The camel book editions are also noted for being written in an approachable and humorous style. History The first edition, which gained the nickname "the pink camel" due to its pink spine, was originally published in January 1991 and covered version 4 of the Perl language. It was the work of Larry Wall and Randal L. Schwartz. The second edition, published in August 1996, included updates for the release of Perl 5, among them references, objects, packages and other modern programming constructs. This edition was written from scratch by the original authors and Tom Christiansen. In July 2000, the third edition of Programming Perl was published. This version was again rewritten, this time by Wall, Christiansen and Jon Orwant, and covered the Perl 5.6 language. The fourth edition constitutes a major update and rewrite of the book for Perl version 5.14, and improves the coverage of Unicode usage in Perl. The fourth edition was published in February 2012. This edition is written by Tom Christiansen, brian d foy, Larry Wall and Jon Orwant. Programming Perl has also been made available electronically by O'Reilly, both through its inclusion in various editions of The Perl CD Bookshelf and through the "Safari" service (a subscription-based website containing technical ebooks). Th
https://en.wikipedia.org/wiki/New%20Zealand%20Dairy%20Workers%20Union
The New Zealand Dairy Workers Union (NZDWU) is a national trade union in New Zealand. It represents 7,000 workers active in dairy factories, town milk supply, processing plants, stores and warehousing, packing, can-making, and other ancillary activities including drivers. The NZDWU is affiliated with the New Zealand Council of Trade Unions, the IUF and the New Zealand Labour Party. External links NZDWU official site. Dairy Workers Union International Union of Food, Agricultural, Hotel, Restaurant, Catering, Tobacco and Allied Workers' Associations Dairy Workers Union Food processing trade unions
https://en.wikipedia.org/wiki/N-nucleotide
N-nucleotides, or nontemplated nucleotides are believed to exist only to create diversity at V(D)J junctions (see V(D)J recombination) during lymphocyte development. The addition of these nucleotides is aided by an enzyme called Terminal deoxynucleotidyl transferase (TdT)
https://en.wikipedia.org/wiki/International%20Journal%20of%20Applied%20Mathematics%20and%20Computer%20Science
The International Journal of Applied Mathematics and Computer Science is a peer-reviewed quarterly scientific journal published since 1991 by the University of Zielona Góra in partnership with De Gruyter Poland and Lubuskie Scientific Society, under the auspices of the Committee on Automatic Control and Robotics of the Polish Academy of Sciences. The editor-in-chief is Józef Korbicz. The journal covers various fields related to control theory, applied mathematics, scientific computing, and computer science. Indexing and abstracting The journal is abstracted and indexed, e.g., in: The full list of indexing services is available on the journal's website. According to the Journal Citation Reports, the journal has a 2021 impact factor of 2.157.
https://en.wikipedia.org/wiki/Intrasporangiaceae
Intrasporangiaceae is an actinomycete family. The family is named after the type genus Intrasporangium. The type species of Intrasporangium (I. calvum) was originally thought to form endospores; however, the mycelium of this strain may bear intercalary vesicles that were originally identified as spores. No members of Intrasporangiaceae are known to form spores. Phylogeny The currently accepted taxonomy is based on the List of Prokaryotic names with Standing in Nomenclature and the phylogeny is based on whole-genome sequences. Notes
https://en.wikipedia.org/wiki/Syllable%20%28computing%29
In computing, a syllable is a name for a platform-dependent unit of information storage. Depending on the target hardware, various bit widths (and sometimes internal groupings) are associated with it. Commonly used in the 1960s and 1970s, the term has mostly fallen into disuse in favour of terms like byte or word. Examples: 3-bit syllables: some experimental CISC designs 8-bit syllables: English Electric KDF9 (represented as syllabic octals and also called slob-octals or slobs in this context) and Burroughs large systems (see also: Burroughs B6x00-7x00 instruction set) 12-bit syllables: NCR computers such as the NCR 315 (also called slabs in this context) and Burroughs large systems 13-bit syllables: Saturn Launch Vehicle Digital Computer (LVDC) and Gemini Spacecraft On-Board Computer (OBC) See also Byte Catena (computing) Instruction syllable Nibble Opcode Opstring Parcel (computing) Syllable (in linguistics) Word (computer architecture)
https://en.wikipedia.org/wiki/Ribosome%20profiling
Ribosome profiling, or Ribo-Seq (also named ribosome footprinting), is an adaptation of a technique developed by Joan Steitz and Marilyn Kozak almost 50 years ago that Nicholas Ingolia and Jonathan Weissman adapted to work with next generation sequencing that uses specialized messenger RNA (mRNA) sequencing to determine which mRNAs are being actively translated. A related technique that can also be used to determine which mRNAs are being actively translated is the Translating Ribosome Affinity Purification (TRAP) methodology, which was developed by Nathaniel Heintz at Rockefeller University (in collaboration with Paul Greengard and Myriam Heiman). TRAP does not involve ribosome footprinting but provides cell type-specific information. Description It produces a “global snapshot” of all the ribosomes actively translating in a cell at a particular moment, known as a translatome. Consequently, this enables researchers to identify the location of translation start sites, the complement of translated ORFs in a cell or tissue, the distribution of ribosomes on a messenger RNA, and the speed of translating ribosomes. Ribosome profiling targets only mRNA sequences protected by the ribosome during the process of decoding by translation unlike RNA-Seq, which sequences all of the mRNA of a given sequence present in a sample. This technique is also different from polysome profiling. History Ribosome profiling is based on the discovery that the mRNA within a ribosome can be isolated through the use of nucleases that degrade unprotected mRNA regions. This technique analyzes the regions of mRNAs being converted to protein, as well as the levels of translation of each region to provide insight into global gene expression. Prior to its development, efforts to measure translation in vivo included microarray analysis on the RNA isolated from polysomes, as well as translational profiling through the affinity purification of epitope tagged ribosomes. These are useful and complementary m
https://en.wikipedia.org/wiki/Buchholz%27s%20ordinal
In mathematics, ψ0(Ωω), widely known as Buchholz's ordinal, is a large countable ordinal that is used to measure the proof-theoretic strength of some mathematical systems. In particular, it is the proof theoretic ordinal of the subsystem -CA0 of second-order arithmetic; this is one of the "big five" subsystems studied in reverse mathematics (Simpson 1999). It is also the proof-theoretic ordinal of , the theory of finitely iterated inductive definitions, and of , a fragment of Kripke-Platek set theory extended by an axiom stating every set is contained in an admissible set. Buchholz's ordinal is also the order type of the segment bounded by in Buchholz's ordinal notation . Lastly, it can be expressed as the limit of the sequence: , , , ... Definition , and for n > 0. is the closure of under addition and the function itself (the latter of which only for and ). is the smallest ordinal not in . Thus, ψ0(Ωω) is the smallest ordinal not in the closure of under addition and the function itself (the latter of which only for and ).
https://en.wikipedia.org/wiki/Algebraic%20operation
In mathematics, a basic algebraic operation is any one of the common operations of elementary algebra, which include addition, subtraction, multiplication, division, raising to a whole number power, and taking roots (fractional power). These operations may be performed on numbers, in which case they are often called arithmetic operations. They may also be performed, in a similar way, on variables, algebraic expressions, and more generally, on elements of algebraic structures, such as groups and fields. An algebraic operation may also be defined simply as a function from a Cartesian power of a set to the same set. The term algebraic operation may also be used for operations that may be defined by compounding basic algebraic operations, such as the dot product. In calculus and mathematical analysis, algebraic operation is also used for the operations that may be defined by purely algebraic methods. For example, exponentiation with an integer or rational exponent is an algebraic operation, but not the general exponentiation with a real or complex exponent. Also, the derivative is an operation that is not algebraic. Notation Multiplication symbols are usually omitted, and implied, when there is no operator between two variables or terms, or when a coefficient is used. For example, 3 × x2 is written as 3x2, and 2 × x × y is written as 2xy. Sometimes, multiplication symbols are replaced with either a dot or center-dot, so that x × y is written as either x . y or x · y. Plain text, programming languages, and calculators also use a single asterisk to represent the multiplication symbol, and it must be explicitly used; for example, 3x is written as 3 * x. Rather than using the ambiguous division sign (÷), division is usually represented with a vinculum, a horizontal line, as in . In plain text and programming languages, a slash (also called a solidus) is used, e.g. 3 / (x + 1). Exponents are usually formatted using superscripts, as in x2. In plain text, the TeX mark-up l
https://en.wikipedia.org/wiki/Danish%20Union%20of%20Slaughterhouse%20Workers
The Danish Union of Slaughterhouse Workers (, DSA) was a trade union representing workers in the meat industry in Denmark. The union was founded in 1895, and soon affiliated to the Danish Confederation of Trade Unions. By 1979, it had 22,902 members. The following year, it merger with the Danish Tobacco Workers' Union, the Bakery, Pastry and Mill Workers' Union, and the Confectionery and Chocolate Workers' Union, to form the Danish Food and Allied Workers' Union.
https://en.wikipedia.org/wiki/Carbapenem-resistant%20enterobacteriaceae
Carbapenem-resistant Enterobacteriaceae (CRE) or carbapenemase-producing Enterobacteriaceae (CPE) are Gram-negative bacteria that are resistant to the carbapenem class of antibiotics, considered the drugs of last resort for such infections. They are resistant because they produce an enzyme called a carbapenemase that disables the drug molecule. The resistance can vary from moderate to severe. Enterobacteriaceae are common commensals and infectious agents. Experts fear CRE as the new "superbug". The bacteria can kill up to half of patients who get bloodstream infections. Tom Frieden, former head of the Centers for Disease Control and Prevention has referred to CRE as "nightmare bacteria". Examples of enzymes found in certain types of CRE are KPC (Klebsiella pneumoniae carbapenemase) and NDM (New Delhi Metallo-beta-lactamase). KPC and NDM are enzymes that break down carbapenems and make them ineffective. Both of these enzymes, as well as the enzyme VIM (Verona Integron-Mediated Metallo-β-lactamase) have also been reported in Pseudomonas. Definition Carbapenem-resistant Enterobacteriaceae (CRE) have been defined as carbapenem-nonsusceptible and extended-spectrum cephalosporin-resistant Escherichia coli, Enterobacter aerogenes, Enterobacter cloacae complex, Klebsiella pneumoniae, or Klebsiella oxytoca. Some exclude ertapenem resistance from the definition. Risk factors Hospitals are primary transmission sites for CRE-based infections. Up to 75% of hospital admissions attributed to CRE were from long-term care facilities or transferred from another hospital. Suboptimal maintenance practices are the largest cause of CRE transmission. This includes the failure to adequately clean and disinfect medication cabinets, other surfaces in patient rooms, and portable medical equipment, such as X-ray and ultrasound machines that are used for both CRE and non-CRE patients. Thus far, CRE have primarily been nosocomial infectious agents. Almost all CRE infections occur in people r
https://en.wikipedia.org/wiki/Daina%20Taimi%C5%86a
Daina Taimiņa (born August 19, 1954) is a Latvian mathematician, retired adjunct associate professor of mathematics at Cornell University, known for developing a way of modeling hyperbolic geometry with crocheted objects. Education and career Taimiņa received all of her formal education in Riga, Latvia, where in 1977 she graduated summa cum laude from the University of Latvia and completed her graduate work in Theoretical Computer Science (with thesis advisor Prof. Rūsiņš Mārtiņš Freivalds) in 1990. As one of the restrictions of the Soviet system at that time, a doctoral thesis was not allowed to be defended in Latvia, so she defended hers in Minsk, receiving the title of Candidate of Sciences. This explains the fact that Taimiņa's doctorate was formally issued by the Institute of Mathematics of the National Academy of Sciences of Belarus. After Latvia regained independence in 1991, Taimiņa received her higher doctoral degree (doktor nauk) in mathematics from the University of Latvia, where she taught for 20 years. Daina Taimiņa joined the Cornell Math Department in December 1996. Combining her interests in mathematics and crocheting, she is one of 24 mathematicians and artists who make up the Mathemalchemy Team. Hyperbolic crochet While attending a geometry workshop at Cornell University about teaching geometry for university professors in 1997, Taimiņa was presented with a fragile paper model of a hyperbolic plane, made by the professor in charge of the workshop, David Henderson (designed by geometer William Thurston.) It was made «out of thin, circular strips of paper taped together». She decided to make more durable models, and did so by crocheting them. The first night after first seeing the paper model at the workshop she began experimenting with algorithms for a crocheting pattern, after visualising hyperbolic planes as exponential growth. The following fall, Taimiņa was scheduled to teach a geometry class at Cornell. She was determined to find what she
https://en.wikipedia.org/wiki/Zoological%20Society%20of%20Ireland
The Zoological Society of Ireland (ZSI) is the body responsible for running Dublin Zoo, where it is based, and Fota Wildlife Park in County Cork. It is the successor to the Royal Zoological Society of Ireland (RZSI), a learned society for the study of zoology. History The original ZSI was founded in Dublin on 10 May 1830 at a meeting in the Rotunda Hospital called and chaired by the Duke of Leinster, "to form a collection of living animals on the plan to the Zoological Society of London". Dublin Zoo opened in September 1831 in the Phoenix Park. The Society's first general meeting was held in November 1832. In 1838, the Zoo held an open day to mark the coronation of Queen Victoria and the Society was rewarded with the prefix "Royal" in its name. In October 1993, the members of the RZSI voted to dissolve the society and transfer its assets to a new non-profit limited company called "Zoological Society of Ireland Limited". This was to facilitate government and private funding arrangements for future development. Responsibility for government assistance was transferred at the same time from the Department of Education to the Office of Public Works.
https://en.wikipedia.org/wiki/Three%20prime%20untranslated%20region
In molecular genetics, the three prime untranslated region (3′-UTR) is the section of messenger RNA (mRNA) that immediately follows the translation termination codon. The 3′-UTR often contains regulatory regions that post-transcriptionally influence gene expression. During gene expression, an mRNA molecule is transcribed from the DNA sequence and is later translated into a protein. Several regions of the mRNA molecule are not translated into a protein including the 5' cap, 5' untranslated region, 3′ untranslated region and poly(A) tail. Regulatory regions within the 3′-untranslated region can influence polyadenylation, translation efficiency, localization, and stability of the mRNA. The 3′-UTR contains binding sites for both regulatory proteins and microRNAs (miRNAs). By binding to specific sites within the 3′-UTR, miRNAs can decrease gene expression of various mRNAs by either inhibiting translation or directly causing degradation of the transcript. The 3′-UTR also has silencer regions which bind to repressor proteins and will inhibit the expression of the mRNA. Many 3′-UTRs also contain AU-rich elements (AREs). Proteins bind AREs to affect the stability or decay rate of transcripts in a localized manner or affect translation initiation. Furthermore, the 3′-UTR contains the sequence AAUAAA that directs addition of several hundred adenine residues called the poly(A) tail to the end of the mRNA transcript. Poly(A) binding protein (PABP) binds to this tail, contributing to regulation of mRNA translation, stability, and export. For example, poly(A) tail bound PABP interacts with proteins associated with the 5' end of the transcript, causing a circularization of the mRNA that promotes translation. The 3′-UTR can also contain sequences that attract proteins to associate the mRNA with the cytoskeleton, transport it to or from the cell nucleus, or perform other types of localization. In addition to sequences within the 3′-UTR, the physical characteristics of the region,
https://en.wikipedia.org/wiki/43%2C112%2C609
43,112,609 (forty-three million, one hundred twelve thousand, six hundred nine) is the natural number following 43,112,608 and preceding 43,112,610. In mathematics 43,112,609 is a prime number. Moreover, it is the exponent of the 47th Mersenne prime, equal to M43,112,609 = 243,112,609 − 1, a prime number with 12,978,189 decimal digits. It was discovered on August 23, 2008 by Edson Smith, a volunteer of the Great Internet Mersenne Prime Search. The 45th Mersenne prime, M37,156,667 = 237,156,667 − 1, was discovered two weeks later on September 6, 2008, marking the shortest chronological gap between discoveries of Mersenne primes since the formation of the online collaborative project in 1996. It was the first time since 1963 when two Mersenne primes were discovered less than 30 days apart from each other. Less than a year later, on June 4, 2009, the 46th Mersenne prime, M42,643,801 = 242,643,801 − 1, was discovered by Odd Magnar Strindmo, a GIMPS participant from Norway. The result for this prime was first reported to the server in April 2009, but due to a bug, remained unnoticed for nearly two months. Having 12,837,064 decimal digits, it is only 141,125 digits, or 1.09%, shorter than M43,112,609. These two Mersenne primes hold the record for the ones with the smallest ratio between their exponents. 43,112,609 is the degree of four of the seven largest primitive binary trinomials over GF(2) found in 2016. and were the four largest in 2011. 43,112,609 is a Sophie Germain prime, the largest of only eight known Mersenne prime indexes to have this property. 43,112,609 is not a Gaussian prime, the largest of only 28 known Mersenne prime indexes to have this property.
https://en.wikipedia.org/wiki/KvLQT1
Kv7.1 (KvLQT1) is a potassium channel protein whose primary subunit in humans is encoded by the KCNQ1 gene. Kv7.1 is a voltage and lipid-gated potassium channel present in the cell membranes of cardiac tissue and in inner ear neurons among other tissues. In the cardiac cells, Kv7.1 mediates the IKs (or slow delayed rectifying K+) current that contributes to the repolarization of the cell, terminating the cardiac action potential and thereby the heart's contraction. It is a member of the KCNQ family of potassium channels. Structure KvLQT1 is made of six membrane-spanning domains S1-S6, two intracellular domains, and a pore loop. The KvLQT1 channel is made of four KCNQ1 subunits, which form the actual ion channel. Function This gene encodes a protein for a voltage-gated potassium channel required for the repolarization phase of the cardiac action potential. The gene product can form heteromultimers with two other potassium channel proteins, KCNE1 and KCNE3. The gene is located in a region of chromosome 11 that contains a large number of contiguous genes that are abnormally imprinted in cancer and the Beckwith-Wiedemann syndrome. Two alternative transcripts encoding distinct isoforms have been described. Clinical significance Mutations in the gene can lead to a defective protein and several forms of inherited arrhythmias as Long QT syndrome which is a prolongation of the QT interval of heart repolarization, Short QT syndrome, and Familial Atrial Fibrillation. KvLQT1 are also expressed in the pancreas, and KvLQT1 Long QT syndrome patients has been shown to have hyperinsulinemic hypoglycaemia following an oral glucose load. Currents arising from Kv7.1 in over-expression systems have never been recapitulated in native tissues - Kv7.1 is always found in native tissues with a modulatory subunit. In cardiac tissue, these subunits comprise KCNE1 and yotiao. Though physiologically irrelevant, homotetrameric Kv7.1 channels also display a unique form of C-type ina
https://en.wikipedia.org/wiki/Shrink%E2%80%93swell%20capacity
The shrink–swell capacity of soils refers to the extent certain clay minerals will expand when wet and retract when dry. Soil with a high shrink–swell capacity is problematic and is known as shrink–swell soil, or expansive soil. The amount of certain clay minerals that are present, such as montmorillonite and smectite, directly affects the shrink-swell capacity of soil. This ability to drastically change volume can cause damage to existing structures, such as cracks in foundations or the walls of swimming pools. Description Due to the physical and chemical properties of some clays (such as the Lias Group) large swelling occurs when water is absorbed. Conversely when the water dries up these clays contract (shrink). The presence of these clay minerals is what allows soils to have the capacity to shrink and swell. Some of these clay minerals are: smectite, nontronite, bentonite, chlorite, montmorillonite, beidellite, attapulgite, illite and vermiculite. The amount of these minerals in a particular soil will also determine the severity of the shrink-swell capacity. For instance, soils with a small amount of expansive clay minerals will not expand as much when exposed to moisture as a soil with a large amount of the same clay minerals. If a soil is composed of at least 5 percent of these clay minerals by weight, it could have the ability to shrink and swell. This property is measured using coefficient of linear extensibility (COLE) values. If a soil has a COLE value greater than 0.06, then it can cause structural damage. A COLE value of 0.06 means that 100 inches of soil will expand by 6 inches when wet. Soils with this shrink-swell capacity fall under the soil order of Vertisols. As these soils dry, deep cracks can form on the surface, which then allows water to penetrate to deeper levels of the soil. This can cause the swelling of these soils to become cyclical, with periods of both shrinking and swelling. Damage Clay groups with a high shrink–swell capacity ten