source stringlengths 31 227 | text stringlengths 9 2k |
|---|---|
https://en.wikipedia.org/wiki/National%20Poisons%20Information%20Service | The National Poisons Information Service is an information service commissioned by Public Health England on behalf of the UK health departments. Poisoning accounts for 1% of NHS admissions. Pesticides used in agriculture (particularly organophosphorus insecticides) are extremely toxic, but 87% of around 120,000 annual poisoning cases in the UK take place in the home.
History
In August 1962 the Ministry of Health announced it was forming a poisons information service. This was after the Emergency Treatment in Hospital of Cases of Acute Poisoning published by the Central Health Services Council in March 1962. Many more household chemicals were on the market, and the chemical composition was only known to the manufacturers. 4,000 to 5,000 people each year were lethally poisoned, with 6,085 in 1962; however, many of the deaths were (non-accidental) suicides.
It started in 1963 by Dr Roy Goulding at the Medical Toxicology Unit of Guy's Hospital, with a staff of 65. By the late 1960s, recreational drugs were presenting a widespread danger. Other centres were soon set up in Edinburgh, Belfast and Cardiff.
UKTIS was based in Newcastle from 1995.
Structure
It has four units in Birmingham, Cardiff, Edinburgh and Newcastle. It is a member of the European Association of Poisons Centres and Clinical Toxicologists.
NPIS Birmingham Unit (West Midlands Poisons Unit), City Hospital, Birmingham - Dr Sally Bradberry
NPIS Cardiff Unit, Llandough Hospital - Dr Laurence Gray
NPIS Edinburgh Unit (Scottish Poisons Information Bureau), Royal Infirmary of Edinburgh - Dr. Euan Sandilands.
NPIS Newcastle Unit, Regional Drug and Therapeutics Centre, Newcastle University - Professor Simon Thomas
NPIS Birmingham has the NPIS Product Data Centre, provided by manufacturer's information, and deals with around 2,300 companies.
NPIS supports (24 hours) NHS Direct (for England and Wales) and (24 hours) NHS 24 for Scotland. Northern Ireland has its Northern Ireland Regional Medicines and Pois |
https://en.wikipedia.org/wiki/Temporal%20feedback | Within molecular and cell biology, Temporal feedback, also referred to as interlinked or interlocked feedback, is a biological regulatory motif in which fast and slow positive feedback loops are interlinked to create "all or none" switches. This interlinking produces separate, adjustable activation and de-activation times. This type of feedback is thought to be important in cellular processes in which an "all or none" decision is a necessary response to a specific input. The mitotic trigger, polarization in budding yeast, mammalian calcium signal transduction, EGF receptor signaling, platelet activation, and Xenopus oocyte maturation are examples for interlinked fast and slow multiple positive feedback systems.
In biological systems, temporal feedback is a ubiquitous signal transduction motif that allows systems to convert graded inputs into decisive, all-or-none digital outputs. A system with interlinked fast and slow feedback loops produces a dual-time switch, which is rapidly inducible and robust to noise during stimulus. In contrast, a single fast or slow loop is separately responsible for the speed of switching and the stability of switches. Computer simulation studies have shown that linking two loops of the same kind brings no overall advantage over having a single loop, however the dual-loop switch performs in a monostable regime. Both single and dual loops can behave as a bistable switch. Several computational models have been produced to demonstrate the responses of single and dual positive feedback loop switches to stimuli.
Biological examples
The transcription factor NF-κB regulates various genes that play essential roles in signaling, stress responses, cell growth and apoptosis. The temporal control of NF-κB activation by the degradation and synthesis of its inhibitor isoforms, I-κBα, -β, - ε has been computationally modeled. The model suggested that I-κBα results in robust negative feedback that leads to a fast turn off of NF-κB response. On the ot |
https://en.wikipedia.org/wiki/Dual%20wavelet | In mathematics, a dual wavelet is the dual to a wavelet. In general, the wavelet series generated by a square-integrable function will have a dual series, in the sense of the Riesz representation theorem. However, the dual series is not itself in general representable by a square-integrable function.
Definition
Given a square-integrable function , define the series by
for integers .
Such a function is called an R-function if the linear span of is dense in , and if there exist positive constants A, B with such that
for all bi-infinite square summable series . Here, denotes the square-sum norm:
and denotes the usual norm on :
By the Riesz representation theorem, there exists a unique dual basis such that
where is the Kronecker delta and is the usual inner product on . Indeed, there exists a unique series representation for a square-integrable function f expressed in this basis:
If there exists a function such that
then is called the dual wavelet or the wavelet dual to ψ. In general, for some given R-function ψ, the dual will not exist. In the special case of , the wavelet is said to be an orthogonal wavelet.
An example of an R-function without a dual is easy to construct. Let be an orthogonal wavelet. Then define for some complex number z. It is straightforward to show that this ψ does not have a wavelet dual.
See also
Multiresolution analysis |
https://en.wikipedia.org/wiki/Hydraulic%20fluid | A hydraulic fluid or hydraulic liquid is the medium by which power is transferred in hydraulic machinery. Common hydraulic fluids are based on mineral oil or water. Examples of equipment that might use hydraulic fluids are excavators and backhoes, hydraulic brakes, power steering systems, automatic transmissions, garbage trucks, aircraft flight control systems, lifts, and industrial machinery.
Hydraulic systems like the ones mentioned above will work most efficiently if the hydraulic fluid used has zero compressibility.
Functions and properties
The primary function of a hydraulic fluid is to convey power. In use, however, there are other important functions of hydraulic fluid such as protection of the hydraulic machine components. The table below lists the major functions of a hydraulic fluid and the properties of a fluid that affect its ability to perform that function:
Composition
Base stock
The original hydraulics fluid, dating back to the time of ancient Egypt, was water. Beginning in the 1920s, mineral oil began to be used more than water as a base stock due to its inherent lubrication properties and ability to be used at temperatures above the boiling point of water. Today most hydraulic fluids are based on mineral oil base stocks.
Natural oils such as rapeseed are used as base stocks for fluids where biodegradability and renewable sources are considered important.
Other base stocks are used for specialty applications, such as for fire resistance and extreme temperature applications. Some examples include: glycol ethers, organophosphate ester, polyalphaolefin, propylene glycol, and silicone oils.
NaK-77, a eutectic alloy of sodium and potassium, can be used as a hydraulic fluid in high-temperature and high-radiation environments, for temperature ranges of 10 to 1400 °F (-12 to 760 °C). Its bulk modulus at 1000 °F (538 °C) is 310,000 psi (2.14 GPa), higher than of a hydraulic oil at room temperature. Its lubricity is poor, so positive-displacement pum |
https://en.wikipedia.org/wiki/Lateralization%20of%20bird%20song | Passerine birds produce song through the vocal organ, the syrinx, which is composed of bilaterally symmetric halves located where the trachea separates into the two bronchi. Using endoscopic techniques, it has been observed that song is produced by air passing between a set of medial and lateral labia on each side of the syrinx. Song is produced bilaterally, in both halves, through each separate set of labia unless air is prevented from flowing through one side of the syrinx. Birds regulate the airflow through the syrinx with muscles—M. syringealis dorsalis and M. tracheobronchialis dorsalis—that control the medial and lateral labia in the syrinx, whose action may close off airflow. Song may, hence, be produced unilaterally through one side of the syrinx when the labia are closed in the opposite side.
Early experiments discover lateralization
Lateral dominance of the hypoglossal nerve conveying messages from the brain to the syrinx was first observed in the 1970s. This lateral dominance was determined in a breed of canary, the waterschlager canary, bred for its long and complex song, by lesioning the ipsilateral tracheosyringeal branch of the hypoglossal nerve, disabling either the left or right syrinx. The numbers of song elements in the birds’ repertoires were greatly attenuated when the left side was cut, but only modestly attenuated when the right side was disabled, indicating left syringeal dominance of song production in these canaries. Similar lateralized effects have been observed in other species such as the white-crowned sparrow (Zonotrichia leucophrys), the Java sparrow (Lonchura oryzivora) and the zebra finch (Taeniopygia guttata), which is right-side dominant. However, denervation in these birds does not entirely silence the affected syllables but creates qualitative changes in phonology and frequency.
Respiratory control and neurophysiology
In waterslager canaries, which produce most syllables using the left syrinx, as soon as a unilaterally produced |
https://en.wikipedia.org/wiki/KCNJ6 | G protein-activated inward rectifier potassium channel 2 is a protein that in humans is encoded by the KCNJ6 gene. Mutation in KCNJ6 gene has been proposed to be the cause of Keppen-Lubinsky Syndrome (KPLBS).
Function
Potassium channels are present in most mammalian cells, where they participate in a wide range of physiologic responses. The protein encoded by this gene is an integral membrane protein and inward-rectifier type potassium channel. The encoded protein, which has a greater tendency to allow potassium to flow into a cell rather than out of a cell, is controlled by G-proteins and may be involved in the regulation of insulin secretion by glucose. It associates with two other G-protein-activated potassium channels to form a heteromultimeric pore-forming complex.
Interactions
KCNJ6 has been shown to interact with KCNJ9 and DLG1.
See also
G protein-coupled inwardly-rectifying potassium channel
Inward-rectifier potassium ion channel |
https://en.wikipedia.org/wiki/Centre%20for%20Human%20Ecology | The Centre for Human Ecology is an independent academic institute based in Glasgow, Scotland. It was founded in 1972 by Conrad Hal Waddington at the University of Edinburgh. |
https://en.wikipedia.org/wiki/Mucous%20membrane%20pemphigoid | Mucous membrane pemphigoid is a rare chronic autoimmune subepithelial blistering disease characterized by erosive lesions of the mucous membranes and skin. It is one of the pemphigoid diseases that can result in scarring.
Signs and symptoms
The autoimmune reaction most commonly affects the oral mucosa in the mouth, causing lesions in the gums (gingiva), known as desquamative gingivitis. More severe cases can also affect areas of mucous membrane elsewhere in the body, such as the sinuses, genitals, anus, and cornea. When the cornea of the eye is affected, repeated scarring may result in blindness.
Brunsting–Perry cicatricial pemphigoid is a rare variant of mucous membrane pemphigoid involving the scalp and the neck without mucosal involvement. It is proposed by some authors that this be called a variant of epidermolysis bullosa acquisita.
Nikolsky's sign (gentle lateral pressure) on unaffected mucosa or skin raises a bulla. If no lesions are present on examination it may be useful way of demonstrating reduced epithelial adhesion. In contrast, in Pemphigus, the epithelium tends to disintegrate rather than form a bulla.
Nikolsky's sign is present in pemphigus and mucous membrane pemphigoid, but not in bullous pemphigoid.
Pathophysiology
In mucous membrane pemphigoid, the autoimmune reaction occurs in the skin, specifically at the level of the basement membrane, which connects the lower skin layer (dermis) to the upper skin layer (epidermis) and keeps it attached to the body.
When the condition is active, the basement membrane is dissolved by the antibodies produced, and areas of skin lift away at the base, causing hard blisters which scar if they burst. In other words, this is a desquamating/blistering disease in which the epithelium "unzips" from the underlying connective tissue, allowing fluid to gather that subsequently manifest as bullae, or blisters.
Diagnosis
Diagnostic techniques:
antibodies (IgG) precipitate complement (C3) in the lamina lucida of the |
https://en.wikipedia.org/wiki/ETP-1 | ETP-1 (or Electronic Test Pattern One) was a test card designed and used by the Independent Broadcasting Authority (IBA). After test transmissions from the IBA's Engineering Regional Operations Centre (ROC) in Croydon from 1978 it was phased in on ITV over a period starting from 1979, replacing, in different ITV regions: Test Card F, Test Card G and full screen height EBU colour bars. After ITV went 24 hours in 1988, the card ceased to be seen on the channel. It was used for both 625-line PAL and 405-line monochrome broadcasts.
It was also seen extensively on both Channel 4 and S4C during both their pre-launch tests and their downtime of those channels, due to their limited hours early on. After launch, it was alternated with in-vision teletext services from ORACLE and 4-Tel on View/Sbectel. On these channels, it was captioned either IBA:CH4 or IBA:S4C, with lines above and below this indicating the card was being generated by the channel, the absence of these lines meant it was generated at the transmitter.
After the splitting up of the IBA in 1990, the captioning was changed to NTL:CH4/NTL:S4C. ETP-1 was last broadcast on 31 December 1992 after which Channel 4 showed 4-Tel on View during closedown periods as the channel gradually increased its late night programming, eventually going 24 hours a day early in 1997. After the conclusion of its late-night teleshopping block, S4C shows a closedown slide promoting its S4C Clic VOD service with library music playing in the background during off-air hours.
On 18 April 2012, ETP-1 made one final appearance on Channel 4 to announce the loss of analogue television services in the London region. The card was shown from the Crystal Palace transmitter only, and was the last thing broadcast by analogue Channel 4 before the signal was switched off.
Operation
The test card was generated electronically by a video generator, and includes features for various adjustments and checks:
Crosshatch pattern - For CRT color convergence |
https://en.wikipedia.org/wiki/Proceedings%20of%20the%20Combustion%20Institute | The Proceedings of the Combustion Institute are the proceedings of the biennial Combustion Symposium put on by The Combustion Institute. The publication contains the most significant contributions in fundamentals and applications fundamental research of combustion science combustion phenomena. Research papers and invited topical reviews are included on topics of reaction kinetics, soot, PAH and other large molecules, diagnostics, laminar flames, turbulent flames, heterogenous combustion, spray and droplet combustion, detonations, explosions & supersonic combustion, fire research, stationary combustion systems, internal combustion engine and gas turbine combustion, and new technology concepts. The editors-in-chief are Daniel C. Haworth (Pennsylvania State University) and Terese Løvås (no) (Norwegian University of Science and Technology).
History
The need for development of automotive engines, fuels, and aviation formed the basis for the organization which became The Combustion Institute. The first three symposiums were held in 1928, 1937, and 1948. Since 1952, symposiums have been held every second year. The first combustion symposium with published proceedings was in 1948.
Abstracting and indexing
The journal is abstracted and indexed in:
According to the Journal Citation Reports, the journal has a 2015 impact factor of 4.120.
See also |
https://en.wikipedia.org/wiki/High-efficiency%20hybrid%20cycle | The high-efficiency hybrid cycle (HEHC) is a new 4-stroke thermodynamic cycle combining elements of the Otto cycle, Diesel cycle, Atkinson cycle and Rankine cycle.
HEHC engines
The 3rd generation design of the Liquidpiston Engine currently in development is the only engine currently designed around the HEHC. It is a rotary combustion engine. |
https://en.wikipedia.org/wiki/RV%20Gloria%20Michelle | RV Gloria Michelle (F7201), sometimes rendered as R/V Gloria Michelle, is an American fisheries research vessel in non-commissioned service in the National Oceanic and Atmospheric Administration (NOAA) fleet since 1980. Prior to her NOAA career, she was a commercial shrimp boat that came into the possession of the United States Government after her seizure for carrying marijuana.
Registered as NOAA F7201, Gloria Michelle conducts operations in coastal waters along the northeastern coast of North America.
Construction and acquisition
The Diesel Shipbuilding Company built Gloria Michelle at Jacksonville, Florida, in 1974 as a steel-hulled, diesel-powered commercial shrimp boat for use in the Gulf of Mexico. In 1979, the United States Customs Service seized Gloria Michelle after discovering 16 short tons (14.25 long tons; 14.5 metric tons/tonnes) of marijuana in her fish hold, at the time the largest seizure of illegal drugs in the history of Mississippi. After the conclusion of the trial of the people involved in smuggling the marijuana, Gloria Michelle was laid up in a bayou near Biloxi, Mississippi, to await her fate.
At about the same time, the Conservation Engineering Group at the Northeast Fisheries Science Center (NEFSC) at NOAA's National Marine Fisheries Service Gloucester Laboratory in Gloucester, Massachusetts, was looking for a replacement for RV Rorqual, a 40-year-old former United States Army harbor tug the NEFSC used to support its fisheries science activities and its work designing and testing commercial fishing gear. Finding Gloria Michelle well-suited for these endeavors, NOAA acquired her and assigned her to the Conservation Engineering Group at Gloucester. Too small to meet NOAA's criteria for entering commissioned service because she is under in length, Gloria Michelle instead entered non-commissioned service with NOAA as RV Gloria Michelle in 1980 after undergoing extensive modifications and upgrades necessary for her to operate as a researc |
https://en.wikipedia.org/wiki/Reverse-delete%20algorithm | The reverse-delete algorithm is an algorithm in graph theory used to obtain a minimum spanning tree from a given connected, edge-weighted graph. It first appeared in , but it should not be confused with Kruskal's algorithm which appears in the same paper. If the graph is disconnected, this algorithm will find a minimum spanning tree for each disconnected part of the graph. The set of these minimum spanning trees is called a minimum spanning forest, which contains every vertex in the graph.
This algorithm is a greedy algorithm, choosing the best choice given any situation. It is the reverse of Kruskal's algorithm, which is another greedy algorithm to find a minimum spanning tree. Kruskal’s algorithm starts with an empty graph and adds edges while the Reverse-Delete algorithm starts with the original graph and deletes edges from it. The algorithm works as follows:
Start with graph G, which contains a list of edges E.
Go through E in decreasing order of edge weights.
For each edge, check if deleting the edge will further disconnect the graph.
Perform any deletion that does not lead to additional disconnection.
Pseudocode
function ReverseDelete(edges[] E) is
sort E in decreasing order
Define an index i ← 0
while i < size(E) do
Define edge ← E[i]
delete E[i]
if graph is not connected then
E[i] ← edge
i ← i + 1
return edges[] E
In the above the graph is the set of edges E with each edge containing a weight and connected vertices v1 and v2.
Example
In the following example green edges are being evaluated by the algorithm and red edges have been deleted.
Running time
The algorithm can be shown to run in O(E log V (log log V)3) time (using big-O notation), where E is the number of edges and V is the number of vertices. This bound is achieved as follows:
Sorting the edges by weight using a comparison sort takes O(E log E) time, which can be simplified to O(E log V) using the fac |
https://en.wikipedia.org/wiki/Keystone%20%28architecture%29 | A keystone (or capstone) is the wedge-shaped stone at the apex of a masonry arch or typically round-shaped one at the apex of a vault. In both cases it is the final piece placed during construction and locks all the stones into position, allowing the arch or vault to bear weight. In arches and vaults (such as quasi-domes) keystones are often enlarged beyond the structural requirements and decorated. A variant in domes and crowning vaults is a lantern.
Keystones, as a hallmark of strength or good architecture, or their suggested form are sometimes placed in the centre of the flat top of doors, recesses and windows for decorative effect, so as to form an upward projection of a lintel.
Although a masonry arch or vault cannot be self-supporting until the keystone is placed, the keystone experiences the least stress of any of the voussoirs, due to its position at the apex. Old keystones can decay due to vibration, a condition known as bald arch.
Architecture
In a rib-vaulted ceiling, keystones commonly mark the intersections of any two or more arched ribs. For aesthetics, keystones are often larger than ribs in vaults and many of the voussoirs (arch stones) in arches, or embellished with a boss.
A "dropped keystone" is one where the keystone projects lower than the other voussoirs. Following Giulio Romano, Mannerist architects of the 16th century often designed arches with enlarged and slightly dropped keystones, as in the "church house" entrance portal at Colditz Castle. Numerous examples are found in the work of Sebastiano Serlio, a 16th-century Italian Mannerist architect.
Metaphor
Keystone is often used metaphorically for an essential part on which the whole depends or as an acme of the whole.
The U.S. state of Pennsylvania calls itself the "Keystone State", because during early American history, it held a crucial central position among the Thirteen Colonies geographically, economically, and politically, like the keystone in an arch. A keystone is a symbol |
https://en.wikipedia.org/wiki/Toad%20%28software%29 | Toad is a database management toolset from Quest Software for managing relational and non-relational databases using SQL aimed at database developers, database administrators, and data analysts. The Toad toolset runs against Oracle, SQL Server, IBM DB2 (LUW & z/OS), SAP and MySQL. A Toad product for data preparation supports many data platforms.
History
A practicing Oracle DBA, Jim McDaniel, designed Toad for his own use in the mid-1990s. He called it Tool for Oracle Application Developers, shortened to "TOAD". McDaniel initially distributed the tool as shareware and later online as freeware.
Quest Software acquired TOAD in October 1998. Quest Software itself was acquired by Dell in 2012 to form Dell Software. In June 2016, Dell announced the sale of their software division, including the Quest business, to Francisco Partners and Elliott Management Corporation. On October 31, 2016, the sale was finalized. On November 1, 2016, the sale of Dell Software to Francisco Partners and Elliott Management was completed, and the company re-launched as Quest Software.
Features
Connection Manager - Allow users to connect natively to the vendor’s database whether on-premise or DBaaS.
Browser - Allow users to browse all the different database/schema objects and their properties effective management.
Editor - A way to create and maintain scripts and database code with debugging and integration with source control.
Unit Testing (Oracle) - Ensures code is functionally tested before it is released into production.
Static code review (Oracle) - Ensures code meets required quality level using a rules-based system.
SQL Optimization - Provides developers with a way to tune and optimize SQL statements and database code without relying on a DBA. Advanced optimization enables DBAs to tune SQL effectively in production.
Scalability testing and database workload replay - Ensures that database code and SQL will scale properly before it gets released into production.
Books
Toad Po |
https://en.wikipedia.org/wiki/Stopped-flow | Stopped-flow is an experimental technique for studying chemical reactions with a half time of the order of 1 ms, introduced by Britton Chance and extended by Quentin Gibson (Other techniques, such as the temperature-jump method, are available for much faster processes.)
Description of the method
Summary
Stopped-flow spectrometry allows chemical kinetics of fast reactions (with half times of the order of milliseconds) to be studied in solution. It was first used primarily to study enzyme-catalyzed reactions. Then the stopped-flow rapidly found its place in almost all biochemistry, biophysics, and chemistry laboratories with a need to follow chemical reactions in the millisecond time scale.
In its simplest form, a stopped-flow mixes two solutions. Small volumes of solutions are rapidly and continuously driven into a high-efficiency mixer. This mixing process then initiates an extremely fast reaction. The newly mixed solution travels to the observation cell and pushes out the contents of the cell (the solution remaining from the previous experiment or from necessary washing steps). The time required for this solution to pass from the mixing point to the observation point is known as dead time. The minimum injection volume will depend on the volume of the mixing cell. Once enough solution has been injected to completely remove the previous solution, the instrument reaches a stationary state and the flow can be stopped. Depending on the syringe drive technology, the flow stop is achieved by using a stop valve called the hard-stop or by using a stop syringe. The stopped-flow also sends a ‘start signal’ to the detector called the trigger so the reaction can be observed. The timing of the trigger is usually software controlled so the user can trigger at the same time the flow stops or a few milliseconds before the stop to check the stationary state has been reached.
So this is a very economical technique.
Reactant syringes
Two syringes are filled with solutions that |
https://en.wikipedia.org/wiki/List%20of%20secondary%20school%20sports%20team%20names%20and%20mascots%20derived%20from%20Indigenous%20peoples | Among the categories of names for sports teams in the United States and Canada, those referring to Indigenous peoples are lesser in popularity only to the names of various animals. In a list of the top 100 team names, "Indians" is 14th, "Braves" is 38th, "Chiefs" is 57th. The typical logo is an image of a stereotypical Native American man in profile, wearing a Plains Indians headdress; and are often cartoons or caricatures. Other imagery include dreamcatchers, feathers, spears, and arrows. Individual schools may have performance traditions, such as the tomahawk chop, a mascot or cheerleaders in stereotypical Native attire, and chants adapted from Hollywood movies. These fictional representations stand in the way of any authentic understanding of contemporary Indigenous peoples, and promote racism.
The documents often cited to justifying the trend for change are an advisory opinion by the United States Commission on Civil Rights in 2001 and a resolution by the American Psychological Association in 2005. Both support the views of Native American organizations and individuals that such mascots maintain harmful stereotypes that are discriminatory and cause harm by distorting the past and preventing understanding of Native American/First Nations peoples in the present.
The trend towards the elimination of Indigenous names and mascots in local schools has been steady, with two-thirds having been eliminated during the 50 years prior to 2013 according to the National Congress of American Indians (NCAI). In more recent years, the trend has accelerated, particularly in July 2020, following a wave of racial awareness and reforms in wake of national protests after the murder of George Floyd, and the decision by the Washington Commanders to change their Redskins name and logo.
In a few states with significant Native American populations; such as Colorado, Maine, Nevada, Oregon,
Washington, and Wisconsin, change has been mandated by law. A law was passed in Connecticut whic |
https://en.wikipedia.org/wiki/Price%20of%20stability | In game theory, the price of stability (PoS) of a game is the ratio between the best objective function value of one of its equilibria and that of an optimal outcome. The PoS is relevant for games in which there is some objective authority that can influence the players a bit, and maybe help them converge to a good Nash equilibrium. When measuring how efficient a Nash equilibrium is in a specific game we often also talk about the price of anarchy (PoA), which is the ratio between the worst objective function value of one of its equilibria and that of an optimal outcome.
Examples
Another way of expressing PoS is:
In particular, if the optimal solution is a Nash equilibrium, then the PoS is 1.
In the following prisoner’s dilemma game, since there is a single equilibrium we have PoS = PoA = 1/2.
On this example which is a version of the battle of sexes game, there are two equilibrium points, and , with values 3 and 15, respectively. The optimal value is 15. Thus, PoS = 1 while PoA = 1/5.
Background and milestones
The price of stability was first studied by A. Schulz and N. Stier-Moses while the term was coined by E. Anshelevich et al. Schulz and Stier-Moses focused on equilibria in a selfish routing game in which edges have capacities. Anshelevich et al. studied network design games and showed that a pure strategy Nash equilibrium always exists and the price of stability of this game is at most the nth harmonic number in directed graphs. For undirected graphs Anshelevich and others presented a tight bound on the price of stability of 4/3 for a single source and two players case. Jian Li has proved that for undirected graphs with a distinguished destination to which all players must connect the price of stability of the Shapely network design game is where is the number of players. On the other hand, the price of anarchy is about in this game.
Network design games
Setup
Network design games have a very natural motivation for the Price of Stability.
In the |
https://en.wikipedia.org/wiki/Institut%20Jacques%20Monod | The Institut Jacques Monod, funded jointly by the CNRS and the University Paris Diderot, is one of the main centres for basic research in biology in Paris, France. It is headed by Valerie Doye.
There are 3 broad research topics (Genome and chromosome dynamics, Cellular dynamics and signalling, Development and evolution) and 2 main transverse axes (Quantitative biology and modelling, Molecular and cellular pathologies). Research at the interface of biology with physics, mathematics, chemistry and medicine is strongly encouraged.
Some 300 people work at the Institute (tenured investigators, Ph.D. students, post-docs, technicians, engineers, French and foreign visitors, and administrative staff).
History
In the early 60s, Jacques Monod was entrusted with the creation of an institute of molecular biology located on a university campus. The original goal was to create a major research centre that would combine teaching and research.
The “Institut de Biologie Moléculaire” (IBM) thus saw the day in 1966, and Raymond Dedonder was appointed as its first director. The steering committee comprised, in particular, Jacques Monod and François Jacob, who were to receive the Nobel Prize for Medicine in 1965, with their colleague Andre Lwoff from the Institut Pasteur. This committee set the scientific guidelines that were to be developed around the main theme of biochemistry of heredity: the study of replication, protein biosynthesis, the mechanisms and control of transcription and translation in bacteria and in cells of higher organisms, the mechanisms of differentiation, the study of conformation and conformational changes in biological macromolecules, the study of the functional associations of macromolecules and the conformational problems that are posed by these associations.
IBM’s premises (8500 m2) in the Faculty of Sciences of Paris, on the site of the "Halle aux Vins", were formally occupied in December 1969. During the following year, 13 laboratories moved in, and ad |
https://en.wikipedia.org/wiki/Design-Oriented%20Programming | Design-oriented programming is a way to author computer applications using a combination of text, graphics, and style elements in a unified code-space. The goal is to improve the experience of program writing for software developers, boost accessibility, and reduce eye-strain. Good design helps computer programmers to quickly locate sections of code using visual cues typically found in documents and web page authoring.
User interface design and graphical user interface builder research are the conceptual precursors to design-oriented programming languages. The former focus on the software experience for end users of the software application and separate editing of the user interface from the code-space. The important distinction is that design-oriented programming involves user experience of programmers themselves and fully merges all elements into a single unified code-space.
See also
User interface design
Graphical user interface builder
Elements of graphical user interfaces
Visual programming language
Experience design
User experience design
Usability |
https://en.wikipedia.org/wiki/Com21 | Com21 was an early pioneer in developing cable modem networks in the era before the standard DOCSIS was introduced for Internet access via cable television networks. The company filed for bankruptcy in 2003.
The company
Com21 Inc. was an American public company. Com21 shares traded on Nasdaq using CMTO as symbol. The headquarters of the company was located in Milpitas, California. Com21 had offices in 13 countries. The European head office was located in Delft and had a development centre in Cork, Ireland.
In 2000 the company had over 260 employees.
Products
Com21 was quite successful with their cable modem line of products in the years before the new standard DOCSIS was available. The Com21 portfolio can be divided into three segments: the central or head end equipment, cable modems or CPE, and the management platform.
ComController
The central system, typically installed in the head end or CATV hub locations were called ComControllers. A ComController provided the same function as a CMTS in modern Docsis networks. A ComcCntroller was a 19" rack module that was built around an integrated ATM switch and needed several modules to become a functional cable modem head end system.
The main components or modules in a ComController were:
the management module (one per system)
the downstream HF controller (one per system)
upstream HF controllers, each providing two upstream channels (at least one per system)
network controllers, providing interconnection with the internet (or private IP network)
The management module provided the interface between each ComController system and the Com21 network management platform NMAPS. The management module offered a single Ethernet interface to an out of band management network.
Each ComController had one HF Downstream module. The downstream controller had two F-connector interfaces to connect a coax cable towards the distribution section of a hybrid fibre coax cable TV network. Below the primary interface a test interface conne |
https://en.wikipedia.org/wiki/Nitrogen%20cycle | The nitrogen cycle is the biogeochemical cycle by which nitrogen is converted into multiple chemical forms as it circulates among atmospheric, terrestrial, and marine ecosystems. The conversion of nitrogen can be carried out through both biological and physical processes. Important processes in the nitrogen cycle include fixation, ammonification, nitrification, and denitrification. The majority of Earth's atmosphere (78%) is atmospheric nitrogen, making it the largest source of nitrogen. However, atmospheric nitrogen has limited availability for biological use, leading to a scarcity of usable nitrogen in many types of ecosystems.
The nitrogen cycle is of particular interest to ecologists because nitrogen availability can affect the rate of key ecosystem processes, including primary production and decomposition. Human activities such as fossil fuel combustion, use of artificial nitrogen fertilizers, and release of nitrogen in wastewater have dramatically altered the global nitrogen cycle. Human modification of the global nitrogen cycle can negatively affect the natural environment system and also human health.
Processes
Nitrogen is present in the environment in a wide variety of chemical forms including organic nitrogen, ammonium (), nitrite (), nitrate (), nitrous oxide (), nitric oxide (NO) or inorganic nitrogen gas (). Organic nitrogen may be in the form of a living organism, humus or in the intermediate products of organic matter decomposition. The processes in the nitrogen cycle is to transform nitrogen from one form to another. Many of those processes are carried out by microbes, either in their effort to harvest energy or to accumulate nitrogen in a form needed for their growth. For example, the nitrogenous wastes in animal urine are broken down by nitrifying bacteria in the soil to be used by plants. The diagram alongside shows how these processes fit together to form the nitrogen cycle.
Nitrogen fixation
The conversion of nitrogen gas () into nitrates |
https://en.wikipedia.org/wiki/Ephemeral%20plant | An ephemeral plant is a plant with a very short life cycle or very short period of active growth, often one that grows only during brief periods when conditions are favorable. Several types of ephemeral plants exist. The first, spring ephemeral, refers to perennial plants that emerge quickly in the spring and die back to their underground parts after a short growth and reproduction phase. Desert ephemerals are plants which are adapted to take advantage of the short wet periods in arid climates. Mud-flat ephemerals take advantage of short periods of low water. In areas subjected to recurring human disturbance, such as plowing, weedy ephemerals are very short-lived plants whose entire life cycle takes less than a growing season. In each case, the species has a life cycle timed to exploit a short period when resources are freely available. An evergreen plant could be considered the opposite of an ephemeral plant.
Spring ephemerals
Spring ephemerals are perennial woodland wildflowers which develop aerial parts (i.e. stems, leaves, and flowers) of the plant early each spring and then quickly bloom, and produce seed. The leaves often wither leaving only underground structures (i.e. roots, rhizomes, and bulbs) for the remainder of the year. This strategy is very common in herbaceous communities of deciduous forests as it allows small herbaceous plants to take advantage of the high levels of sunlight reaching the forest floor prior to the formation of a canopy by woody plants. Examples include: spring beauties, trilliums, harbinger of spring and the genus of Dicentra particularly D. cucullaria, Dutchman's breeches and D. canadensis, squirrel corn.
In the herb layer of beech forest and hornbeam-sessile oak forest, tuberous, bulbous and rhizomous plants are abundant. They comprise the spring geophytes (tuberous, bulbous and rhizomous).
Desert ephemerals
Desert ephemerals, such as Arabidopsis thaliana, are plants which are adapted to take advantage of the very short favou |
https://en.wikipedia.org/wiki/Pelargonic%20acid | Pelargonic acid, also called nonanoic acid, is an organic compound with structural formula . It is a nine-carbon fatty acid. Nonanoic acid is a colorless oily liquid with an unpleasant, rancid odor. It is nearly insoluble in water, but very soluble in organic solvents. The esters and salts of pelargonic acid are called pelargonates or nonanoates.
The acid is named after the pelargonium plant, since oil from its leaves contains esters of the acid.
Preparation
Together with azelaic acid, it is produced industrially by ozonolysis of oleic acid.
Alternatively, pelargonic acid can be produced in a two-step process beginning with coupled dimerization and hydroesterification of 1,3-butadiene. This step produces a doubly unsaturated C9-ester, which can be hydrogenated to give esters of pelargonic acid.
\overset{1,3-butadiene}{2CH2=CH-CH=CH2}{} + CO + CH3OH -> CH2=CH(CH2)3CH=CHCH2CO2CH3
CH2=CH(CH2)3CH=CHCH2CO2CH3{} + 2 H2 -> \underset{pelargonic\ acid\ ester}{CH3(CH2)7CO2CH3}
A laboratory preparation involves permanganate oxidation of 1-decene.
Occurrence and uses
Pelargonic acid occurs naturally as esters in the oil of pelargonium.
Synthetic esters of pelargonic acid, such as methyl pelargonate, are used as flavorings. Pelargonic acid is also used in the preparation of plasticizers and lacquers. The derivative 4-nonanoylmorpholine is an ingredient in some pepper sprays. The ammonium salt of pelargonic acid, ammonium pelargonate, is a herbicide. It is commonly used in conjunction with glyphosate, a non-selective herbicide, for a quick burn-down effect in the control of weeds in turfgrass.
The methyl form and ethylene glycol pelargonate act as nematicides against Meloidogyne javanica on Solanum lycopersicum, and the methyl against Heterodera glycines and M. incognita on Glycine max.
Esters of pelargonic acid are precursors to lubricants.
Pharmacological effects
Pelargonic acid may be more potent than valproic acid in treating seizures. Moreover, in contrast to val |
https://en.wikipedia.org/wiki/Network%20segmentation | Network segmentation in computer networking is the act or practice of splitting a computer network into subnetworks, each being a network segment. Advantages of such splitting are primarily for boosting performance and improving security.
Advantages
Reduced congestion: On a segmented network, there are fewer hosts per subnetwork and the traffic and thus congestion per segment is reduced
Improved security:
Broadcasts will be contained to local network. Internal network structure will not be visible from outside.
There is a reduced attack surface available to pivot in if one of the hosts on the network segment is compromised. Common attack vectors such as LLMNR and NetBIOS poisoning can be partially alleviated by proper network segmentation as they only work on the local network. For this reason it is recommended to segment the various areas of a network by usage. A basic example would be to split up web servers, databases servers and standard user machines each into their own segment.
By creating network segments containing only the resources specific to the consumers that you authorise access to, you are creating an environment of least privilege
Containing network problems: Limiting the effect of local failures on other parts of network
Controlling visitor access: Visitor access to the network can be controlled by implementing VLANs to segregate the network
Improved security
When a cyber-criminal gains unauthorized access to a network, segmentation or “zoning” can provide effective controls to limit further movement across the network. PCI-DSS (Payment Card Industry Data Security Standard), and similar standards, provide guidance on creating clear separation of data within the network, for example separating the network for Payment Card authorizations from those for Point-of-Service (till) or customer wi-fi traffic. A sound security policy entails segmenting the network into multiple zones, with varying security requirements, and rigorously enforcing th |
https://en.wikipedia.org/wiki/Global%20variable | In computer programming, a global variable is a variable with global scope, meaning that it is visible (hence accessible) throughout the program, unless shadowed. The set of all global variables is known as the global environment or global state. In compiled languages, global variables are generally static variables, whose extent (lifetime) is the entire runtime of the program, though in interpreted languages (including command-line interpreters), global variables are generally dynamically allocated when declared, since they are not known ahead of time.
In some languages, all variables are global, or global by default, while in most modern languages variables have limited scope, generally lexical scope, though global variables are often available by declaring a variable at the top level of the program. In other languages, however, global variables do not exist; these are generally modular programming languages that enforce a module structure, or class-based object-oriented programming languages that enforce a class structure.
Use
Interaction mechanisms with global variables are called global environment (see also global state) mechanisms. The global environment paradigm is contrasted with the local environment paradigm, where all variables are local with no shared memory (and therefore all interactions can be reconducted to message passing).
Global variables are used extensively to pass information between sections of code that do not share a caller/callee relation like concurrent threads and signal handlers. Languages (including C) where each file defines an implicit namespace eliminate most of the problems seen with languages with a global namespace though some problems may persist without proper encapsulation. Without proper locking (such as with a mutex), code using global variables will not be thread-safe except for read only values in protected memory.
Environment variables
Environment variables are a facility provided by some operating systems. Within t |
https://en.wikipedia.org/wiki/Morphogen | A morphogen is a substance whose non-uniform distribution governs the pattern of tissue development in the process of morphogenesis or pattern formation, one of the core processes of developmental biology, establishing positions of the various specialized cell types within a tissue. More specifically, a morphogen is a signaling molecule that acts directly on cells to produce specific cellular responses depending on its local concentration.
Typically, morphogens are produced by source cells and diffuse through surrounding tissues in an embryo during early development, such that concentration gradients are set up. These gradients drive the process of differentiation of unspecialised stem cells into different cell types, ultimately forming all the tissues and organs of the body. The control of morphogenesis is a central element in evolutionary developmental biology (evo-devo).
History
The term was coined by Alan Turing in the paper "The Chemical Basis of Morphogenesis", where he predicted a chemical mechanism for biological pattern formation, decades before the formation of such patterns was demonstrated.
The concept of the morphogen has a long history in developmental biology, dating back to the work of the pioneering Drosophila (fruit fly) geneticist, Thomas Hunt Morgan, in the early 20th century. Lewis Wolpert refined the morphogen concept in the 1960s with the French flag model, which described how a morphogen could subdivide a tissue into domains of different target gene expression (corresponding to the colours of the French flag). This model was championed by the leading Drosophila biologist, Peter Lawrence. Christiane Nüsslein-Volhard was the first to identify a morphogen, Bicoid, one of the transcription factors present in a gradient in the Drosophila syncitial embryo. She was awarded the 1995 Nobel Prize in Physiology and Medicine for her work explaining the morphogenic embryology of the common fruit fly. Groups led by Gary Struhl and Stephen Cohen then de |
https://en.wikipedia.org/wiki/Fish%20finger | Fish fingers (British English) or fish sticks (American English) are a processed food made using a whitefish, such as cod, hake, haddock, shark, or pollock, which has been battered or breaded. They are commonly available in the frozen food section of supermarkets. They can be baked in an oven, grilled, shallow fried, or deep-fried.
History
The term "fish finger" is first referenced in a recipe given in a popular British magazine in 1900, and the dish is often considered symbolic of the United Kingdom.
The food restrictions during and after WWII expanded the consumption of fish fingers, but companies struggled to maintain decent quality. The commercialization of fish fingers may be traced to 1953 when the American company Gorton-Pew Fisheries, now known as Gorton's, was the first company to introduce a frozen ready-to-cook fish finger; the product, named Gorton's Fish Sticks, won the Parents magazine Seal of Approval in 1956. The developer of those fish sticks was Aaron L. Brody.
There was an abundance of herring in the United Kingdom after World War II. Clarence Birdseye test-marketed herring fish fingers, a product he had discovered in the United States, under the name "herring savouries". These were tested in Southampton and South Wales against "cod fingers", a comparatively bland product used as a control. Shoppers, however, confounded expectations by showing an overwhelming preference for the cod. The snack was nearly called Battered Cod Pieces, until a poll of Birds Eye workers opted for the snappier Fish Fingers.
Varieties
Minced fish comes in industry standard 7.5 kg frozen blocks for further slicing and battering. These are more commonly used in store brand economy products. They may have either batter or breadcrumbs around the outside as casing, although the coating is normally breadcrumbs.
In addition to white fish, fish fingers are sometimes made with salmon.
See also
Crab stick
Chicken fingers
Chicken nugget
Fishcake
Fish finger sandwich |
https://en.wikipedia.org/wiki/Komar%20mass | The Komar mass (named after Arthur Komar) of a system is one of several formal concepts of mass that are used in general relativity. The Komar mass can be defined in any stationary spacetime, which is a spacetime in which all the metric components can be written so that they are independent of time. Alternatively, a stationary spacetime can be defined as a spacetime which possesses a timelike Killing vector field.
The following discussion is an expanded and simplified version of the motivational treatment in (Wald, 1984, pg 288).
Motivation
Consider the Schwarzschild metric. Using the Schwarzschild basis, a frame field for the Schwarzschild metric, one can find that the radial acceleration required to hold a test mass stationary at a Schwarzschild coordinate of r is:
Because the metric is static, there is a well-defined meaning to "holding a particle stationary".
Interpreting this acceleration as being due to a "gravitational force", we can then compute the integral of normal acceleration multiplied by area to get a "Gauss law" integral of:
While this approaches a constant as r approaches infinity, it is not a constant independent of r. We are therefore motivated to introduce a correction factor to make the above integral independent of the radius r of the enclosing shell. For the Schwarzschild metric, this correction factor is just , the "red-shift" or "time dilation" factor at distance r. One may also view this factor as "correcting" the local force to the "force at infinity", the force that an observer at infinity would need to apply through a string to hold the particle stationary. (Wald, 1984).
To proceed further, we will write down a line element for a static metric.
where gtt and the quadratic form are functions only of the spatial coordinates x, y, z and are not functions of time. In spite of our choices of variable names, it should not be assumed that our coordinate system is Cartesian. The fact that none of the metric coefficients are functi |
https://en.wikipedia.org/wiki/Uniporter | A uniporter is a membrane transport protein that transports a single species of substrate (charged or uncharged) across a cell membrane. It may use either facilitated diffusion and transport along a diffusion gradient or transport against one with an active transport process. Uniporters include both carriers and ion channels, and are referred to as facilitated transporters, suggesting movement down a concentration or electrochemical gradient.
Uniporter carrier proteins work by binding to one molecule of substrate at a time. Uniporter channels open in response to a stimulus and allow the free flow of specific molecules.
There are several ways in which the opening of uniporter channels may be regulated:
Voltage – Regulated by the difference in voltage across the membrane
Stress – Regulated by physical pressure on the transporter (as in the cochlea of the ear)
Ligand – Regulated by the binding of a ligand to either the intracellular or extracellular side of the cell
Uniporters are found in mitochondria and neurons. The uniporter in the mitochondria is responsible for calcium uptake. The calcium channels are used for cell signaling and triggering apoptosis. The calcium uniporter transports calcium across the inner mitochondrial membrane and is activated when calcium rises above a certain concentration. Voltage-gated potassium channels are also uniporters that can be found in neurons and are essential for action potentials. This channel is activated by a voltage gradient created by sodium-potassium pumps. When the membrane reaches a certain voltage, the channels open which depolarizes the membrane which leads to the action potential being sent down the membrane.
See also
Antiporter
Symporter |
https://en.wikipedia.org/wiki/Trust%20boundary | Trust boundary is a term used in computer science and security which describes a boundary where program data or execution changes its level of "trust," or where two principals with different capabilities exchange data or commands. The term refers to any distinct boundary where within a system all sub-systems (including data) have equal trust. An example of an execution trust boundary would be where an application attains an increased privilege level (such as root). A data trust boundary is a point where data comes from an untrusted source--for example, user input or a network socket.
A "trust boundary violation" refers to a vulnerability where computer software trusts data that has not been validated before crossing a boundary. |
https://en.wikipedia.org/wiki/Autonomous%20things | Autonomous things, abbreviated AuT, or the Internet of autonomous things, abbreviated as IoAT, is an emerging term for the technological developments that are expected to bring computers into the physical environment as autonomous entities without human direction, freely moving and interacting with humans and other objects.
Self-navigating drones are the first AuT technology in (limited) deployment. It is expected that the first mass-deployment of AuT technologies will be the autonomous car, generally expected to be available around 2020. Other currently expected AuT technologies include home robotics (e.g., machines that provide care for the elderly, infirm or young), and military robots (air, land or sea autonomous machines with information-collection or target-attack capabilities).
AuT technologies share many common traits, which justify the common notation. They are all based on recent breakthroughs in the domains of (deep) machine learning and artificial intelligence. They all require extensive and prompt regulatory developments to specify the requirements from them and to license and manage their deployment (see the further reading below). And they all require unprecedented levels of safety (e.g., automobile safety) and security, to overcome concerns about the potential negative impact of the new technology.
As an example, the autonomous car both addresses the main existing safety issues and creates new issues. It is expected to be much safer than existing vehicles, by eliminating the single most dangerous elementthe driver. The US's National Highway Traffic Safety Administration estimates 94 percent of US accidents were the result of human error and poor decision-making, including speeding and impaired driving, and the Center for Internet and Society at Stanford Law School claims that "Some ninety percent of motor vehicle crashes are caused at least in part by human error". So while safety standards like the ISO 26262 specify the required safety, there is |
https://en.wikipedia.org/wiki/Agroinfiltration | Agroinfiltration is a method used in plant biology and especially lately in plant biotechnology to induce transient expression of genes in a plant, or isolated leaves from a plant, or even in cultures of plant cells, in order to produce a desired protein. In the method, a suspension of Agrobacterium tumefaciens is introduced into a plant leaf by direct injection or by vacuum infiltration, or brought into association with plant cells immobilised on a porous support (plant cell packs), whereafter the bacteria transfer the desired gene into the plant cells via transfer of T-DNA. The main benefit of agroinfiltration when compared to the more traditional plant transformation is speed and convenience, although yields of the recombinant protein are generally also higher and more consistent.
The first step is to introduce a gene of interest to a strain of Agrobacterium tumefaciens. Subsequently, the strain is grown in a liquid culture and the resulting bacteria are washed and suspended into a suitable buffer solution. For injection, this solution is then placed in a syringe (without a needle). The tip of the syringe is pressed against the underside of a leaf while simultaneously applying gentle counterpressure to the other side of the leaf. The Agrobacterium suspension is then injected into the airspaces inside the leaf through stomata, or sometimes through a tiny incision made to the underside of the leaf.
Vacuum infiltration is another way to introduce Agrobacterium deep into plant tissue. In this procedure, leaf disks, leaves, or whole plants are submerged in a beaker containing the solution, and the beaker is placed in a vacuum chamber. The vacuum is then applied, forcing air out of the intercellular spaces within the leaves via the stomata. When the vacuum is released, the pressure difference forces the "Agrobacterium" suspension into the leaves through the stomata into the mesophyll tissue. This can result in nearly all of the cells in any given leaf being in c |
https://en.wikipedia.org/wiki/Texas%20Instruments%20DaVinci | The Texas Instruments DaVinci is a family of system on a chip processors that are primarily used in embedded video and vision applications. Many processors in the family combine a DSP core based on the TMS320 C6000 VLIW DSP family and an ARM CPU core into a single system on chip. By using both a general-purpose processor and a DSP, the control and media portions can both be executed by separate processors.
Later chips in the family included DSP-only and ARM-only processors. All the later chips integrate several accelerators to offload commodity application specific processing from the processor cores to dedicated accelerators. Most notable among these are HDVICP, an H.264, SVC and MPEG-4 compression and decompression engine, ISP, an accelerator engine with methods for enhancing video primarily input from camera sensors and an OSD engine for display acceleration. Some of the newest processors also integrate a vision coprocessor in the SoC.
History
DaVinci processors were introduced at a time when embedded processors with homogeneous processor cores were widely used. These processors were based either on cores that could do signal processing optimally, like DSPs or GPUs or based on cores that could do general purpose processing optimally, like, powerPC, ARM, and StrongARM. By using both a general-purpose processor and a DSP on a single chip, the control and media portions can both be executed by processors that excel at their respective tasks. By providing a bundled offering with system and application software, evaluation modules and debug tools based on Code Composer Studio, TI DaVinci processors were intended to win over a broader set of customers looking to add video feature to their electronic products.
TI announced its first DaVinci branded video processors, the DM6443 and DM6446, on 5 December 2005. A year later, TI followed up with DSP only versions of the chips in the family, called DM643x (DM6431, DM6433, DM6435, DM6437). On January 15, 2007, TI announc |
https://en.wikipedia.org/wiki/Dromedary | The dromedary (Camelus dromedarius or ;), also known as the dromedary camel, Somalian Camel, Arabian camel, or one-humped camel, is a large even-toed ungulate, of the genus Camelus, with one hump on its back.
It is the tallest of the three species of camel; adult males stand at the shoulder, while females are tall. Males typically weigh between , and females weigh between .
The species' distinctive features include its long, curved neck, narrow chest, a single hump (compared with two on the Bactrian camel and wild Bactrian camel), and long hairs on the throat, shoulders and hump. The coat is generally a shade of brown. The hump, tall or more, is made of fat bound together by fibrous tissue.
Dromedaries are mainly active during daylight hours. They form herds of about 20 individuals, which are led by a dominant male. They feed on foliage and desert vegetation; several adaptations, such as the ability to tolerate losing more than 30% of its total water content, allow it to thrive in its desert habitat. Mating occurs annually and peaks in the rainy season; females bear a single calf after a gestation of 15 months.
The dromedary has not occurred naturally in the wild for nearly 2,000 years. It was probably first domesticated in the Arabian Peninsula about 4,000 years ago, or in Somalia where there are paintings in Laas Geel that figure it from 5,000 to 9,000 years ago. In the wild, the dromedary inhabited arid regions, including the Sahara Desert. The domesticated dromedary is generally found in the semi-arid to arid regions of the Old World, mainly in Africa and the Arabian Peninsula, and a significant feral population occurs in Australia. Products of the dromedary, including its meat and milk, support several North African tribes; it is also commonly used for riding and as a pack animal.
Etymology
The common name "dromedary" comes from the Old French or the Late Latin . These originated from the Greek word , δρομάς (ο, η) (GEN (γενική) , δρομάδος), meaning " |
https://en.wikipedia.org/wiki/Dolphin%20Research%20Center | The Dolphin Research Center (DRC) is a dolphinarium on Grassy Key, Florida. The series of saltwater lagoons is home to several dolphins and California sea lions.
History
Fisherman Milton Santini opened Santini's Porpoise School in 1958, which he operated until 1972.
Santini blasted lagoons into the seashore with dynamite to create deeper pens for the dolphins.
One of Santini's dolphins, Mitzi, starred in the film Flipper.
The property was sold and renamed as Flipper's Sea School in 1972 and operated under this name until 1977.
At this point, it was purchased by Jean Paul Fortom-Gouin.
Fortom-Gouin closed the property to the public, running it as the Institute for Delphinid Research until 1983.
The Dolphin Research Center was founded in 1984 by Armando ("Mandy") and Jayne Rodriguez.
Mandy had earlier been a dolphin trainer at Miami Seaquarium and decided to interact with dolphins in a manner less "circus-y".
Animals
As of 2022, DRC listed 25 dolphins as currently residing at its facilities. Twenty-four of the dolphins are bottlenose dolphins, while the other dolphin is an Atlantic spotted dolphin. Twenty-one of the dolphins were born in captivity. Four of the bottlenose dolphins and the Atlantic spotted dolphins were rescued from the wild.
It is also home to three sea lions, several exotic birds, and two African spurred tortoises.
Guest experience
Visitors to the DRC can come in under general admission to see the dolphins and sea lions demonstrate behaviors, research, training, and other activities in narrated behavior sessions throughout the day. They can also pay to experience a variety of interactive programs either in the water or from a floating dock.
In addition to the general admission price, guests can pay more to interact with dolphins in the water.
DRC also has an "Ultimate Trainer for the Day" program where guests can shadow and participate in activities with trainers and a "Researcher Experience" where they can shadow the DRC research team.
Other gue |
https://en.wikipedia.org/wiki/Back-Up%20Interceptor%20Control | Backup Interceptor Control (BUIC, ) was the Electronic Systems Division 416M System to backup the SAGE 416L System in the United States and Canada. BUIC deployed Cold War command, control, and coordination systems to SAGE radar stations to create dispersed NORAD Control Centers.
Background
Prior to the SAGE Direction Centers becoming operational, the USAF deployed data link systems at NORAD Control Centers with ground computers for controlling manned interceptors. After SAGE IBM AN/FSQ-7 Combat Direction Centrals became operational and the Super Combat Centers with improved (digital) computers were cancelled, a backup to SAGE was planned in the event the above-ground SAGE Air Defense Direction Center failed.
General Electric AN/GPA-37 Course Directing Group
BUIC began with deployment of General Electric AN/GPA-37 Course Directing Groups to several Long Range Radar stations. Units designated included the "U.S. Air Force 858th Air Defense Group (BUIC) [which became] a permanent operating facility" at Naval Air Station Fallon in Nevada.
BUIC II
BUIC II was used to command and control sites using the Burroughs AN/GSA-51 Radar Course Directing Group. North Truro AFS became the first ADC installation configured for BUIC II.
BUIC III
The AN/GYK-19 (initially AN/GSA-51A) was an upgraded version of the BUIC II system designated AN/GSA-51A and required a larger building than the AN/GSA-51. The first BUIC III site was Fort Fisher AFS, and Air Defense Command's was first installed at Fort Fisher Air Force Station, North Carolina.
Although more advanced systems were contemplated, the final design of the BUIC III system was an upgraded version of the BUIC II with around twice the performance.
Closure and upgrade
In 1972 the USAF decided to shut down most of the BUIC sites; most of the sites mothballed by 1974, except for the BUIC III site at Tyndall Air Force Base. In Canada the BUIC site at Senneterre was shut down, but St Margarets remained open. The remaining sites |
https://en.wikipedia.org/wiki/Microarchitecture | In electronics, computer science and computer engineering, microarchitecture, also called computer organization and sometimes abbreviated as µarch or uarch, is the way a given instruction set architecture (ISA) is implemented in a particular processor. A given ISA may be implemented with different microarchitectures; implementations may vary due to different goals of a given design or due to shifts in technology.
Computer architecture is the combination of microarchitecture and instruction set architecture.
Relation to instruction set architecture
The ISA is roughly the same as the programming model of a processor as seen by an assembly language programmer or compiler writer. The ISA includes the instructions, execution model, processor registers, address and data formats among other things. The microarchitecture includes the constituent parts of the processor and how these interconnect and interoperate to implement the ISA.
The microarchitecture of a machine is usually represented as (more or less detailed) diagrams that describe the interconnections of the various microarchitectural elements of the machine, which may be anything from single gates and registers, to complete arithmetic logic units (ALUs) and even larger elements. These diagrams generally separate the datapath (where data is placed) and the control path (which can be said to steer the data).
The person designing a system usually draws the specific microarchitecture as a kind of data flow diagram. Like a block diagram, the microarchitecture diagram shows microarchitectural elements such as the arithmetic and logic unit and the register file as a single schematic symbol. Typically, the diagram connects those elements with arrows, thick lines and thin lines to distinguish between three-state buses (which require a three-state buffer for each device that drives the bus), unidirectional buses (always driven by a single source, such as the way the address bus on simpler computers is always driven by |
https://en.wikipedia.org/wiki/Biomin | Biomin is an animal health and nutrition company headquartered in Inzersdorf-Getzersdorf, Austria that develops and produces feed additives and premixes for livestock animals including swine, poultry, dairy and beef cattle as well as aquaculture.
The firm supplies customers in more than 100 countries throughout the world.
The Biomin Research Center (BRC) at Campus Tulln in Austria, employs 80 researchers engaged in applied basic research to lead the firm’s in-house R&D efforts, supported by a research network of 150 academic and research institutions worldwide.
Company history
1983: Erich Erber opened Biomin in a small warehouse in Pottenbrunn, Austria with a staff of three, starting with a livestock premix product line containing probiotics. Founder Erich Erber named the product and company based on the idea of combining “biological minerals” and so the brand name "Biomin" was born.
1985: Biomin bought Interpremix, a small premix factory, as its first production facility. The acquired firm also made a product called Antitox Plus, the first product available that ‘binds’ mycotoxins through a process known as adsorption.
1988: Biomin signed a research agreement with the University of Veterinary Medicine Vienna, Austria, and Dr. Josef Leibetseder to further improve the efficacy of Antitox Plus to a wider range of mycotoxins. Only several years later the second generation of the mycotoxin deactivation product was launched under the brand name Mycofix.
1994: The Biomin brand name was separated from the Interpremix product line. Erber AG became the holding company of Biomin.
2012: Biomin finalized acquisition of Microplus, a Germany-based feed additive producer, and rolled out Digestarom, a phytogenic feed additive.
2013: Biomin receives EU authorization for its multi-species probiotic, PoultryStar.
2015: Global Product introduction of Mycofix 5.0.
2016: Product introduction of Digestarom DC – TheFeedConverter
2017: Opening of first ERBER Group North American |
https://en.wikipedia.org/wiki/Ordered%20set%20operators | In mathematical notation, ordered set operators indicate whether an object precedes or succeeds another. These relationship operators are denoted by the unicode symbols U+227A-F, along with symbols located unicode blocks U+228x through U+22Ex.
Examples
The relationship x precedes y is written x ≺ y. The relation x precedes or is equal to y is written x ≼ y.
The relationship x succeeds (or follows) y is written x ≻ y. The relation x succeeds or is equal to y is written x ≽ y.
Use in political science
In Political science and Decision theory, order relations are typically used in the context of an agent's choice, for example the preferences of a voter over several political candidates.
x ≺ y means that the voter prefers candidate y over candidate x.
x ∼ y means the voter is indifferent between candidates x and y.
x ≲ y means the voter is indifferent or prefers candidate y. |
https://en.wikipedia.org/wiki/Microtubule-associated%20protein | In cell biology, microtubule-associated proteins (MAPs) are proteins that interact with the microtubules of the cellular cytoskeleton. MAPs are integral to the stability of the cell and its internal structures and the transport of components within the cell.
Function
MAPs bind to the tubulin subunits that make up microtubules to regulate their stability. A large variety of MAPs have been identified in many different cell types, and they have been found to carry out a wide range of functions. These include both stabilizing and destabilizing microtubules, guiding microtubules towards specific cellular locations, cross-linking microtubules and mediating the interactions of microtubules with other proteins in the cell.
Within the cell, MAPs bind directly to the tubulin dimers of microtubules. This binding can occur with either polymerized or depolymerized tubulin, and in most cases leads to the stabilization of microtubule structure, further encouraging polymerization. Usually, it is the C-terminal domain of the MAP that interacts with tubulin, while the N-terminal domain can bind with cellular vesicles, intermediate filaments or other microtubules. MAP-microtubule binding is regulated through MAP phosphorylation. This is accomplished through the function of the microtubule-affinity-regulating-kinase (MARK) protein. Phosphorylation of the MAP by the MARK causes the MAP to detach from any bound microtubules. This detachment is usually associated with a destabilization of the microtubule causing it to fall apart. In this way the stabilization of microtubules by MAPs is regulated within the cell through phosphorylation.
Types
MAPs have been divided into several different categories and sub-categories. There are "structural" MAPs which bind along the microtubules and "+TIP" MAPs which bind to the growing end of the microtubules. Structural MAPs have been divided into MAP1, MAP2, MAP4, and Tau families. +TIP MAPs are motor proteins such as kinesin, dyneins, and other MAPs |
https://en.wikipedia.org/wiki/Fruitless%20%28gene%29 | The fruitless gene (fru) is a Drosophila melanogaster gene that encodes several variants of a putative transcription factor protein. Normal fruitless function is required for proper development of several anatomical structures necessary for courtship, including motor neurons which innervate muscles needed for fly sexual behaviors. The gene does not have an obvious mammalian homolog, but appears to function in sex determination in species as distant as the mosquito Anopheles gambiae.
fruitless serves as an example of how a gene or a group of genes may regulate the development and/or function of neurons involved in innate behavior. Research on fruitless has received attention in the popular press, since it provokes discussion on genetics of human sexual orientation, and behaviors such as gender-specific aggression.
Function
Male flies with mutations in the fruitless gene display altered sexual behavior. Fruitfly courtship, which involves a complex male-initiated ritual, may be disrupted in many ways by mutated fru alleles; fru is necessary for every step in the ritual. Some alleles prevent courting entirely, while others disrupt individual components. Notably, some loss-of-function alleles change or remove sexual preference.
Although many genes are known to be involved in male courtship behavior, the fruitless gene has been considered noteworthy because it exhibits sex-specific alternative splicing. When females produce the male-spliced gene product, they behave as males. Males that do not produce the male-specific product do not court females and are infertile. In the brain, a subset (ca. 2,000) of neurons express fruitless and fruitless expression is sufficient to instruct sexually dimorphic connectivity.
fruitless has at least four promoters, each encoding proteins containing both a BTB (Broad complex/tramtrack/bric-a-brac) domain and a zinc finger motif. Alternative splicing occurs at both the 5' and 3' ends, and there are several variants (other than the m |
https://en.wikipedia.org/wiki/Important%20ecological%20areas | Important ecological areas (IEAs) are habitat areas which, either by themselves or in a network, contribute significantly to an ecosystem’s productivity, biodiversity, and resilience. Appropriate management of key ecological features delineates the management boundaries of an IEA. The identification and protection of IEAs is an element of an ecosystem-based management approach.
Important ecological areas may have varying levels of management of extractive activities, from monitoring up to and including marine reserve. IEAs have management measures tailored to the ecological features within the area with consideration of socioeconomic factors. Whereas marine reserves generally have a fixed management policy of no extraction or ‘no-take’. Nonetheless, a marine reserve may be the appropriate management policy for an IEA.
The identification and management of IEAs is a form of ocean zoning. In the event that there are a series of linked IEAs within a large marine ecosystem, a collective action to manage the network, such as a marine sanctuary or national monument, may be warranted.
Examples are tropical rainforests, oceans, forests, etc. |
https://en.wikipedia.org/wiki/Chemical%20communication%20in%20insects | Chemical communication in insects is social signalling between insects of the same or different species, using chemicals. These chemicals may be volatile, to be detected at a distance by other insects' sense of smell, or non-volatile, to be detected on an insect's cuticle by other insects' sense of taste. Many of these chemicals are pheromones, acting like hormones outside the body.
Among the many functions of chemical communication are attracting mates, aggregating conspecific individuals of both sexes, deterring other individuals from approaching, announcing a new food source, marking a trail, recognizing nest-mates, marking territory and triggering aggression.
History of research
In 1965, the entomologist Edward O. Wilson published a paper on chemical communication in the social insects, arguing that their societies were principally organised by "complex systems of chemical signals". By 1990, Mahmoud Ali and David Morgan noted that the field had grown too large to review comprehensively.
Semiochemicals
In addition to the use of means such as making sounds, generating light, and touch for communication, a wide range of insects have evolved chemical signals, semiochemicals. Types of semiochemicals include pheromones and kairomones. Chemoreception is the physiological response of a sense organ to a chemical stimulus where the chemicals act as signals to regulate the state or activity of a cell.
Semiochemicals are often derived from plant metabolites. Pheromones, a type of semiochemical, are used for attracting mates, for aggregating conspecific individuals of both sexes, for deterring other individuals from approaching, to mark a trail, and to trigger aggression in nearby individuals. Allomones benefit their producer by the effect they have upon the receiver. Kairomones benefit their receiver instead of their producer. Synomones benefit the producer and the receiver. While some chemicals are targeted at individuals of the same species, others are used for |
https://en.wikipedia.org/wiki/Peruvian%20inti | The inti was the currency of Peru between 1985 and 1991. Its ISO 4217 code was PEI and its abbreviation was I/. The inti was divided into 100 céntimos. The inti replaced the inflation-stricken sol. The new currency was named after Inti, the Inca sun god.
History
The inti was introduced on 1 February 1985, replacing the sol which had suffered from high inflation. One inti was equivalent to 1,000 soles. Coins denominated in the new unit were put into circulation from May 1985 and banknotes followed in June of that year.
By 1990, the inti had itself suffered from high inflation. As an interim measure, from January to July 1991, the "inti millón" (I/m.) was used as a unit of account. One inti millón was equal to 1,000,000 intis and hence to one new sol. The nuevo sol ("new sol") was adopted on 1 July 1991, replacing the inti at an exchange rate of a million to one. Thus: 1 new sol = 1,000,000 intis = 1,000,000,000 old soles.
Inti notes and coins are no longer legal tender in Peru, nor can they be exchanged for notes and coins denominated in the current nuevo sol.
Coins
Coins were introduced in 1985 in denominations of 1, 5, 10, 20, and 50 centimos (designs were taken from the previous 10, 50, 100, and 500 soles de oro coins), plus 1 and 5 intis. The 1 céntimo coin was issued only in 1985. The 5 céntimo coins were issued until 1986. All the other denominations were issued until 1988. All coins featured Navy Admiral Miguel Grau: cent coins on the reverse, Inti coins on the obverse.
Banknotes
In June 1985, notes were introduced in denominations of I/.10, I/.50 (taken from previous 10,000 and 50,000 soles de oro notes) and I/.100, followed by I/.500 in December of the same year. The next year, I/.1,000 notes were added, followed by I/.5,000 and I/.10,000 in 1988. 50,000 and I/.100,000 notes were added in 1989. I/.500,000 denominations were added early in 1990, I/.1,000,000 denominations were added in mid-1990, and I/.5,000,000 intis in August 1990. The obverses featur |
https://en.wikipedia.org/wiki/Binary%20number | A binary number is a number expressed in the base-2 numeral system or binary numeral system, a method of mathematical expression which uses only two symbols: typically "0" (zero) and "1" (one).
The base-2 numeral system is a positional notation with a radix of 2. Each digit is referred to as a bit, or binary digit. Because of its straightforward implementation in digital electronic circuitry using logic gates, the binary system is used by almost all modern computers and computer-based devices, as a preferred system of use, over various other human techniques of communication, because of the simplicity of the language and the noise immunity in physical implementation.
Negative numbers are commonly represented in binary using two's complement.
History
The modern binary number system was studied in Europe in the 16th and 17th centuries by Thomas Harriot, Juan Caramuel y Lobkowitz, and Gottfried Leibniz. However, systems related to binary numbers have appeared earlier in multiple cultures including ancient Egypt, China, and India. Leibniz was specifically inspired by the Chinese I Ching.
Egypt
The scribes of ancient Egypt used two different systems for their fractions, Egyptian fractions (not related to the binary number system) and Horus-Eye fractions (so called because many historians of mathematics believe that the symbols used for this system could be arranged to form the eye of Horus, although this has been disputed). Horus-Eye fractions are a binary numbering system for fractional quantities of grain, liquids, or other measures, in which a fraction of a hekat is expressed as a sum of the binary fractions 1/2, 1/4, 1/8, 1/16, 1/32, and 1/64. Early forms of this system can be found in documents from the Fifth Dynasty of Egypt, approximately 2400 BC, and its fully developed hieroglyphic form dates to the Nineteenth Dynasty of Egypt, approximately 1200 BC.
The method used for ancient Egyptian multiplication is also closely related to binary numbers. In this meth |
https://en.wikipedia.org/wiki/Irish%20Nutrition%20and%20Dietetic%20Institute | The Irish Nutrition & Dietetic Institute (INDI) is the national organisation in Ireland for nutrition and dietetic professionals. The INDI's mission is to encourage, foster and maintain the highest possible standards in the science and practice of human nutrition and dietetics, to positively influence the nutrition status and health of the individual and the population in general. This incorporates clinical nutrition, community nutrition, business and industry, research, education and private practice.
External links
Official web site
Food science institutes
Nutrition organizations
Government agencies of the Republic of Ireland |
https://en.wikipedia.org/wiki/MSU-DOE%20Plant%20Research%20Laboratory | The MSU-DOE Plant Research Laboratory (PRL), commonly referred to as Plant Research Lab,
is a research institute funded to a large extent by the U.S. Department of Energy Office of Science and located at Michigan State University (MSU)
in East Lansing, Michigan. The Plant Research Lab was founded in 1965, and it currently includes twelve laboratories that conduct collaborative basic research
into the biology of diverse photosynthetic organisms, including plants, bacteria, and algae, in addition to developing new technologies towards
addressing energy and food challenges.
History
1964-1978
The contract for the establishment of the MSU-DOE Plant Research Laboratory was signed on March 6, 1964, between the U.S. Atomic Energy Commission (AEC) and Michigan State University. The institute was initially funded by the AEC's Division of Biology and Medicine, which saw a need for improving the state of plant sciences in the United States. The Division aimed to create a new program at one or more universities where student interest in plant research could be fostered.
The contract signed between AEC and Michigan State University provided for a comprehensive research program in plant biology and related education and training at the graduate and postgraduate levels. The program was to draw strongly on related disciplines such as biochemistry, biophysics, genetics, microbiology, and others.
In 1966, personnel of the new program - called MSU-AEC Plant Research Laboratory at that time - moved into their new quarters in the Plant Biology Laboratories building at Michigan State University. The first research projects generally focused on problems specific to plants, such as cell growth and its regulation by plant hormones, cell wall structure and composition, and the physiology of flower formation; other research projects addressed general biological problems, such as the regulation of enzyme formation during development and cellular and genetic aspects of hormone action.
I |
https://en.wikipedia.org/wiki/Uniparental%20disomy | Uniparental disomy (UPD) occurs when a person receives two copies of a chromosome, or of part of a chromosome, from one parent and no copy from the other. UPD can be the result of heterodisomy, in which a pair of non-identical chromosomes are inherited from one parent (an earlier stage meiosis I error) or isodisomy, in which a single chromosome from one parent is duplicated (a later stage meiosis II error). Uniparental disomy may have clinical relevance for several reasons. For example, either isodisomy or heterodisomy can disrupt parent-specific genomic imprinting, resulting in imprinting disorders. Additionally, isodisomy leads to large blocks of homozygosity, which may lead to the uncovering of recessive genes, a similar phenomenon seen in inbred children of consanguineous partners.
UPD has been found to occur in about 1 in 2,000 births.
Pathophysiology
UPD can occur as a random event during the formation of egg cells or sperm cells or may happen in early fetal development. It can also occur during trisomic rescue.
When the child receives two (different) homologous chromosomes (inherited from both grandparents) from one parent, this is called heterodisomic UPD. Heterodisomy (heterozygous) indicates a meiosis I error if the gene loci in question didn't cross over.
When the child receives two (identical) replica copies of a single homologue of a chromosome, this is called an isodisomic UPD. Isodisomy (homozygous) indicates either a meiosis II (if the gene loci in question didn't cross over) or postzygotic chromosomal duplication.
A meiosis I error can result in isodisomic UPD if the gene loci in question crossed over, for example, a distal isodisomy would be due to duplicated gene loci from the maternal grandmother that crossed over and due to an error during meiosis I, ended up in the same gamete.
A meiosis II error can result in heterodisomy UPD if the gene loci crossed over in a similar fashion.
Phenotype
Most occurrences of UPD result in no phenotypical ano |
https://en.wikipedia.org/wiki/Functional%20flow%20block%20diagram | A functional flow block diagram (FFBD) is a multi-tier, time-sequenced, step-by-step flow diagram of a system's functional flow. The term "functional" in this context is different from its use in functional programming or in mathematics, where pairing "functional" with "flow" would be ambiguous. Here, "functional flow" pertains to the sequencing of operations, with "flow" arrows expressing dependence on the success of prior operations. FFBDs may also express input and output data dependencies between functional blocks, as shown in figures below, but FFBDs primarily focus on sequencing.
The FFBD notation was developed in the 1950s, and is widely used in classical systems engineering. FFBDs are one of the classic business process modeling methodologies, along with flow charts, data flow diagrams, control flow diagrams, Gantt charts, PERT diagrams, and IDEF.
FFBDs are also referred to as functional flow diagrams, functional block diagrams, and functional flows.
History
The first structured method for documenting process flow, the flow process chart, was introduced by Frank Gilbreth to members of American Society of Mechanical Engineers (ASME) in 1921 as the presentation “Process Charts—First Steps in Finding the One Best Way”. Gilbreth's tools quickly found their way into industrial engineering curricula.
In the early 1930s, an industrial engineer, Allan H. Mogensen began training business people in the use of some of the tools of industrial engineering at his Work Simplification Conferences in Lake Placid, New York. A 1944 graduate of Mogensen's class, Art Spinanger, took the tools back to Procter and Gamble where he developed their Deliberate Methods Change Program. Another 1944 graduate, Ben S. Graham, Director of Formcraft Engineering at Standard Register Industrial, adapted the flow process chart to information processing with his development of the multi-flow process chart to display multiple documents and their relationships. In 1947, ASME adopted a symbo |
https://en.wikipedia.org/wiki/Bulbul%20coronavirus%20HKU11 | Bulbul coronavirus HKU11 (Bulbul-CoV HKU11) is a positive-sense single-stranded RNA Deltacoronavirus of avian origin found in Chinese bulbuls. |
https://en.wikipedia.org/wiki/Adiabatic%20shear%20band | In physics, mechanics and engineering, an adiabatic shear band is one of the many mechanisms of failure that occur in metals and other materials that are deformed at a high rate in processes such as metal forming, machining and ballistic impact. Adiabatic shear bands are usually very narrow bands, typically 5-500 μm wide and consisting of highly sheared material. Adiabatic is a thermodynamic term meaning an absence of heat transfer – the heat produced is retained in the zone where it is created. (The opposite extreme, where all heat that is produced is conducted away, is isothermal.)
Deformation
It is necessary to include some basics of plastic deformation to understand the link between heat produced and the plastic work done. If we carry out a compression test on a cylindrical specimen to, say, 50% of its original height, the stress of the work material will increase usually significantly with reduction. This is called ‘work hardening’. During work hardening, the micro-structure, distortion of grain structure and the generation and glide of dislocations all occur. The remainder of the plastic work done – which can be as much as 90% of the total, is dissipated as heat.
If the plastic deformation is carried out under dynamic conditions, such as by drop forging, then the plastic deformation is localized more as the forging hammer speed is increased. This also means that the deformed material becomes hotter the higher the speed of the drop hammer. Now as metals become warmer, their resistance to further plastic deformation decreases. From this point we can see that there is a type of cascade effect: as more plastic deformation is absorbed by the metal, more heat is produced, making it easier for the metal to deform further. This is a catastrophic effect which almost inevitably leads to failure.
History
The first person to carry out any reported experimental programme to investigate the heat produced as a result of plastic deformation was Henri Tresca in June 1878 T |
https://en.wikipedia.org/wiki/C%C4%81ng%20zh%C3%BA | Cāng zhú (苍 术 or 蒼 术 or 蒼 朮), also known as black atractylodes rhizome or Rhizoma Atractylodes, is a Chinese herbal medicine. It is the dried rhizome of Atractylodes lancea (Thunb.) DC., synonyms Atractylodes chinensis (DC.) Koidz, and Atractylodes japonica Koidz. (One study suggested that A. chinensis is a subspecies of A. lancea, and A. chinensis var. liaotungensis is a subspecies of A. coreana) The medicine is distinguished from bái zhú (白 术 or 白 朮, white atractylodes rhizome from Atractylodes macrocephala), which is typically cultivated, whereas cāng zhú more often tends to be collected from the wild. It is believed that the distinction between cāng zhú and bái zhú emerged in relatively modern times; a single drug "zhú" described in the Shen nong ben cao jing probably included many Atractylodes species.
Production
Cultivation
A. lancea is grown mainly in Hubei and Jiangsu.
A. chinensis and A. japonica are grown in Jilin, Liaoning, Shandong, Inner Mongolia, and Korea.
Harvesting
The rhizome is dug up in the spring. After cleaning, it can be sliced and stir baked to a yellow brown color.
Traditional attributes
In traditional Chinese medicine the herb is described as spicy or pungent, bitter, warm, and aromatic, acting on the spleen and stomach meridians.
A number of effects of the herb are described as ways of "drying dampness":
As a stomachic - for "Damp obstruction or accumulation in the Middle Jiao", with symptoms such as low appetite, abdominal distension, epigastric distress and fullness, indigestion, dyspepsia, diarrhea, nausea and vomiting, weariness, a heavy sensation in the body, and a thick greasy tongue coating.
To eliminate exopathogens - to "dispel wind-damp-cold (bi-syndrome)", explained as "headaches and body aches, fever, chills, blocked nasal passages, and an absence of sweating"
To treat "damp heat conditions" in the lower Jiao, including "Damp Leg Qi, aching and swollen joints, and vaginal discharge" (leukorrhea). This includes r |
https://en.wikipedia.org/wiki/Delaware%20lunar%20sample%20displays | The Delaware lunar sample displays are two commemorative plaques consisting of small fragments of Moon specimen brought back with the Apollo 11 and Apollo 17 lunar missions and given in the 1970s to the people of the state of Delaware by United States President Richard Nixon as goodwill gifts.
Description
Apollo 11
Apollo 17
History
The Delaware Apollo 11 "goodwill Moon rocks" plaque display was stolen in 1976.
The Delaware Apollo 17 lunar samples plaque display is held in storage by the Delaware Division of Historical and Cultural Affairs. Exhibits of the display are rare.
See also
List of Apollo lunar sample displays |
https://en.wikipedia.org/wiki/Director%20string | In mathematics, in the area of lambda calculus and computation, directors or director strings are a mechanism for keeping track of the free variables in a term. Loosely speaking, they can be understood as a kind of memoization for free variables; that is, as an optimization technique for rapidly locating the free variables in a term algebra or in a lambda expression. Director strings were introduced by Kennaway and Sleep in 1982 and further developed by Sinot, Fernández and Mackie as a mechanism for understanding and controlling the computational complexity cost of beta reduction.
Motivation
In beta reduction, one defines the value of the expression on the left to be that on the right:
or (Replace all x in E(body) by y)
While this is a conceptually simple operation, the computational complexity of the step can be non-trivial: a naive algorithm would scan the expression E for all occurrences of the free variable x. Such an algorithm is clearly O(n) in the length of the expression E. Thus, one is motivated to somehow track the occurrences of the free variables in the expression. One may attempt to track the position of every free variable, wherever it may occur in the expression, but this can clearly become very costly in terms of storage; furthermore, it provides a level of detail that is not really needed. Director strings suggest that the correct model is to track free variables in a hierarchical fashion, by tracking their use in component terms.
Definition
Consider, for simplicity, a term algebra, that is, a collection of free variables, constants, and operators which may be freely combined. Assume that a term t takes the form
where f is a function, of arity n, with no free variables, and the are terms that may or may not contain free variables. Let V denote the set of all free variables that may occur in the set of all terms. The director is then the map
from the free variables to the power set of the set . The values taken by are simply a list of |
https://en.wikipedia.org/wiki/Premature%20convergence | In evolutionary algorithms (EA), the term of premature convergence means that a population for an optimization problem converged too early, resulting in being suboptimal. In this context, the parental solutions, through the aid of genetic operators, are not able to generate offspring that are superior to, or outperform, their parents. Premature convergence is a common problem found in evolutionary algorithms in general and genetic algorithms in particular, as it leads to a loss, or convergence of, a large number of alleles, subsequently making it very difficult to search for a specific gene in which the alleles were present. An allele is considered lost if, in a population, a gene is present, where all individuals are sharing the same value for that particular gene. An allele is, as defined by De Jong, considered to be a converged allele, when 95% of a population share the same value for a certain gene (see also convergence).
Strategies for preventing premature convergence
Strategies to regain genetic variation can be:
a mating strategy called incest prevention,
uniform crossover,
favored replacement of similar individuals (preselection or crowding),
segmentation of individuals of similar fitness (fitness sharing),
increasing population size.
The genetic variation can also be regained by mutation though this process is highly random.
One way to reduce the risk of premature convergence is to use structured populations instead of the commonly used panmictic ones, see below.
Identification of the occurrence of premature convergence
It is hard to determine when premature convergence has occurred, and it is equally hard to predict its presence in the future. One measure is to use the difference between the average and maximum fitness values, as used by Patnaik & Srinivas, to then vary the crossover and mutation probabilities. Population diversity is another measure which has been extensively used in studies to measure premature convergence. However, although i |
https://en.wikipedia.org/wiki/SahysMod | SahysMod is a computer program for the prediction of the salinity of soil moisture, groundwater and drainage water, the depth of the watertable, and the drain discharge in irrigated agricultural lands, using different hydrogeologic and aquifer conditions, varying water management options, including the use of ground water for irrigation, and several crop rotation schedules, whereby the spatial variations are accounted for through a network of polygons.
Rationale
There is a need for a computer program that is easier to operate and that requires a simpler data structure then most currently available models. Therefore, the SahysMod program was designed keeping in mind a relative simplicity of operation to facilitate the use by field technicians, engineers and project planners instead of specialized geo-hydrologists.
It aims at using input data that are generally available, or that can be estimated with reasonable accuracy, or that can be measured with relative ease. Although the calculations are done numerically and have to be repeated many times, the final results can be checked by hand using the formulas in this manual.
SahysMod's objective is to predict the long-term hydro-salinity in terms of general trends, not to arrive at exact predictions of how, for example, the situation would be on the first of April in ten years from now.
Further, SahysMod gives the option of the re-use of drainage and well water (e.g. for irrigation) and it can account for farmers' responses to waterlogging, soil salinity, water scarcity and over-pumping from the aquifer. Also it offers the possibility to introduce subsurface drainage systems at varying depths and with varying capacities so that they can be optimized.
Other features of SahysMod are found in the next section.
Methods
Calculation of aquifer conditions in polygons
The model calculates the ground water levels and the incoming and outgoing ground water flows between the polygons by a numerical solution of the well-known |
https://en.wikipedia.org/wiki/Hierarchical%20temporal%20memory | Hierarchical temporal memory (HTM) is a biologically constrained machine intelligence technology developed by Numenta. Originally described in the 2004 book On Intelligence by Jeff Hawkins with Sandra Blakeslee, HTM is primarily used today for anomaly detection in streaming data. The technology is based on neuroscience and the physiology and interaction of pyramidal neurons in the neocortex of the mammalian (in particular, human) brain.
At the core of HTM are learning algorithms that can store, learn, infer, and recall high-order sequences. Unlike most other machine learning methods, HTM constantly learns (in an unsupervised process) time-based patterns in unlabeled data. HTM is robust to noise, and has high capacity (it can learn multiple patterns simultaneously). When applied to computers, HTM is well suited for prediction, anomaly detection, classification, and ultimately sensorimotor applications.
HTM has been tested and implemented in software through example applications from Numenta and a few commercial applications from Numenta's partners.
Structure and algorithms
A typical HTM network is a tree-shaped hierarchy of levels (not to be confused with the "layers" of the neocortex, as described below). These levels are composed of smaller elements called regions (or nodes). A single level in the hierarchy possibly contains several regions. Higher hierarchy levels often have fewer regions. Higher hierarchy levels can reuse patterns learned at the lower levels by combining them to memorize more complex patterns.
Each HTM region has the same basic function. In learning and inference modes, sensory data (e.g. data from the eyes) comes into bottom-level regions. In generation mode, the bottom level regions output the generated pattern of a given category. The top level usually has a single region that stores the most general and most permanent categories (concepts); these determine, or are determined by, smaller concepts at lower levels—concepts that are more res |
https://en.wikipedia.org/wiki/Sysctl | sysctl is a software utility of some Unix-like operating systems that reads and modifies the attributes of the system kernel such as its version number, maximum limits, and security settings. It is available both as a system call for compiled programs, and an administrator command for interactive use and scripting. Linux additionally exposes sysctl as a virtual file system.
BSD
In BSD, these parameters are generally objects in a management information base (MIB) that describe tunable limits such as the size of a shared memory segment, the number of threads the operating system will use as an NFS client, or the maximum number of processes on the system; or describe, enable or disable behaviors such as IP forwarding, security restrictions on the superuser (the "securelevel"), or debugging output.
In OpenBSD and DragonFly BSD, sysctl is also used as the transport layer for the hw.sensors framework for hardware monitoring, whereas NetBSD uses the ioctl system call for its sysmon envsys counterpart. Both sysctl and ioctl are the two system calls which can be used to add extra functionality to the kernel without adding yet another system call; for example, in 2004 with OpenBSD 3.6, when the tcpdrop utility was introduced, sysctl was used as the underlying system call. In FreeBSD, although there is no sensors framework, the individual temperature and other sensors are still commonly exported through the sysctl tree through Newbus, for example, as is the case with the aibs(4) driver that's available in all the 4 BSD systems, including FreeBSD.
In BSD, a system call or system call wrapper is usually provided for use by programs, as well as an administrative program and a configuration file (for setting the tunable parameters when the system boots).
This feature first appeared in 4.4BSD. It has the advantage over hardcoded constants that changes to the parameters can be made dynamically without recompiling the kernel.
Historically, although kernel variables themselves |
https://en.wikipedia.org/wiki/Australian%20Society%20for%20Biochemistry%20and%20Molecular%20Biology | The Australian Society for Biochemistry and Molecular Biology (ASBMB) is an academic society founded in 1955. Originally named Australian Biochemical Society, it was renamed to its current title in 1990. Its main activities include hosting scientific conferences, supporting ancillary symposia, workshops and publishing an educational magazine.
History
Biochemical research in Australia began in the 1920s in the Australian national science agency Council for Scientific and Industrial Research (now the CSIRO). The first university biochemistry department then started at the University of Adelaide under Thorburn Brailsford Robertson. There was initially a joint biochemistry and physiology section within the Australian and New Zealand Association for the Advancement of Science.
The society began in 1955 as the Australian Biochemical Society, with Rudi Lemberg as its founding president. It was based on Lemberg's experience with the British Biochemical Society and Hugh Ennor's meetings with the International Union of Biochemistry and relevant university department heads in Australia. Additional key initial members included Frederick Collins as treasurer and Victor Trikojus in a recruitment role.
Meetings and activities
ASBMB has hosted a yearly meeting each year since 1956. It has also coordinated the joint ComBio meeting with societies in related research fields since 1999. It also supports smaller special interest group meetings, symposia, workshops, conferences, and school science competitions.
Publications
The society publishes the magazine Australian Biochemist with three issues per year. The publication started in 1998, following on from the ABS/ASBMB newsletter, which was started in 1970.
Awards
ASBMB gives out an array of annual awards in different categories.
Lemberg Medal - lifetime contribution
Shimadzu Research Medal - within 15 years post-PhD graduation
Eppendorf Edman ECR Award - within 7 years post-PhD
SDR Scientific Education Award - education |
https://en.wikipedia.org/wiki/Electropositive%20shark%20repellent | Electropositive metals (EPMs) are a new class of shark repellent materials that produce a measurable voltage when immersed in an electrolyte such as seawater. The voltages produced are as high as 1.75 VDC in seawater. It is hypothesized that this voltage overwhelms the ampullary organ in sharks, producing a repellent action. Since bony fish lack the ampullary organ, the repellent is selective to sharks and rays. The process is electrochemical, so no external power input is required. As chemical work is done, the metal is lost in the form of corrosion. Depending on the alloy or metal utilized and its thickness, the electropositive repellent effect lasts up to 48 hours. The reaction of the electropositive metal in seawater produces hydrogen gas bubbles and an insoluble nontoxic hydroxide as a precipitate which settles downward in the water column.
History
SharkDefense made the discovery of electrochemical shark repellent effects on May 1, 2006 at South Bimini, Bahamas at the Bimini Biological Field Station. An electropositive metal, which was a component of a permanent magnet, was chosen as an experimental control for a tonic immobility experiment by Eric Stroud using a juvenile lemon shark (Negaprion brevirostris). It was anticipated that this metal would produce no effect, since it was not ferromagnetic. However, a violent rousing response was observed when the metal was brought within 50 cm of the shark's nose. The experiment was repeated with three other juvenile lemon sharks and two other juvenile nurse sharks (Ginglymostoma cirratum), and care was taken to eliminate all stray metal objects in the testing site. Patrick Rice, Michael Herrmann, and Eric Stroud were present at this first trial. Mike Rowe, from Discovery Channel’s Dirty Jobs series, subsequently witnessed and participated in a test using an electropositive metal within 24 hours after the discovery.
In the next three months, a variety of transition metals, lanthanides, post-transition metals, |
https://en.wikipedia.org/wiki/Wiring%20diagram | A wiring diagram is a simplified conventional pictorial representation of an electrical circuit. It shows the components of the circuit as simplified shapes, and the power and signal connections between the devices.
A wiring diagram usually gives information about the relative position and arrangement of devices and terminals on the devices, to help in building or servicing the device. This is unlike a schematic diagram, where the arrangement of the components' interconnections on the diagram usually does not correspond to the components' physical locations in the finished device. A pictorial diagram would show more detail of the physical appearance, whereas a wiring diagram uses a more symbolic notation to emphasize interconnections over physical appearance.
A wiring diagram is often used to troubleshoot problems and to make sure that all the connections have been made and that everything is present.
Architectural wiring diagrams
Architectural wiring diagrams show the approximate locations and interconnections of receptacles, lighting, and permanent electrical services in a building. Interconnecting wire routes may be shown approximately, where particular receptacles or fixtures must be on a common circuit.
Wiring diagrams use standard symbols for wiring devices, usually different from those used on schematic diagrams. The electrical symbols not only show where something is to be installed, but also what type of device is being installed. For example, a surface ceiling light is shown by one symbol, a recessed ceiling light has a different symbol, and a surface fluorescent light has another symbol. Each type of switch has a different symbol and so do the various outlets. There are symbols that show the location of smoke detectors, the doorbell chime, and thermostat. On large projects symbols may be numbered to show, for example, the panel board and circuit to which the device connects, and also to identify which of several types of fixture are to be installe |
https://en.wikipedia.org/wiki/H-infinity%20loop-shaping | H-infinity loop-shaping is a design methodology in modern control theory. It combines the traditional intuition of classical control methods, such as Bode's sensitivity integral, with H-infinity optimization techniques to achieve controllers whose stability and performance properties hold despite bounded differences between the nominal plant assumed in design and the true plant encountered in practice. Essentially, the control system designer describes the desired responsiveness and noise-suppression properties by weighting the plant transfer function in the frequency domain; the resulting 'loop-shape' is then 'robustified' through optimization. Robustification usually has little effect at high and low frequencies, but the response around unity-gain crossover is adjusted to maximise the system's stability margins. H-infinity loop-shaping can be applied to multiple-input multiple-output (MIMO) systems.
H-infinity loop-shaping can be carried out using commercially available software.
H-infinity loop-shaping has been successfully deployed in industry. In 1995, R. Hyde, K. Glover and G. T. Shanks published a paper describing the successful application of the technique to a VTOL aircraft. In 2008, D. J. Auger, S. Crawshaw and S. L. Hall published another paper describing a successful application to a steerable marine radar tracker, noting that the technique had the following benefits:
Easy to apply – commercial software handles the hard math.
Easy to implement – standard transfer functions and state-space methods can be used.
Plug and play – no need for re-tuning on an installation-by-installation basis.
A closely related design methodology, developed at about the same time, was based on the theory of the gap metric. It was applied in 1993 for designing controllers to dampen vibrations in large flexible structures at Wright-Patterson Air Force Base and Jet Propulsion Laboratory
See also
Control theory
H-infinity control |
https://en.wikipedia.org/wiki/Gold%20plating | Gold plating is a method of depositing a thin layer of gold onto the surface of another metal, most often copper or silver (to make silver-gilt), by chemical or electrochemical plating. This article covers plating methods used in the modern electronics industry; for more traditional methods, often used for much larger objects, see gilding.
Types
There are several types of gold plating used in the electronics industry:
Soft, pure gold plating is used in the semiconductor industry. The gold layer is easily soldered and wire bonded. Its Knoop hardness ranges between 60 and 85. The plating baths have to be kept free of contamination.
Soft, pure gold is deposited from special electrolytes. Entire printed circuit boards can be plated. This technology can be used for depositing layers suitable for wire bonding.
Bright hard gold on contacts, with Knoop hardness between 120–300 and purity of 99.7–99.9% gold. Often contains a small amount of nickel and/or cobalt; these elements interfere with die bonding, therefore the plating baths cannot be used for semiconductors.
Bright hard gold on printed circuit board tabs is deposited using lower concentration of gold in the baths. Usually contains nickel and/or cobalt as well. Edge connectors are often made by controlled-depth immersion of only the edge of the boards.
Gold plating chemistry
There are five recognized classes of gold plating chemistry:
Alkaline gold cyanide, for gold and gold alloy plating
Neutral gold cyanide, for high-purity plating
Acid gold plating for bright hard gold and gold alloy plating
Non-cyanide, generally sulphite or chloride-based for gold and gold alloy plating
Miscellaneous
Jewelry
Gold plating of silver is used in the manufacture of jewelry. The thickness of gold plating on jewellery is noted in microns (or micro-meters). The microns of thickness determines how long the gold plating lasts with usage. The jewellery industry denotes different qualities of gold plating in the following ter |
https://en.wikipedia.org/wiki/ZNF207 | Zinc finger protein 207 is a protein in humans that is encoded by the ZNF207 gene. |
https://en.wikipedia.org/wiki/Cryptoloop | Cryptoloop is a Linux kernel's disk encryption module that relies on the Crypto API, which is a cryptography framework introduced in version 2.5.45 of the Linux kernel mainline. Cryptoloop was first introduced in the 2.5.x kernel series; its functionality was later incorporated into the device mapper, a generic framework used to map one block device onto another.
Cryptoloop can create an encrypted file system within a partition or from within a regular file in the regular file system. Once a file is encrypted, it can be moved to another storage device. This is accomplished by making use of a loop device, a pseudo device that enables a normal file to be mounted as if it were a physical device. By encrypting I/O to the loop device, any data being accessed must first be decrypted before passing through the regular file system; conversely, any data being stored will be encrypted.
Cryptoloop is vulnerable to watermarking attacks, making it possible to determine presence of watermarked data on the encrypted filesystem:
This attack exploits weakness in IV computation and knowledge of how file systems place files on disk. This attack works with file systems that have soft block size of 1024 or greater. At least ext2, ext3, reiserfs and minix have such property. This attack makes it possible to detect presence of specially crafted watermarked files. Watermarked files contain special bit patterns that can be detected without decryption.
Newer versions of cryptoloop's successor, dm-crypt, are less vulnerable to this type of attack if used correctly.
See also
Comparison of disk encryption software
Disk encryption |
https://en.wikipedia.org/wiki/Spring%20house | A spring house, or springhouse, is a small building, usually of a single room, constructed over a spring. While the original purpose of a springhouse was to keep the spring water clean by excluding fallen leaves, animals, etc., the enclosing structure was also used for refrigeration before the advent of ice delivery and, later, electric refrigeration. The water of the spring maintains a constant cool temperature inside the spring house throughout the year. Food that would otherwise spoil, such as meat, fruit, or dairy products, could be kept there, safe from animal depredations as well. Springhouses thus often also served as pumphouses, milkhouses, and root cellars.
The Tomahawk Spring spring house at Tomahawk, West Virginia, was listed on the National Register of Historic Places in 1994.
Gallery
See also
Ice house (building)
Smokehouse
Windcatcher |
https://en.wikipedia.org/wiki/Glycyrrhizol | Glycyrrhizol A is a prenylated pterocarpan and an isoflavonoid derivative. It is a compound isolated from the root of the Chinese licorice plant (Glycyrrhiza uralensis).
It may has in vitro antibacterial properties. In one study, the strongest antibacterial activity was observed against Streptococcus mutans, an organism known to cause tooth decay in humans. |
https://en.wikipedia.org/wiki/Manually%20Annotated%20Sub-Corpus | Manually Annotated Sub-Corpus (MASC) is a balanced subset of 500K words of written texts and transcribed speech drawn primarily from the Open American National Corpus (OANC). The OANC is a 15 million word (and growing) corpus of American English produced since 1990, all of which is in the public domain or otherwise free of usage and redistribution restrictions.
All of MASC includes manually validated annotations for logical structure (headings, sections, paragraphs, etc.), sentence boundaries, three different tokenizations with associated part of speech tags, shallow parse (noun and verb chunks), named entities (person, location, organization, date and time), and Penn Treebank syntax. Additional manually produced or validated annotations have been produced by the MASC project for portions of the sub-corpus, including full-text annotation for FrameNet frame elements and a 100K+ sentence corpus with WordNet 3.1 sense tags, of which one-tenth are also annotated for FrameNet frame elements. Annotations of all or portions of the sub-corpus for a wide variety of other linguistic phenomena have been contributed by other projects, including PropBank, TimeBank, MPQA opinion, and several others. Co-reference annotations and clause boundaries of the entire MASC corpus are scheduled to be released by the end of 2016.
WordNet sense annotations for all occurrences of 114 words are also included in the MASC distribution, as well as FrameNet annotations for 50-100 occurrences of each of the 114 words. The sentences with WordNet and FrameNet annotations are also distributed as a part of the MASC Sentence Corpus.
Genres
Unlike most freely available corpora including a wide variety of linguistic annotations, MASC contains a balanced selection of texts from a broad range of genres:
Annotations
At present, MASC includes seventeen different types of linguistic annotation (* = in production; ** currently available in original format only):
All MASC annotations, whether contribute |
https://en.wikipedia.org/wiki/Microscopic%20scale | The microscopic scale () is the scale of objects and events smaller than those that can easily be seen by the naked eye, requiring a lens or microscope to see them clearly. In physics, the microscopic scale is sometimes regarded as the scale between the macroscopic scale and the quantum scale. Microscopic units and measurements are used to classify and describe very small objects. One common microscopic length scale unit is the micrometre (also called a micron) (symbol: μm), which is one millionth of a metre.
History
Whilst compound microscopes were first developed in the 1590s, the significance of the microscopic scale was only truly established in the 1600s when Marcello Malphigi and Antonie van Leeuwenhoek microscopically observed frog lungs and microorganisms. As microbiology was established, the significance of making scientific observations at a microscopic level increased.
Published in 1665, Robert Hooke’s book Micrographia details his microscopic observations including fossils insects, sponges, and plants, which was possible through his development of the compound microscope. During his studies of cork, he discovered plant cells and coined the term ‘cell’.
Prior to the use of the micro- prefix, other terms were originally incorporated into the International metric system in 1795, such as centi- which represented a factor of 10^-2, and milli-, which represented a factor of 10^-3.
Over time the importance of measurements made at the microscopic scale grew, and an instrument named the Millionometre was developed by watch-making company owner Antoine LeCoultre in 1844. This instrument had the ability to precisely measure objects to the nearest micrometre.
The British Association for the Advancement of Science committee incorporated the micro- prefix into the newly established CGS system in 1873.
The micro- prefix was finally added to the official SI system in 1960, acknowledging measurements that were made at an even smaller level, denoting a factor of 10 |
https://en.wikipedia.org/wiki/Electromagnetic%20absorption%20by%20water | The absorption of electromagnetic radiation by water depends on the state of the water.
The absorption in the gas phase occurs in three regions of the spectrum. Rotational transitions are responsible for absorption in the microwave and far-infrared, vibrational transitions in the mid-infrared and near-infrared. Vibrational bands have rotational fine structure. Electronic transitions occur in the vacuum ultraviolet regions.
Liquid water has no rotational spectrum but does absorb in the microwave region. Its weak absorption in the visible spectrum results in the pale blue color of water.
Overview
The water molecule, in the gaseous state, has three types of transition that can give rise to absorption of electromagnetic radiation:
Rotational transitions, in which the molecule gains a quantum of rotational energy. Atmospheric water vapour at ambient temperature and pressure gives rise to absorption in the far-infrared region of the spectrum, from about 200 cm−1 (50 μm) to longer wavelengths towards the microwave region.
Vibrational transitions in which a molecule gains a quantum of vibrational energy. The fundamental transitions give rise to absorption in the mid-infrared in the regions around 1650 cm−1 (μ band, 6 μm) and 3500 cm−1 (so-called X band, 2.9 μm)
Electronic transitions in which a molecule is promoted to an excited electronic state. The lowest energy transition of this type is in the vacuum ultraviolet region.
In reality, vibrations of molecules in the gaseous state are accompanied by rotational transitions, giving rise to a vibration-rotation spectrum. Furthermore, vibrational overtones and combination bands occur in the near-infrared region. The HITRAN spectroscopy database lists more than 37,000 spectral lines for gaseous H216O, ranging from the microwave region to the visible spectrum.
In liquid water the rotational transitions are effectively quenched, but absorption bands are affected by hydrogen bonding. In crystalline ice the vibrational spec |
https://en.wikipedia.org/wiki/Chinese%20Journal%20of%20Chemical%20Physics | The Chinese Journal of Chemical Physics (CJCP) is a peer-reviewed journal published by the Chinese Physical Society and hosted by the American Institute of Physics. It publishes experimental, computational and theoretical research on the interdisciplinary fields of physics, chemistry, biology and materials sciences. The journal is currently edited by Xue-ming Yang (杨学明) of the Dalian Institute of Chemical Physics, Chinese Academy of Sciences. CJCP publishes around 120 articles per year via bimonthly issues and has an impact factor of 0.496 (2015).
History
CJCP was established in 1988. From 2006 to 2012 its content was hosted by IOP Publishing, and from 2013 by the American Institute of Physics (AIP). As of 2006, the journal started publishing abstracts in Chinese as well.
External links
Chinese Journal of Chemical Physics website |
https://en.wikipedia.org/wiki/Gilbert%20Ames%20Bliss | Gilbert Ames Bliss, (9 May 1876 – 8 May 1951), was an American mathematician, known for his work on the calculus of variations.
Life
Bliss grew up in a Chicago family that eventually became affluent; in 1907, his father became president of the company supplying all of Chicago's electricity. The family was not affluent, however, when Bliss entered the University of Chicago in 1893 (its second year of operation). Hence he had to support himself while a student by winning a scholarship, and by playing in a student professional mandolin quartet.
After obtaining the B.Sc. in 1897, he began graduate studies at Chicago in mathematical astronomy (his first publication was in that field), switching in 1898 to mathematics. He discovered his life's work, the calculus of variations, via the lecture notes of Weierstrass's 1879 course, and Bolza's teaching. Bolza went on to supervise Bliss's Ph.D. thesis, The Geodesic Lines on the Anchor Ring, completed in 1900 and published in the Annals of Mathematics in 1902. After two years as an instructor at the University of Minnesota, Bliss spent the 1902–03 academic year at the University of Göttingen, interacting with Felix Klein, David Hilbert, Hermann Minkowski, Ernst Zermelo, Erhard Schmidt, Max Abraham, and Constantin Carathéodory.
Upon returning to the United States, Bliss taught one year each at the University of Chicago and the University of Missouri. In 1904, he published two more papers on the calculus of variations in the Transactions of the American Mathematical Society. Bliss was a Preceptor at Princeton University, 1905–08, joining a strong group of young mathematicians that included Luther P. Eisenhart, Oswald Veblen, and Robert Lee Moore. While at Princeton he was also an associate editor of the Annals of Mathematics.
In 1908, Chicago's Maschke died and Bliss was hired to replace him; Bliss remained at Chicago until his 1941 retirement. While at Chicago, he was an editor of the Transactions of the American Mathematica |
https://en.wikipedia.org/wiki/Server%20Base%20System%20Architecture | The Server Base System Architecture (SBSA) is a hardware system architecture for servers based on 64-bit ARM processors.
Rationale
Historically, ARM-based products have often been tailored for specific applications and power profiles. Variation between ARM-based hardware platforms has been an impediment requiring operating system adjustments for each product.
The SBSA seeks to strengthen the ARM ecosystem by specifying a minimal set of standardized features so that an OS built for this standard platform should function correctly without modification on all hardware products compliant with the specification.
Features
CPU features
Memory management
Peripheral access
Interrupts
Watchdog (errant system detection)
Existing specifications for USB, PCIe, ACPI, TPM, and other standards are incorporated to solidify the specification.
Server Base Boot Requirements
Firmware issues are addressed separately in the Server Base Boot Requirements (SBBR) specification.
Platform validation
The Architecture Compliance Suite (ACS) checks whether an environment is compliant with the SBSA specification, and is provided under an Apache 2 open source license. It is available at https://github.com/ARM-software/sbsa-acs.
Compliance levels
The specification defines levels of compliance, with level 0 being the most basic, and successive levels building on prior levels. In the words of the spec, "Unless explicitly stated, all specification items belonging to level N apply to levels greater than N."
Level 0, 1, and 2
Levels 0, 1, and 2 have been deprecated and folded into level 3.
Level 3
Level 3 contains base-level specifications for:
PE (Processing Element--a core) features
Memory map
Interrupt controller
PPI (peripheral interrupt) assignments
MMU behavior
Clock and timer subsystem
Wake up semantics
Power state semantics
Watchdogs
Peripheral subsystems
Level 4
Extends level 3, e.g. with support for RAS fault recovery extensions of ARMv8.2 spec.
Leve |
https://en.wikipedia.org/wiki/Proxy%20list | A proxy list is a list of open HTTP/HTTPS/SOCKS proxy servers all on one website. Proxies allow users to make indirect network connections to other computer network services. Proxy lists include the IP addresses of computers hosting open proxy servers, meaning that these proxy servers are available to anyone on the internet. Proxy lists are often organized by the various proxy protocols the servers use. Many proxy lists index, which can be used without changing browser settings.
Proxy Anonymity Levels
Elite proxies - Such proxies do not change request fields and look like a real browser, and your real IP address is hidden. Server administrators will commonly be fooled into believing that you are not using a proxy.
Anonymous proxies - These proxies do not show a real IP address, however, they do change the request fields, therefore it is very easy to detect that a proxy is being used by log analysis. You are still anonymous, but some server administrators may restrict proxy requests.
Transparent proxies - (not anonymous, simply HTTP) - These change the request fields and they transfer the real IP. Such proxies are not applicable for security or privacy uses while surfing the web, and should only be used for network speed improvement.
SOCKS is a protocol that relays TCP sessions through a firewall host to allow application users transparent access across the firewall. Because the protocol is independent of application protocols, it can be (and has been) used for many different services, such as telnet, FTP, finger, whois, gopher, WWW, etc. Access control can be applied at the beginning of each TCP session; thereafter the server simply relays the data between the client and the application server, incurring minimum processing overhead. Since SOCKS never has to know anything about the application protocol, it should also be easy for it to accommodate applications that use encryption to protect their traffic from nosy snoopers. No information about the client is se |
https://en.wikipedia.org/wiki/INTLAB | INTLAB (INTerval LABoratory) is an interval arithmetic library using MATLAB and GNU Octave, available in Windows and Linux, macOS. It was developed by S.M. Rump from Hamburg University of Technology. INTLAB was used to develop other MATLAB-based libraries such as VERSOFT and INTSOLVER, and it was used to solve some problems in the Hundred-dollar, Hundred-digit Challenge problems.
Version history
12/30/1998 Version 1
03/06/1999 Version 2
11/16/1999 Version 3
03/07/2002 Version 3.1
12/08/2002 Version 4
12/27/2002 Version 4.1
01/22/2003 Version 4.1.1
11/18/2003 Version 4.1.2
04/04/2004 Version 5
06/04/2005 Version 5.1
12/20/2005 Version 5.2
05/26/2006 Version 5.3
05/31/2007 Version 5.4
11/05/2008 Version 5.5
05/08/2009 Version 6
12/12/2012 Version 7
06/24/2013 Version 7.1
05/10/2014 Version 8
01/22/2015 Version 9
12/07/2016 Version 9.1
05/29/2017 Version 10
07/24/2017 Version 10.1
12/15/2017 Version 10.2
01/07/2019 Version 11
03/06/2020 Version 12
Functionality
INTLAB can help users to solve the following mathematical/numerical problems with interval arithmetic.
Works cited by INTLAB
INTLAB is based on the previous studies of the main author, including his works with co-authors.
External links
See also
List of numerical analysis software
Comparison of linear algebra libraries |
https://en.wikipedia.org/wiki/British%20Physics%20Olympiad | The British Physics Olympiad (BPhO) is a series of ten competitions in the study of physics for students in years 11–13 in the United Kingdom.
On a day-to-day basis, the BPhO is run by five committee members as well as a large team of volunteer physics teachers and academics from across the UK. The BPhO's administrative office is based in the Denys Wilkinson Building of the Department of Physics at the University of Oxford. The BPhO is a charitable trust.
History
Founded in 1979, the BPhO has been used as a qualifier for selection into the UK Physics Team for the International Physics Olympiad. The organisation aims to encourage the study of Physics and through ten annual Physics and Astronomy & Astrophysics competitions. Competitions include: Round 1, Round 2, Physics Challenge (Year 13), Experimental Project, Intermediate Physics Challenge (Year 11), Senior Physics Challenge (Y12) and BAAO.
BPhO Round 1
The BPhO Round 1 is a two-sectioned paper consisting of Section A and Section B, both lasting for 1 hour and 20 minutes (+5 mins reading for section B). Section A has only one question, with many independent parts that test the candidate on a variety of topics in physics. Section A can be awarded a maximum of 50 marks, even though the total marks is higher than 50.
Section B consists of several questions, each question involves a more specific topic in physics. Candidates are required to choose 2 questions out of Section B. The maximum marks awarded is also 50.
BPhO Round 2
The second round consists of several questions, in which all should be attempted. Total time allowed for the exam is 3 hours.
International Physics Olympiad
Following Round 2, the best students eligible to represent the UK at the International Physics Olympiad (IPhO) will be invited to attend the Training Camp. Problem solving skills will be developed, practical skills enhanced, as well as some coverage of new material (Thermodynamics, Relativity, etc.). At the Training Camp a practical |
https://en.wikipedia.org/wiki/Computer%20access%20control | In computer security, general access control includes identification, authorization, authentication, access approval, and audit. A more narrow definition of access control would cover only access approval, whereby the system makes a decision to grant or reject an access request from an already authenticated subject, based on what the subject is authorized to access. Authentication and access control are often combined into a single operation, so that access is approved based on successful authentication, or based on an anonymous access token. Authentication methods and tokens include passwords, biometric scans, physical keys, electronic keys and devices, hidden paths, social barriers, and monitoring by humans and automated systems.
Software entities
In any access-control model, the entities that can perform actions on the system are called subjects, and the entities representing resources to which access may need to be controlled are called objects (see also Access Control Matrix). Subjects and objects should both be considered as software entities, rather than as human users: any human users can only have an effect on the system via the software entities that they control.
Although some systems equate subjects with user IDs, so that all processes started by a user by default have the same authority, this level of control is not fine-grained enough to satisfy the principle of least privilege, and arguably is responsible for the prevalence of malware in such systems (see computer insecurity).
In some models, for example the object-capability model, any software entity can potentially act as both subject and object.
, access-control models tend to fall into one of two classes: those based on capabilities and those based on access control lists (ACLs).
In a capability-based model, holding an unforged-able reference or capability to an object, that provides access to the object (roughly analogous to how possession of one's house key grants one access to one's hous |
https://en.wikipedia.org/wiki/IXS%20Enterprise | IXS Enterprise is a conceptual interstellar superluminal spacecraft designed by NASA scientist Dr. Harold G. White, revealed at SpaceVision 2013, designed for the goal of achieving warp travel. The conceptual spacecraft would make use of a modified version of the Alcubierre drive. Dr. White is currently running the White–Juday warp-field interferometer experiment in order to develop a proof of concept for Alcubierre-style warp travel, when possible. The Alcubierre drive uses exotic matter (not to be confused with antimatter) to travel faster than light.
While the concept had been out since 2013 the design of IXS Enterprise was popularized in June 2014 after a series of media outlets reported on the conceptual artwork done by Dutch artist Mark Rademaker in collaboration with NASA. According to Mark Rademaker, over 1,600 hours have been spent on the conceptual artwork that he created.
In 2012, NASA reported that it was experimenting with the concept of warp drive and the loop hole within Einstein's Theory of Relativity. By 2014, it was announced that designer Mark Rademaker had created a CGI representation of a new vessel that would achieve warp velocity. The vessel he designed was the IXS Enterprise, named after the famed ship of the Star Trek franchise. The energy required to power the warp drive, according to White, is approximately the negative (negative energy is required for the Alcubierre drive concept to function) mass–energy equivalence of Voyager 1, which has a mass of approximately 700 kilograms. Using E=mc2, −700 kilograms of mass is equivalent to ~−63 exajoules of energy (this number is not definitive and can be further reduced). The ship has two thick outer rings (to reduce required energy) that generate the warp field—a contraction of space ahead, and expansion of space behind it. The space inside the rings is optimized to fit more space for cargo, crew and equipment.
See also
Breakthrough Starshot
Interstellar probe
Interstellar travel
Project |
https://en.wikipedia.org/wiki/Sentinus | Sentinus is a educational charity based in Lisburn, Northern Ireland that provides educational programs for young people interested in science, technology, engineering and mathematics (STEM).
History
Northern Ireland produces around 2,000 qualified IT workers each year; there are around 16,000 IT jobs in the Northern Ireland economy.
Function
It works with EngineeringUK and the Council for the Curriculum, Examinations & Assessment (CCEA). It works with primary and secondary schools in Northern Ireland.
It runs summer placements for IT workshops for those of sixth form age (16-18). It offers Robotics Roadshows for primary school children.
Sentinus Young Innovators
Sentinus hosts the annual Big Bang Northern Ireland Fair which incorporates Sentinus Young Innovators. This is a one day science and engineering project exhibition for post-primary students. It is one of largest such events in the United Kingdom. In 2019 over 3,000 students participated from 130 schools across both Northern Ireland and the Republic of Ireland.
The competition is affiliated with the International Science and Engineering Fair (ISEF) and the Broadcom MASTERS program. The overall winner represents Northern Ireland at the following year's ISEF.
Past Overall Winners
See also
Discover Science & Engineering, equivalent in the Republic of Ireland
Science Week Ireland
The Big Bang Fair
Young Scientist and Technology Exhibition |
https://en.wikipedia.org/wiki/Regular%20grid | A regular grid is a tessellation of n-dimensional Euclidean space by congruent parallelotopes (e.g. bricks).
Its opposite is irregular grid.
Grids of this type appear on graph paper and may be used in finite element analysis, finite volume methods, finite difference methods, and in general for discretization of parameter spaces. Since the derivatives of field variables can be conveniently expressed as finite differences, structured grids mainly appear in finite difference methods. Unstructured grids offer more flexibility than structured grids and hence are very useful in finite element and finite volume methods.
Each cell in the grid can be addressed by index (i, j) in two dimensions or (i, j, k) in three dimensions, and each vertex has coordinates in 2D or in 3D for some real numbers dx, dy, and dz representing the grid spacing.
Related grids
A Cartesian grid is a special case where the elements are unit squares or unit cubes, and the vertices are points on the integer lattice.
A rectilinear grid is a tessellation by rectangles or rectangular cuboids (also known as rectangular parallelepipeds) that are not, in general, all congruent to each other. The cells may still be indexed by integers as above, but the mapping from indexes to vertex coordinates is less uniform than in a regular grid. An example of a rectilinear grid that is not regular appears on logarithmic scale graph paper.
A skewed grid is a tessellation of parallelograms or parallelepipeds. (If the unit lengths are all equal, it is a tessellation of rhombi or rhombohedra.)
A curvilinear grid or structured grid is a grid with the same combinatorial structure as a regular grid, in which the cells are quadrilaterals or [general] cuboids, rather than rectangles or rectangular cuboids.
See also |
https://en.wikipedia.org/wiki/Stem%20cell%20secretome | The stem cell secretome (also referred to as the stromal cell secretome) is a collective term for the paracrine soluble factors produced by stem cells and utilized for their inter-cell communication. In addition to inter-cell communication, the paracrine factors are also responsible for tissue development, homeostasis and (re-)generation. The stem cell secretome consists of extracellular vesicles, specifically exosomes, microvesicles, membrane particles, peptides and small proteins (cytokines). The paracrine activity of stem cells, i.e. the stem cell secretome, has been found to be the predominant mechanism by which stem cell-based therapies mediate their effects in degenerative, auto-immune and/or inflammatory diseases. Though not only stem cells possess a secretome which influences their cellular environment, their secretome currently appears to be the most relevant for therapeutic use.
Extracellular Vesicles
The Extracellular Vesicles are small partials that are normally discharged and have boundaries that are formed by a lipid bilayer. Although cells can replicate, extracellular vesicle is not able to. In the extracellular vesical, things that consist of the stem cell secretome and are being packed are organelles, mRNA, miRNA, and proteins. Exosomes are discharged from the extracellular vesicles, which are found in biological fluid. Biological fluid like the cerebrospinal fluid, which can be used for treatment. Most impotently, exosomes can be found in between the eukaryotic organism's cell, also known as the tissue matrix.
Research
Stem Cell therapies, here referred to as therapies employing non-hematopoietic, mesenchymal stem cells, have a wide range of potential therapeutic benefits for different diseases, most of which are currently investigated in clinical trials. Stem cell therapies can benefit as a regenerative medicine for patients that have or been diagnosed with disease that affect the mid part of the brain, strokes and heart disease, joint disease |
https://en.wikipedia.org/wiki/Tversky%20index | The Tversky index, named after Amos Tversky, is an asymmetric similarity measure on sets that compares a variant to a prototype. The Tversky index can be seen as a generalization of the Sørensen–Dice coefficient and the Jaccard index.
For sets X and Y the Tversky index is a number between 0 and 1 given by
Here, denotes the relative complement of Y in X.
Further, are parameters of the Tversky index. Setting produces the Tanimoto coefficient; setting produces the Sørensen–Dice coefficient.
If we consider X to be the prototype and Y to be the variant, then corresponds to the weight of the prototype and corresponds to the weight of the variant. Tversky measures with are of special interest.
Because of the inherent asymmetry, the Tversky index does not meet the criteria for a similarity metric. However, if symmetry is needed a variant of the original formulation has been proposed using max and min functions
.
,
,
This formulation also re-arranges parameters and . Thus, controls the balance between and in the denominator. Similarly, controls the effect of the symmetric difference versus in the denominator.
Notes
Eponymous indices
Index numbers
Measure theory
Similarity measures
Asymmetry |
https://en.wikipedia.org/wiki/Zarya%20of%20the%20Dawn | Zarya of the Dawn is a short comic book written by Kris Kashtanova and illustrated using Midjourney. It is illustrated entirely using artificial intelligence, which resulted in a copyright dispute.
Plot
Zarya awakens in an abandoned New York City with no memories. A postcard from a person named Rusty falls out of their pocket, allowing them to remember their name and home address. After returning home and getting new clothes, Zarya meets Raya, their "inter-world assistant", who tells Zarya that a mental health crisis in 2023 led to the almost complete destruction of life on earth. Raya then takes Zarya to Zatura World, the world of acceptance. There, Zarya meets a mysterious woman and learns to accept their feelings. When they return to Central Park, it is covered in greenhouses. Zarya remarks that "acceptance is the first step of letting go".
Copyright dispute
In September 2022, Kashtanova applied for the comic's copyright protection with the United States Copyright Office, but they did not disclose that the illustrations were created using Midjourney, an artificial intelligence image generator. The comic was granted copyright protection, but the Copyright Office initiated a proceeding to revoke the protection of the artwork after discovering the fact. The artwork's copyright protection was revoked in February 2023, and the Copyright Office explained that only human-created works can receive protection. Although the images themselves are not protected by copyright, the arrangement of the images and the text and story of the book are, as they are the creative work of Kashtanova and not the artificial intelligence.
See also
Artificial intelligence and copyright
Artificial intelligence art
Alice and Sparkle |
https://en.wikipedia.org/wiki/Plant%20for%20Pakistan | Plant for Pakistan (Plant4Pakistan), also known as 10 Billion Tree Tsunami, is a five-year project to plant 10 billion trees across Pakistan from 2018 to 2023. Prime Minister Imran Khan started the drive on 2 September 2018 with approximately 1.5 million trees planted on the first day. The campaign was based on the successful Billion Tree Tsunami campaign of the former Pakistan Tehreek-e-Insaf government, also led by Imran Khan, in the province of Khyber Pakhtunkhwa in 2014. This different initiative is often confused with the initiative of the Prime Minister Mian Muhammad Nawaz Sharif who launched a national campaign of Green Pakistan and also allocated budget funding from the PSDP.
In 2020, the program tripled its number of workers to 63,600 after being momentarily halted following the coronavirus pandemic, aiming to enlist those left unemployed by its economic consequences. Most of the work, which pays between 500 and 800 rupees (US$3–5) a day, takes place in rural areas, with people setting up nurseries, planting saplings, and serving as forest protection guards. The plan was awarded 7.5 billion rupees ($46m) in funding.
Saplings planted during the initiative included mulberry, acacia, moringa and other indigenous species.
Planting Billions of Trees
While Pakistan's economy is 135th in terms of greenhouse gas emissions, it is among the top ten countries in the world to be affected by global warming according to the annual report of Global Climate Risk Index of German Watch, which ranks Pakistan as eighth among the countries most at risk of climate change. As of November 2021, Lahore is the number one in the world in terms of air pollution. Major cities have experienced major heat waves and high levels of pollution in the 21st century, including Karachi and Islamabad.
Pakistan has a wide variety of ecosystems within its borders, including the Arabian sea, several deserts and major rivers, and more than seven thousand glaciers. Pakistan is also at the confluen |
https://en.wikipedia.org/wiki/Statistical%20regions%20of%20Serbia | The statistical regions of Serbia () are regulated by the Law of the Regional Development and the Law of the Official Statistics. Serbia is divided into five statistical regions which are chiefly used for statistical purposes, such as census data. The regions encompass one or multiple districts each.
Introduction
In 2009, National Assembly of Serbia adopted the Law on Equal Territorial Development that formed seven statistical regions in the territory of Serbia. The Law was amended on 7 April 2010, so that the number of regions was reduced to five. The previously formed region of Eastern Serbia was merged with Southern Serbia and the region of Šumadija was merged with Western Serbia.
The five statistical regions of Serbia are:
Vojvodina
Belgrade
Šumadija and Western Serbia
Southern and Eastern Serbia
Kosovo and Metohija
Statistical regional classification
In a bylaw from 2010, the Government of Serbia specified a nomenclature of statistic territorial units in the country. The act was an attempt to synchronize the existing statistical division of the country with the Nomenclature of Territorial Units for Statistics of the European Union. According to the act, an additional top level of grouping was introduced, with the territory of Serbia divided into two NUTS 1 regions:
Serbia-North, comprising
Vojvodina
Belgrade, and
Serbia-South, comprising
Šumadija and Western Serbia
Southern and Eastern Serbia
Kosovo and Metohija
The five statistical regions would therefore become NUTS level 2 regions, while the Districts of Serbia would correspond with NUTS level 3. However, the classification has remained largely in internal, and limited, use within Serbia. , it has not been sanctioned by the European Union. According to a 2011 whitepaper by ESPON, which discusses the possibility to include Albania, Serbia, Montenegro and Bosnia and Herzegovina into NUTS nomenclature, "the statistical NUTS1 and NUTS2 regions created by the government in order to meet the NUTS criter |
https://en.wikipedia.org/wiki/Segregation%20%28materials%20science%29 | In materials science, segregation is the enrichment of atoms, ions, or molecules at a microscopic region in a materials system. While the terms segregation and adsorption are essentially synonymous, in practice, segregation is often used to describe the partitioning of molecular constituents to defects from solid solutions, whereas adsorption is generally used to describe such partitioning from liquids and gases to surfaces. The molecular-level segregation discussed in this article is distinct from other types of materials phenomena that are often called segregation, such as particle segregation in granular materials, and phase separation or precipitation, wherein molecules are segregated in to macroscopic regions of different compositions. Segregation has many practical consequences, ranging from the formation of soap bubbles, to microstructural engineering in materials science, to the stabilization of colloidal suspensions.
Segregation can occur in various materials classes. In polycrystalline solids, segregation occurs at defects, such as dislocations, grain boundaries, stacking faults, or the interface between two phases. In liquid solutions, chemical gradients exist near second phases and surfaces due to combinations of chemical and electrical effects.
Segregation which occurs in well-equilibrated systems due to the instrinsic chemical properties of the system is termed equilibrium segregation. Segregation that occurs due to the processing history of the sample (but that would disappear at long times) is termed non-equilibrium segregation.
History
Equilibrium segregation is associated with the lattice disorder at interfaces, where there are sites of energy different from those within the lattice at which the solute atoms can deposit themselves. The equilibrium segregation is so termed because the solute atoms segregate themselves to the interface or surface in accordance with the statistics of thermodynamics in order to minimize the overall free energy of t |
https://en.wikipedia.org/wiki/Timeline%20of%20solar%20cells | In the 19th century, it was observed that the sunlight striking certain materials generates detectable electric current – the photoelectric effect. This discovery laid the foundation for solar cells. Solar cells have gone on to be used in many applications. They have historically been used in situations where electrical power from the grid was unavailable.
As the invention was brought out it made solar cells as a prominent utilization for power generation for satellites. Satellites orbit the Earth, thus making solar cells a prominent source for power generation through the sunlight falling on them. Solar cells are commonly used in satellites in today's times.
1800s
1839 - Edmond Becquerel observes the photovoltaic effect via an electrode in a conductive solution exposed to light.
1873 - Willoughby Smith finds that selenium shows photoconductivity.
1874 - James Clerk Maxwell writes to fellow mathematician Peter Tait of his observation that light affects the conductivity of selenium.
1877 - William Grylls Adams and Richard Evans Day observed the photovoltaic effect in solidified selenium, and published a paper on the selenium cell. 'The action of light on selenium,' in "Proceedings of the Royal Society, A25, 113.
1883 - Charles Fritts develops a solar cell using selenium on a thin layer of gold to form a device giving less than 1% efficiency.
1887 - Heinrich Hertz investigates ultraviolet light photoconductivity and discovers the photoelectric effect
1887 - James Moser reports dye sensitized photoelectrochemical cell.
1888 - Edward Weston receives patent US389124, "Solar cell," and US389125, "Solar cell."
1888–91 - Aleksandr Stoletov creates the first solar cell based on the outer photoelectric effect
1894 - Melvin Severy receives patent US527377, "Solar cell," and US527379, "Solar cell."
1897 - Harry Reagan receives patent US588177, "Solar cell."
1899 - Weston Bowser receives patent US598177, "Solar storage."
1900–1929
1901 - Philipp von Lenard |
https://en.wikipedia.org/wiki/EPICS | The Experimental Physics and Industrial Control System (EPICS) is a set of software tools and applications used to develop and implement distributed control systems to operate devices such as particle accelerators, telescopes and other large scientific facilities. The tools are designed to help develop systems which often feature large numbers of networked computers delivering control and feedback. They also provide SCADA capabilities.
History
EPICS was initially developed as the Ground Test Accelerator Controls System (GTACS) at Los Alamos National Laboratory (LANL) in 1988 by Bob Dalesio, Jeff Hill, et al. In 1989, Marty Kraimer from Argonne National Laboratory (ANL) came to work alongside the GTA controls team for 6 months, bringing his experience from his work on the Advanced Photon Source (APS) Control System to the project. The resulting software was renamed EPICS and was presented at the International Conference on Accelerator and Large Experimental Physics Control Systems (ICALEPCS) in 1991.
EPICS was originally available under a commercial license, with enhanced versions sold by Tate & Kinetic Systems. Licenses for collaborators were free, but required a legal agreement with LANL and APS. An EPICS community was established and development grew as more facilities joined in with the collaboration. In February 2004, EPICS became freely distributable after its release under the EPICS Open License.
It is now used and developed by over 50 large science institutions worldwide, as well as by several commercial companies.
Architecture
EPICS uses client–server and publish–subscribe techniques to communicate between computers. Servers, the “input/output controllers” (IOCs), collect experiment and control data in real time, using the measurement instruments attached to them. This information is then provided to clients, using the high-bandwidth Channel Access (CA) or the recently added pvAccess networking protocols that are designed to suit real-time applications |
https://en.wikipedia.org/wiki/Strahl%20%28astronomy%29 | Strahl, from the German for "ray" or "beam" such as radiation is, is the electron component of the solar wind that is the most closely aligned with the magnetic field of the Sun. These electrons are part of the high speed component of the solar wind and have sufficient velocity to escape from the electrostatic potential of the Sun. They carry the electron heat flux of the solar wind and always move away from the Sun along the magnetic field lines. Strahl are distinguished from the 'core' and 'halo' populations of electrons, which form part of the low and high speed solar winds and are normally anisotropic with respect to the Sun's magnetic field.
Electrons precipitating into the polar regions of the Sun form the source of the strahl. When the magnetic lines of force are connected with the interplanetary magnetic field (IMF) in the appropriate direction, the strahl electrons are transported from the solar corona with energies of up to several hundred eV. When the Sun's magnetic field lines are not connected with the IMF, the strahl effect is blocked at that pole. |
https://en.wikipedia.org/wiki/Chiral%20inversion | Chiral inversion is the process of conversion of one enantiomer of a chiral molecule to its mirror-image version with no other change in the molecule.
Chiral inversion happens depending on various factors (viz. biological-, solvent-, light-, temperature- induced, etc.) and the energy barrier associated with the stereogenic element present in the chiral molecule. 2-Arylpropionic acid nonsteroidal anti-inflammatory drugs (NSAIDs) provide one of the best pharmaceutical examples of chiral inversion. Chirality is attributed to a molecule due to the presence of a stereogenic element (viz. center, planar, helical, or axis). Many pharmaceutical drugs are chiral and have a labile (configurationally unstable) stereogenic element. Chiral compounds with stereogenic center are found to have high energy barriers for inversion and generally undergo biologically mediated chiral inversion. While compounds with helical or planar chirality have low energy barriers and chiral inversions are often caused by solvent, light, temperature. When this happens, the configuration of the chiral molecule may rapidly change reversibly or irreversibly depending on the conditions. The chiral inversion has been intensively studied in the context of the pharmacological and toxicological consequences. Other than NSAIDs, chiral drugs with different chemical structures can also show this effect.
Chiral drugs have different effects on the body depending on whether one enantiomer or both enantiomers act on different biological targets. As a result, chiral inversion can change how a pharmaceutical drug works in the body. From a pharmacological and toxicological point of view, it is very important to learn more about chiral inversion, the things that make it happen, and the tools used to figure out chiral inversion.
Types
Essentially there are two types of chiral inversion, unidirectional and bidirectional. Inversion process is dependent on species and substrate.
Unidirectional chiral inversion (enzym |
https://en.wikipedia.org/wiki/Laminarin | The molecule laminarin (also known as laminaran) is a storage glucan (a polysaccharide of glucose) found in brown algae. It is used as a carbohydrate food reserve in the same way that chrysolaminarin is used by phytoplankton, especially in diatoms. It is created by photosynthesis and is made up of β(1→3)-glucan with β(1→6)-branches. It is a linear polysaccharide, with a β(1→3):β(1→6) ratio of 3:1. Its hydrolysis is catalyzed by enzymes such as laminarinase (EC 3.2.1.6) that breaks the β(1→3) bonds. It has been suggested that the annual production of algae laminarin amounts to 12 ± 8 gigatons, i.e., about three times the annual atmospheric CO2 increase by fossil fuel burning, that its concentration is driven by light variability and that it contributes substantially to the carbon export from surface waters, as it may account for up to half of organic carbon in sinking diatom-containing particles. |
https://en.wikipedia.org/wiki/Premio%20Presidente%20della%20Repubblica%20%28prize%29 | The Premio Presidente della Repubblica is an Italian award introduced by the former president and academic Luigi Einaudi. Since 1949 it has been awarded on a regular basis by the Accademia dei Lincei, the Accademia di San Luca, and the Accademia Nazionale di Santa Cecilia. It is among the most distinguished awards of the three prestigious academies.
History
The award was established on 11 October 1948 by Luigi Einaudi with a letter to the president of the Lincei National Academy to continue the tradition of royal awards. The prize was first introduced to the class of physical, mathematical, and natural sciences and the class of moral, historical, and philological sciences.
In the same year, Einaudi established a national prize for artists and architects awarded by the academies of San Luca and Santa Cecilia. The prize is given by the President of Italy in charge in an official ceremony. Among the people awarded, there are several winners of other important awards such as the Nobel Prize, the Wolf Prize, and the Academy Award.
Prize recipients
See also
Accademia dei Lincei
Accademia di San Luca
Accademia Nazionale di Santa Cecilia
President of Italy |
https://en.wikipedia.org/wiki/Krogh%27s%20principle | Krogh's principle states that "for such a large number of problems there will be some animal of choice, or a few such animals, on which it can be most conveniently studied." This concept is central to those disciplines of biology that rely on the comparative method, such as neuroethology, comparative physiology, and more recently functional genomics.
History
Krogh's principle is named after the Danish physiologist August Krogh, winner of the Nobel Prize in Physiology for his contributions to understanding the anatomy and physiology of the capillary system, who described it in The American Journal of Physiology in 1929. However, the principle was first elucidated nearly 60 years prior to this, and in almost the same words as Krogh, in 1865 by Claude Bernard, the French instigator of experimental medicine, on page 27 of his "Introduction à l'étude de la médecine expérimentale":
Krogh wrote the following in his 1929 treatise on the then current 'status' of physiology (emphasis added):
"Krogh's principle" was not utilized as a formal term until 1975 when the biochemist Hans Adolf Krebs (who initially described the Citric Acid Cycle), first referred to it.
More recently, at the International Society for Neuroethology meeting in Nyborg, Denmark in 2004, Krogh's principle was cited as a central principle by the group at their 7th Congress. Krogh's principle has also been receiving attention in the area of functional genomics, where there has been increasing pressure and desire to expand genomics research to a more wide variety of organisms beyond the traditional scope of the field.
Philosophy and applications
A central concept to Krogh's principle is evolutionary adaptation. Evolutionary theory maintains that organisms are suited to particular niches, some of which are highly specialized for solving particular biological problems. These adaptations are typically exploited by biologists in several ways:
Methodology: (e.g. Taq polymerase and PCR): The need to manipulat |
https://en.wikipedia.org/wiki/Quorn | Quorn is a brand of meat substitute products, or the company that makes them. Quorn originated in the UK and is sold primarily in Europe, but is available in 14 countries. The brand is owned by parent company Monde Nissin.
Quorn is sold as both a cooking ingredient and as a meat substitute used in a range of prepackaged meals.
All Quorn foods contain mycoprotein as an ingredient, which is derived from the Fusarium venenatum fungus. In most Quorn products, the fungus culture is dried and mixed with egg albumen, which acts as a binder, and then is adjusted in texture and pressed into various forms. A vegan formulation also exists that uses potato protein as a binder instead of egg albumen.
History
Quorn was launched in 1985 by Marlow Foods, a joint venture between Rank Hovis McDougall (RHM) and Imperial Chemical Industries (ICI)
Microbial biomass is produced commercially as single-cell protein (SCP) for human food or animal feed and as viable yeast cells for the baking industry. The industrial production of bakers' yeast started in the early 1900s, and yeast biomass was used as human food in Germany during World War I. The development of large-scale processes for the production of microbial biomass as a source of commercial protein began in earnest in the late 1960s.
Several of the processes investigated did not come to fruition owing to political and economic problems, but the establishment of the ICI Pruteen process for the production of bacterial SCP for animal feed was a milestone in the development of the fermentation industry. This process used continuous culture on a large scale . The economics of the production of SCP as animal feed were marginal, which eventually led to the discontinuation of the Pruteen process.
The technical expertise gained from the Pruteen process assisted ICI in collaborating with company Rank Hovis McDougall on a process for the production of fungal biomass for human food. A continuous fermentation process for the production o |
https://en.wikipedia.org/wiki/Expanded%20Program%20on%20Immunization%20%28Philippines%29 | The Expanded Program on Immunization (EPI) in the Philippines began in July 1979. And, in 1986, made a response to the Universal Child Immunization goal. The four major strategies include:
sustaining high routine Full Immunized Child (FIC) coverage of at least 90% in all provinces and cities;
sustaining the polio-free country for global certification;
eliminating measles by 2008; and
eliminating neonatal tetanus by 2008.
Routine Schedule of Immunization
Every Wednesday is designated as immunization day and is adopted in all parts of the country. Immunization is done monthly in barangay health stations, quarterly in remote areas of the country.
Routine Immunization Schedule for Infants
The standard routine immunization schedule for infants in the Philippines is adopted to provide maximum immunity against the seven vaccine preventable diseases in the country before the child's first birthday. The fully immunized child must have completed BCG 1, DPT 1, DPT 2, DPT 3, OPV 1, OPV 2, OPV 3, HB 1, HB 2, HB 3 and measles vaccines before the child is 12 months of age.
General Principles in Infants/Children Immunization
Because measles kills, every infant needs to be vaccinated against measles at the age of 9 months or as soon as possible after 9 months as part of the routine infant vaccination schedule. It is safe to vaccinate a sick child who is suffering from a minor illness (cough, cold, diarrhea, fever or malnutrition) or who has already been vaccinated against measles.
If the vaccination schedule is interrupted, it is not necessary to restart. Instead, the schedule should be resumed using minimal intervals between doses to catch up as quickly as possible.
Vaccine combinations (few exceptions), antibiotics, low-dose steroids (less than 20 mg per day), minor infections with low fever (below 38.5º Celsius), diarrhea, malnutrition, kidney or liver disease, heart or lung disease, non-progressive encephalopathy, well controlled epilepsy or advanced age, are not co |
https://en.wikipedia.org/wiki/Kombucha | Kombucha (also tea mushroom, tea fungus, or Manchurian mushroom when referring to the culture; Latin name Medusomyces gisevii) is a fermented, lightly effervescent, sweetened black tea drink commonly consumed for its purported health benefits. Sometimes the beverage is called kombucha tea to distinguish it from the culture of bacteria and yeast. Juice, spices, fruit or other flavorings are often added.
Kombucha is thought to have originated in China, where the drink is traditional. By the early 20th century it spread to Russia, then other parts of Eastern Europe and Germany. Kombucha is now homebrewed globally, and also bottled and sold commercially. The global kombucha market was worth approximately billion .
Kombucha is produced by symbiotic fermentation of sugared tea using a symbiotic culture of bacteria and yeast (SCOBY) commonly called a "mother" or "mushroom". The microbial populations in a SCOBY vary. The yeast component generally includes Saccharomyces cerevisiae, along with other species; the bacterial component almost always includes Gluconacetobacter xylinus to oxidize yeast-produced alcohols to acetic acid (and other acids). Although the SCOBY is commonly called "tea fungus" or "mushroom", it is actually "a symbiotic growth of acetic acid bacteria and osmophilic yeast species in a zoogleal mat ". The living bacteria are said to be probiotic, one of the reasons for the popularity of the drink.
Numerous health benefits have been claimed to correlate with drinking kombucha; there is little evidence to support any of these claims. The beverage has caused rare serious adverse effects, possibly arising from contamination during home preparation. It is not recommended for therapeutic purposes.
History
Kombucha most likely originated in the Bohai Sea district in China. The drink was consumed in Russia and from there entered the rest of Europe. Its consumption increased in the United States during the early 21st century. Having an alcohol content of less th |
https://en.wikipedia.org/wiki/TB6Cs1H3%20snoRNA | TB6Cs1H3 is a member of the H/ACA-like class of non-coding RNA (ncRNA) molecule that guide the sites of modification of uridines to pseudouridines of substrate RNAs. It is known as a small nucleolar RNA (snoRNA) thus named because of its cellular localization in the nucleolus of the eukaryotic cell. TB6Cs1H3 is predicted to guide the pseudouridylation of SSU ribosomal RNA (rRNA) at residue Ψ662. |
https://en.wikipedia.org/wiki/Circuit%20quantum%20electrodynamics | Circuit quantum electrodynamics (circuit QED) provides a means of studying the fundamental interaction between light and matter (quantum optics). As in the field of cavity quantum electrodynamics, a single photon within a single mode cavity coherently couples to a quantum object (atom). In contrast to cavity QED, the photon is stored in a one-dimensional on-chip resonator and the quantum object is no natural atom but an artificial one. These artificial atoms usually are mesoscopic devices which exhibit an atom-like energy spectrum. The field of circuit QED is a prominent example for quantum information processing and a promising candidate for future quantum computation.
In the late 2010s decade, experiments involving cQED in 3 dimensions have demonstrated deterministic gate teleportation and other operations on multiple qubits.
Resonator
The resonant devices used for circuit QED are superconducting coplanar waveguide microwave resonators, which are two-dimensional microwave analogues of the Fabry–Pérot interferometer. Coplanar waveguides consist of a signal carrying centerline flanked by two grounded planes. This planar structure is put on a dielectric substrate by a photolithographic process. Superconducting materials used are mostly aluminium (Al) or niobium (Nb). Dielectrics typically used as substrates are either surface oxidized silicon (Si) or sapphire (Al2O3).
The line impedance is given by the geometric properties, which are chosen to match the 50 of the peripheric microwave equipment to avoid partial reflection of the signal.
The electric field is basically confined between the center conductor and the ground planes resulting in a very small mode volume which gives rise to very high electric fields per photon (compared to three-dimensional cavities). Mathematically, the field can be found as
,
where is the reduced Planck constant, is the angular frequency, and is the permittivity of free space.
One can distinguish between two different types o |
https://en.wikipedia.org/wiki/William%20W.%20Tait | William Walker Tait (born 1929) is an emeritus professor of philosophy at the University of Chicago, where he served as a faculty member from 1972 to 1996, and as department chair from 1981 to 1987.
Education and career
Tait received his B.A. from Lehigh University in 1952, and his Ph.D. from Yale University in 1958. Frederic Fitch served as his doctoral advisor.
Prior to teaching at Chicago he held positions at Stanford University from 1958 to 1964, the University of Illinois at Chicago from 1965 to 1971, and the University of Aarhus from 1971 to 1972. In 1966, he signed a tax resistance vow to protest the Vietnam War. In 2002, he was elected as a Fellow of the American Academy of Arts and Sciences.
Publications
Early analytic philosophy : Frege, Russell, Wittgenstein : essays in honor of Leonard Linsky / edited by William W. Tait. Chicago, Ill. : Open Court, c1997. vii, 291 p. : ill. ; 24 cm. (cloth : alk. paper), (pbk. : alk. paper)
Tait, William W. The provenance of pure reason : essays in the philosophy of mathematics and its history / William Tait. Oxford ; New York : Oxford University Press, c2005. viii, 332 p. : ill., ; 25 cm.
See also
American philosophy
List of American philosophers |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.