source stringlengths 31 227 | text stringlengths 9 2k |
|---|---|
https://en.wikipedia.org/wiki/Integrin%20alpha%204 | CD49d is an integrin alpha subunit. It makes up half of the α4β1 lymphocyte homing receptor.
Function
The product of this gene belongs to the integrin alpha chain family of proteins. Integrins are heterodimeric integral membrane proteins composed of an alpha chain and a beta chain. This gene encodes an alpha 4 chain. Unlike other integrin alpha chains, alpha 4 neither contains an I-domain, nor undergoes disulfide-linked cleavage. Alpha 4 chain associates with either beta 1 chain or beta 7 chain.
Interactions
CD49d has been shown to interact with LGALS8 and Paxillin.
See also
Carotegrast methyl, an integrin alpha 4 antagonist used for the treatment of ulcerative colitis |
https://en.wikipedia.org/wiki/Watson%20Wyatt%20Worldwide | Watson Wyatt Worldwide, Inc. was a global consulting firm that merged in January 2010 with Towers Perrin to form Towers Watson. The firm's services included managing the cost and effectiveness of employee benefit programs; developing attraction, retention and reward strategies; advising pension plan sponsors and other institutions on optimal investment strategies; providing strategic and financial advice to insurance and financial service companies; and delivering related technology, outsourcing and data services. Its principal operating subsidiary, Watson Wyatt & Company, was a human capital consulting firm with operations in the Americas, Europe and Asia Pacific. Its corporate offices were in Arlington, Virginia.
History
Reuben Watson formed the UK actuarial firm R. Watson & Sons in 1878. In 1943, Birchard Wyatt established The Wyatt Company as an actuarial consulting firm.
In 1995, the two firms formed a global alliance under the brand Watson Wyatt Worldwide. The U.S.-based arm of the alliance was renamed Watson Wyatt & Company and the UK firm was called Watson Wyatt LLP. In 2000, Watson Wyatt & Company completed a successful initial public offering and began trading on the New York Stock Exchange. In August 2005, the two firms formally merged and the new company was named Watson Wyatt Worldwide.
2007 transactions
During 2007,dear Watson Wyatt acquired three companies in Europe. It acquired its Netherlands-based alliance partner, Watson Wyatt Brans & Co. in February. In July, Watson Wyatt bought a German human resources consulting firm, Heissmann GmbH. It purchased Oakbridge Consulting Group in Sweden in October.
Also during 2007, Watson Wyatt acquired the talent management technology and consulting firm WisdomNet.
Also during 2007, Watson Wyatt announced that it would spin off its multiemployer retirement practices in the United States and Canada. (A multiemployer retirement plan is set up under the terms of collective bargaining agreements involving more t |
https://en.wikipedia.org/wiki/Heme%20O | Heme O (or haem O) differs from the closely related heme A by having a methyl group at ring position 8 instead of the formyl group. The isoprenoid chain at position 2 is the same.
Heme O, found in the bacterium Escherichia coli, functions in a similar manner to heme A in mammalian oxygen reduction.
See also
Heme |
https://en.wikipedia.org/wiki/Heme%20C | Heme C (or haem C) is an important kind of heme.
History
The correct structure of heme C was published, in mid 20th century, by the Swedish biochemist K.-G. Paul. This work confirmed the structure first inferred by the great Swedish biochemist Hugo Theorell. The structure of heme C, based upon NMR and IR experiments of the reduced, Fe(II), form of the heme, was confirmed in 1975. The structure of heme C including the absolute stereochemical configuration about the thioether bonds was first presented for the vertebrate protein, cytochrome c and is now extended to many other heme C containing proteins.
Properties
Heme C differs from heme B in that the two vinyl side chains of heme B are replaced by covalent, thioether linkages to the apoprotein. The two thioether linkages are typically made by cysteine residues of the protein. These linkages do not allow the heme C to easily dissociate from the holoprotein, cytochrome c, compared with the more easily dissociated heme B that may dissociate from the holoprotein, the heme-protein complex, even under mild conditions. This allows a very wide range of cytochrome c structure and function, with myriad c type cytochromes acting primarily as electron carriers. The redox potential for cytochrome c can also be "fine-tuned" by small changes in protein structure and solvent interaction.
The number of heme C units bound to a holoprotein is highly variable. For vertebrate cells one heme C per protein is the rule but for bacteria this number is often 2, 4, 5, 6 or even 16 heme C groups per holoprotein. It is generally agreed the number and arrangement of heme C groups are related and even required for proper holoprotein function. For instance, those proteins containing several heme C groups are involved with multiple electron transfer reactions, particularly important is the 6 electron reduction required to reduce atmospheric nitrogen into two organic ammonia molecules. It is common for the heme C to amino acid ratio to be h |
https://en.wikipedia.org/wiki/Heme%20B | Heme B or haem B (also known as protoheme IX) is the most abundant heme. Hemoglobin and myoglobin are examples of oxygen transport proteins that contain heme B. The peroxidase family of enzymes also contain heme B. The COX-1 and COX-2 enzymes (cyclooxygenase) of recent fame, also contain heme B at one of two active sites.
Generally, heme B is attached to the surrounding protein matrix (known as the apoprotein) through a single coordination bond between the heme iron and an amino-acid side-chain.
Both hemoglobin and myoglobin have a coordination bond to an evolutionarily-conserved histidine, while nitric oxide synthase and cytochrome P450 have a coordination bond to an evolutionarily-conserved cysteine bound to the iron center of heme B.
Since the iron in heme B containing proteins is bound to the four nitrogens of the porphyrin (forming a plane) and a single electron donating atom of the protein, the iron is often in a pentacoordinate state. When oxygen or the toxic carbon monoxide is bound the iron becomes hexacoordinated.
The correct structures of heme B and heme S were first elucidated by German chemist Hans Fischer. |
https://en.wikipedia.org/wiki/Heme%20A | Heme A (or haem A) is a heme, a coordination complex consisting of a macrocyclic ligand called a porphyrin, chelating an iron atom. Heme A is a biomolecule and is produced naturally by many organisms. Heme A, often appears a dichroic green/red when in solution, is a structural relative of heme B, a component of hemoglobin, the red pigment in blood.
Relationship to other hemes
Heme A differs from heme B in that a methyl side chain at ring position 8 is oxidized to a formyl group and a hydroxyethylfarnesyl group, an isoprenoid chain, has been attached to the vinyl side chain at ring position 2 of the iron tetrapyrrole heme. Heme A is similar to heme o, in that both have this farnesyl addition at position 2 but heme O does not have the formyl group at position 8, still containing the methyl group. The correct structure of heme A, based upon NMR and IR experiments of the reduced, Fe(II) form of the heme, was published in 1975. The structure was confirmed by synthesis of the dimethyl ester of the iron-free form.
History
Heme A was first isolated by the German biochemist Otto Warburg in 1951 and shown by him to be the active component of the integral membrane metalloprotein cytochrome c oxidase.
Stereochemistry
The final structural question of the exact geometric configuration about the first carbon at ring position 3 of ring I, the carbon bound to the hydroxyl group, has been shown to be the chiral S configuration.
Like heme B, heme A is often attached to the apoprotein through a coordinate bond between the heme iron and a conserved amino acid side-chain. In the important respiratory protein cytochrome c oxidase (CCO) this ligand 5 for the heme A at the oxygen reaction center is a histidyl group. Histidine is a common ligand for many hemeproteins including hemoglobin and myoglobin.
Heme A in the cytochrome a portion of cytochrome c oxidase, bound by two histidine residues (shown in pink)
An example of a metalloprotein that contains heme A is cytochrome c oxidase. |
https://en.wikipedia.org/wiki/GDF7 | Growth differentiation factor 7 (GDF7) is a protein that in humans is encoded by the GDF7 gene.
GDF7 belongs to the transforming growth factor beta superfamily that is specifically found in a signaling center known as the roof plate that is located in the developing nervous system of embryos. The roof plate is required for the generation of several classes of spinal cord dorsal interneurons; GDF7 specifically induces the formation of sensory neurons in the dorsal spinal cord from neural crest cells by generating signals within the roof plate.
GDF7 is also known as bone morphogenic protein 12 (BMP-12). |
https://en.wikipedia.org/wiki/New%20Zealand%20Journal%20of%20Ecology | The New Zealand Journal of Ecology is a biannual peer-reviewed scientific journal publishing ecological research relevant to New Zealand and the South Pacific. It has been published since 1952, firstly as a 1952 issue of New Zealand Science Review and then as the Proceedings of the New Zealand Ecological Society until 1977. The Journal is published by the New Zealand Ecological Society, and is covered by Current Contents/Agriculture, Biology and Environmental Science, GEOBASE, and Geo Abstracts. George Perry is the journal's current editor, with Katherine Russell as technical editor.
Free access is available to all issues in PDF format on the journal website. The compilation of PDF files from issues dating back to 1953 was funded by the New Zealand Government's Terrestrial and Freshwater Biodiversity Information System. |
https://en.wikipedia.org/wiki/Integrin%20alpha-1 | Integrin alpha-1 also CD49a is an integrin alpha subunit encoded in humans by the gene ITGA1. It makes up half of the α1β1 integrin duplex. Though CD49a can bind a number of ligands including collagen IV, collagen I, and others.
CD49a has been implicated as a marker of tissue resident memory T cells, where it may be coexpressed with other markers CD103 and CD69. It has been shown to affect the motility of T cells. |
https://en.wikipedia.org/wiki/Integrin%20alpha%202 | Integrin alpha-2, or CD49b (cluster of differentiation 49b), is a protein which in humans is encoded by the CD49b gene.
The CD49b protein is an integrin alpha subunit. It makes up half of the α2β1 integrin duplex. Integrins are heterodimeric integral membrane glycoproteins composed of a distinct alpha chain and a common beta chain. They are found on a wide variety of cell types including T cells (the NKT cells), NK cells, fibroblasts and platelets. Integrins are involved in cell adhesion and also participate in cell-surface-mediated signalling.
Expression of CD49b in conjunction with LAG-3 has been used to identify type 1 regulatory (Tr1) cells.
The DX5 monoclonal antibody recognizes mouse CD49b.
Interactions
CD49b has been shown to interact with MMP1. |
https://en.wikipedia.org/wiki/Integrin%20alpha%203 | Integrin alpha-3 is a protein that in humans is encoded by the ITGA3 gene.
ITGA3 is an integrin alpha subunit. Together with beta-1 subunit, it makes up half of the α3β1 integrin duplex that plays a role in neural migration and corticogenesis, acted upon by such factors as netrin-1 and reelin.
ITGA3 encodes the integrin alpha 3 chain. Integrins are heterodimeric integral membrane proteins composed of an alpha chain and a beta chain. Alpha chain 3 undergoes post-translational cleavage in the extracellular domain to yield disulfide-linked light and heavy chains that join with beta 1 to form an integrin that interacts with many extracellular matrix proteins.
Alternative names
The alpha 3 beta 1 integrin is known variously as: very late (activation) antigen 3 ('VLA-3'), very common antigen 2 ('VCA-2'), extracellular matrix receptor 1 ('ECMR1'), and galactoprotein b3 ('GAPB3').
Interactions
CD49c has been shown to interact with:
CD9
FHL2,
LGALS8, and
TSPAN4. |
https://en.wikipedia.org/wiki/Fritillaria%20pudica | Fritillaria pudica, the yellow fritillary, is a small perennial plant found in the sagebrush country in the western United States (Idaho, Montana, Oregon, Washington, Wyoming, very northern California, Nevada, northwestern Colorado, North Dakota and Utah) and Canada (Alberta and British Columbia). It is a member of the lily family Liliaceae. Another common (but somewhat ambiguous) name is "yellow bells", since it has a bell-shaped yellow flower. It may be found in dryish, loose soil; it is amongst the first plants to flower after the snow melts, but the flower does not last very long; as the petals age, they turn a brick-red colour and begin to curl outward. The flowers grow singly or in pairs on the stems, and the floral parts grow in multiples of threes. The species produces a small corm, which forms corms earning the genus the nickname 'riceroot'. During his historic journey, Meriwether Lewis collected a specimen while passing through Idaho in 1806.
The corm can be dug up and eaten fresh or cooked; it served Native Americans as a good source of food in times past, and is still eaten occasionally. Today these plants are not common, so digging and eating the corms is not encouraged. The plant is called in Sahaptin. |
https://en.wikipedia.org/wiki/Multi-objective%20optimization | Multi-objective optimization or Pareto optimization (also known as multi-objective programming, vector optimization, multicriteria optimization, or multiattribute optimization) is an area of multiple-criteria decision making that is concerned with mathematical optimization problems involving more than one objective function to be optimized simultaneously. Multi-objective is a type of vector optimization that has been applied in many fields of science, including engineering, economics and logistics where optimal decisions need to be taken in the presence of trade-offs between two or more conflicting objectives. Minimizing cost while maximizing comfort while buying a car, and maximizing performance whilst minimizing fuel consumption and emission of pollutants of a vehicle are examples of multi-objective optimization problems involving two and three objectives, respectively. In practical problems, there can be more than three objectives.
For a multi-objective optimization problem, it is not guaranteed that a single solution simultaneously optimizes each objective. The objective functions are said to be conflicting. A solution is called nondominated, Pareto optimal, Pareto efficient or noninferior, if none of the objective functions can be improved in value without degrading some of the other objective values. Without additional subjective preference information, there may exist a (possibly infinite) number of Pareto optimal solutions, all of which are considered equally good. Researchers study multi-objective optimization problems from different viewpoints and, thus, there exist different solution philosophies and goals when setting and solving them. The goal may be to find a representative set of Pareto optimal solutions, and/or quantify the trade-offs in satisfying the different objectives, and/or finding a single solution that satisfies the subjective preferences of a human decision maker (DM).
Bicriteria optimization denotes the special case in which there are tw |
https://en.wikipedia.org/wiki/Choquet%20integral | A Choquet integral is a subadditive or superadditive integral created by the French mathematician Gustave Choquet in 1953. It was initially used in statistical mechanics and potential theory, but found its way into decision theory in the 1980s, where it is used as a way of measuring the expected utility of an uncertain event. It is applied specifically to membership functions and capacities. In imprecise probability theory, the Choquet integral is also used to calculate the lower expectation induced by a 2-monotone lower probability, or the upper expectation induced by a 2-alternating upper probability.
Using the Choquet integral to denote the expected utility of belief functions measured with capacities is a way to reconcile the Ellsberg paradox and the Allais paradox.
Definition
The following notation is used:
– a set.
– a collection of subsets of .
– a function.
– a monotone set function.
Assume that is measurable with respect to , that is
Then the Choquet integral of with respect to is defined by:
where the integrals on the right-hand side are the usual Riemann integral (the integrands are integrable because they are monotone in ).
Properties
In general the Choquet integral does not satisfy additivity. More specifically, if is not a probability measure, it may hold that
for some functions and .
The Choquet integral does satisfy the following properties.
Monotonicity
If then
Positive homogeneity
For all it holds that
Comonotone additivity
If are comonotone functions, that is, if for all it holds that
.
which can be thought of as and rising and falling together
then
Subadditivity
If is 2-alternating, then
Superadditivity
If is 2-monotone, then
Alternative representation
Let denote a cumulative distribution function such that is integrable. Then this following formula is often referred to as Choquet Integral:
where .
choose to get ,
choose to get
Applications
The Choquet integral was applied in image processing, |
https://en.wikipedia.org/wiki/Coulure | Coulure (pronounced coo-LYUR) is a viticultural hazard that is the result of metabolic reactions to weather conditions that causes a failure of grapes to develop after flowering. In English the word shatter is sometimes used. Coulure is triggered by periods of cold, cloudy, rainy weather or very high out-of-season temperatures. The condition is most often manifested in the spring. It also occurs in vines that have little sugar content in their tissue. Flowers stay closed and are not fertilized. Thus the vines are not pollinated as the grape fails to develop and falls off. Coulure can also cause irregular bunches of grapes which are less compact than normal. These bunches are more sensitive to developing various grape diseases. The yield of a vine with coulure will decrease substantially. Grape varieties with high proclivity to coulure are Grenache, Malbec, Merlot, and Muscat Ottonel. Other causes of coulure may be vineyard conditions and practices, pruning too early or too severely, excessively fertile soils or overuse of fertilizers, and improper selection of rootstocks or clones.
During the flowering part of the growing season (May–June in the Northern Hemisphere, November–December in the Southern Hemisphere), grapevines often need dry conditions with sufficient sunlight and ambient air temperature around for pollination to go smoothly. Less ideal conditions, particularly wet, rainy weather, increases the odds that a higher than normal numbers of flowers go unpollinated and coulure to occur.
Coulure is a distinct phenomena unrelated to another viticultural hazard, millerandage, where the flowers are pollinated but the resulting berries develop without seeds and remain small. Like coulure, millerandage is often caused by inclement weather during the flowering and fruit set period and cause reduced yields.
Cause and effect
Coulure is caused by a carbohydrate deficiency in the plant tissues that causes the vine to conserve resources that would otherwise be funne |
https://en.wikipedia.org/wiki/Rigid%20unit%20modes | Rigid unit modes (RUMs) represent a class of lattice vibrations or phonons that exist in network materials such as quartz, cristobalite or zirconium tungstate. Network materials can be described as three-dimensional networks of polyhedral groups of atoms such as SiO4 tetrahedra or TiO6 octahedra. A RUM is a lattice vibration in which the polyhedra are able to move, by translation and/or rotation, without distorting. RUMs in crystalline materials are the counterparts of floppy modes in glasses, as introduced by Jim Phillips and Mike Thorpe.
The interest in rigid unit modes
The idea of rigid unit modes was developed for crystalline materials to enable an understanding of the origin of displacive phase transitions in materials such as silicates, which can be described as infinite three-dimensional networks of corner-lined SiO4 and AlO4 tetrahedra. The idea was that rigid unit modes could act as the soft modes for displacive phase transitions.
The original work in silicates showed that many of the phase transitions in silicates could be understood in terms of soft modes that are RUMs.
After the original work on displacive phase transitions, the RUM model was also applied to understanding the nature of the disordered high-temperature phases of materials such as cristobalite, the dynamics and localised structural distortions in zeolites, and negative thermal expansion.
Why rigid unit modes can exist
The simplest way to understand the origin of RUMs is to consider the balance between the numbers of constraints and degrees of freedom of the network, an engineering analysis that dates back to James Clerk Maxwell and which was introduced to amorphous materials by Jim Phillips and Mike Thorpe. If the number of constraints exceeds the number of degrees of freedom, the structure will be rigid. On the other hand, if the number of degrees of freedom exceeds the number of constraints, the structure will be floppy.
For a structure that consists of corner-linked tetrahedra (su |
https://en.wikipedia.org/wiki/Shekel%20sign | The shekel sign ⟨₪⟩ is a currency sign used for the Israeli new shekel, which is the currency of Israel.
Israeli new shekel (1986–present)
The Israeli new shekel (, ), also known by the acronym NIS ( ), was announced officially on 22 September 1985, when the first new shekel banknotes and coins were introduced. It is constructed by combining the two Hebrew letters that constitute the acronym (the first letter of each of the two words, Hebrew being written from right to left): ⟨ש⟩ and ⟨ח⟩. Sometimes the ⟨₪⟩ symbol (Unicode 20AA) is used following the number, other times the acronym .
The shekel sign, like the dollar sign ⟨$⟩, is usually placed left of the number (i.e. "₪12,000" and not "12,000₪"), but since Hebrew is written from right to left, this means that the symbol is actually written after the number. It is either not separated from the preceding number, or is separated only by a thin space.
According to the Academy of the Hebrew Language recommendation for writing numbers in Hebrew, the sign should be written to the left of the number without a space between them.
Unlike the dollar sign, the new shekel sign is not used that often when handwriting monetary amounts.
The road sign announcing the entrance to an Israeli toll road, such as Highway 6 or the Carmel Tunnels, is a shekel symbol with a road in the background.
Unicode and input
The symbol has the Unicode code point . It has been in Unicode since June 1993, version 1.1.0.
Under the Unicode bidirectional algorithm, typing the sign after the number will cause it to be displayed to the right of the number in any text directions. This contradicts the recommendation of the Academy of the Hebrew Language to place the sign to the left of the number in the Hebrew caption.
Using the standard Hebrew keyboard (SI 1452), it must be typed as (the letter ש appears on the same key in regular Hebrew mode). The Shekel sign, however, is not engraved on most keyboards sold in Israel and the sign is rarely used in |
https://en.wikipedia.org/wiki/Journal%20of%20the%20Physical%20Society%20of%20Japan | Journal of the Physical Society of Japan (JPSJ) is a monthly, peer-reviewed, scientific journal published by the Physical Society of Japan (JPS). It was first published in July 1946 (volume 1). The editor-in-chief was A. Kawabata until August 2010. The impact factor for JPSJ in 2017 is 1.485, according to Journal Citation Reports.
Volume 1 consists of a single issue designated, on the cover, from July to December 1946. Between 1967 and 1980, this journal published at a rate of two volumes per year. The other (Japanese) title for this journal is Nihon Butsuri Gakkai ōji hōkoku. Volumes for 1967 to the present day are accompanied by an annual supplement.
Research paper formats include full papers, letters, short notes, comments, addenda, errata, invited papers and special topics.
Organizational structure
The organizational structure of the journal is described as follows:
The Full Papers, Letters and Short Notes sections of the journal comprise the published original research results. Furthermore, the Full Papers section is intended to be self-contained, original research papers. The Short Notes are brief reporting on recent breakthroughs. Finally, invited reviews from a notable researcher in the field, and a collection of relevant subjects under Special Topics are occasionally included for publication.
All articles are published online in advance, before they are printed on paper. The online version of JPSJ is updated twice a month (on the 10th and 25th). The paper version of JPSJ is printed once per month (on the 15th). This version comprises the two groups of articles that are published online on two different dates.
Overview
The journal was established in 1946, succeeding its predecessor publication, Proceedings of the Physico-Mathematical Society of Japan. In its present state, JPSJ is an international journal, with submissions from authors worldwide. Additionally, financial support is available to those authors in need from developing countries.
The on |
https://en.wikipedia.org/wiki/Pluteus%20salicinus | Pluteus salicinus is a European psychedelic mushroom that grows on wood. It is an edible mushroom after parboiling.
Taxonomy
The species was originally described by Christian Hendrik Persoon as Agaricus salicinus in 1798. Paul Kummer transferred it to the genus Pluteus in 1871.
Description
Cap: 2 — 5(8) cm in diameter, convex becoming broadly convex to plane, silver-gray to brownish-gray, often with blue or greenish tint in age, smooth, with tiny scales near the center, darker at the margin, slightly translucent-striate when moist, unlined cap margin, flesh white with a grayish tinge, thin to moderate. Cap skin fibrous.
Gills: Crowded, broad, free, at first white, becoming pink-flesh colored; ventricose. Edges discoloring or bruising grayish.
Stipe: 3 — 5(10) long, 0.2 — 0.6 cm thick, more or less equal or slightly swollen at the base, flesh white with grayish-green to bluish-green tones, especially near the base. Ring absent. Firm, full or stuffed.
Taste: Unpleasant, indefinite or somewhat raphanoid (like radish).
Odor: Unpleasant, indefinite or somewhat raphanoid.
Spores: pink, smooth, 7 — 8.5 x 5 - 6 µm. Spore print pink-flesh colored to brown-pink.
Microscopic features: Pleurocystidia fusiform with slightly thickened walls 50 — 70 x 11 — 18 µm; with 3 — 5 horn-like projections.
Habitat and distribution
This mushroom is widely distributed across western Europe and Siberia. It is found on hardwoods - Alnus, Eucalyptus, Fagus, Populus and Quercus.
It is always found growing on wood. Summer-fall, solitary or gregarious on dead wood of hardwoods, in damp forests on flood-plains.
Common name
The 'knackers crumpet' is a localised, common name referring to Pluteus salicinus. Its use is most prominent in the North of England.
Chemistry
The concentration of psilocybin and psilocin in the dried sample of P. salicinus has been reported in the range of 0.21-0.35 and 0.011-0.05%, respectively.
See also
List of Pluteus species
List of Psilocybin mushrooms |
https://en.wikipedia.org/wiki/Poultry%20CRC | The Poultry Co-operative Research Center, or Poultry CRC, was a joint venture established and supported under the Australian Government's Cooperative Research Centres Programme.
The Poultry CRC's major challenge was to help Australia achieve sustainable, ethical poultry production in the face of population growth and climate change.
In July 2008, the Poultry CRC won the World's Poultry Science Association (WPSA) Industry/Organisation Award at the World's Poultry Congress in Brisbane in recognition of an outstanding contribution to the development of the poultry industry. In August 2012, the Poultry CRC was awarded the WPSA's Education Award at the World's Poultry Congress in Brazil for their exceptional contribution to poultry education. In addition, the Poultry CRC received an Australian Collaborative Innovation Award in May 2012.
Structure of the CRC
The Poultry CRC was an unincorporated joint venture between seven essential participants and was governed by a skills-based board. It managed its research and development programmes through a public company, Poultry CRC Ltd. The Poultry CRC was headquartered at the University of New England in Armidale, New South Wales, and had an extensive collaborative network comprising researchers, educators, and support staff from its participating organisations. The original Poultry CRC was established on 1 July 2003, with the subsequent CRC being extended to 2017.
Essential Participants
Australian Egg Corporation Limited
Bioproperties Pty Ltd
Commonwealth Scientific and Industrial Research Organisation (CSIRO) Livestock Industries
Department of Agriculture, Fisheries and Forestry (formerly DEEDI) Queensland
Rural Industries Research and Development Corporation (RIRDC) Chicken Meat Program
University of Melbourne
University of New England
Other Participants
Active Research Pty Ltd
Alltech Biotechnology
Aviagen
Baiada Poultry
Cordina Chicken Farms
D.A. Hall & Co.
Deakin University
Feedworks
FSA Consulting
Golden Cocke |
https://en.wikipedia.org/wiki/The%20Banksias | The Banksias, by Celia Rosser, is a three-volume series of monographs containing paintings of every Banksia species. Its publication represented the first time such a large genus had been entirely painted by a single botanical artist. It has been described as "one of the outstanding botanical works of this century."
The paintings themselves are watercolours on Arches rag paper. The three volumes comprise plates reproduced using offset printing, and bound in green leather. Alex George wrote the accompanying text.
Rosser began working on the series in 1974. Volume I of The Banksias, containing 24 plates, was published in 1981. The edition comprised 730 books and 100 portfolios. Volume II, published in 1988, also contained 24 plates and was also released in an edition of 730 books and 100 portfolios. Volume III, completed in 2000, contained 28 plates, and was released in an edition of 530 books and 300 portfolios. Since the publication of Volume III, a new Banksia, B. rosserae has been described; Rosser subsequently painted it and released a set of prints. In 2007, the genus Dryandra was transferred to Banksia, so there are now a great many Banksia species that have not been painted by Rosser.
Each volume of The Banksias has been presented to the Queen by the Australian Government as a gift of the Australian people. Largely on the basis of her work for The Banksias, Rosser was awarded a Medal of the Order of Australia 1995 for her contribution to botanical art, and the Jill Smythies Award for botanical art in 1997. |
https://en.wikipedia.org/wiki/ACEnet | Established in 2003 as a shared service to provide advanced computing support and services to the Atlantic Canadian research community, ACENET was a consortium of five universities. Since then, its membership has grown and so has its mission. ACENET has 14 Atlantic university and community college members, and provides access to advanced computing infrastructure, technical support, and digital skills development to academic researchers and their students at any post-secondary institution in the region, as well as government departments and industry.
Publicly funded through the federal and provincial governments, it provides these services at no charge to researchers and students. Through its federally incorporated not-for-profit, ACENET Solutions Inc., it serves government departments and industry in the region on a cost-recovery basis. It is the only organization in Atlantic Canada providing these resources, expertise, support and training with the depth of scientific and technical expertise demanded by researchers and industry. ACENET currently support more than 300 projects totaling over 1000 users in Atlantic Canada.
ACENET connects users in the region to digital research infrastructure (DRI) resources across the country through its partnership with the national Digital Research Alliance of Canada (the Alliance) and the four other regional consortia - Calcul Quebec, Compute Ontario, DRI Prairies and DRI BC. The Alliance was created as a result of the federal government’s 2019 investment of $572M in a new Digital Research Infrastructure Strategy for Canada. It assumed the national responsibility for Advanced Research Computing (ARC) from Compute Canada on 1 April, 2022, with an expanded role that also includes Research Data Management (previously under the Portage Network) and Research Software (previously under the domain of CANARIE).
The Alliance and its five regional partners together comprise the Alliance Federation. Through the federation, ACENET represe |
https://en.wikipedia.org/wiki/Chemical%20library | A chemical library or compound library is a collection of stored chemicals usually used ultimately in high-throughput screening or industrial manufacture. The chemical library can consist in simple terms of a series of stored chemicals. Each chemical has associated information stored in some kind of database with information such as the chemical structure, purity, quantity, and physiochemical characteristics of the compound.
Purpose
In drug discovery high-throughput screening, it is desirable to screen a drug target against a selection of chemicals that try to take advantage of as much of the appropriate chemical space as possible. The chemical space of all possible chemical structures is extraordinarily large. Most stored chemical libraries do not typically have a fully represented or sampled chemical space mostly because of storage and cost concerns. However, since many molecular interactions cannot be predicted, the wider the chemical space that is sampled by the chemical library, the better the chance that high-throughput screening will find a "hit"—a chemical with an appropriate interaction in a biological model that might be developed into a drug.
An example of a chemical library in drug discovery would be a series of chemicals known to inhibit kinases, or in industrial processes, a series of catalysts known to polymerize resins.
Generation of chemical libraries
Chemical libraries are usually generated for a specific goal and larger chemical libraries could be made of several groups of smaller libraries stored in the same location. In the drug discovery process for instance, a wide range of organic chemicals are needed to test against models of disease in high-throughput screening. Therefore, most of the chemical synthesis needed to generate chemical libraries in drug discovery is based on organic chemistry. A company that is interested in screening for kinase inhibitors in cancer may limit their chemical libraries and synthesis to just those types of ch |
https://en.wikipedia.org/wiki/HDCP%20repeater%20bit | The HDCP repeater bit is a part of the High-bandwidth Digital Content Protection specification and applies to intermediate devices (HDCP Repeaters) between the Source device and the Presentation device. For example, a Blu-ray connected by HDMI to an AV Receiver which in turn connected to a TV using HDMI makes the AV Receiver an HDCP Repeater.
The AV Receiver reports to the Transmitter whether it is a Repeater or a Receiver (no downstream devices) using the REPEATER bit.
Details
Following the above example the AV Receiver must report whether it has downstream devices or not using the REPEATER bit. E.g. when the TV is turned on, the AV Receiver should set REPEATER=TRUE, and when TV is turned off it sets REPEATER=FALSE. Since there is no proper way to notify on REPEATER status change, the AV Receiver must reset the connection with the Source. In order to avoid this redundant reset cycles on the connections, which has annoying user-experience, HDCP v1.x allows devices that are potentially Repeater to set the REPEATER all the time, even when no downstream devices are connected (e.g. while the TV is off).
In HDCP v2.x that smart has been lost and the Repeater must not keep the REPEATER flag set once it has no downstream devices.
The repeater bit is a part of the HDCPv1 and HDCPv2 specifications, which is available on the digital content protection LLC web site.
The specification defines repeater devices:
HDCP Repeater. An HDCP device that can receive and decrypt HDCP content through one or more of its HDCP-protected interface ports, and can also re-encrypt and emit said HDCP content through one or more of its HDCP-protected interface ports, is referred to as an HDCP repeater. An HDCP repeater may also be referred to as either an HDCP receiver or an HDCP transmitter when referring to either the upstream side or the downstream side, respectively.
Issues with REPEATER
Some HDCPv1.x devices (usually old) may not agree to work with Repeater that has REPEATER=TRUE but not |
https://en.wikipedia.org/wiki/Heterosigma%20akashiwo | Heterosigma akashiwo is a species of microscopic algae of the class Raphidophyceae. It is a swimming marine alga that episodically forms toxic surface aggregations known as harmful algal bloom. The species name akashiwo is from the Japanese for "red tide".
Synonyms include Olisthodiscus luteus (Hulburt 1965), and Entomosigma akashiwo (Hada 1967). H. akashiwo and H. inlandica have been recognized as two species of Heterosigma. However, Hara and Chihara (1987) described both specimens as one species, validly describing them as H. akashiwo.
Description
H. akashiwo cells are relatively small, ranging in size from 18 to 34 μm in diameter. They appear golden brown, and appear in clusters. Morphology is highly variable, but does not appear to vary significantly between locations. One culture may contain flat or round individual cells. Molecular techniques for identification (including quantitative PCR) are preferred over traditional microscope fixing, which may lyse the cells.
Distribution
Heterosigma akashiwo has been identified off the coasts of the United States, Canada, Chile, the Netherlands, Scotland, Ireland, Sweden, Norway, Japan, S.Korea, Hong Kong, Australia, New Zealand and South Africa. Most of the literature suggests H. akashiwo is associated with shallow water within 10 m of the surface, but this is not a universal rule.
Physiology
Heterosigma akashiwo is a mixotrophic alga, supplementing nutrient uptake and photosynthesis with ingestion of bacteria. Each cell may contain 18-27 chloroplasts. These cells have been observed to glide and twirl under microscopic examination, but nonmotile cells have been associated with toxic blooms. Blooms are clearly visible by air, appearing as a red area in otherwise blue water. Optimal growth occurs at 25 °C and 100 μE m−2s−1, conditions which are associated with very low toxicity. Maximum toxicity occurs (and relatively slow growth) occurs at 20 °C and 200 μE m−2s−1. H. akashiwo reproduces asexually by |
https://en.wikipedia.org/wiki/Ossia | () is a musical term for an alternative passage which may be played instead of the original passage. The word ossia comes from the Italian for "alternatively" and was originally spelled o sia, meaning "or be it".
Ossia passages are very common in opera and solo-piano works. They are usually an easier version of the preferred form of passage, but in Mily Balakirev's Islamey, for instance, the urtext has ossia passages of both types (simpler and more difficult). Bel canto vocal music also frequently uses ossia, also called oppure, passages to illustrate a more embellished version of the vocal line.
On the other hand, an ossia marking does not always indicate a change in difficulty; the piano solo music of Franz Liszt is typically full of alternative passages, often no easier or more difficult than the rest of the piece. This reflects Liszt's desire to leave his options open during a performance. Many of his ossia passages are cadenzas.
An unusual use of ossia is found in Alban Berg's Violin Concerto where several ossia parts are included for the solo violin. If the soloist chooses to play these, the concertmaster is required to play a different ossia (which takes part of the solo violin line that is lost in favor of the soloist's ossia). |
https://en.wikipedia.org/wiki/Elsie%20Widdowson | Elsie Widdowson (21 October 1906 – 14 June 2000), was a British dietitian and nutritionist. She and Dr Robert McCance, a pediatrician, physiologist, biochemist, and nutritionist, were responsible for overseeing the government-mandated addition of vitamins to food and wartime rationing in Britain during World War II.
Early life
Widdowson was born in Wallington, Surrey on 21 October 1906 to Rose Elphick and Harry Widdowson. Her father, Thomas Henry (known as Harry), was from Grantham in Lincolnshire and moved to Battersea as a grocer's assistant and eventually owned a stationery business, whilst her mother Rose, originally from Dorking, worked as a dressmaker. Her younger sister Eva Crane trained as a nuclear physicist but became a world-renowned authority on bees. The family were Plymouth Brethren.
Elsie lived in Dulwich as a child and attended Sydenham County Grammar School for Girls where both she and her sister won prizes. During the 1920s and 1930s, professional opportunities for women, apart from nursing or teaching, were limited. Educated women such as Widdowson had to develop skills that offered employment potential; therefore, Widdowson trained as a chemist.
She studied chemistry at Imperial College, London and although she completed her degree in two year, she had to wait until 1928 to be awarded her BSc, when she became one of the first women graduates of Imperial College. She did postgraduate work at the Department of Plant Physiology at Imperial College, developing methods for separating and measuring the fructose, glucose, sucrose, and hemicellulose of fruit. She would measure individual changes in the carbohydrates in fruit from the time it appeared on the tree to when it ripened. Once a fortnight, she took a train to Kentish apple orchard and picked apples, measuring their carbohydrate levels. In 1931, she received her PhD in chemistry from the Imperial College for her thesis on the carbohydrate content of apples. This work would go on to have inte |
https://en.wikipedia.org/wiki/Hot%20chocolate%20effect | The hot chocolate effect is a phenomenon of wave mechanics in which the pitch heard from tapping a cup of hot liquid rises after the addition of a soluble powder. It was first documented in 1980 by Frank Crawford of the Lawrence Berkeley Laboratory. It was initially observed in the making of hot chocolate or instant coffee, but also occurs in other situations such as adding salt to supersaturated hot water or cold beer. Recent research has found many more substances which create the effect, even in initially non-supersaturated liquids. The effect is thought to happen because upon initial stirring, entrained gas bubbles reduce the speed of sound in the liquid, lowering the frequency. As the bubbles clear, sound travels faster in the liquid and the frequency increases.
The effect can be observed by pouring hot milk or hot water into a mug, stirring in chocolate powder, and tapping the bottom of the mug with a spoon. The pitch of the taps will increase progressively with no relation to the speed or force of tapping. Subsequent stirring of the same solution (without adding more chocolate powder) will gradually decrease the pitch again, followed by another increase. This process can be repeated a number of times, until equilibrium has been reached. Musical effects can be achieved by varying the strength and timing of the stirring action along with the timing of the tapping action.
Origin of the phenomenon
The phenomenon is explained by the effect of bubble density on the speed of sound in the liquid. The note heard is the frequency of a standing wave where a quarter wavelength is the distance between the base of the mug and the liquid surface. This frequency f is equal to the speed v of the wave divided by four times the height of the water column h:
The speed of sound v in a homogeneous liquid or gas is dependent on the fluid's mass density () and adiabatic bulk modulus (), according to the Newton-Laplace formula:
Water is approximately 800 times denser than air, a |
https://en.wikipedia.org/wiki/History%20of%20biochemistry | The history of biochemistry can be said to have started with the ancient Greeks who were interested in the composition and processes of life, although biochemistry as a specific scientific discipline has its beginning around the early 19th century. Some argued that the beginning of biochemistry may have been the discovery of the first enzyme, diastase (today called amylase), in 1833 by Anselme Payen, while others considered Eduard Buchner's first demonstration of a complex biochemical process alcoholic fermentation in cell-free extracts to be the birth of biochemistry. Some might also point to the influential work of Justus von Liebig from 1842, Animal chemistry, or, Organic chemistry in its applications to physiology and pathology, which presented a chemical theory of metabolism, or even earlier to the 18th century studies on fermentation and respiration by Antoine Lavoisier.
The term biochemistry itself is derived from the combining form bio-, meaning 'life', and chemistry. The word is first recorded in English in 1848, while in 1877, Felix Hoppe-Seyler used the term ( in German) in the foreword to the first issue of Zeitschrift für Physiologische Chemie (Journal of Physiological Chemistry) as a synonym for physiological chemistry and argued for the setting up of institutes dedicate to its studies. Nevertheless, several sources cite German chemist Carl Neuberg as having coined the term for the new discipline in 1903, and some credit it to Franz Hofmeister.
The subject of study in biochemistry is the chemical processes in living organisms, and its history involves the discovery and understanding of the complex components of life and the elucidation of pathways of biochemical processes. Much of biochemistry deals with the structures and functions of cellular components such as proteins, carbohydrates, lipids, nucleic acids and other biomolecules; their metabolic pathways and flow of chemical energy through metabolism; how biological molecules give rise to the p |
https://en.wikipedia.org/wiki/Anisotropic%20conductive%20film | Anisotropic conductive film (ACF) is an adhesive interconnect system that is commonly used in liquid crystal display manufacturing to make the electrical and mechanical connections from the driver electronics to the glass substrates of the LCD. The material is also available in a paste form referred to as anisotropic conductive paste (ACP), and both are grouped together as anisotropic conductive adhesives (ACAs). ACAs have more recently been used to perform the flex-to-board or flex-to-flex connections used in handheld electronic devices such as mobile phones, MP3 players, or in the assembly of CMOS camera modules.
History
ACAs developed in the late 1970s and early 1980s, with heat seal connectors by Nippon Graphite Industries, and ACFs by Hitachi Chemicals and Dexerials (formerly known as Sony Chemicals & Information Devices). Currently there are many manufacturers of heat seal connectors and ACAs, but Hitachi and Sony continue to dominate the industry in terms of market share. Other manufacturers of ACAs include 3M, Loctite, DELO, Creative Materials, Henkel, Sun Ray Scientific, Kyocera, Three Bond, Panacol, and Btech.
In the very early years, ACAs were made from rubber, acrylic, and other adhesive compounds, but they rapidly converged on several different variations of thermoset biphenyl type epoxy resins. The temperatures required were relatively high at 170-180C, however, and the market leaders Sony and Hitachi developed and released acrylic-based materials in the early 2000s that brought the curing temperatures down below 150C while keeping the curing times in the 10–12 second range. Further advances in the acrylic compounds used decreased the curing cycle times to below 5 seconds in many cases, which is where they remain as of this writing. Specification sheets are available at all of the manufacturers' sites listed above.
Current market
ACF continues to be the most popular form factor for ACAs, largely due to the ability to precisely control the volume of |
https://en.wikipedia.org/wiki/Authbind | authbind is an open-source system utility written by Ian Jackson and is distributed under the GNU General Public License. The authbind software allows a program that would normally require superuser privileges to access privileged network services to run as a non-privileged user. authbind allows the system administrator to permit specific users and groups access to bind to TCP and UDP ports below 1024. Ports 0 - 1023 are normally privileged and reserved for programs that are run as the root user. Allowing regular users limited access to privileged ports helps prevent possible privilege escalation and system compromise if the software happens to contain software bugs or is found to be vulnerable to unknown exploits.
authbind achieves this by defining the LD_PRELOAD environment variable which loads a libauthbind library. This library overrides the bind() call with a version that executes a setuid helper program (/usr/lib/authbind/helper) with the socket as file descriptor 0. The helper validates its arguments and checks its configuration, calls the real bind() system call on file descriptor 0 (which also affects the original process's socket), and exits, allowing the original process to continue with the socket bound to the requested address and port.
authbind is currently distributed with the Debian and Ubuntu Linux distributions.
Alternatives
The Linux kernel's implementation of POSIX capabilities includes the CAP_NET_BIND_SERVICE which allows either explicitly enabled binaries (with "setcap CAP_NET_BIND_SERVICE+ep /path/to/binary") or binaries configured to accept the capability from the invoking user's capability set ("setcap CAP_NET_BIND_SERVICE+ei /path/to/binary") if available, making userland software unnecessary for binding to lower numeral ports. Linux capabilities, however were not introduced until the latter half of 1999, more than a year after authbind's release, and (similar to setuid/setgid) cannot be set on scripts. Both these explain why the so |
https://en.wikipedia.org/wiki/Marker-and-cell%20method | The marker-and-cell method is commonly used in computer graphics to discretize functions for fluid and other simulations. It was developed by Francis Harlow and his collaborators at the Los Alamos National Laboratory.
See also
Immersed boundary method
Stokesian dynamics
Volume of fluid method
Level-set method |
https://en.wikipedia.org/wiki/Suicide%20of%20Kevin%20Whitrick | Kevin Neil Whitrick (17 August 1964 – 21 March 2007) was a British electrical engineer who committed suicide by hanging himself while in an online chat room.
Background
Whitrick had been married to his wife Paula since 1988, and they had two children. At the time of his death, the marriage broke down about two years prior, and he was living separately from his family. In 2006, Whitrick was severely injured in a car accident and suffered long-term health consequences.
Suicide
In March 2007, Whitrick was using the Paltalk video chat service, in a special "insult" chatroom with about 60 other users where people "have a go at each other". He stood on a chair, punched a hole in his ceiling and placed a rope around a joist, then tied the other end around his neck and stepped off the chair in order to asphyxiate himself. Some people thought this was a prank, until his face started turning blue. Some people in the chat room encouraged him by saying things like "just get on with it", while others tried desperately to find his address. A member in the room contacted the police, who arrived at the scene two minutes later. Whitrick was pronounced dead at 11:15 pm GMT.
Aftermath
The death was widely reported in the media. Concerns were raised over the possibility that it could inspire further suicides, as well as the webcam footage becoming available in perpetuity on the internet. Detectives traced about 100 chatroom users to question them about their role in the suicide, though the Crown Prosecution Service did not pursue any criminal charges against them.
See also
Abraham Biggs
Suicide of Ronnie McNutt
List of people who died by suicide by hanging |
https://en.wikipedia.org/wiki/Notation%20for%20differentiation | In differential calculus, there is no single uniform notation for differentiation. Instead, various notations for the derivative of a function or variable have been proposed by various mathematicians. The usefulness of each notation varies with the context, and it is sometimes advantageous to use more than one notation in a given context. The most common notations for differentiation (and its opposite operation, the antidifferentiation or indefinite integration) are listed below.
Leibniz's notation
The original notation employed by Gottfried Leibniz is used throughout mathematics. It is particularly common when the equation is regarded as a functional relationship between dependent and independent variables and . Leibniz's notation makes this relationship explicit by writing the derivative as
Furthermore, the derivative of at is therefore written
Higher derivatives are written as
This is a suggestive notational device that comes from formal manipulations of symbols, as in,
The value of the derivative of at a point may be expressed in two ways using Leibniz's notation:
.
Leibniz's notation allows one to specify the variable for differentiation (in the denominator). This is especially helpful when considering partial derivatives. It also makes the chain rule easy to remember and recognize:
Leibniz's notation for differentiation does not require assigning a meaning to symbols such as or (known as differentials) on their own, and some authors do not attempt to assign these symbols meaning. Leibniz treated these symbols as infinitesimals. Later authors have assigned them other meanings, such as infinitesimals in non-standard analysis, or exterior derivatives. Commonly, is left undefined or equated with , while is assigned a meaning in terms of , via the equation
which may also be written, e.g.
(see below). Such equations give rise to the terminology found in some texts wherein the derivative is referred to as the "differential coefficie |
https://en.wikipedia.org/wiki/Sergio%20Verd%C3%BA | Sergio Verdú (born Barcelona, Spain, August 15, 1958) is a former professor of electrical engineering and specialist in information theory. Until September 22, 2018, he was the Eugene Higgins Professor of Electrical Engineering at Princeton University, where he taught and conducted research on information theory in the Information Sciences and Systems Group. He was also affiliated with the program in Applied and Computational Mathematics. He was dismissed from the faculty following a university investigation of alleged sexual misconduct.
Verdu received the Telecommunications Engineering degree from the Polytechnic University of Catalonia, Barcelona, Spain, in 1980 and the PhD degree in Electrical Engineering from the University of Illinois at Urbana-Champaign in 1984. Conducted at the Coordinated Science Laboratory of the University of Illinois, his doctoral research was supervised by Vincent Poor and pioneered the field of multiuser detection. In 1998, his book Multiuser Detection was published by Cambridge University Press.
Sexual harassment incident and dismissal from tenured position
A Title IX investigation by Princeton, made public in 2017 by the Huffington Post, concluded that Verdú had sexually harassed one of his graduate students, a South Korean woman. According to the student, Verdú was required only to attend an 8-hour training session as a consequence. The student changed advisers and changed her research topic. A university spokesperson denied the claim that additional training was the only consequence for Verdú, stating that "penalties were imposed in addition to the required counseling", but did not identify what those penalties were. According to the Princeton Dean of Faculty, there were allegations that Verdú had also harassed others, but only the one student was willing to make a formal complaint. Verdú denied the findings of the investigation, stating: "The university advised me not to reply but I categorically deny that there were any advances |
https://en.wikipedia.org/wiki/GDF10 | Growth differentiation factor 10 (GDF10) also known as bone morphogenetic protein 3B (BMP-3B) is a protein that in humans is encoded by the GDF10 gene.
GDF10 belongs to the transforming growth factor beta superfamily that is closely related to bone morphogenetic protein-3 (BMP3). It plays a role in head formation and may have multiple roles in skeletal morphogenesis. GDF10 is also known as BMP-3b, with GDF10 and BMP3 regarded as a separate subgroup within the TGF-beta superfamily.
In mice, GDF10 mRNA is abundant in the brain, inner ear, uterus, prostate, neural tissues, blood vessels and adipose tissue with low expression in spleen and liver. It is also present in bone of both adults and neonatal mice. Human GDF10 mRNA is found in the cochlea and lung of foetuses, and in testis, retina, pineal gland, and other neural tissues of adults. |
https://en.wikipedia.org/wiki/GDF11 | Growth differentiation factor 11 (GDF11) also known as bone morphogenetic protein 11 (BMP-11) is a protein that in humans is encoded by the growth differentiation factor 11 gene. GDF11 is a member of the Transforming growth factor beta family.
GDF11 acts as a cytokine and its molecular structure is identical in humans, mice and rats. The bone morphogenetic protein group is characterized by a polybasic proteolytic processing site, which is cleaved to produce a protein containing seven conserved cysteine residues.
Tissue distribution
GDF11 is expressed in many tissues, including skeletal muscle, pancreas, kidney, nervous system, and retina.
Function
Gene deletion and over-expression studies indicate that GDF11 primarily regulates the embryological development of the skeletal system. It may also help regulate development of the central nervous system, blood vessels, the kidney and other tissues.
GDF11 improves neurodegenerative and neurovascular disease outcomes, increases skeletal muscle volume, and enhances muscle strength. Its wide-ranging biological effects may include the reversal of senescence in clinical applications, as well as the ability to reverse age-related pathological changes and regulate organ regeneration after injury.
Effects on cell growth and differentiation
GDF11 belongs to the transforming growth factor beta superfamily that controls anterior-posterior patterning by regulating the expression of Hox genes. It determines Hox gene expression domains and rostrocaudal identity in the caudal spinal cord.
During mouse development, GDF11 expression begins in the tail bud and caudal neural plate region. GDF knock-out mice display skeletal defects as a result of patterning problems with anterior-posterior positioning. This cytokine also inhibits the proliferation of olfactory receptor neural progenitors to regulate the number of neurons in the olfactory epithelium, and controls the competence of progenitor cells to regulate numbers of retinal ga |
https://en.wikipedia.org/wiki/Autophagin | Autophagin-1 (Atg4/Apg4) is a unique cysteine protease responsible for the cleavage of the carboxyl terminus of Atg8/Apg8/Aut7, a reaction essential for its lipidation during autophagy. Human Atg4 homologues cleave the carboxyl termini of the three human Atg8 homologues, microtubule-associated protein light chain 3 (LC3), GABARAP, and GATE-16.
The rapid advancement in our understanding of the mechanisms and regulation of autophagy has placed this process in the center of current research in major human disorders. The future challenge is to develop easy methods to separately manipulate the activity of each of the autophagic pathways. This would allow researchers to further understand their contribution to disease such as cancer, neurodegeneration, infectious disease, muscular disorders and possibly will provide therapeutic tools.
See also
Autophagy
Apoptosis
Ubiquitin |
https://en.wikipedia.org/wiki/Implicit%20cognition | Implicit cognition refers to cognitive processes that occur outside conscious awareness or conscious control. This includes domains such as learning, perception, or memory which may influence a person's behavior without their conscious awareness of those influences.
Overview
Implicit cognition is everything one does and learns unconsciously or without any awareness that one is doing it. An example of implicit cognition could be when a person first learns to ride a bike: at first they are aware that they are learning the required skills. After having stopped for many years, when the person starts to ride the bike again they do not have to relearn the motor skills required, as their implicit knowledge of the motor skills takes over and they can just start riding the bike as if they had never stopped. In other words, they do not have to think about the actions that they are performing in order to ride the bike. It can be seen from this example that implicit cognition is involved with many of the different mental activities and everyday situations in people's daily lives. There are many processes in which implicit memory works, which include learning, our social cognition, and our problem-solving skills.
History
Implicit cognition was first discovered in the year of 1649 by Descartes in his Passions of the Soul. He said in one of his writings that he saw that unpleasant childhood experiences remain imprinted in a child's brain until its death without any conscious memory of it remaining. Even though this idea was never accepted by any of his peers, in 1704 Gottfried Wilhelm Leibniz in his New Essays Concerning Human Understanding stressed the importance of unconscious perceptions which he said were the ideas that we are not consciously aware of yet still influence people's behavior. He claimed that people have residual effects of prior impressions without any remembrance of them. In 1802 French philosopher Maine de Biran in his The Influence of Habit on the Faculty of |
https://en.wikipedia.org/wiki/Allan%20Hobson | John Allan Hobson (June 3, 1933 – July 7, 2021) was an American psychiatrist and dream researcher. He was known for his research on rapid eye movement sleep. He was Professor of Psychiatry, Emeritus, Harvard Medical School, and Professor, Department of Psychiatry, Beth Israel Deaconess Medical Center.
Biography
Hobson grew up in Hartford, Connecticut. In 1955 he obtained his A.B. degree from Wesleyan University. Four years later he earned his MD degree at Harvard Medical School in 1959.
For the following two years he interned at Bellevue Hospital Center, New York. Then in 1960, he was a resident in Psychiatry at Massachusetts Mental Health Center in Boston for a year. Hobson then traveled to France where he was a Special Fellow of the National Institute of Mental Health for the Department of Physiology at the University of Lyon.
Upon returning to the United States, he went back to the Psychiatry at Massachusetts Mental Health Center in Boston until 1966.
He worked in numerous hospitals and research laboratories over the years and became the Director of the Laboratory of Neurophysiology at the Massachusetts Mental Health Center.
Hobson received four awards for his work:
Admission to the Boylston Medical Society
The Benjamin Rush Gold Medal for Best Scientific Exhibit
Honorary Member of the American Psychiatric Association since 1978.
Recipient of the 1998 Distinguished Scientist Award of the Sleep Research Society
Hobson's sense of humor was shown in such quips as: “The only known function of sleep is to cure sleepiness".
Work
In addition to his many paid appointments, Hobson was actively involved with four groups relating to his neurological sleep research: the Society Memberships, the Society for Neuroscience, the Society for Sleep Research, the AAAS, and the International Association for the Study of Dreams (IASD), for which he used to be president.
Dream theories
Hobson's research specialty was quantifying mental events and correlating them with |
https://en.wikipedia.org/wiki/Positive-definite%20kernel | In operator theory, a branch of mathematics, a positive-definite kernel is a generalization of a positive-definite function or a positive-definite matrix. It was first introduced by James Mercer in the early 20th century, in the context of solving integral operator equations. Since then, positive-definite functions and their various analogues and generalizations have arisen in diverse parts of mathematics. They occur naturally in Fourier analysis, probability theory, operator theory, complex function-theory, moment problems, integral equations, boundary-value problems for partial differential equations, machine learning, embedding problem, information theory, and other areas.
Definition
Let be a nonempty set, sometimes referred to as the index set. A symmetric function is called a positive-definite (p.d.) kernel on if
holds for any , given .
In probability theory, a distinction is sometimes made between positive-definite kernels, for which equality in (1.1) implies , and positive semi-definite (p.s.d.) kernels, which do not impose this condition. Note that this is equivalent to requiring that any finite matrix constructed by pairwise evaluation, , has either entirely positive (p.d.) or nonnegative (p.s.d.) eigenvalues.
In mathematical literature, kernels are usually complex valued functions, but in this article we assume real-valued functions, which is the common practice in applications of p.d. kernels.
Some general properties
For a family of p.d. kernels
The conical sum is p.d., given
The product is p.d., given
The limit is p.d. if the limit exists.
If is a sequence of sets, and a sequence of p.d. kernels, then both and are p.d. kernels on .
Let . Then the restriction of to is also a p.d. kernel.
Examples of p.d. kernels
Common examples of p.d. kernels defined on Euclidean space include:
Linear kernel: .
Polynomial kernel: .
Gaussian kernel (RBF kernel): .
Laplacian kernel: .
Abel kernel: .
Kernel generating Sobolev spa |
https://en.wikipedia.org/wiki/Interleaved%20memory | In computing, interleaved memory is a design which compensates for the relatively slow speed of dynamic random-access memory (DRAM) or core memory, by spreading memory addresses evenly across memory banks. That way, contiguous memory reads and writes use each memory bank in turn, resulting in higher memory throughput due to reduced waiting for memory banks to become ready for the operations.
It is different from multi-channel memory architectures, primarily as interleaved memory does not add more channels between the main memory and the memory controller. However, channel interleaving is also possible, for example in freescale i.MX6 processors, which allow interleaving to be done between two channels.
Overview
With interleaved memory, memory addresses are allocated to each memory bank in turn. For example, in an interleaved system with two memory banks (assuming word-addressable memory), if logical address 32 belongs to bank 0, then logical address 33 would belong to bank 1, logical address 34 would belong to bank 0, and so on. An interleaved memory is said to be n-way interleaved when there are banks and memory location resides in bank .
Interleaved memory results in contiguous reads (which are common both in multimedia and execution of programs) and contiguous writes (which are used frequently when filling storage or communication buffers) actually using each memory bank in turn, instead of using the same one repeatedly. This results in significantly higher memory throughput as each bank has a minimum waiting time between reads and writes.
Interleaved DRAM
Main memory (random-access memory, RAM) is usually composed of a collection of DRAM memory chips, where a number of chips can be grouped together to form a memory bank. It is then possible, with a memory controller that supports interleaving, to lay out these memory banks so that the memory banks will be interleaved.
Data in DRAM is stored in units of pages. Each DRAM bank has a row buffer that se |
https://en.wikipedia.org/wiki/Projective%20cover | In the branch of abstract mathematics called category theory, a projective cover of an object X is in a sense the best approximation of X by a projective object P. Projective covers are the dual of injective envelopes.
Definition
Let be a category and X an object in . A projective cover is a pair (P,p), with P a projective object in and p a superfluous epimorphism in Hom(P, X).
If R is a ring, then in the category of R-modules, a superfluous epimorphism is then an epimorphism such that the kernel of p is a superfluous submodule of P.
Properties
Projective covers and their superfluous epimorphisms, when they exist, are unique up to isomorphism. The isomorphism need not be unique, however, since the projective property is not a full fledged universal property.
The main effect of p having a superfluous kernel is the following: if N is any proper submodule of P, then . Informally speaking, this shows the superfluous kernel causes P to cover M optimally, that is, no submodule of P would suffice. This does not depend upon the projectivity of P: it is true of all superfluous epimorphisms.
If (P,p) is a projective cover of M, and P' is another projective module with an epimorphism , then there is a split epimorphism α from P' to P such that
Unlike injective envelopes and flat covers, which exist for every left (right) R-module regardless of the ring R, left (right) R-modules do not in general have projective covers. A ring R is called left (right) perfect if every left (right) R-module has a projective cover in R-Mod (Mod-R).
A ring is called semiperfect if every finitely generated left (right) R-module has a projective cover in R-Mod (Mod-R). "Semiperfect" is a left-right symmetric property.
A ring is called lift/rad if idempotents lift from R/J to R, where J is the Jacobson radical of R. The property of being lift/rad can be characterized in terms of projective covers: R is lift/rad if and only if direct summands of the R module R/J (as a right or le |
https://en.wikipedia.org/wiki/Stress%20majorization | Stress majorization is an optimization strategy used in multidimensional scaling (MDS) where, for a set of -dimensional data items, a configuration of points in -dimensional space is sought that minimizes the so-called stress function . Usually is or , i.e. the matrix lists points in or dimensional Euclidean space so that the result may be visualised (i.e. an MDS plot). The function is a cost or loss function that measures the squared differences between ideal (-dimensional) distances and actual distances in r-dimensional space. It is defined as:
where is a weight for the measurement between a pair of points , is the euclidean distance between and and is the ideal distance between the points (their separation) in the -dimensional data space. Note that can be used to specify a degree of confidence in the similarity between points (e.g. 0 can be specified if there is no information for a particular pair).
A configuration which minimizes gives a plot in which points that are close together correspond to points that are also close together in the original -dimensional data space.
There are many ways that could be minimized. For example, Kruskal recommended an iterative steepest descent approach. However, a significantly better (in terms of guarantees on, and rate of, convergence) method for minimizing stress was introduced by Jan de Leeuw. De Leeuw's iterative majorization method at each step minimizes a simple convex function which both bounds from above and touches the surface of at a point , called the supporting point. In convex analysis such a function is called a majorizing function. This iterative majorization process is also referred to as the SMACOF algorithm ("Scaling by MAjorizing a COmplicated Function").
The SMACOF algorithm
The stress function can be expanded as follows:
Note that the first term is a constant and the second term is quadratic in (i.e. for the Hessian matrix the second term is equivalent to tr) an |
https://en.wikipedia.org/wiki/Rice%20Institute%20Computer | The Rice Institute Computer, also known as the Rice Computer or R1, was a 54-bit tagged architecture digital computer built during 1958–1961 (partially operational beginning in 1959) on the campus of Rice University, Houston, Texas, United States. Operating as Rice's primary computer until the middle 1960s, the Rice Institute Computer was decommissioned in 1971. The system initially used vacuum tubes and semiconductor diodes for its logic circuits; some later peripherals were built in solid-state emitter-coupled logic. It was designed by Martin H. Graham.
A copy of the machine called OSAGE was built and operated at the University of Oklahoma.
Memory
Memory was implemented using a variety of technologies over the lifetime of the R1. Originally a cathode ray tube or "Williams tube" array, RCA core memory was introduced in 1966, followed by Ampex core memory in 1967. Following those two upgrades, the R1 had reached its full 32k word capacity, although the original electrostatic memory was soon decommissioned due to falling reliability in its old age.
Architecture
The R1 had seven memory-mapped general-purpose processor registers, each 54 bits in size, in addition to a constant zero register. For memory addressing, seven 16-bit "B-Registers" were used. The program counter was also held in a writable "B-Register". See the table below for conventions and hardware-enforced usage of these registers.
See also
List of vacuum-tube computers |
https://en.wikipedia.org/wiki/Active%20immunization | Active immunization is the induction of immunity after exposure to an antigen. Antibodies are created by the recipient and may be stored permanently.
Active immunization can occur naturally when microbes or other antigen are received by a person who has not yet come into contact with the microbes and has no pre-made antibodies for defense. The immune system will eventually create antibodies for the microbes, but this is a slow process and, if the microbes are dead, there may not be enough time for the antibodies to be used.
Artificial active immunization is where the microbe is injected into the person before they are able to take it in naturally. The microbe is treated, so that it will not harm the infected person. Depending on the type of disease, this technique also works with dead microbes, parts of the microbe, or treated toxins from the microbe. A common example of this form of active immunization is vaccinations.
See also
Immunization
Vaccination
Passive immunity |
https://en.wikipedia.org/wiki/GDF15 | Growth/differentiation factor 15 is a protein that in humans is encoded by the GDF15 gene. GDF15 was first identified as Macrophage inhibitory cytokine-1 or MIC-1.
It is a protein belonging to the transforming growth factor beta superfamily. Under normal conditions, GDF15 is expressed in low concentrations in most organs and upregulated because of injury of organs such as liver, kidney, heart and lung.
Function
The function of GDF15 is not fully clear but it seems to have a role in regulating inflammatory pathways and to be involved in regulating apoptosis, angiogenesis, cell repair and cell growth, which are biological processes observed in cardiovascular and neoplastic disorders.
Clinical significance
GDF15 has shown to be a strong prognostic protein in patients with different diseases such as heart diseases and cancer.
However, elevated GDF15 levels in diseases such as cancer and heart disease may be the result of inflammation caused by these diseases. Note that GDF15 is necessary for surviving both bacterial and viral infections, as well as sepsis. The protective effects of GDF15 were largely independent of pathogen control or the magnitude of inflammatory response, suggesting a role in disease tolerance.
Metformin was shown to cause increased levels of GDF15. This increase mediates the effect of body weight loss by metformin. Further study has shown weight loss is promoted by maintaining energy expenditure in addition to appetite suppression.
Elevations in GDF15 reduce food intake and body mass in animal models through binding to glial cell-derived neurotrophic factor family receptor alpha-like (GFRAL) and the recruitment of the receptor tyrosine kinase RET in the hindbrain.
In both mice and humans have shown that metformin and exercise increase circulating levels of GDF15. GDF15 might also exert anti-inflammatory effects through mechanisms that are not fully understood. These unique and distinct mechanisms for suppressing food intake and inflammati |
https://en.wikipedia.org/wiki/Above%20and%20Beyond%3A%20The%20Encyclopedia%20of%20Aviation%20and%20Space%20Sciences | Above and Beyond: The Encyclopedia of Aviation and Space Sciences was the first-ever attempt at creating an encyclopedia of all matters related to the history, technology and aims of the aerospace industry as it existed in the late 1960s. Published in 1967 by New Horizons Publishers, Inc., of Chicago, this fourteen-volume collection was aimed primarily at teens and young adults.
Encyclopedias of science
Children's encyclopedias
Aerospace
American encyclopedias
1967 non-fiction books |
https://en.wikipedia.org/wiki/Methoxypyrazine | Methoxypyrazines are a class of chemical compounds that produce odors. The odors tend to be undesirable, as in the case of certain wines, or as defensive chemicals used by insects such as Harmonia axyridis which produces isopropyl methoxy pyrazine (IPMP). They have also been identified as additives in cigarette manufacture. Detection thresholds are very low, typically near 2 parts per trillion (1 ng/L).
Examples of methoxypyrazines in wine grapes
Cabernet Sauvignon has high levels of methoxypyrazines. Two methoxypyrazine compounds, 3-isobutyl-2-methoxypyrazine (IBMP) and 3-isopropyl-2-methoxypyrazine (IPMP), are considered to be important determinants of green flavours in Sauvignon blanc wines.
See also
Alkylpyrazines |
https://en.wikipedia.org/wiki/Greg%20Jein | Greg Jein (born October 31, 1945, in Los Angeles, US; died May 22, 2022, in Los Angeles) was a Chinese American model designer who created miniatures for use in the special effects portions of many films and television series, beginning in the 1970s. Jein was nominated for an Academy Award for Best Visual Effects for his work on the films Close Encounters of the Third Kind (1977) and 1941 (1979), and also nominated for an Outstanding Special Visual Effects Emmy for his work on Angels in America.
Biography
One of Jein's first jobs was building models for the sex comedy spoof Flesh Gordon; this was followed by work on a number of television series, commercials and movies including Wonder Woman and The UFO Incident. In 1975 he was contacted by Douglas Trumbull's office and asked to do some work on Steven Spielberg's Close Encounters of the Third Kind. For that film Jein contributed a number of models including miniature landscapes for UFOs to fly over, but most significantly he and his crew built the detailed mothership model that features heavily in the final sequence of the film after Spielberg decided he wanted "a more flamboyant design". For their work Jein, Trumbull, Roy Arbogast, Matthew Yuricich, and Richard Yuricich were nominated for an Academy Award for Best Visual Effects at the 50th Academy Awards, but lost to the team who produced the effects for Star Wars. Jein then went on to work on Spielberg's next film, 1941, where he and his team constructed a number of models including a twelve-foot model of the Ferris wheel that's dislodged from its mount and rolls down the pier and into the water. For their work on 1941 Jein, William A. Fraker and A. D. Flowers were nominated for an Academy Award for Best Visual Effects at the 52nd Academy Awards but lost this time to the team who provided the effects for Ridley Scott's Alien.
After working on 1941, Jein was invited by Douglas Trumbull to work on Star Trek: The Motion Picture building planetary models for Spock' |
https://en.wikipedia.org/wiki/Aristotelian%20physics | Aristotelian physics is the form of natural science or natural philosophy described in the works of the Greek philosopher Aristotle (384–322 BC). In his work Physics, Aristotle intended to establish general principles of change that govern all natural bodies, both living and inanimate, celestial and terrestrialincluding all motion (change with respect to place), quantitative change (change with respect to size or number), qualitative change, and substantial change ("coming to be" [coming into existence, 'generation'] or "passing away" [no longer existing, 'corruption']). To Aristotle, 'physics' was a broad field that included subjects that would now be called the philosophy of mind, sensory experience, memory, anatomy and biology. It constitutes the foundation of the thought underlying many of his works.
Key concepts of Aristotelian physics include the structuring of the cosmos into concentric spheres, with the Earth at the centre and celestial spheres around it. The terrestrial sphere was made of four elements, namely earth, air, fire, and water, subject to change and decay. The celestial spheres were made of a fifth element, an unchangeable aether. Objects made of these elements have natural motions: those of earth and water tend to fall; those of air and fire, to rise. The speed of such motion depends on their weights and the density of the medium. Aristotle argued that a vacuum could not exist as speeds would become infinite.
Aristotle described four causes or explanations of change as seen on earth: the material, formal, efficient, and final causes of things. As regards living things, Aristotle's biology relied on observation of natural kinds, both the basic kinds and the groups to which these belonged. He did not conduct experiments in the modern sense, but relied on amassing data, observational procedures such as dissection, and making hypotheses about relationships between measurable quantities such as body size and lifespan.
Methods
While consistent wi |
https://en.wikipedia.org/wiki/Prodromus%20Florae%20Novae%20Hollandiae%20et%20Insulae%20Van%20Diemen | Prodromus Florae Novae Hollandiae et Insulae Van Diemen (Prodromus of the Flora of New Holland and Van Diemen's Land) is book dealing with the flora of Australia written by botanist Robert Brown and published in 1810. Often referred to as Prodromus Flora Novae Hollandiae, or by its standard botanical abbreviation Prodr. Fl. Nov. Holland., it was the first attempt at a survey of the Australian flora. It described over 2040 species, over half of which were published for the first time.
Brown's Prodromus was originally published as Volume One, and following the Praemonenda (Preface), page numbering commences on page 145. Sales of the Prodromus were so poor, however, that Brown withdrew it from sale. Due to the commercial failure of the first volume, pages 1 to 144 were never issued, and Brown never produced the additional volumes that he had planned.
In 1813, a book of illustrations for the Prodromus was published separately by Ferdinand Bauer under the title Ferdinandi Bauer Illustrationes florae Novae Hollandiae sive icones generum quae in Prodromo florae Novae Hollandiae et Insulae Van Diemen descripsit Robertus Brown, usually referred to as Illustrationes Florae Novae Hollandiae. The Prodromus itself was eventually reprinted in 1819, and a slightly modified second edition released in 1821. In 1830, Brown published a short supplement to the Prodromus, entitled Supplementum Primum Prodromi Florae Novae Hollandiae. |
https://en.wikipedia.org/wiki/Lee%20algorithm | The Lee algorithm is one possible solution for maze routing problems based on breadth-first search.
It always gives an optimal solution, if one exists, but is slow and requires considerable memory.
Algorithm
1) Initialization
- Select start point, mark with 0
- i := 0
2) Wave expansion
- REPEAT
- Mark all unlabeled neighbors of points marked with i with i+1
- i := i+1
UNTIL ((target reached) or (no points can be marked))
3) Backtrace
- go to the target point
REPEAT
- go to next node that has a lower mark than the current node
- add this node to path
UNTIL (start point reached)
4) Clearance
- Block the path for future wirings
- Delete all marks
Of course the wave expansion marks only points in the routable area of the chip, not in the blocks or already wired parts, and to minimize segmentation you should keep in one direction as long as possible.
External links
http://www.eecs.northwestern.edu/~haizhou/357/lec6.pdf |
https://en.wikipedia.org/wiki/Flora%20Australiensis | Flora Australiensis: a description of the plants of the Australian Territory, more commonly referred to as Flora Australiensis, and also known by its standard abbreviation Fl. Austral., is a seven-volume flora of Australia published between 1863 and 1878 by George Bentham, with the assistance of Ferdinand von Mueller. It was one of the famous Kew series of colonial floras, and the first flora of any large continental area that had ever been finished. In total the flora included descriptions of 8125 species.
Bentham prepared the flora from Kew; with Mueller, the first plant taxonomist residing permanently in Australia, loaning the entire collection of the National Herbarium of Victoria to Bentham over the course of several years. Mueller had been dissuaded from preparing a flora from Australia while in Australia by Bentham and Joseph Dalton Hooker since historic collections of Australian species were all held in European herbaria which Mueller could not access from Australia. Mueller did eventually produce his own flora of Australia, the Systematic Census of Australian Plants published in 1882 extended the work of Bentham with the addition of new species and taxonomic revisions.
Flora Australiensis was the standard reference work on the Australian flora for more than a century. As late as 1988, James Willis wrote that "Flora Australiensis still remains the only definitive work on the vascular vegetation of the whole continent." According to Nancy Burbidge, "it represents a prodigious intellectual effort never equalled."
Flora Australiensis is credited with forming the basis of subsequently published regional floras; 19th century floras were published for all states except Western Australia, they were for the most part extracts of this work. |
https://en.wikipedia.org/wiki/Isothermal%20coordinates | In mathematics, specifically in differential geometry, isothermal coordinates on a Riemannian manifold are local coordinates where the metric is conformal to the Euclidean metric. This means that in isothermal coordinates, the Riemannian metric locally has the form
where is a positive smooth function. (If the Riemannian manifold is oriented, some authors insist that a coordinate system must agree with that orientation to be isothermal.)
Isothermal coordinates on surfaces were first introduced by Gauss. Korn and Lichtenstein proved that isothermal coordinates exist around any point on a two dimensional Riemannian manifold.
By contrast, most higher-dimensional manifolds do not admit isothermal coordinates anywhere; that is, they are not usually locally conformally flat. In dimension 3, a Riemannian metric is locally conformally flat if and only if its Cotton tensor vanishes. In dimensions > 3, a metric is locally conformally flat if and only if its Weyl tensor vanishes.
Isothermal coordinates on surfaces
In 1822, Carl Friedrich Gauss proved the existence of isothermal coordinates on an arbitrary surface with a real-analytic Riemannian metric, following earlier results of
Joseph Lagrange in the special case of surfaces of revolution. The construction used by Gauss made use of the Cauchy–Kowalevski theorem, so that his method is fundamentally restricted to the real-analytic context. Following innovations in the theory of two-dimensional partial differential equations by Arthur Korn, Leon Lichtenstein found in 1916 the general existence of isothermal coordinates for Riemannian metrics of lower regularity, including smooth metrics and even Hölder continuous metrics.
Given a Riemannian metric on a two-dimensional manifold, the transition function between isothermal coordinate charts, which is a map between open subsets of , is necessarily angle-preserving. The angle-preserving property together with orientation-preservation is one characterization (among many) of |
https://en.wikipedia.org/wiki/Digital%20biquad%20filter | In signal processing, a digital biquad filter is a second order recursive linear filter, containing two poles and two zeros. "Biquad" is an abbreviation of "biquadratic", which refers to the fact that in the Z domain, its transfer function is the ratio of two quadratic functions:
The coefficients are often normalized such that a0 = 1:
High-order infinite impulse response filters can be highly sensitive to quantization of their coefficients, and can easily become unstable. This is much less of a problem with first and second-order filters; therefore, higher-order filters are typically implemented as serially-cascaded biquad sections (and a first-order filter if necessary). The two poles of the biquad filter must be inside the unit circle for it to be stable. In general, this is true for all discrete filters i.e. all poles must be inside the unit circle in the Z-domain for the filter to be stable.
Implementation
Direct form 1
The most straightforward implementation is the direct form 1, which has the following difference equation:
or, if normalized:
Here the , and coefficients determine zeros, and , determine the position of the poles.
Flow graph of biquad filter in direct form 1:
When these sections are cascaded for filters of order greater than 2, efficiency of implementation can be improved by noticing the delay of a section output is cloned in the next section input. Two storage delay components may be eliminated between sections.
Direct form 2
The direct form 2 implements the same normalized transfer function as direct form 1, but in two parts:
and using the difference equation:
Flow graph of biquad filter in direct form 2:
The direct form 2 implementation only needs N delay units, where N is the order of the filter – potentially half as much as direct form 1. This structure is obtained by reversing the order of the numerator and denominator sections of direct Form 1, since they are in fact two linear systems, and the commutativity property a |
https://en.wikipedia.org/wiki/Online%20office%20suite | An online office suite, online productivity suite or cloud office suite is an office suite offered in the form of a web application, accessed online using a web browser. This allows people to work together worldwide and at any time, thereby leading to web-based collaboration and virtual teamwork. Some online office suites can be installed either on-premise or online and some are offered only as online as a software as a service. Some versions can be free of charge, some have a subscription fee. Some online office suites can run as progressive web applications which no longer require an online connection to function. Online office suites exist as both open-source and proprietary software.
Components
Online office suites typically includes this base set of applications
Word processor
Spreadsheet
Presentation program
Other frequently available applications
Webmail
Calendar
Instant messaging including voice and video conferencing
Diagramming
Raster graphics editor
Publishing applications
Content management system
Web portal
Wiki
Blog
Forums
Other applications
Data management
Project management
Customer relationship management
Enterprise resource planning
Accounting
Maps
Notetaking
News
Considerations for weighing Advantages and Disadvantages
Advantages and disadvantages between locally installed office suites and online office suites vary significantly, here are some considerations:
Online office suites provide the ability for a group of people to collaboratively edit a document from anywhere easily using just a web browser, without the need for special servers, or networking customizations, or devices running specific operating systems
They provide varying levels of functionality. Some have nearly full desktop application functionality, whereas some have less functionality and will prompt to use their desktop office suites for the functionality that is missing, which creates extra costs and problems, such as their desktop office suite may not be compatible, or |
https://en.wikipedia.org/wiki/List%20of%20electronics%20brands | This list of electronics brands is specialized as the list of brands of companies that provide electronics equipment.
Categories
Electronics equipment includes the following categories (abbreviations used in parentheses):
audio system (AS) (includes home audio)
avionics (AV)
car audio (CA)
car navigation (CN)
copy machine (CM)
computer (CP) (except personal computer (PC))
digital camera (DC)
display device (DD)
digital video camera (DVC)
digital video player (DVP)
digital video recorder (DVR)
fax (FAX)
global positioning system (GPS)
hard disk drive (HDD)
multifunction printer (MFP)
mechatronics (MN)
mobile phone (MP)
list of video game companies (VG/Electronics)
network device (NW)
personal computer (PC)
portable media player (PMP)
printer (PR)
semiconductor (SC)
video cassette recorder (VHS)
video game (VG)
video game developer (VGD)
video game publisher (VGP)
indie game developer (IGD)
transportation electronics system (TES)
television (TV)
wireless devices (WD)
other electronics equipment (OEE)
Other indications:
( )company name
(( ))parent company name
< >previous company name
<< >>company name in local language
Asia
Bangladesh
BMTF
Doel
Jamuna
Walton
Minister
Marcel
Rangs
Transcom
Butterfly Group
Fair Group
China
Aigo (Beijing Huaqi Information Digital Technology Co. Ltd.)
Amoi
BYD Electronic
Changhong
Gionee
Haier
Hasee
Hisense
Huawei
Konka Group
Meizu
Ningbo Bird
Oppo
Panda
Skyworth
TCL
TP-Link/intex
Vivo Electronics
Zopo Mobile
ZTE
Xiaomi
OnePlus
Hong Kong
Lenovo
India
Amkette
Beetel
Bharat Electronics
BPL
Celkon
Electronics Corporation of India
Godrej
HCL
Havells
IBALL
Intex
Karbonn
Micromax
Myzornis
Moser Baer
Notion Ink
Onida
Surya Roshni Limited
Simmtronics
Sterlite Technologies
Voltas
Videocon
Videotex
Wipro
Indonesia
Axioo
Maspion
Nexian
Zyrex
Iran
Maadiran Group
Snowa
Japan
Allied TelesisNW, OEE
AlpineCA, CN
Atari
Brother IndustriesCM, CP, FAX, MFP, PR, OEE
Buffalo (Melco)HDD, NW, OEE
CanonCM, DC, DVC, FAX, MFP, PR, OEE
C |
https://en.wikipedia.org/wiki/Lie%20bracket%20of%20vector%20fields | In the mathematical field of differential topology, the Lie bracket of vector fields, also known as the Jacobi–Lie bracket or the commutator of vector fields, is an operator that assigns to any two vector fields X and Y on a smooth manifold M a third vector field denoted .
Conceptually, the Lie bracket is the derivative of Y along the flow generated by X, and is sometimes denoted ("Lie derivative of Y along X"). This generalizes to the Lie derivative of any tensor field along the flow generated by X.
The Lie bracket is an R-bilinear operation and turns the set of all smooth vector fields on the manifold M into an (infinite-dimensional) Lie algebra.
The Lie bracket plays an important role in differential geometry and differential topology, for instance in the Frobenius integrability theorem, and is also fundamental in the geometric theory of nonlinear control systems.
Definitions
There are three conceptually different but equivalent approaches to defining the Lie bracket:
Vector fields as derivations
Each smooth vector field on a manifold M may be regarded as a differential operator acting on smooth functions (where and of class ) when we define to be another function whose value at a point is the directional derivative of f at p in the direction X(p). In this way, each smooth vector field X becomes a derivation on C∞(M). Furthermore, any derivation on C∞(M) arises from a unique smooth vector field X.
In general, the commutator of any two derivations and is again a derivation, where denotes composition of operators. This can be used to define the Lie bracket as the vector field corresponding to the commutator derivation:
Flows and limits
Let be the flow associated with the vector field X, and let D denote the tangent map derivative operator. Then the Lie bracket of X and Y at the point can be defined as the Lie derivative:
This also measures the failure of the flow in the successive directions to return to the point x:
In coordinates
Though t |
https://en.wikipedia.org/wiki/Backtesting | Backtesting is a term used in modeling to refer to testing a predictive model on historical data. Backtesting is a type of retrodiction, and a special type of cross-validation applied to previous time period(s).
Financial analysis
In the economic and financial field, backtesting seeks to estimate the performance of a strategy or model if it had been employed during a past period. This requires simulating past conditions with sufficient detail, making one limitation of backtesting the need for detailed historical data. A second limitation is the inability to model strategies that would affect historic prices. Finally, backtesting, like other modeling, is limited by potential overfitting. That is, it is often possible to find a strategy that would have worked well in the past, but will not work well in the future. Despite these limitations, backtesting provides information not available when models and strategies are tested on synthetic data.
Backtesting has historically only been performed by large institutions and professional money managers due to the expense of obtaining and using detailed datasets. However, backtesting is increasingly used on a wider basis, and independent web-based backtesting platforms have emerged. Although the technique is widely used, it is prone to weaknesses. Basel financial regulations require large financial institutions to backtest certain risk models.
For a Value at Risk 1-day at 99% backtested 250 days in a row, the test is considered green (0-95%), orange (95-99.99%) or red (99.99-100%) depending on the following table:
For a Value at Risk 10-day at 99% backtested 250 days in a row, the test is considered green (0-95%), orange (95-99.99%) or red (99.99-100%) depending on the following table:
Hindcast
In oceanography and meteorology, backtesting is also known as hindcasting: a hindcast is a way of testing a mathematical model; researchers enter known or closely estimated inputs for past events into the model to see how wel |
https://en.wikipedia.org/wiki/XLT%20agar | XLT Agar (Xylose Lysine Tergitol-4) is a selective culture medium for the isolation and identification of salmonellae from food and environmental samples. It is similar to XLD agar; however, the agar is supplemented with the surfactant, Tergitol 4, which causes inhibition of Proteus spp. and other non-Salmonellae.
Successful growth of Salmonella will result in growth of red colonies with a black centre.
XLT Agar contains:
See also
Agar plate
XLD agar
R2a agar
MRS agar |
https://en.wikipedia.org/wiki/Princeton%20Ocean%20Model | The Princeton Ocean Model (POM) is a community general numerical model for ocean circulation that can be used to simulate and predict oceanic currents, temperatures, salinities and other water properties. POM-WEB and POMusers.org
Development
The model code was originally developed at Princeton University (G. Mellor and Alan Blumberg) in collaboration with Dynalysis of Princeton (H. James Herring, Richard C. Patchen). The model incorporates the Mellor–Yamada turbulence scheme developed in the early 1970s by George Mellor and Ted Yamada; this turbulence sub-model is widely used by oceanic and atmospheric models. At the time, early computer ocean models such as the Bryan–Cox model (developed in the late 1960s at the Geophysical Fluid Dynamics Laboratory, GFDL, and later became the Modular Ocean Model, MOM)), were aimed mostly at coarse-resolution simulations of the large-scale ocean circulation, so there was a need for a numerical model that can handle high-resolution coastal ocean processes. The Blumberg–Mellor model (which later became POM) thus included new features such as free surface to handle tides, sigma vertical coordinates (i.e., terrain-following) to handle complex topographies and shallow regions, a curvilinear grid to better handle coastlines, and a turbulence scheme to handle vertical mixing. At the early 1980s the model was used primarily to simulate estuaries such as the Hudson–Raritan Estuary (by Leo Oey) and the Delaware Bay (Boris Galperin), but also first attempts to use a sigma coordinate model for basin-scale problems have started with the coarse resolution model of the Gulf of Mexico (Blumberg and Mellor) and models of the Arctic Ocean (with the inclusion of ice-ocean coupling by Lakshmi Kantha and Sirpa Hakkinen).
In the early 1990s when the web and browsers started to be developed, POM became one of the first ocean model codes that were provided free of charge to users through the web. The establishment of the POM users group and its web sup |
https://en.wikipedia.org/wiki/Foundational%20Model%20of%20Anatomy | The Foundational Model of Anatomy Ontology (FMA) is a reference ontology for the domain of human anatomy. It is a symbolic representation of the canonical, phenotypic structure of an organism; a spatial-structural ontology of anatomical entities and relations which form the physical organization of an organism at all salient levels of granularity.
FMA is developed and maintained by the Structural Informatics Group at the University of Washington.
Description
FMA ontology contains approximately 75,000 classes and over 120,000 terms, over 2.1 million relationship instances from over 168 relationship types.
See also
Terminologia Anatomica
Anatomography |
https://en.wikipedia.org/wiki/James%20A.%20Shapiro | James Alan Shapiro (born May 18, 1943) is an American biologist, an expert in bacterial genetics and a professor in the Department of Biochemistry and Molecular Biology at the University of Chicago.
Academic biography
Shapiro obtained his Bachelor's degree in English from Harvard College in 1964. Then, inspired by a genetics course he had taken as a senior, he shifted from English to science. He was awarded a Marshall scholarship for postgraduate research at Corpus Christi College, Cambridge from 1964 to 1967, spending his final year at Hammersmith hospital under the supervision of William Hayes, and being awarded a PhD in genetics in 1968.
His thesis, The Structure of the Galactose Operon in Escherichia coli K12, contains the first suggestion of transposable elements in bacteria. He confirmed this hypothesis in 1968 during his postdoctoral tenure as a Jane Coffin Childs fellow in the laboratory of François Jacob at the Institut Pasteur in Paris.
As an American Cancer Society fellow in Jon Beckwith’s laboratory at the Harvard Medical School 1968-70, he and his colleagues used in vivo genetic manipulations to clone and purify the lac operon of E. coli.
He was troubled by the potential genetic engineering applications of his research.
He served as Invited Professor in the School of Biological Sciences at the University of Havana, Cuba 1970-1972, before returning to another postdoctorate with Harlyn Halvorson at Brandeis University. Since 1973, he has worked as a professor of microbiology at the University of Chicago.
In 1975 Shapiro attended the ICN-UCLA Squaw Valley Symposium on Bacterial Plasmids, where his interest in DNA restructuring in bacteria was heightened by learning about the movements of antibiotic resistance transposons to new genomic locations. This prompted him to organize, in collaboration with Sankar Adhya and the late Ahmed Bukhari, the first meeting on the topic of DNA insertion elements at Cold Spring Harbour Laboratory in 1976. Although they |
https://en.wikipedia.org/wiki/Tetramer%20assay | A tetramer assay (also known as a tetramer stain) is a procedure that uses tetrameric proteins to detect and quantify T cells that are specific for a given antigen within a blood sample. The tetramers used in the assay are made up of four major histocompatibility complex (MHC) molecules, which are found on the surface of most cells in the body. MHC molecules present peptides to T-cells as a way to communicate the presence of viruses, bacteria, cancerous mutations, or other antigens in a cell. If a T-cell's receptor matches the peptide being presented by an MHC molecule, an immune response is triggered. Thus, MHC tetramers that are bioengineered to present a specific peptide can be used to find T-cells with receptors that match that peptide.
The tetramers are labeled with a fluorophore, allowing tetramer-bound T-cells to be analyzed with flow cytometry. Quantification and sorting of T-cells by flow cytometry enables researchers to investigate immune response to viral infection and vaccine administration as well as functionality of antigen-specific T-cells. Generally, if a person's immune system has encountered a pathogen, the individual will possess T cells with specificity toward some peptide on that pathogen. Hence, if a tetramer stain specific for a pathogenic peptide results in a positive signal, this may indicate that the person's immune system has encountered and built a response to that pathogen.
History
This methodology was first published in 1996 by a lab at Stanford University. Previous attempts to quantify antigen-specific T-cells involved the less accurate limiting dilution assay, which estimates numbers of T-cells at 50-500 times below their actual levels. Stains using soluble MHC monomers were also unsuccessful due to the low binding affinity of T-cell receptors and MHC-peptide monomers. MHC tetramers can bind to more than one receptor on the target T-cell, resulting in an increased total binding strength and lower dissociation rates.
Uses
CD8+ T-ce |
https://en.wikipedia.org/wiki/Maxwell%20Finland%20Award | The Maxwell Finland Award for Scientific Achievement is an award given annually by the National Foundation for Infectious Diseases to a scientist who has made "outstanding contributions to the understanding of infectious diseases or public health," based on criteria that include "excellence in clinical and/or research activities; participation in the training of future leaders in the field; and positive impact on the health of humankind." The award is named after epidemiologist Maxwell Finland, who investigated antimicrobial resistance. The first award was given in 1988.
Past winners
See also
List of biomedical science awards |
https://en.wikipedia.org/wiki/Robert%20Fricke | Karl Emanuel Robert Fricke (24 September 1861 – 18 July 1930) was a German mathematician, known for his work in complex analysis, especially on elliptic, modular and automorphic functions. He was one of the main collaborators of Felix Klein, with whom he produced two classic, two-volume monographs on elliptic modular functions and automorphic functions.
In 1893 in Chicago, his paper Die Theorie der automorphen Functionen und die Arithmetik was read (but not by Fricke) at the International Mathematical Congress held in connection with the World's Columbian Exposition. From 1894 to 1930 Fricke was professor of Higher Mathematics at the Technische Hochschule Carolo-Wilhelmina in Braunschweig.
See also
Fricke involution
Bibliography
;
; |
https://en.wikipedia.org/wiki/Portable%20operation%20%28amateur%20radio%29 | Amateur radio operators take part in portable operations using radio equipment when traveling. "Portable" equipment indicates a configuration that allows for relatively rapid collection, transportation, and deployment of amateur radio gear. A portable station can be anything from a small QRP (Low Power) radio and antenna, to a large transceiver. On long-distance expeditions, such equipment allows them to report progress, arrivals and sometimes exchanging safety messages along the way.
Portable operations
'Portable' operation is usually signified by amateur radio operators appending the suffix '/P' to their callsign. Operating '/P' normally means that stations are operating away from their licensed station address.
The advantages of /P operation include the use of large empty spaces where full size beam and wire antennas can be erected on tall trailer mounted masts. If operating on VHF/UHF, this can mean a location on the top of a hill or cliff, with clear line of sight to the horizon. The main disadvantage is normally the power supply available. As normal mains grid power is unavailable, the /P operator may have to resort to batteries, portable generators, solar panels. and wind turbines.
Operating amateur radio at sea is known as 'maritime mobile', as is signified by the suffix '/MM' on the call.
Operating amateur radio from a vehicle is known as 'Mobile', as is signified by the suffix '/M' on the call.
A popular pastime for portable operation is the Summits on the Air programme, part of which involves portable operation from a worldwide list of over 73,500 summits.
Some countries allow the direct connection of amateur transceivers to telephone lines called "phone patching". Thus, a traveler may be able to call another amateur station and, via a phone patch, speak directly with someone else by telephone.
Notes
Portable electronics
Amateur radio |
https://en.wikipedia.org/wiki/PHLPP | The PHLPP isoforms (PH domain and Leucine rich repeat Protein Phosphatases) are a pair of protein phosphatases, PHLPP1 and PHLPP2, that are important regulators of Akt serine-threonine kinases (Akt1, Akt2, Akt3) and conventional/novel protein kinase C (PKC) isoforms. PHLPP may act as a tumor suppressor in several types of cancer due to its ability to block growth factor-induced signaling in cancer cells.
PHLPP dephosphorylates Ser-473 (the hydrophobic motif) in Akt, thus partially inactivating the kinase.
In addition, PHLPP dephosphorylates conventional and novel members of the protein kinase C family at their hydrophobic motifs, corresponding to Ser-660 in PKCβII.
Domain structure
PHLPP is a member of the PPM family of phosphatases, which requires magnesium or manganese for their activity and are insensitive to most common phosphatase inhibitors, including [okadaic acid]. PHLPP1 and PHLPP2 have a similar domain structure, which includes a putative Ras association domain, a pleckstrin homology domain, a series of leucine-rich repeats, a PP2C phosphatase domain, and a C-terminal PDZ ligand. PHLPP1 has two splice variants, PHLPP1α and PHLPP1β, of which PHLPP1β is larger by approximately 1.5 kilobase pairs. PHLPP1α, which was the first PHLPP isoform to be characterized, lacks the N-terminal portion of the protein, including the Ras association domain. PHLPP's domain structure influences its ability to dephosphorylate its substrates. A PHLPP construct lacking the PH domain is unable to decrease PKC phosphorylation, while PHLPP lacking the PDZ ligand is unable to decrease Akt phosphorylation.
Dephosphorylation of Akt
The phosphatases in the PHLPP family, PHLPP1 and PHLPP2 have been shown to directly dephosphorylate, and therefore inactivate, distinct Akt isoforms, at one of the two critical phosphorylation sites required for activation: Serine473. PHLPP2 dephosphorylates AKT1 and AKT3, whereas PHLPP1 is specific for AKT2 and AKT3. Lack of PHLPP appears |
https://en.wikipedia.org/wiki/Chen%20Tze-chiang | Dr. Chen Tze-chiang, or T. C. Chen (), joined the IBM Thomas J. Watson Research Center in 1984. He is currently an IBM Fellow and the Vice President of Science and Technology at Thomas J. Watson Research Center, IBM Research Division in Yorktown Heights, New York.
During the period of Feb. 1999 - Feb. 2003, Dr. Chen was the Director of Advanced Logic/Memory Technology Development at the Semiconductor Research and Development Center, IBM Microelectronics Division in East Fishkill, New York. During the period of 1992–1999, he was the Senior Manager responsible for the 64Mb/256Mb/1Gb DRAM Technology Development in IBM/Siemens/Toshiba DRAM Development Alliance. Before assuming his role as the Project Manager in DRAM Development, he held a variety of managerial assignments in his career at IBM, including Functional Manager of High-Performance BiCMOS Technology, Manufacturing Engineering Manager in the Bipolar VLSI Line, and Manager of Optical Lithography Development at Watson Research Center.
Dr. Chen was born in Taiwan in 1951. He received the B.Sc. and M.S. degrees in physics from National Cheng Kung University, Taiwan, in 1974 and 1976, respectively, and the M.S. and Ph.D. Degrees in electrical engineering from Yale University in 1979 and 1985, respectively. He has published more than 60 papers in technical journals and conferences. His contributions to advanced bipolar technology had a major impact on IBM S/390 mainframe systems. More recently, his contributions to CMOS miniaturization and DRAM devices have had a profound impact on IBM's leadership in CMOS process technology and DRAM manufacturing. Dr. Chen was elected as a Fellow of the Institute of Electrical and Electronics Engineers (IEEE) in 1999 for contributions to silicon bipolar and DRAM technology development. Dr.Chen has been recognized with several IBM Technical Innovation Awards and was named IBM Distinguished Engineer and IBM Fellow in 1996 and 1999, respectively.
Awards and honors
In 2011 he receive |
https://en.wikipedia.org/wiki/Combined%20drug%20intoxication | Combined drug intoxication (CDI), or multiple drug intake (MDI), is a cause of death by drug overdose from poly drug use, often implicated in polysubstance dependence.
Deaths by combined drug intoxication are relatively rare (one in several million). In 2004, there were 3,800 deaths in the US resulting from a fatal medication error involving alcohol, while in 1983 there were less than 100 such deaths. It is more of a risk for older patients.
Risk factors
People who engage in polypharmacy and other hypochondriac behaviors are at an elevated risk of death from CDI. Other dangers of combining drugs such as "brain damage, heart problems, seizures, stomach bleeding, liver damage/ liver failure, heatstroke, coma, suppressed breathing, and respiratory failure", along with many other complications. Disorders like depression and anxiety can also stem from polydrug use. Elderly people are at the highest risk of CDI, because of having many age-related and health problems requiring many medications combined with age-impaired judgment, leading to confusion in taking medications. Elderly patients are often prescribed more than one drug within the same drug class, and doctors may treat the side effects of prescribed drugs with even more drugs, which can overwhelm the patient.
Prevention
In general, the simultaneous use of multiple drugs should be carefully monitored by a qualified individual such as board certified and licensed medical doctor, either an MD or DO. Close association between prescribing physicians and pharmacies, along with the computerization of prescriptions and patients' medical histories, aim to avoid the occurrence of dangerous drug interactions. Lists of contraindications for a drug are usually provided with it, either in monographs, package inserts (accompanying prescribed medications), or in warning labels (for OTC drugs). CDI/MDI might also be avoided by physicians requiring their patients to return any unused prescriptions. Patients should ask their do |
https://en.wikipedia.org/wiki/Coupon-eligible%20converter%20box | A coupon-eligible converter box (CECB) was a digital television adapter that met eligibility specifications for subsidy "coupons" from the United States government. The subsidy program was enacted to provide terrestrial television viewers with an affordable way to continue receiving free digital terrestrial television services after the nation's television service transitioned to digital transmission and analog transmissions ceased. The specification was developed by the National Telecommunications and Information Administration (NTIA), with input from the broadcast and consumer electronics industries as well as public interest groups.
History
Early proposals
In March 2005, United States House Commerce Committee chairman Joe Barton of Texas said he would introduce a bill requiring the transition to digital television "sometime in the spring", saying he wanted analog broadcasting to end on December 31, 2006. Included in his plan was a $400–$500 million subsidy for converter boxes, which were expected to cost $50 each. The subsidies were intended only for people who could not afford a pay service such as cable or satellite television. Each home would receive a rebate coupon for one box, which could be mailed to the United States Treasury for redemption. Barton estimated that 8 to 10 million converters would be needed.
Digital Transition and Public Safety Act of 2005
The United States Digital Transition and Public Safety Act of 2005, part of the Deficit Reduction Act of 2005, required that the Federal Communications Commission (FCC) direct all full-power television stations to cease analog TV broadcasting before midnight on February 17, 2009. (This deadline later changed to June 12.) Recognizing that consumers might wish to continue receiving broadcast programming over-the-air using analog-only televisions, the Act authorized the NTIA to create a digital-to-analog converter box assistance program. Consumer education plans for the subsidy program were targeted |
https://en.wikipedia.org/wiki/Elliptic%20cohomology | In mathematics, elliptic cohomology is a cohomology theory in the sense of algebraic topology. It is related to elliptic curves and modular forms.
History and motivation
Historically, elliptic cohomology arose from the study of elliptic genera. It was known by Atiyah and Hirzebruch that if acts smoothly and non-trivially on a spin manifold, then the index of the Dirac operator vanishes. In 1983, Witten conjectured that in this situation the equivariant index of a certain twisted Dirac operator is at least constant. This led to certain other problems concerning -actions on manifolds, which could be solved by Ochanine by the introduction of elliptic genera. In turn, Witten related these to (conjectural) index theory on free loop spaces. Elliptic cohomology, invented in its original form by Landweber, Stong and Ravenel in the late 1980s, was introduced to clarify certain issues with elliptic genera and provide a context for (conjectural) index theory of families of differential operators on free loop spaces. In some sense it can be seen as an approximation to the K-theory of the free loop space.
Definitions and constructions
Call a cohomology theory even periodic if for i odd and there is an invertible element . These theories possess a complex orientation, which gives a formal group law. A particularly rich source for formal group laws are elliptic curves. A cohomology theory with
is called elliptic if it is even periodic and its formal group law is isomorphic to a formal group law of an elliptic curve over . The usual construction of such elliptic cohomology theories uses the Landweber exact functor theorem. If the formal group law of is Landweber exact, one can define an elliptic cohomology theory (on finite complexes) by
Franke has identified the condition needed to fulfill Landweber exactness:
needs to be flat over
There is no irreducible component of , where the fiber is supersingular for every
These conditions can be checked in many cases r |
https://en.wikipedia.org/wiki/Fiber%20disk%20laser | A fiber disk laser is a fiber laser with transverse delivery of the pump light. They are characterized by the pump beam not being parallel to the active core of the optical fiber (as in a double-clad fiber), but directed to the coil of the fiber at an angle (usually, between 10 and 40 degrees). This allows use of the specific shape of the pump beam emitted by the laser diode, providing the efficient use of the pump.
Fiber disk lasers should not be confused with the LaserDiscs (disk-shaped devices for storage and reading of information with laser beam) nor the disk laser or "active mirror", which is a laser with a thin active layer where the heat sink is realized in a direction opposite to that of propagation of the output beam.
Realizations of fiber disk lasers
First disk lasers were developed in the Institute for Laser Science, Japan.
Several realizations of fiber disk lasers were reported. The fiber disk laser is so named because the fiber is tightly coiled. Typically, no special feedback for the laser frequency is required, as the small reflection at end of the fiber is sufficient to provide efficient operation. In this case, both ends of the coiled fiber can be used as output.
Application and power scaling
Fiber disk lasers are used for cutting of metal (up to few mm thick), welding and folding. The disk-shaped configuration allows efficient heat dissipation (usually, the disks are cooled with flowing water)); allowing power scaling. When the increase of the length of the fiber becomes limited by stimulated scattering, additional power scaling can be achieved by combining several fiber disk lasers into a stack.
The spiral-coiled configuration is not the only possible arrangement; any other scheme of stacking of optical fibers with lateral delivery of pump can also be called a fiber disk laser, even if the resulting shape of the device is not circular. The term fiber disk laser applies to the concept of lateral delivery of pump to the active optical fiber ra |
https://en.wikipedia.org/wiki/Gravity%20gradiometry | Gravity gradiometry is the study of variations (anomalies) in the Earth's gravity field via measurements of the spatial gradient of gravitational acceleration. The gravity gradient tensor is a 3x3 tensor representing the partial derivatives, along each coordinate axis, of each of the three components of the acceleration vector (), totaling 9 scalar quantities:
It has dimension of square reciprocal time, in units of s-2 (or mm-1s-2).
Gravity gradiometry is used by oil and mineral prospectors to measure the density of the subsurface, effectively by measuring the rate of change of gravitational acceleration due to underlying rock properties. From this information it is possible to build a picture of subsurface anomalies which can then be used to more accurately target oil, gas and mineral deposits. It is also used to image water column density, when locating submerged objects, or determining water depth (bathymetry). Physical scientists use gravimeters to determine the exact size and shape of the earth and they contribute to the gravity compensations applied to inertial navigation systems.
Gravity gradient
Gravity measurements are a reflection of the earth's gravitational attraction, its centripetal force, tidal accelerations due to the sun, moon, and planets, and other applied forces. Gravity gradiometers measure the spatial derivatives of the gravity vector. The most frequently used and intuitive component is the vertical gravity gradient, Gzz, which represents the rate of change of vertical gravity (gz) with height (z). It can be deduced by differencing the value of gravity at two points separated by a small vertical distance, l, and dividing by this distance.
The two gravity measurements are provided by accelerometers which are matched and aligned to a high level of accuracy.
Units
The unit of gravity gradient is the eotvos (abbreviated as E), which is equivalent to 10−9 s−2 (or 10−4 mGal/m). A person walking past at a distance of 2 metres would provide a grav |
https://en.wikipedia.org/wiki/Melittology | Melittology (from Greek , melitta, "bee"; and -logia) is a branch of entomology concerning the scientific study of bees. It may also be called apicology. Melittology covers the species found in the clade Anthophila within the superfamily Apoidea, comprising more than 20,000 species, including bumblebees and honey bees.
Subdivisions
Apiology – (from Latin , "bee"; and Ancient Greek , -logia) is the scientific study of honey bees. Honey bees are often chosen as a study group to answer questions on the evolution of social systems.
Apidology is a variant spelling of apiology used outside of the Western Hemisphere, primarily in Europe; it is sometimes used interchangeably with melittology.
Melittological societies
Melittologists and apiologists are served by a number of scientific societies, both national and international in scope. Their main role is to encourage the study of bees and apicultural research.
International Bee Research Association
National Bee Association of New Zealand
British Beekeepers Association
German Beekeepers Association
Federation of Irish Beekeepers' Associations
Melittological journals
Apidologie
American Bee Journal
Journal of Apicultural Research
Journal of Melittology
See also
Bee
Honey bee
Beekeeping
Entomology
I Have a Bee
Vespology
Melissopalynology |
https://en.wikipedia.org/wiki/African%20Society%20for%20Bioinformatics%20and%20Computational%20Biology | The African Society for Bioinformatics and Computational Biology (ASBCB) is a non-profit professional association dedicated to the advancement of bioinformatics and computational biology in Africa. Transformed from the African Bioinformatics Network (ABioNET), ASBCB was established in February 2004 at a meeting in Cape Town, South Africa. The Society serves as an international forum and resource devoted to developing competence and expertise in bioinformatics and computational biology in Africa. It complements its activities with those of other international and national societies, associations and institutions, public and private, that have similar aims. It also promotes the standing of African bioinformatics and computational biology in the global arena through liaison and cooperation with other international bodies.
It is an affiliated regional group of the International Society for Computational Biology (ISCB).
Vision
The Society sees itself as conduit to promote the exchange of ideas, infrastructure and resources in the fields of bioinformatics and computational biology and facilitate the interaction and collaboration among scientists and educators around the world.
It also strives to measurably advance the awareness and understanding of the science of bioinformatics and computational biology. The society represents the bioinformatics and computational biology community in Africa and will be the most respected and reliable international non-profit organization representing this community.
Mission
Be a scholarly society dedicated to advancing, developing and promoting bioinformatics and computational biology in Africa.
Serve a global membership by impacting government and scientific policies, providing high quality publications and meetings, and through distribution of valuable information about training, education, employment and relevant news from related fields.
Develop the application of bioinformatics in Africa in collaboration with individuals, grou |
https://en.wikipedia.org/wiki/Reduced%20residue%20system | In mathematics, a subset R of the integers is called a reduced residue system modulo n if:
gcd(r, n) = 1 for each r in R,
R contains φ(n) elements,
no two elements of R are congruent modulo n.
Here φ denotes Euler's totient function.
A reduced residue system modulo n can be formed from a complete residue system modulo n by removing all integers not relatively prime to n. For example, a complete residue system modulo 12 is {0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11}. The so-called totatives 1, 5, 7 and 11 are the only integers in this set which are relatively prime to 12, and so the corresponding reduced residue system modulo 12 is {1, 5, 7, 11}. The cardinality of this set can be calculated with the totient function: φ(12) = 4. Some other reduced residue systems modulo 12 are:
{13,17,19,23}
{−11,−7,−5,−1}
{−7,−13,13,31}
{35,43,53,61}
Facts
If {r1, r2, ... , rφ(n)} is a reduced residue system modulo n with n > 2, then .
Every number in a reduced residue system modulo n is a generator for the additive group of integers modulo n.
If {r1, r2, ... , rφ(n)} is a reduced residue system modulo n, and a is an integer such that gcd(a, n) = 1, then {ar1, ar2, ... , arφ(n)} is also a reduced residue system modulo n.
See also
Complete residue system modulo m
Multiplicative group of integers modulo n
Congruence relation
Euler's totient function
Greatest common divisor
Least residue system modulo m
Modular arithmetic
Number theory
Residue number system
Notes |
https://en.wikipedia.org/wiki/Death%20and%20the%20Maiden%20%28motif%29 | Death and the Maiden (Der Tod und das Mädchen in German) was a common motif in Renaissance art, especially painting and prints in Germany. The usual form shows just two figures, with a young woman being seized by a personification of Death, often shown as a skeleton. Variants may include other figures. It developed from the Danse Macabre with an added erotic subtext. The German artist Hans Baldung depicted it several times.
The motif was revived during the romantic era in the arts, a notable example being Franz Schubert's song "Der Tod und das Mädchen", setting a poem by the German poet Matthias Claudius. Part of the piano part was re-used in Schubert's famous String Quartet No. 14, which is therefore also known by this title, in either English or German.
Selected versions
Painting: Death and the Maiden (Der Tod und das Mädchen) by Hans Baldung (1517)
Painting: Death and the Maiden (Der Tod und das Mädchen) by Niklaus Manuel Deutsch I (1517)
Painting: Young Woman and Death (La jeune fille et la mort) by Henri-Léopold Lévy (1876)
Engraving: Death and the Maiden (Døden og Piken) by Edvard Munch (1894)
Painting: Death and the Maiden (Der Tod und das Mädchen) by Adolf Hering (1900) - private collection, location unknown
Painting: Death and the Maiden by Marianne Stokes (1908)
Painting: Death and the Maiden (Tod und Mädchen) by Egon Schiele (1915)
Drawing: Death and the Maiden (Der Tod und das Mädchen) by Clara Siewert (1920s)
Drawing: Death and the Maiden (Der Tod und das Mädchen) by Joseph Beuys (1959)
Painting: Death and the Maiden, Ballet for Two by Herbert Lautman (1995)
Gallery
Notes
External links
Death and the Maiden @ La Mort dans l'Art
Iconography
Women and death
Renaissance art
Romanticism
Personifications of death
Visual motifs |
https://en.wikipedia.org/wiki/Systematic%20Census%20of%20Australian%20Plants | The Systematic census of Australian plants, with chronologic, literary and geographic annotations, more commonly known as the Systematic Census of Australian Plants, also known by its standard botanic abbreviation Syst. Census Austral. Pl., is a survey of the vascular flora of Australia prepared by Government botanist for the state of Victoria Ferdinand von Mueller and published in 1882.
Von Mueller describes the development of the census in the preface of the volume as an extension of the seven volumes of the Flora Australiensis written by George Bentham. A new flora was necessary since as more areas of Australia were explored and settled, the flora of the island-continent became better collected and described. The first census increased the number of described species from the 8125 in Flora Australiensis to 8646. The book records all the known species indigenous to Australia and Norfolk Island; with records of species distribution.
Von Mueller noted that by 1882 it had become difficult to distinguish some introduced species from native ones:
The lines of demarkation between truly indigenous and more recently immigrated plants can no longer in all cases be drawn with precision; but whereas Alchemilla vulgaris and Veronica serpyllifolia were found along with several European Carices in untrodden parts of the Australian Alps during the author's earliest explorations, Alchemilla arvensis and Veronica peregrina were at first only noticed near settlements. The occurrence of Arabis glabra, Geum urbanum, Agiimonia eupatoria, Eupatorium cannabinum, Cavpesium cernuum and some others may therefore readily be disputed as indigenous, and some questions concerning the nativity of various of our plants will probably remain for ever involved in doubts.
In 1889 an updated edition of the census was published, the Second Systematic Census increased the number of described species to 8839. Von Mueller dedicated both works to Joseph Dalton Hooker and Augustin Pyramus de Candolle.
|
https://en.wikipedia.org/wiki/Quantum%20healing | Quantum healing is a pseudoscientific mixture of ideas which purportedly draws from quantum mechanics, psychology, philosophy, and neurophysiology. Advocates of quantum healing assert that quantum phenomena govern health and wellbeing. There are different versions, which allude to various quantum ideas including wave particle duality and virtual particles, and more generally to "energy" and to vibrations. Quantum healing is a form of alternative medicine.
Deepak Chopra coined the term "quantum healing" when he published the first edition of his book with that title in 1989. His discussions of quantum healing have been characterised as technobabble - "incoherent babbling strewn with scientific terms" which drives those who actually understand physics "crazy" and as "redefining Wrong".
Quantum healing has a number of vocal followers, but the scientific community widely regards it as nonsensical. The main criticism revolves around its systematic misinterpretation of modern physics, especially of the fact that macroscopic objects (such as the human body or individual cells) are much too large to exhibit inherently quantum properties like interference and wave function collapse.
Physicist and science communicator Brian Cox argues that misuse of the word "quantum", such as its use in the phrase quantum healing, has a negative effect on society as it undermines genuine science and discourages people from engaging with conventional medicine. He states that "for some scientists, the unfortunate distortion and misappropriation of scientific ideas that often accompanies their integration into popular culture is an unacceptable price to pay."
See also
List of esoteric healing articles
Quantum mysticism
Quantum mind |
https://en.wikipedia.org/wiki/Pleocytosis | In medicine, pleocytosis (or pleiocytosis) is an increased cell count (from Greek pleion, "more"), particularly an increase in white blood cell count, in a bodily fluid, such as cerebrospinal fluid. It is often defined specifically as an increased white blood cell count in cerebrospinal fluid.
Increased white blood cell count in the blood is called leukocytosis. |
https://en.wikipedia.org/wiki/Running%20economy | Running economy (RE) a complex, multifactorial concept that represents the sum of metabolic, cardiorespiratory, biomechanical and neuromuscular efficiency during running. Oxygen consumption (VO2) is the most commonly used method for measuring running economy, as the exchange of gases in the body, specifically oxygen and carbon dioxide, closely reflects energy metabolism. Those who are able to consume less oxygen while running at a given velocity are said to have a better running economy. However, straightforward oxygen usage does not account for whether the body is metabolising lipids or carbohydrates, which produce different amounts of energy per unit of oxygen; as such, accurate measurements of running economy must use and data to estimate the calorific content of the substrate that the oxygen is being used to respire.
In distance running, an athlete may attempt to improve performance through training designed to improve running economy. Running economy has been found to be a good predictor of race performance; it has been found to be a stronger correlate of performance than maximal oxygen uptake (VO2 max) in trained runners with the same values.
The idea of running economy is increasingly used to understand performance, as new technology can drastically lower running times over marathon distances, independently of physiology or even training. Factors affecting running economy include a runner’s biology, training regimens, equipment, and environment. The recent accomplishment of Eliud Kipchoge running a marathon in under two hours has enhanced interest in the subject.
Measurement and values
Measurement
Running Economy is calculated by measuring VO₂ while running on a treadmill at various constant speeds for anywhere between three and fifteen minutes. VO₂ is the amount of oxygen consumed in milliliters over one minute and normalized by kilogram of body weight. To compare running economies between individuals, VO₂ is interpolated to common running velocities |
https://en.wikipedia.org/wiki/Oxoguanine%20glycosylase | 8-Oxoguanine glycosylase, also known as OGG1, is a DNA glycosylase enzyme that, in humans, is encoded by the OGG1 gene. It is involved in base excision repair. It is found in bacterial, archaeal and eukaryotic species.
Function
OGG1 is the primary enzyme responsible for the excision of 8-oxoguanine (8-oxoG), a mutagenic base byproduct that occurs as a result of exposure to reactive oxygen species (ROS). OGG1 is a bifunctional glycosylase, as it is able to both cleave the glycosidic bond of the mutagenic lesion and cause a strand break in the DNA backbone. Alternative splicing of the C-terminal region of this gene classifies splice variants into two major groups, type 1 and type 2, depending on the last exon of the sequence. Type 1 alternative splice variants end with exon 7 and type 2 end with exon 8. One set of spliced forms are designated 1a, 1b, 2a to 2e. All variants have the N-terminal region in common. Many alternative splice variants for this gene have been described, but the full-length nature for every variant has not been determined. In eukaryotes, the N-terminus of this gene contains a mitochondrial targeting signal, essential for mitochondrial localization. However, OGG1-1a also has a nuclear location signal at its C-terminal end that suppresses mitochondrial targeting and causes OGG1-1a to localize to the nucleus. The main form of OGG1 that localizes to the mitochondria is OGG1-2a. A conserved N-terminal domain contributes residues to the 8-oxoguanine binding pocket. This domain is organised into a single copy of a TBP-like fold.
Despite the presumed importance of this enzyme, mice lacking Ogg1 have been generated and found to have a normal lifespan, and Ogg1 knockout mice have a higher probability to develop cancer, whereas MTH1 gene disruption concomitantly suppresses lung cancer development in Ogg1-/- mice. Mice lacking Ogg1 have been shown to be prone to increased body weight and obesity, as well as high-fat-diet-induced insulin resistan |
https://en.wikipedia.org/wiki/Gordon%20Royle | Gordon F. Royle is a professor at the School of Mathematics and Statistics at The University of Western Australia.
Royle is the co-author (with Chris Godsil) of the book Algebraic Graph Theory (Springer Verlag, 2001, ).
Royle is also known for his research into the mathematics of Sudoku and his search for the Sudoku puzzle with the smallest number of entries that has a unique solution.
Royle earned his Ph.D. in 1987 from the University of Western Australia under the supervision of Cheryl Praeger and Brendan McKay. |
https://en.wikipedia.org/wiki/Generic%20views | The principal of generic views in the study of cognition stipulates that the interpretation made by an observer of a distal phenomenon should be such as to not require that the observer be in a special position to, or relationship with, the phenomenon in question. The principal is a fairly general account of the inductive bias that allows an observer to reconstruct distal phenomena from an impoverished proximal datum. This principle has been advanced particularly in vision research as an account of how, for example, three-dimensional structure is extracted from an inadequate two-dimensional projection.
The principal of generic views has been discussed by Richards and Hoffman, and has been given a sophisticated Bayesian formalization by Freeman.
Relation to Bayesian inference
Another expression of the generic views principal is that the inference of distal structure should be such that the inference would remain substantially the same if the "position" of the observer were moderately altered (perturbed). If the inference made would have been qualitatively or categorically different under a perturbation of the observer, then the inference does not satisfy the generic views assumption, and should be rejected. (The question of what constitutes a qualitative or categorical difference is an interesting point of detail.) On this view, it can be argued that the principal of generic views is nothing more than an inference based on the maximum posterior probability (MAP) which accounts for aspects of observation. Thus, we infer the distal phenomenon which possess the highest probability of having generated the observations in question, and this probability incorporates (in addition to relevant priors) both the likelihood of the distal phenomenon generating certain observable signals, and the likelihood of the observer transducing those signals in a manner consistent with the observations. On such an analysis (and with various assumptions invoked), one can obtain a b |
https://en.wikipedia.org/wiki/Abraham%20Pais%20Prize | The Abraham Pais Prize for History of Physics is an award given each year since 2005 jointly by the American Physical Society and the American Institute of Physics for "outstanding scholarly achievements in the history of physics". The prize is named after Abraham Pais (1918-2000), science historian and particle physicist; as of 2023 it is valued at $10,000.
Recipients
Source:
See also
List of American Physical Society prizes and awards
List of physics awards
External links
Abraham Pais Prize for History of Physics, American Physical Society |
https://en.wikipedia.org/wiki/Andrei%20Sakharov%20Prize%20%28APS%29 | The Andrei Sakharov Prize is a prize that is to be awarded every second year by the American Physical Society since 2006. The recipients are chosen for "outstanding leadership and/or achievements of scientists in upholding human rights". The prize is named after Andrei Sakharov (1921-1989), Soviet nuclear physicist, dissident and human rights activist; since 2007 it has been valued at $10,000. The first Sakharov Prize was awarded to physicist and former Soviet gulag prisoner Yuri Orlov.
Recipients
Source:
2006 Yuri Orlov (Cornell University)
2008 Liangying Xu (Chinese Academy of Sciences)
2010 Herman Winick (Stanford Linear Accelerator Center), Joseph Birman (City University of New York), and Morris (Moishe) Pripstein (National Science Foundation)
2012 Mulugeta Bekele (University of Addis Ababa) and Richard Wilson (Harvard University)
2014 Boris Altshuler (P.N. Lebedev Physical Institute) and Omid Kokabee (University of Texas at Austin)
2016 Zafra M. Lerman (Malta Conferences Foundation)
2018 Narges Mohammadi (Iran Engineering Inspection Corporation) and Ravi Kuchimanchi (Association for India's Development)
2020 Ayşe Erzan (Istanbul Technical University) and Xiaoxing Xi (Temple University)
2022 John C. Polanyi (University of Toronto)
See also
List of American Physical Society prizes and awards
List of physics awards |
https://en.wikipedia.org/wiki/Monas%20Hieroglyphica | Monas Hieroglyphica (or The Hieroglyphic Monad) is a book by John Dee, the Elizabethan magus and court astrologer of Elizabeth I of England, published in Antwerp in 1564. It is an exposition of the meaning of an esoteric symbol that he invented.
Dee's Monas Hieroglyphica presents a complex emblem constructed from various astrological symbols, with elements of Latin wordplay, capitalization, spacing, and diacritics, rendering its interpretation challenging. The symbol is intended to embody a profound concept, representing the unity of all creation influenced by celestial forces. Dee believed that this symbol contained the essence of alchemical transformation and spiritual evolution, and by meditating upon it, he aimed to access hidden knowledge transcending linguistic barriers. In merging astrology, alchemy, mysticism, and metaphysics, the Hieroglyphic Monad serves as a visual manifestation of Dee's interconnected worldview.
Content
Understanding the text is difficult because of Dee's Latin wordplay, unexplained capitalization, odd spacing and diacritics.
Meaning of the symbol
John Dee intended the Monad to incorporate a wide range of mystical and esoteric concepts. This complex symbol was meant to symbolize the unity of all creation, influenced by astrological and planetary forces. Dee believed it held the essence of alchemical transformation and spiritual growth. By meditating on the Monad, he thought to gain insights into hidden knowledge about the universe, transcending language barriers and tapping into profound truths. Overall, Dee's intention was to encapsulate his interconnected worldview, combining elements of astrology, alchemy, mysticism, and metaphysics.
Reception and influence
The book received little notice in English sources, though it is praised in the 1591 edition of George Ripley's The Compound of Alchymy as well as in Elias Ashmole’s Theatrum Chemicum Britannicum (1652). A number of references appear in other languages, for example, Jean-Ja |
https://en.wikipedia.org/wiki/Legacy%20Plug%20and%20Play | The term Legacy Plug and Play, also shortened to Legacy PnP, describes a series of specifications and Microsoft Windows features geared towards operating system configuration of devices, and some device IDs are assigned by UEFI Forum. The standards were primarily aimed at the IBM PC standard bus, later dubbed Industry Standard Architecture (ISA). Related specifications are also defined for the common external or specialist buses commonly attached via ISA at the time of development, including RS-232 and parallel port devices.
As a Windows feature, Plug and Play refers to operating system functionality that supports connectivity, configuration and management with native plug and play devices. Originally considered part of the same feature set as the specifications, Plug and Play in this context refers primarily to the responsibilities and interfaces associated with Windows driver development.
Plug and Play allows for detection of devices without user intervention, and occasionally for minor configuration of device resources, such as I/O ports and device memory maps. PnP is a specific set of standards, not be confused with the generic term plug and play, which describes any hardware specification that alleviates the need for user configuration of device resources.
ACPI is the successor to Legacy Plug and Play.
Overview
The Plug and Play standard requires configuration of devices to be handled by the PnP BIOS, which then provides details of resources allocations to the operating system. The process is invoked at boot time. When the computer is first turned on, compatible devices are identified and assigned non-conflicting IO addresses, interrupt request numbers and DMA channels.
The term was adopted by Microsoft in reference to their Windows 95 product. Other operating systems, such as AmigaOS Autoconfig and the Mac OS NuBus system, had already supported such features for some time (under various names, or no name). Even Yggdrasil Linux advertised itself as "Plug |
https://en.wikipedia.org/wiki/Aneesur%20Rahman%20Prize | The Aneesur Rahman Prize for Computational Physics is a prize that has been awarded annually by the American Physical Society since 1993. The recipient is chosen for "outstanding achievement in computational physics research" and it is the highest award given by the APS for work in computational physics. The prize is named after Aneesur Rahman (1927–1987), pioneer of the molecular dynamics simulation method. The prize was valued at $5,000 from 2007 to 2014, and is currently valued at $10,000.
Recipients
Source: American Physical Society
2023 Pablo G. Debenedetti
2022 Giulia Galli
2021 Anders W. Sandvik
2020 Antoine Georges and Gabriel Kotliar
2019 Sharon C. Glotzer
2018
2017 Sauro Succi
2016
2015 John D. Joannopoulos
2014 Robert Swendsen
2013 James R. Chelikowsky
2012 Kai-Ming Ho
2011 James M. Stone
2010 Frans Pretorius
2009
2008 Gary S. Grest
2007 Daan Frenkel
2006 David Vanderbilt
2005 Uzi Landman
2004 Farid Abraham
2003 Steven R. White
2002 David P. Landau
2001 Alex Zunger
2000 Michael John Creutz
1999 Michael L. Klein
1998 David Matthew Ceperley
1997 Donald H. Weingarten
1996 Steven Gwon Sheng Louie
1995 Roberto Car and Michele Parrinello
1994 John M. Dawson
1993 Kenneth G. Wilson
See also
List of American Physical Society prizes and awards
List of physics awards |
https://en.wikipedia.org/wiki/Arthur%20L.%20Schawlow%20Prize%20in%20Laser%20Science | The Arthur L. Schawlow Prize in Laser Science is a prize that has been awarded annually by the American Physical Society since 1991. The recipient is chosen for "outstanding contributions to basic research which uses lasers to advance our knowledge of the fundamental physical properties of materials and their interaction with light". The prize is named after Arthur L. Schawlow (1921–1999), laser pioneer and a Nobel laureate in physics, and as of 2007 is valued at $10,000.
Recipients
Source:
2023 Demetrios N. Christodoulides
2022 Tony F. Heinz
2021 Peter J. Delfyett, Jr.
2020 Shaul Mukamel
2019 Steven T. Cundiff
2018 Gérard Albert Mourou
2017 Louis F. DiMauro
2016 Robert W. Boyd
2015 Christopher Monroe
2014 Mordechai Segev
2013 Robert Alfano
2012 Michael D. Fayer
2011
2010 , Margaret M. Murnane
2009 Robert W. Field
2008
2007 Szymon Suckewer
2006 Paul B. Corkum
2005 Marlan O. Scully
2004 Federico Capasso
2003 David E. Pritchard
2002 Stephen E. Harris
2001 David J. Wineland
2000 Richard Neil Zare
1999 Carl E. Wieman
1998 William D. Phillips
1997 Charles V. Shank, Erich Peter Ippen
1996 Theodor W. Hänsch
1995 Richart E. Slusher
1994 Steven Chu
1993 John L. Hall
1992 Yuen-Ron Shen
1991 Peter P. Sorokin
See also
List of physics awards
List of prizes named after people |
https://en.wikipedia.org/wiki/Differentiation%20rules | This is a summary of differentiation rules, that is, rules for computing the derivative of a function in calculus.
Elementary rules of differentiation
Unless otherwise stated, all functions are functions of real numbers (R) that return real values; although more generally, the formulae below apply wherever they are well defined — including the case of complex numbers (C).
Constant term rule
For any value of , where , if is the constant function given by , then .
Proof
Let and . By the definition of the derivative,
This shows that the derivative of any constant function is 0.
Intuitive (geometric) explanation
The derivative of the function at a point is the slope of the line tangent to the curve at the point. Slope of the constant function is zero, because the tangent line to the constant function is horizontal and it's angle is zero.
In other words, the value of the constant function, y, will not change as the value of x increases or decreases.
Differentiation is linear
For any functions and and any real numbers and , the derivative of the function with respect to is:
In Leibniz's notation this is written as:
Special cases include:
The constant factor rule
The sum rule
The difference rule
The product rule
For the functions f and g, the derivative of the function h(x) = f(x) g(x) with respect to x is
In Leibniz's notation this is written
The chain rule
The derivative of the function is
In Leibniz's notation, this is written as:
often abridged to
Focusing on the notion of maps, and the differential being a map , this is written in a more concise way as:
The inverse function rule
If the function has an inverse function , meaning that and then
In Leibniz notation, this is written as
Power laws, polynomials, quotients, and reciprocals
The polynomial or elementary power rule
If , for any real number then
When this becomes the special case that if then
Combining the power rule with the sum and constant multiple rules permit |
https://en.wikipedia.org/wiki/UEFI%20Platform%20Initialization | The Platform Initialization Specification (PI Specification) is a specification published by the Unified EFI Forum that describes the internal interfaces between different parts of computer platform firmware. This allows for more interoperability between firmware components from different sources. This specification is normally, but not by requirement, used in conjunction with the UEFI specification.
Current version
Platform Initialization Specification 1.7, Released January 2019.
Contents
As of version 1.3, the PI specification contains five volumes:
Volume 1: Pre-EFI Initialization Core Interface
Volume 2: Driver Execution Environment Core Interface
Volume 3: Shared Architectural Elements
Volume 4: System Management Mode Core Interface
Volume 5: Standards |
https://en.wikipedia.org/wiki/Davisson%E2%80%93Germer%20Prize | The Davisson–Germer Prize in Atomic or Surface Physics is an annual prize that has been awarded by the American Physical Society since 1965. The recipient is chosen for "outstanding work in atomic physics or surface physics". The prize is named after Clinton Davisson and Lester Germer, who first measured electron diffraction, and as of 2007 it is valued at $5,000.
Recipients
2023: Feng Liu
2022: David S. Weiss
2021: Michael F. Crommie
2020: Klaas Bergmann
2019: Randall M. Feenstra
2018:
2017: and Stephen Kevan
2016: Randall G. Hulet
2015: and
2014: Nora Berrah
2013: Geraldine L. Richmond
2012: Jean Dalibard
2011: Joachim Stohr
2010: Chris H. Greene
2009: and Krishnan Raghavachari
2008:
2007:
2006:
2005: Ernst G. Bauer
2004:
2003: Rudolf M. Tromp
2002: Gerald Gabrielse
2001: Donald M. Eigler
2000: William Happer
1999: Steven Gwon Sheng Louie
1998: Sheldon Datz
1997: Jerry D. Tersoff
1996:
1995: Max G. Lagally
1994: Carl Weiman [sic]
1993:
1992:
1991:
1990: David Wineland
1989:
1988: John L. Hall
1987:
1986: Daniel Kleppner
1985:
1984: and
1983: E. W. Plummer
1982: Llewellyn H. Thomas
1981: Robert Gomer
1980: Alexander Dalgarno
1979: and
1978: Vernon Hughes
1977: Walter Kohn and
1976: Ugo Fano
1975: and Homer D. Hagstrum
1974: Norman Ramsey
1972: Erwin Wilhelm Müller
1970: Hans Dehmelt
1967: Horace Richard Crane
1965:
Source:
See also
List of physics awards |
https://en.wikipedia.org/wiki/Earle%20K.%20Plyler%20Prize | The Earle K. Plyler Prize for Molecular Spectroscopy and Dynamics is a prize that has been awarded annually by the American Physical Society since 1977. The recipient is chosen for "notable contributions to the field of molecular spectroscopy and dynamics". The prize is named after Earle K. Plyler, who was a leading experimenter in the field of infrared spectroscopy; as of 2007 it is valued at $10,000. The prize is currently sponsored by the AIP Journal of Chemical Physics.
Recipients
Source: American Physical Society
See also
List of physics awards
List of chemistry awards |
https://en.wikipedia.org/wiki/Anomalous%20cancellation | An anomalous cancellation or accidental cancellation is a particular kind of arithmetic procedural error that gives a numerically correct answer. An attempt is made to reduce a fraction by cancelling individual digits in the numerator and denominator. This is not a legitimate operation, and does not in general give a correct answer, but in some rare cases the result is numerically the same as if a correct procedure had been applied. The trivial cases of cancelling trailing zeros or where all of the digits are equal are ignored.
Examples of anomalous cancellations which still produce the correct result include (these and their inverses are all the cases in base 10 with the fraction different from 1 and with two digits):
The article by Boas analyzes two-digit cases in bases other than base 10, e.g., = and its inverse are the only solutions in base 4 with two digits.
An example of anomalous cancellation with more than two digits is = , and an example with different numbers of digits is =.
Elementary properties
When the base is prime, no two-digit solutions exist. This can be proven by contradiction: suppose a solution exists. Without loss of generality, we can say that this solution is
where the double vertical line indicates digit concatenation. Thus, we have
But , as they are digits in base ; yet divides , which means that . Therefore. the right hand side is zero, which means the left hand side must also be zero, i.e., , a contradiction by the definition of the problem. (If , the calculation becomes , which is one of the excluded trivial cases.)
Another property is that the numbers of solutions in a base is odd if and only if is an even square. This can be proven similarly to the above: suppose that we have a solution
Then, doing the same manipulation, we get
Suppose that . Then note that is also a solution to the equation. This almost sets up an involution from the set of solutions to itself. But we can also substitute in to get , which onl |
https://en.wikipedia.org/wiki/25-Hydroxyvitamin%20D%201-alpha-hydroxylase | 25-Hydroxyvitamin D 1-alpha-hydroxylase (VD 1A hydroxylase) also known as calcidiol 1-monooxygenase or cytochrome p450 27B1 (CYP27B1) or simply 1-alpha-hydroxylase is a cytochrome P450 enzyme that in humans is encoded by the CYP27B1 gene.
VD 1A hydroxylase is located in the proximal tubule of the kidney and a variety of other tissues, including skin (keratinocytes), immune cells, and bone (osteoblasts).
Reactions
The enzyme catalyzes the hydroxylation of calcifediol to calcitriol (the bioactive form of Vitamin D):
calcidiol + 2 reduced adrenodoxin + 2 H+ + O2 calcitriol + 2 oxidized adrenodoxin + H2O
The enzyme is also able to oxidize ercalcidiol (25-OH D2) to ercalcitriol, secalciferol to calcitetrol, and 25-hydroxy-24-oxocalciol to (1S)-1,25-dihydroxy-24-oxocalciol.
Clinical significance
Loss-of-function mutations in CYP27B1 cause Vitamin D-dependent rickets, type IA.
Interactive pathway map |
https://en.wikipedia.org/wiki/De%20Bruijn%20index | In mathematical logic, the De Bruijn index is a tool invented by the Dutch mathematician Nicolaas Govert de Bruijn for representing terms of lambda calculus without naming the bound variables. Terms written using these indices are invariant with respect to α-conversion, so the check for α-equivalence is the same as that for syntactic equality. Each De Bruijn index is a natural number that represents an occurrence of a variable in a λ-term, and denotes the number of binders that are in scope between that occurrence and its corresponding binder. The following are some examples:
The term λx. λy. x, sometimes called the K combinator, is written as λ λ 2 with De Bruijn indices. The binder for the occurrence x is the second λ in scope.
The term λx. λy. λz. x z (y z) (the S combinator), with De Bruijn indices, is λ λ λ 3 1 (2 1).
The term λz. (λy. y (λx. x)) (λx. z x) is λ (λ 1 (λ 1)) (λ 2 1). See the following illustration, where the binders are coloured and the references are shown with arrows.
De Bruijn indices are commonly used in higher-order reasoning systems such as automated theorem provers and logic programming systems.
Formal definition
Formally, λ-terms (M, N, ...) written using De Bruijn indices have the following syntax (parentheses allowed freely):
M, N, ... ::= n | M N | λ M
where n—natural numbers greater than 0—are the variables. A variable n is bound if it is in the scope of at least n binders (λ); otherwise it is free. The binding site for a variable n is the nth binder it is in the scope of, starting from the innermost binder.
The most primitive operation on λ-terms is substitution: replacing free variables in a term with other terms. In the β-reduction (λ M) N, for example, we must
find the instances of the variables n1, n2, ..., nk in M that are bound by the λ in λ M,
decrement the free variables of M to match the removal of the outer λ-binder, and
replace n1, n2, ..., nk with N, suitably incrementing the free variables occurring in N each |
https://en.wikipedia.org/wiki/Daseian%20notation | Daseian notation (or dasian notation) is the type of musical notation used in the ninth century anonymous musical treatises Musica enchiriadis and Scolica enchiriadis. The music of the Musica enchiriadis and Scolica enchiriadis, written in Daseian notation, are the earliest known examples of written polyphonic music in history.
Usage
Musicologist Willi Apel has called the notation "a mediaeval imitation of the ancient Greek notation". The treatises themselves refer to it as "dasia"; the word derives from the Greek daseia, which refers to "rough breathing" at the start of a word in spoken prosody.
Daseian notation makes use of a staff of varying numbers of lines, from four to as many as eighteen, as well as a system of four shapes which are rotated in various ways to represent the full gamut of eighteen pitches used in the treatises. These eighteen pitches are based on a system of four repeating tetrachords, resulting in the following scale: G A B♭ c | d e f g | a b c' d' | e' f♯' g' a' | b' c♯''. This scale does not correspond to any known performance practice. When it is used to construct polyphonic music, as directed in the treatises, it results in a number of written tritones, which were considered undesirable by theorists in performance and were probably mistakes of the author or else implied the system was transposable to C D E♭ F | G A B♭ c | d e f g | a b c' d' | e' f♯' or d e f g | a b c' d' | e' f♯' g' a' | b' c♯'' d” e” | f♯'' g♯'' when necessary.
The notational signs were then placed at the far left of the staff (similar in placement to modern clef), and some illustrations are supplemented with "T" and "S" in between the signs so as to clarify the placement of semitones. Syllables of the spoken words were then written on the staff lines (see example above). If the pitch changed, the word syllables would be raised or lowered to a different staff line. This was used to notate organum in two, three and four-voice styles.
In addition to the Enchiriadis t |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.