source stringlengths 31 227 | text stringlengths 9 2k |
|---|---|
https://en.wikipedia.org/wiki/Faunal%20assemblage | In archaeology and paleontology a faunal assemblage is a group of associated animal fossils found together in a given stratum.
The principle of faunal succession is used in biostratigraphy to determine each biostratigraphic unit, or biozone. The biostratigraphic unit being a section of geological strata that is defined on the basis of its characteristic fossil taxa or faunal assemblage.
For example, in East Africa, a distinctive group of animal species, mostly pigs, is characteristic of the fossils preserved from a particular period of time. This faunal assemblage has been used effectively to chronologically correlate the East African early hominid sites.
Faunal assemblages are useful in determining the foraging patterns of hominids. One such assemblage at Lang Rongrien in Thailand indicated a hunter-gatherer group which was highly flexible when it came to finding food. They relied heavily on turtle and tortoise to supply the meat portion of their diet when hunting large game was unpredictable. This assemblage also suggested the paleoenvironment was drier and cooler than today because of a distinct lack of pig bones. |
https://en.wikipedia.org/wiki/Cryptographic%20accelerator | In computing, a cryptographic accelerator is a co-processor designed specifically to perform computationally intensive cryptographic operations, doing so far more efficiently than the general-purpose CPU. Because many servers' system loads consist mostly of cryptographic operations, this can greatly increase performance.
Intel's AES-NI is by far the most common cryptographic accelerator in commodity hardware. VIA PadLock is another recent example.
Operating system support
Several operating systems provide some support for cryptographic hardware. The BSD family of systems has the OpenBSD Cryptographic Framework (OCF), Linux systems have the Crypto API, Solaris OS has the Solaris Cryptographic Framework (SCF) and Microsoft Windows has the Microsoft CryptoAPI.
Some cryptographic accelerators offer new machine instructions and can therefore be used directly by programs. Libraries such as OpenSSL and LibreSSL support some such cryptographic accelerators.
Almost all Unix-like operating systems use OpenSSL or the fork LibreSSL as their cryptography library. These libraries use cryptographic accelerators such as AES-NI if available.
See also
TLS acceleration
Hardware-based Encryption
Hardware acceleration
Computer optimization
Coprocessors
Cryptographic hardware |
https://en.wikipedia.org/wiki/JEDEC%20memory%20standards | The JEDEC memory standards are the specifications for semiconductor memory circuits and similar storage devices promulgated by the Joint Electron Device Engineering Council (JEDEC) Solid State Technology Association, a semiconductor trade and engineering standardization organization.
JEDEC Standard 100B.01 specifies common terms, units, and other definitions in use in the semiconductor industry. JESC21-C specifies semiconductor memories from the 256 bit static RAM to DDR4 SDRAM modules.
JEDEC standardization goals
The Joint Electron Device Engineering Council characterizes its standardization efforts as follows:
JEDEC Standard 100B.01
The December 2002 JEDEC Standard 100B.01 is entitled Terms, Definitions, and Letter Symbols for Microcomputers, Microprocessors, and Memory Integrated Circuits. The purpose of the standard is to promote the uniform use of symbols, abbreviations, terms, and definitions throughout the semiconductor industry.
Units of information
The specification defines the two common units of information:
The bit (b) is the smallest unit of information in the binary numeration system and is represented by the digits 0 and 1.
The byte (B) is a binary character string typically operated upon as one unit. It is usually shorter than a computer word.
Unit prefixes for semiconductor storage capacity
The specification contains citations of the commonly used prefixes kilo, mega, and giga "as a prefix to units of semiconductor storage capacity" to designate multiples of the units.
The specification cites three prefixes as follows, with the note that these prefixes are included in the document only to reflect common usage.
kilo (K): A multiplier equal to 1024 (210).
mega (M): A multiplier equal to (220 or K2, where K = 1024).
giga (G): A multiplier equal to (230 or K3, where K = 1024).
It refers to the IEEE/ASTM SI 10-1997 standard as stating that "this practice frequently leads to confusion and is deprecated". The document further refers to the d |
https://en.wikipedia.org/wiki/Edward%20Bromhead | Sir Edward Thomas ffrench Bromhead, 2nd Baronet FRS FRSE (26 March 1789 – 14 March 1855) was a British landowner and mathematician, best remembered as patron of the mathematician and physicist George Green and mentor of George Boole.
Life
Born the son of Gonville Bromhead, 1st Baronet Bromhead (grandfather of the British second in command of the same name at Rorke's Drift) and Lady Jane ffrench, Baroness ffrench, in Dublin. Bromhead was educated at the University of Glasgow and later at Caius College, Cambridge ( B.A. 1812, M.A. 1815) before taking up the study of law at the Inner Temple in London. He was elected a Fellow of the Royal Society in 1817. Returning to Lincolnshire, he became High Steward of Lincoln. He became the 2nd Bromhead baronet, of Thurlby Hall in 1822.
While at Cambridge, Bromhead was a founder of the Analytical Society, a precursor of the Cambridge Philosophical Society, together with John Herschel, George Peacock and Charles Babbage, with whom he maintained a close and lifelong friendship. While he was, by all accounts, a gifted mathematician in his own right (although ill-health prevented him from pursuing his studies further), his greatest contribution to the subject is at second hand: having subscribed to the first publication of self-taught mathematician and physicist George Green, he encouraged Green to continue his research and to write further papers (which Bromhead sent on to be published in the Transactions of the Cambridge Philosophical Society and those of the Royal Society of Edinburgh).
Bromhead repeated his success by encouraging the young George Boole from Lincoln. Bromhead was President of the Lincoln Mechanics Institute in the Lincoln Greyfriars, where George Boole's father was the curator. Boole first came to public notice when he gave a lecture on the work of Sir Isaac Newton on 5 February 1835. The young Boole's development was fed by books that Bromhead supplied.
Bromhead lost his sight when he was old and he died unm |
https://en.wikipedia.org/wiki/Degranulation | Degranulation is a cellular process that releases antimicrobial cytotoxic or other molecules from secretory vesicles called granules found inside some cells. It is used by several different cells involved in the immune system, including granulocytes (neutrophils, basophils, and eosinophils) and mast cells. It is also used by certain lymphocytes such as natural killer (NK) cells and cytotoxic T cells, whose main purpose is to destroy invading microorganisms.
Mast cells
Degranulation in mast cells is part of an inflammatory response, and substances such as histamine are released. Granules from mast cells mediate processes such as "vasodilation, vascular homeostasis, innate and adaptive immune responses, angiogenesis, and venom detoxification."
Antigens interact with IgE molecules already bound to high affinity Fc receptors on the surface of mast cells to induce degranulation, via the activation of tyrosine kinases within the cell. The mast cell releases a mixture of compounds, including histamine, proteoglycans, serotonin, and serine proteases from its cytoplasmic granules.
Eosinophils
In a similar mechanism, activated eosinophils release preformed mediators such as major basic protein, and enzymes such as peroxidase, following interaction between their Fc receptors and IgE molecules that are bound to large parasites like helminths.
Neutrophils
Degranulation in neutrophils can occur in response to infection, and the resulting granules are released in order to protect against tissue damage. Excessive degranulation of neutrophils, sometimes triggered by bacteria, is associated with certain inflammatory disorders, such as asthma and septic shock.
Four kinds of granules exist in neutrophils that display differences in content and regulation. Secretory vesicles are the most likely to release their contents by degranulation, followed by gelatinase granules, specific granules, and azurophil granules.
Cytotoxic T cells and NK cells
Cytotoxic T cells and NK cells rele |
https://en.wikipedia.org/wiki/Perceptive%20Pixel | Perceptive Pixel is a division of Microsoft specializing in research, development and production of multi-touch interfaces. Its technology is used in fields including broadcast, defense, geo-intelligence, energy exploration, industrial design and medical imaging.
Background
The division originated as an independent company with the name Perceptive Pixel, Inc., headquartered in New York City and maintaining offices in Mountain View, Portland, and Washington, D.C. Prior to the company's establishment, founder Jeff Han publicly demonstrated multitouch hardware and software technology at a TED conference in February 2006. The company was founded later in the same year. Subsequently, it shipped its first Multi-Touch Workstation and larger Multi-Touch Collaboration Wall in 2007. The latter gained widespread recognition for transforming the way CNN covered the 2008 US Presidential elections. In 2009, the Smithsonian awarded Perceptive Pixel the National Design Award in the inaugural category of Interaction Design. Throughout its history as a startup company, Perceptive Pixel had focused on ultra-high-end touch displays that were sold to large businesses in key niches.
Microsoft acquisition
On July 9, 2012, Microsoft CEO Steve Ballmer announced that they would acquire Perceptive Pixel Inc. in a transaction that closed three weeks later on July 31. At that year's WPC conference, Ballmer stated the company hoped to use its newly acquired technology to enhance its Windows 8 operating system. Han also stated that the sale could allow their products to be more broadly available.
On January 21, 2015, Microsoft announced the Surface Hub during its Windows 10 launch. The new product was described as a next-generation device under the Surface brand and was developed by members of the Perceptive Pixel team. |
https://en.wikipedia.org/wiki/CoffeeCup%20Software | CoffeeCup Software is an American computer software development company based in Atlanta, Georgia, United States, founded in Corpus Christi, Texas, in 1996. The name comes from the company's origins in an internet cafe owned by its founder.
The company develops software applications for creating, designing, and editing responsive websites and a number of online services for webmasters. The company's third product, CoffeeCup Direct FTP, was the first FTP program to incorporate text editing functionality directly into the interface in a "split-screen" fashion.
In the spring of 2007, CoffeeCup moved to its new headquarters where it employs just over 10 programmers and designers. In addition to a panel of user-advisers, CoffeeCup has a group of around 8000 “Ambassadors” who are invited to test drive new and existing software programs and report bugs and offer suggestions for improvements.
CoffeeCup's Software has won the Shareware Industry Award six years running from 1999 to 2004 for the CoffeeCup HTML Editor. Other awards include being ranked #400 in the Interactive 500, 11 CNet Editors Choice Awards, 18 Tucows 5-Star Awards, and ZDNet Best Pick for Web Design Software.
History
CoffeeCup Software was started in a coffee house called “The Raven & The Sparrow” which was owned by the company's founder, Nicholas Longo. This was the only coffee house that offered free internet access in Corpus Christi, Texas, at the time.
Since the coffee house already had the www.coffeecup.com domain name, the fledgling software company was named CoffeeCup Software and the first program was named CoffeeCup HTML Editor. Longo posted the program on his website. Eventually, CoffeeCup began charging $20 USD for the program. In 1996 Longo decided to put away the espresso machines and devote full-time attention to developing software.
CoffeeCup offers a core group of programs free to schools. CoffeeCup Software's K–12 Donation Program allows public schools to request the Educational Soft |
https://en.wikipedia.org/wiki/Monodromy%20theorem | In complex analysis, the monodromy theorem is an important result about analytic continuation of a complex-analytic function to a larger set. The idea is that one can extend a complex-analytic function (from here on called simply analytic function) along curves starting in the original domain of the function and ending in the larger set. A potential problem of this analytic continuation along a curve strategy is there are usually many curves which end up at the same point in the larger set. The monodromy theorem gives sufficient conditions for analytic continuation to give the same value at a given point regardless of the curve used to get there, so that the resulting extended analytic function is well-defined and single-valued.
Before stating this theorem it is necessary to define analytic continuation along a curve and study its properties.
Analytic continuation along a curve
The definition of analytic continuation along a curve is a bit technical, but the basic idea is that one starts with an analytic function defined around a point, and one extends that function along a curve via analytic functions defined on small overlapping disks covering that curve.
Formally, consider a curve (a continuous function) Let be an analytic function defined on an open disk centered at An analytic continuation of the pair along is a collection of pairs for such that
and
For each is an open disk centered at and is an analytic function.
For each there exists such that for all with one has that (which implies that and have a non-empty intersection) and the functions and coincide on the intersection
Properties of analytic continuation along a curve
Analytic continuation along a curve is essentially unique, in the sense that given two analytic continuations and of along the functions and coincide on Informally, this says that any two analytic continuations of along will end up with the same values in a neighborhood of
If the curve is clo |
https://en.wikipedia.org/wiki/Graph%20cuts%20in%20computer%20vision | As applied in the field of computer vision, graph cut optimization can be employed to efficiently solve a wide variety of low-level computer vision problems (early vision), such as image smoothing, the stereo correspondence problem, image segmentation, object co-segmentation, and many other computer vision problems that can be formulated in terms of energy minimization. Many of these energy minimization problems can be approximated by solving a maximum flow problem in a graph (and thus, by the max-flow min-cut theorem, define a minimal cut of the graph). Under most formulations of such problems in computer vision, the minimum energy solution corresponds to the maximum a posteriori estimate of a solution. Although many computer vision algorithms involve cutting a graph (e.g., normalized cuts), the term "graph cuts" is applied specifically to those models which employ a max-flow/min-cut optimization (other graph cutting algorithms may be considered as graph partitioning algorithms).
"Binary" problems (such as denoising a binary image) can be solved exactly using this approach; problems where pixels can be labeled with more than two different labels (such as stereo correspondence, or denoising of a grayscale image) cannot be solved exactly, but solutions produced are usually near the global optimum.
History
The theory of graph cuts used as an optimization method was first applied in computer vision in the seminal paper by Greig, Porteous and Seheult of Durham University. Allan Seheult and Bruce Porteous were members of Durham's lauded statistics group of the time, led by Julian Besag and Peter Green, with the optimisation expert Margaret Greig notable as the first ever female member of staff of the Durham Mathematical Sciences Department.
In the Bayesian statistical context of smoothing noisy (or corrupted) images, they showed how the maximum a posteriori estimate of a binary image can be obtained exactly by maximizing the flow through an associated image network, |
https://en.wikipedia.org/wiki/Space%20Science%20Institute | The Space Science Institute (SSI) in Boulder, Colorado, is a nonprofit, public-benefit corporation formed in 1992. Its purpose is to create and maintain an environment where scientific research and education programs can flourish in an integrated fashion. SSI is among the four non-profit institutes in the US cited in a 2007 report by Nature, including Southwest Research Institute, Planetary Science Institute, and Eureka Scientific, which manage federal grants for non-tenure-track astronomers.
Description
SSI's research program encompasses the following areas: space physics, earth science, planetary science, and astrophysics. The flight operations branch manages the Cassini-Huygens spacecraft's visible camera instrument and provides spectacular images of Saturn and its moons and rings to the public. SSI participates in mission operations and is home to the Cassini Imaging Central Laboratory for OPerations (CICLOPS).
The primary goal of SSI is to bring together researchers and educators to improve science education. Toward this end, the institute acts as an umbrella for researchers who wish to be independent of universities. In addition, it works with educators directly to improve teaching methods for astronomy. SSI has also produced several traveling exhibits for science museums, including Electric Space, Mars Quest, and Alien Earths. It is currently producing Giant Worlds.
SSI provides management support for research scientists and principal investigators, which help them to submit proposals to major public funding agencies such as National Aeronautics and Space Administration (NASA), National Science Foundation (NSF), Space Telescope Science Institute (STSci), Department of Energy (DOE), and Jet Propulsion Laboratory (JPL) Principal investigators are supported by SSI though proposal budget preparation, proposal submission, and project reporting tools, and have competitive negotiated overhead rates.
The institute is loosely affiliated with the University |
https://en.wikipedia.org/wiki/Skewb%20Ultimate | The Skewb Ultimate, originally marketed as the Pyraminx Ball, is a twelve-sided puzzle derivation of the Skewb, produced by German toy-maker Uwe Mèffert. Most versions of this puzzle are sold with six different colors of stickers attached, with opposite sides of the puzzle having the same color; however, some early versions of the puzzle have a full set of 12 colors.
Description
The Skewb Ultimate is made in the shape of a dodecahedron, like the Megaminx, but cut differently. Each face is cut into four parts, two equal and two unequal. Each cut is a deep cut: it bisects the puzzle. This results in eight smaller corner pieces and six larger "edge" pieces.
The object of the puzzle is to scramble the colors, and then restore them to the original configuration.
Solutions
At first glance, the Skewb Ultimate appears to be much more difficult to solve than the other Skewb puzzles, because of its uneven cuts which cause the pieces to move in a way that may seem irregular or strange.
Mathematically speaking, however, the Skewb Ultimate has exactly the same structure as the Skewb Diamond. The solution for the Skewb Diamond can be used to solve this puzzle, by identifying the Diamond's face pieces with the Ultimate's corner pieces, and the Diamond's corner pieces with the Ultimate's edge pieces. The only additional trick here is that the Ultimate's corner pieces (equivalent to the Diamond's face pieces) are sensitive to orientation, and so may require an additional algorithm for orienting them after being correctly placed.
Similarly, the Skewb Ultimate is mathematically identical to the Skewb, by identifying corners with corners, and the Skewb's face centers with the Ultimate's edges. The solution of the Skewb can be used directly to solve the Skewb Ultimate. The only addition is that the edge pieces of the Skewb Ultimate are sensitive to orientation, and may require an additional algorithm to orient them after being placed correctly.
Number of combinations
The Skewb Ult |
https://en.wikipedia.org/wiki/Sampling%20risk | Sampling risk is one of the many types of risks an auditor may face when performing the necessary procedure of audit sampling. Audit sampling exists because of the impractical and costly effects of examining all or 100% of a client's records or books. As a result, a "sample" of a client's accounts are examined.
Due to the negative effects produced by sampling risk, an auditor may have to perform additional procedures which in turn can impact the overall efficiency of the audit.
Sampling risk represents the possibility that an auditor's conclusion based on a sample is different from that reached if the entire population were subject to audit procedure. The auditor may conclude that material misstatements exist, when in fact they do not; or material misstatements do not exist but in fact they do exist. Auditors can lower the sampling risk by increasing the sampling size.
Although there are many types of risks associated with the audit process, each type primarily has an effect on the overall audit engagement. The effects produced by sampling risk generally can increase audit risk, the risk that an entity's financial statements will contain a material misstatement, though given an unqualified ('clean') audit report. Sampling risk can also increase detection risk which suggests the possibility that an auditor will not find material misstatements relating to the financial statements through substantive tests and analysis.
Types of sampling risk
Typical scenarios
Auditors must often make professional judgments in assessing sampling risk. When testing samples the auditor is primarily concerned with two aspects of sampling risk:
Risk of accepting incorrect data: the sample supports the conclusion that the recorded account balance is not materially misstated when it is materially misstated.
Risk of incorrect rejection: the risk that the sample supports the conclusion that the recorded amount balance is materially misstated when it is not materially misstated.
In add |
https://en.wikipedia.org/wiki/Atkinson%27s%20theorem | In operator theory, Atkinson's theorem (named for Frederick Valentine Atkinson) gives a characterization of Fredholm operators.
The theorem
Let H be a Hilbert space and L(H) the set of bounded operators on H. The following is the classical definition of a Fredholm operator: an operator T ∈ L(H) is said to be a Fredholm operator if the kernel Ker(T) is finite-dimensional, Ker(T*) is finite-dimensional (where T* denotes the adjoint of T), and the range Ran(T) is closed.
Atkinson's theorem states:
A T ∈ L(H) is a Fredholm operator if and only if T is invertible modulo compact perturbation, i.e. TS = I + C1 and ST = I + C2 for some bounded operator S and compact operators C1 and C2.
In other words, an operator T ∈ L(H) is Fredholm, in the classical sense, if and only if its projection in the Calkin algebra is invertible.
Sketch of proof
The outline of a proof is as follows. For the ⇒ implication, express H as the orthogonal direct sum
The restriction T : Ker(T)⊥ → Ran(T) is a bijection, and therefore invertible by the open mapping theorem. Extend this inverse by 0 on Ran(T)⊥ = Ker(T*) to an operator S defined on all of H. Then I − TS is the finite-rank projection onto Ker(T*), and I − ST is the projection onto Ker(T). This proves the only if part of the theorem.
For the converse, suppose now that ST = I + C2 for some compact operator C2. If x ∈ Ker(T), then STx = x + C2x = 0. So Ker(T) is contained in an eigenspace of C2, which is finite-dimensional (see spectral theory of compact operators). Therefore, Ker(T) is also finite-dimensional. The same argument shows that Ker(T*) is also finite-dimensional.
To prove that Ran(T) is closed, we make use of the approximation property: let F be a finite-rank operator such that ||F − C2|| < r. Then for every x in Ker(F),
||S||⋅||Tx|| ≥ ||STx|| = ||x + C2x|| = ||x + Fx +C2x − Fx|| ≥ ||x|| − ||C2 − F||⋅||x|| ≥ (1 − r)||x||.
Thus T is bounded below on Ker(F), which implies that T(Ker(F)) is closed. On the other hand, T(Ker |
https://en.wikipedia.org/wiki/Match%20Day%20%28series%29 | Match Day is an association football video game franchise created by Jon Ritman in 1984 for the 1980s 8-bit home computer market.
Games in the series
The following games are part of the Match Day series. All of the games were published by Ocean, with the exception of Super Match Soccer:
Match Day: the first game in the series. It was created in 1984 and released on most home computers of the era, but is most well known for its Sinclair Spectrum incarnation. It was designed and developed by Jon Ritman with the help of Chris Clarke.
International Match Day is an improved version of Match Day published in 1985 for ZX Spectrum 128KB. It takes advantage of the extra memory available to provide better sound and some full screen images.
Match Day II was released in 1987 for the Amstrad CPC, ZX Spectrum, MSX and Commodore 64 platforms. It was written by Jon Ritman, with graphics by Bernie Drummond and music and sound by Guy Stevens (Commodore version, programmed by John Darnel). The game was the first one to include complete control over ball direction, power and elevation (using a kickometer), and a brand new deflection system (Diamond Deflection System).
Final Whistle was the name of an arcade machine very similar to Match Day 2 (it was previously called Soccerama during development). Although it was finished it was never released.
Super Match Soccer, was released in 1998 by Acclaim. Although its working title was The Net, it was intended to be released as Match Day III (it was even presented as such in a PlayStation preview in Spain), but due to licensing problems the name was finally changed. It was developed in 1998 by Cranberry Source and published by Acclaim. The game is available for PC and PlayStation. |
https://en.wikipedia.org/wiki/Diecast%20Collector | Diecast Collector is a British magazine dedicated to the hobby of collecting diecast metal vehicles. Published monthly, it is a thick, glossy magazine featuring a variety of articles on toy and model cars, trucks, and buses, and aircraft. Beyond information on the collectables themselves, the magazine is also a source of hobby information such as auction results, collectors' fairs schedules, specialized dealer information, etc.
The magazine also issues a yearly price valuation book.
Overview
The magazine was founded and edited by Mike Forbes, a well-known collector who formerly edited trucking magazines. Regular contributors have included Malcolm Bates, Mike Pigott, Horace Dunkley, and Andrew Ralston. Popular columns include the Matchbox "Models of Yesteryear" series by Horace Dunkley, the amusing 'Drive-Past' column by Brian Gower, and regular features on Hot Wheels, Matchbox Superfast and character toys by Mike Pigott. Every issue includes up-to-date news on the model industry, plus reviews of the latest diecast and white metal releases.
Diecast series covered in the magazine include Matchbox, Corgi, Dinky, Hot Wheels, Lone Star Toys, Norev, Ixo, Kenner, Exclusive First Editions, and Johnny Lightning. Articles based on TV, film and comic book character have included Batman, Spider-Man, Looney Tunes, Star Trek, The Magic Roundabout, James Bond, The Simpsons and The Green Hornet.
In October 2007, Diecast Collector celebrated its tenth anniversary with a special issue in which columnists and readers discussed their favourite diecast models from the past decade.
In 2010, Forbes was replaced as editor by Denise Burrows, who also edited companion magazine Collectors' Gazette. Rick Wilson was editor from July 2013 to January 2016, being replaced by Tim Morgan later that year. Rick Wilson returned as editor in September 2018, a position he still holds. During his second stint, Wilson authored and edited the specialist Fast and Fifty publication, celebrating the 5 |
https://en.wikipedia.org/wiki/Orange%20flower%20water | Orange flower water, or orange blossom water, is a clear aromatic by-product of the distillation of fresh bitter-orange blossoms for their essential oil.
Uses
This essential water has traditionally been used as an aromatizer in many Mediterranean traditional dessert dishes, such as in France for the gibassier and pompe à l'huile or in Spain for the Roscón de Reyes (King cake), or in Italy for the pastiera, or the Samsa in Tunisia or in Moroccan coffee, but has more recently found its way into other cuisines. For example, orange flower water is used in Europe to flavor madeleines, in Mexico to flavor little wedding cakes and Pan de muerto, and in the United States to make orange blossom scones and marshmallows. Orange flower water is also used as an ingredient in some cocktails, such as the Ramos Gin Fizz. In Malta and many North African as well as Middle Eastern countries, orange blossom water is widely used as medicine for stomach ache and given to small children as well as adults.
Orange flower water has been a traditional ingredient used often in North African as well as in Middle Eastern cooking. In Arab variants of baklava, orange blossom water is often mixed with the sweet syrup for flavor. Orange blossoms are believed to be used in this manner because they are seen as the traditional bridal flower and, therefore, symbolize purity (white, small and delicate). It is also added to plain water in the Middle East to mask high mineral content and other unpleasant flavors (for example, those arising from storage in a qulla (), a type of clay jug that keeps water cool in a manner similar to the zeer); some add the fragrance irrespective of the taste of the plain water.
Orange blossom water serves two purposes in the Maghreb: one usage is as a perfume or freshener, usually given to guests to wash their hands upon entering the host house or before drinking tea. It is put in a special silver or metal container, recognizable in the typical Maghrebi tea set. This old |
https://en.wikipedia.org/wiki/Persymmetric%20matrix | In mathematics, persymmetric matrix may refer to:
a square matrix which is symmetric with respect to the northeast-to-southwest diagonal (anti-diagonal); or
a square matrix such that the values on each line perpendicular to the main diagonal are the same for a given line.
The first definition is the most common in the recent literature. The designation "Hankel matrix" is often used for matrices satisfying the property in the second definition.
Definition 1
Let A = (aij) be an n × n matrix. The first definition of persymmetric requires that
for all i, j.
For example, 5 × 5 persymmetric matrices are of the form
This can be equivalently expressed as AJ = JAT where J is the exchange matrix.
A third way to express this is seen by post-multiplyingg AJ = JAT with J on both sides, showing that AT rotated 180 degrees is identical to A. A = JATJ.
A symmetric matrix is a matrix whose values are symmetric in the northwest-to-southeast diagonal. If a symmetric matrix is rotated by 90°, it becomes a persymmetric matrix. Symmetric persymmetric matrices are sometimes called bisymmetric matrices.
Definition 2
The second definition is due to Thomas Muir. It says that the square matrix A = (aij) is persymmetric if aij depends only on i + j. Persymmetric matrices in this sense, or Hankel matrices as they are often called, are of the form
A persymmetric determinant is the determinant of a persymmetric matrix.
A matrix for which the values on each line parallel to the main diagonal are constant is called a Toeplitz matrix.
See also
Centrosymmetric matrix |
https://en.wikipedia.org/wiki/Fc%CE%B1/%CE%BCR | Fcα/μR, also known as is CD351 (Cluster of Differentiation 351), is an Fc receptor that binds IgM with high affinity and IgA with a 10-fold lower affinity. In mice the receptor is expressed on macrophages, follicular dendritic cells, marginal zone B cells, follicular B cells, and kidney tubular epithelial cells. In humans expression has been described on intestinal lamina propria cells, Paneth cells, follicular dendritic cells in tonsils, activated macrophages and some types of pre-germinal centre IgD+/CD38+ B cells. |
https://en.wikipedia.org/wiki/Nuisance%20wildlife%20management | Nuisance wildlife management is the selective removal of problem individuals or populations of specific species of wildlife. Other terms for the field include wildlife damage management, wildlife control, and animal damage control. Some wild animal species may get used to human presence, causing property damage or risking the transfer of diseases (zoonoses) to humans or pets. Many wildlife species coexist with humans very successfully, such as commensal rodents which have become more or less dependent on humans.
Common nuisance species
Wild animals that can cause problems in homes, gardens or yards include armadillos, skunks, boars, foxes, squirrels, snakes, rats, groundhogs, beavers, opossums, raccoons, bats, moles, deer, mice, coyotes, bears, ravens, seagulls, woodpeckers and pigeons. In the United States, some of these species are protected, such as bears, ravens, bats, deer, woodpeckers, and coyotes, and a permit may be required to control some species.
Conflicts between people and wildlife arise in certain situations, such as when an animal's population becomes too large for a particular area to support. Human-induced changes in the environment will often result in increased numbers of a species. For example, piles of scrap building material make excellent sites where rodents can nest. Food left out for household pets is often equally attractive to some wildlife species. In these situations, the wildlife have suitable food and habitat and may become a nuisance.
The Humane Society of the United States (HSUS) provides strategies for the control of species such as bats, bears, chipmunks, coyotes, deer, mice, racoons and snakes.
Control methods
The most commonly used methods for controlling nuisance wildlife around homes and gardens include exclusion, habitat modification, repellents, toxic baits, glue boards, traps and frightening.
Exclusion
Exclusion techniques refer to the act of sealing a home to prevent wildlife; such as, rodents (squirrels, rats, mice |
https://en.wikipedia.org/wiki/Oracle%20VM%20Server%20for%20SPARC | Logical Domains (LDoms or LDOM) is the server virtualization and partitioning technology for SPARC V9 processors. It was first released by Sun Microsystems in April 2007. After the Oracle acquisition of Sun in January 2010, the product has been re-branded as Oracle VM Server for SPARC from version 2.0 onwards.
Each domain is a full virtual machine with a reconfigurable subset of hardware resources. Domains can be securely live migrated between servers while running. Operating systems running inside Logical Domains can be started, stopped, and rebooted independently. A running domain can be dynamically reconfigured to add or remove CPUs, RAM, or I/O devices without requiring a reboot. Using Dynamic Resource Management, CPU resources can be automatically reconfigured as needed.
Supported hardware
SPARC hypervisors run in hyperprivileged execution mode, which was introduced in the sun4v architecture. The sun4v processors released as of October 2015 are the UltraSPARC T1, T2, T2+, T3, T4, T5, M5, M6, M10, and M7. Systems based on UltraSPARC T1 support only Logical Domains versions 1.0-1.2. The newer types of T-series servers support both older Logical Domains and newer Oracle VM Server for SPARC product version 2.0 and later. These include:
UltraSPARC T1-based:
Sun / Fujitsu SPARC Enterprise T1000 and T2000 servers
Sun Fire T1000 and T2000 servers
Netra T2000 Server
Netra CP3060 Blade
Sun Blade T6300 Server Module
UltraSPARC T2-based:
Sun / Fujitsu SPARC Enterprise T5120 and T5220 servers
Sun Blade T6320 Server Module
Netra CP3260 Blade
Netra T5220 Rackmount Server
UltraSPARC T2 Plus systems:
Sun / Fujitsu SPARC Enterprise T5140 and T5240 servers (2 sockets)
Sun / Fujitsu SPARC Enterprise T5440 (4 sockets)
Sun Blade T6340 Server Module (2 sockets)
SPARC T3 systems:
Sun / Fujitsu SPARC T3-1 servers (1 socket)
Sun SPARC T3-1B Server Module (1 socket)
Sun / Fujitsu SPARC T3-2 servers (2 sockets)
Sun / Fujitsu SPARC T3-4 servers (4 sockets)
SPARC T |
https://en.wikipedia.org/wiki/Protein%20G | Protein G is an immunoglobulin-binding protein expressed in group C and G Streptococcal bacteria much like Protein A but with differing binding specificities. It is a ~60-kDA (65 kDA for strain G148 and 58 kDa for strain C40) cell surface protein that has found application in purifying antibodies through its binding to the Fab and Fc region. The native molecule also binds albumin, but because serum albumin is a major contaminant of antibody sources, the albumin binding site has been removed from recombinant forms of Protein G. This recombinant Protein G, either labeled with a fluorophore or a single-stranded DNA strand, was used as a replacement for secondary antibodies in immunofluorescence and super-resolution imaging.
Other antibody binding proteins
In addition to Protein G, other immunoglobulin-binding bacterial proteins such as Protein A, Protein A/G and Protein L are all commonly used to purify, immobilize or detect immunoglobulins. Each of these immunoglobulin-binding proteins has a different antibody binding profile in terms of the portion of the antibody that is recognized and the species and type of antibodies it will bind.
Folding of Protein G, B1 Domain
An ab initio simulation of the protein G B1 domain demonstrates that, as earlier results suggested, this protein initiates folding via a nucleation event in the hydrophobic core residues followed by small adjustments. The folding events are as follows:
a β-hairpin is formed, stabilized by residues W43, Y45, and F52.
Residue contacts between residue F30, in an α-helix, and the β-hairpin strengthen.
Nucleation of the β-sheet starting from residues L5 and F52, occurs.
The last nucleation residue, Y3, assists in forming the central part of the β-sheet resulting in a globular protein.
The protein G B1 domain is (aka. GB1) often used as part of a fusion protein to keep other domains in solution during experiments in solution (e.g. NMR). Many previously insoluble domains have become soluble with the fusi |
https://en.wikipedia.org/wiki/Jack%20Hui | Jack Hui (; born 1988) is an alumnus from Queen's College in Hong Kong. He is considered a prominent student because of the outstanding performance in International Mathematical Olympiad (IMO) and various national and regional mathematics competitions. Though accused of sexually assaulting an airline hostess, he was later found not guilty, and was admitted to the Faculty of Mathematics of Peking University.
Maths awards
Hui's talent in mathematics won him many medals and awards in different competitions. In 2003, he participated in the Asia Inter-cities Teenagers Mathematics Invitation Competition, which was held in Fuzhou, Fujian, China, representing Hong Kong. His team won the junior secondary team and group event. He was successful in science competition too, winning the first class award in the individual event in the Hong Kong Physics Olympiad 2004. In 2005, besides winning a bronze medal in China Mathematics Olympiad (CMO), he was successful in the 46th International Mathematical Olympiad (IMO) in Mérida, Mexico, capturing a silver medal.
In 2006, he claimed the bronze medal again in the China Mathematics Olympiad (CMO). In March, he won another bronze in the Asian Pacific Mathematics Olympiad (APMO) 2006. Finally, he won another silver medal in the 2006 IMO, this time in Ljubljana, Slovenia. |
https://en.wikipedia.org/wiki/Petrifilm | The 3M Petrifilm plate is an all-in-one plating system made by the Food Safety Division of the 3M Company. They are heavily used in many microbiology-related industries and fields to culture various micro-organisms and are meant to be a more efficient method for detection and enumeration compared to conventional plating techniques. A majority of its use is for the testing of foodstuffs.
Petrifilm plates are designed to be as accurate as conventional plating methods. Ingredients usually vary from plate to plate depending on what micro-organism is being cultured, but generally a Petrifilm comprises a cold-water-soluble gelling agent, nutrients, and indicators for activity and enumeration.
A typical Petrifilm plate has a 10 cm(H) × 7.5 cm(W) bottom film which contains a foam barrier accommodating the plating surface, the plating surface itself (a circular area of about 20 cm2), and a top film which encloses the sample within the Petrifilm. A 1 cm × 1 cm yellow grid is printed on the back of the plate to assist enumeration. A plastic “spreader” is also used to spread the inoculum evenly.
Comparisons between Petrifilm plates and standard methods
Petrifilm plates have become widely used because of their cost-effectiveness, simplicity, convenience, and ease of use. For example, conventional plating would require preparing agar for pour plating, or using agar plates and vial inoculum loops for streak plating; but for Petrifilm plates, the agar is completely housed in a single unit so that only the sample has to be added, which saves time.
For incubation, Petrifilm plates can be safely stacked and incubated just like Petri dishes. Since they are paper thin, more plates can be stacked together than Petri dishes (although 3M recommends that Petrifilms be stacked no higher than 20).
For enumeration, Petrifilm plates can be used on any colony counter for enumeration just like a Petri dish. Various enumeration experiments have shown very little or no variance betw |
https://en.wikipedia.org/wiki/High-power%20impulse%20magnetron%20sputtering | High-power impulse magnetron sputtering (HIPIMS or HiPIMS, also known as high-power pulsed magnetron sputtering, HPPMS) is a method for physical vapor deposition of thin films which is based on magnetron sputter deposition. HIPIMS utilises extremely high power densities of the order of kW⋅cm−2 in short pulses (impulses) of tens of microseconds at low duty cycle (on/off time ratio) of < 10%. Distinguishing features of HIPIMS are a high degree of ionisation of the sputtered metal and a high rate of molecular gas dissociation which result in high density of deposited films. The ionization and dissociation degree increase according to the peak cathode power. The limit is determined by the transition of the discharge from glow to arc phase. The peak power and the duty cycle are selected so as to maintain an average cathode power similar to conventional sputtering (1–10 W⋅cm−2).
HIPIMS is used for:
adhesion enhancing pretreatment of the substrate prior to coating deposition (substrate etching)
deposition of thin films with high microstructure density
HIPIMS plasma discharge
HIPIMS plasma is generated by a glow discharge where the discharge current density can reach several A⋅cm−2, whilst the discharge voltage is maintained at several hundred volts. The discharge is homogeneously distributed across the surface of the cathode (target) however above a certain threshold of current density it becomes concentrated in narrow ionization zones that move along a path known as the target erosion "racetrack".
HIPIMS generates a high density plasma of the order of 1013 ions⋅cm−3 containing high fractions of target metal ions. The main ionisation mechanism is electron impact, which is balanced by charge exchange, diffusion, and plasma ejection in flares. The ionisation rates depend on the plasma density.
The ionisation degree of the metal vapour is a strong function of the peak current density of the discharge. At high current densities, sputtered ions with charge 2+ and highe |
https://en.wikipedia.org/wiki/Flat%20topology | In mathematics, the flat topology is a Grothendieck topology used in algebraic geometry. It is used to define the theory of flat cohomology; it also plays a fundamental role in the theory of descent (faithfully flat descent). The term flat here comes from flat modules.
There are several slightly different flat topologies, the most common of which are the fppf topology and the fpqc topology. fppf stands for , and in this topology, a morphism of affine schemes is a covering morphism if it is faithfully flat and of finite presentation. fpqc stands for , and in this topology, a morphism of affine schemes is a covering morphism if it is faithfully flat. In both categories, a covering family is defined be a family which is a cover on Zariski open subsets. In the fpqc topology, any faithfully flat and quasi-compact morphism is a cover. These topologies are closely related to descent. The "pure" faithfully flat topology without any further finiteness conditions such as quasi compactness or finite presentation is not used much as is not subcanonical; in other words, representable functors need not be sheaves.
Unfortunately the terminology for flat topologies is not standardized. Some authors use the term "topology" for a pretopology, and there are several slightly different pretopologies sometimes called the fppf or fpqc (pre)topology, which sometimes give the same topology.
Flat cohomology was introduced by Grothendieck in about 1960.
The big and small fppf sites
Let X be an affine scheme. We define an fppf cover of X to be a finite and jointly surjective family of morphisms
(φa : Xa → X)
with each Xa affine and each φa flat, finitely presented. This generates a pretopology: for X arbitrary, we define an fppf cover of X to be a family
(φa : Xa → X)
which is an fppf cover after base changing to an open affine subscheme of X. This pretopology generates a topology called the fppf topology. (This is not the same as the topology we would get if we started with |
https://en.wikipedia.org/wiki/Lobeline | Lobeline is a piperidine alkaloid found in a variety of plants, particularly those in the genus Lobelia, including Indian tobacco (Lobelia inflata), Devil's tobacco (Lobelia tupa), great lobelia (Lobelia siphilitica), Lobelia chinensis, and Hippobroma longiflora. In its pure form, it is a white amorphous powder which is freely soluble in water.
Potential uses
Lobeline has been sold, in tablet form, for use as a smoking cessation aid, but scientific research has not provided supporting evidence for this use. Lobeline has also been studied for the treatment of other drug addictions such as addiction to amphetamines, cocaine, or alcohol; however, there is limited clinical evidence of any efficacy.
Toxicity
Ingestion of lobeline may cause nausea, vomiting, diarrhea, coughing, dizziness, visual disturbances, hearing disturbances, mental confusion, weakness, slowed heart rate, increased blood pressure, increased breathing rate, tremors, and seizures. Lobeline has a narrow therapeutic index: the potentially beneficial dose of lobeline is very close to the toxic dose.
Pharmacology
Lobeline has multiple mechanisms of action, acting as a VMAT2 ligand, which stimulates dopamine release to a moderate extent when administered alone, but reduces the dopamine release caused by methamphetamine. It also inhibits the reuptake of dopamine and serotonin, and acts as a mixed agonist–antagonist at nicotinic acetylcholine receptors to which it binds at the subunit interfaces of the extracellular domain. It is also an antagonist at μ-opioid receptors. It seems to be a P-glycoprotein inhibitor, according to at least one study. It has been hypothesized that P-glycoprotein inhibition reduces chemotherapeutic resistance in cancer, presumably affecting any substrates of P-gp.
Analogous compounds, such as lobelane (a minor alkaloid found in the same plants) and its synthetic derivatives have similar biological effects with somewhat different relative affinities to VMAT and other proteins. |
https://en.wikipedia.org/wiki/Microsoft%20App-V | Microsoft Application Virtualization (also known as App-V; formerly Softricity SoftGrid) is an application virtualization and application streaming solution from Microsoft. It was originally developed by Softricity, a company based in Boston, Massachusetts, acquired by Microsoft on July 17, 2006. App-V represents Microsoft's entry to the application virtualization market, alongside their other virtualization technologies such as Hyper-V, Microsoft User Environment Virtualization (UE-V), Remote Desktop Services, and System Center Virtual Machine Manager.
Overview
App-V allows applications to be deployed ("streamed") in real-time to any client from a virtual application server. It removes the need for traditional local installation of the applications, although a standalone deployment method is also supported. With a streaming-based implementation, the App-V client needs to be installed on the client machines and application data that is stored on the virtual application server is installed (streamed) to the client cache on demand when it is first used, or pre-installed in a local cache. The App-V stack sandboxes the execution environment so that an application does not make changes directly to the underlying operating system's file system and/or Windows Registry, but rather is contained in an application-specific "bubble". App-V applications are also sandboxed from each other, so that different versions of the same application can be run under App-V concurrently and so that mutually exclusive applications can co-exist on the same system. Nevertheless the separation is not a security boundary.
App-V thus allows centralized installation and management of deployed applications. It supports policy based access control; administrators can define and restrict access to the applications by certain users, or on certain computers, by defining policies governing the usage. App-V also features a tracking interface to track the usage of the virtualized application. Servers may |
https://en.wikipedia.org/wiki/Windows%20Vista%20networking%20technologies | In computing, Microsoft's Windows Vista and Windows Server 2008 introduced in 2007/2008 a new networking stack named Next Generation TCP/IP stack,
to improve on the previous stack in several ways.
The stack includes native implementation of IPv6, as well as a complete overhaul of IPv4. The new TCP/IP stack uses a new method to store configuration settings that enables more dynamic control and does not require a computer restart after a change in settings. The new stack, implemented as a dual-stack model, depends on a strong host-model and features an infrastructure to enable more modular components that one can dynamically insert and remove.
Architecture
The Next Generation TCP/IP stack connects to NICs via a Network Driver Interface Specification (NDIS) driver. The network stack, implemented in tcpip.sys implements the Transport, Network and Data link layers of the TCP/IP model. The Transport layer includes implementations for TCP, UDP and unformatted RAW protocols. At the Network layer, IPv4 and IPv6 protocols are implemented in a dual-stack architecture. And the Data link layer (also called Framing layer) implements 802.3, 802.1, PPP, Loopback and tunnelling protocols. Each layer can accommodate Windows Filtering Platform (WFP) shims, which allows packets at that layer to be introspected and also host the WFP Callout API. The networking API is exposed via three components:
Winsock A user mode API for abstracting network communication using sockets and ports. Datagram sockets are used for UDP, whereas Stream sockets are for TCP. While Winsock is a user mode library, it uses a kernel mode driver, called Ancillary Function Driver (AFD) to implement certain functionality.
Winsock Kernel (WSK) A kernel-mode API providing the same socket-and-port abstraction as Winsock, while exposing other features such as Asynchronous I/O using I/O request packets.
Transport Driver Interface (TDI) A kernel-mode API which can be used for legacy protocols like NetBIOS. It includes a |
https://en.wikipedia.org/wiki/Molecular%20property | Molecular properties include the chemical properties, physical properties, and structural properties of molecules, including drugs. Molecular properties typically do not include pharmacological or biological properties of a chemical compound.
See also
Biological activity
Chemical property
Chemical structure
Lipinski's rule of five, describing molecular properties of drugs
Physical property
QSAR, quantitative structure-activity relationship |
https://en.wikipedia.org/wiki/Wave%20dash | Wave dash () is a character represented in Japanese character encoding, usually used to represent a range. The wave dash is similar to, but not the same as, the tilde character (), which is often used interchangeably with it.
The vertical wave dash () is not currently included in Unicode, but there is a similar symbol available called the wavy line (). It is created by rotating right (clockwise) the wavy dash symbol () to form a vertical wave-like pattern.
Wave dash is also written in vertical text layout. Vertical wave dash is the vertical form by rotation and flip in Unicode and JIS C 6226.
See also
Dash#Swung dash
Tilde#Unicode and Shift JIS encoding of wave dash
Japanese punctuation#Wave dash
Code reference |
https://en.wikipedia.org/wiki/Gore%20%28segment%29 | A gore is a sector of a curved surface or the curved surface that lies between two close lines of longitude on a globe and may be flattened to a plane surface with little distortion. The term has been extended to include similarly shaped pieces such as the panels of a hot-air balloon or parachute, or the triangular insert that allows extra movement in a garment (see Gore (fabrics)).
Examples
Globes of the Earth and the celestial sphere were first mass-produced by Johannes Schöner using a process of printing map details on 12 paper gores that were cut out then pasted to a sphere. This process is still often used. The gores are conveniently made to each have a width of 30 degrees of longitude matching the principal meridians from the South Pole and North Pole to the Equator.
Parachutes and hot air balloons are made from gores of lightweight material. The gores are cut from flat material and stitched together to create various shapes.
Pressure suit joints are often constructed of alternating gores and convolutes of material constrained by cables or straps along the sides of the joint, producing an accordion-like structure that flexes with nearly constant volume to minimize the mechanical work which must be done by the suit occupant.
Corners in round duct-work can be created by welding or fixing gores of metal sheet to form a bend.
Some designers use the stretched grid method to design gores that are cut out of weather-resistant fabric and then stitched together to form fabric structures. |
https://en.wikipedia.org/wiki/Bapat%E2%80%93Beg%20theorem | In probability theory, the Bapat–Beg theorem gives the joint probability distribution of order statistics of independent but not necessarily identically distributed random variables in terms of the cumulative distribution functions of the random variables. Ravindra Bapat and Beg published the theorem in 1989, though they did not offer a proof. A simple proof was offered by Hande in 1994.
Often, all elements of the sample are obtained from the same population and thus have the same probability distribution. The Bapat–Beg theorem describes the order statistics when each element of the sample is obtained from a different statistical population and therefore has its own probability distribution.
Statement
Let be independent real valued random variables with cumulative distribution functions respectively . Write for the order statistics. Then the joint probability distribution of the order statistics (with and ) is
where
is the permanent of the given block matrix. (The figures under the braces show the number of columns.)
Independent identically distributed case
In the case when the variables are independent and identically distributed with cumulative probability distribution function for all i the theorem reduces to
Remarks
No assumption of continuity of the cumulative distribution functions is needed.
If the inequalities x1 < x2 < ... < xk are not imposed, some of the inequalities "may be redundant and the probability can be evaluated after making the necessary reduction."
Complexity
Glueck et al. note that the Bapat‒Beg formula is computationally intractable, because it involves an exponential number of permanents of the size of the number of random variables. However, when the random variables have only two possible distributions, the complexity can be reduced to . Thus, in the case of two populations, the complexity is polynomial in for any fixed number of statistics . |
https://en.wikipedia.org/wiki/Bystander%20effect%20%28radiobiology%29 | The radiation-induced bystander effect (bystander effect) is the phenomenon in which unirradiated cells exhibit irradiated effects as a result of signals received from nearby irradiated cells. In November 1992, Hatsumi Nagasawa and John B. Little first reported this radiobiological phenomenon.
Effect
There is evidence that targeted cytoplasmic irradiation results in mutation in the nucleus of the hit cells. Cells that are not directly hit by an alpha particle, but are in the vicinity of one that is hit, also contribute to the genotoxic response of the cell population. Similarly, when cells are irradiated, and the medium is transferred to unirradiated cells, these unirradiated cells show bystander responses when assayed for clonogenic survival and oncogenic transformation. This is also attributed to the bystander effect.
Demonstration
The demonstration of a bystander effect in 3D human tissues and, more recently, in whole organisms have clear implication of the potential relevance of the non-targeted response to human health.
Consequences
This effect may also contribute to the final biological consequences of exposure to low doses of radiation. However, there is currently insufficient evidence to suggest that the bystander effect promotes carcinogenesis in humans at low doses.
Notes
Note that the bystander effect is not the same as the abscopal effect. The abscopal effect is a phenomenon where the response to radiation is seen in an organ/site distant to the irradiated organ/area, that is, the responding cells are not juxtaposed with the irradiated cells. T-cells and dendritic cells have been implicated to be part of the mechanism.
In suicide gene therapy, the "bystander effect" is the ability of the transfected cells to transfer death signals to neighboring tumor cells. |
https://en.wikipedia.org/wiki/Computer%20security%20incident%20management | In the fields of computer security and information technology, computer security incident management involves the monitoring and detection of security events on a computer or computer network, and the execution of proper responses to those events. Computer security incident management is a specialized form of incident management, the primary purpose of which is the development of a well understood and predictable response to damaging events and computer intrusions.
Incident management requires a process and a response team which follows this process. This definition of computer security incident management follows the standards and definitions described in the National Incident Management System (NIMS). The incident coordinator manages the response to an emergency security incident. In a Natural Disaster or other event requiring response from Emergency services, the incident coordinator would act as a liaison to the emergency services incident manager.
Overview
Computer security incident management is an administrative function of managing and protecting computer assets, networks and information systems. These systems continue to become more critical to the personal and economic welfare of our society. Organizations (public and private sector groups, associations and enterprises) must understand their responsibilities to the public good and to the welfare of their memberships and stakeholders. This responsibility extends to having a management program for “what to do, when things go wrong.” Incident management is a program which defines and implements a process that an organization may adopt to promote its own welfare and the security of the public.
Components of an incident
Events
An event is an observable change to the normal behavior of a system, environment, process, workflow or person (components). There are three basic types of events:
Normal—a normal event does not affect critical components or require change controls prior to the implementation |
https://en.wikipedia.org/wiki/WebNFS | WebNFS is an extension to the Network File System (NFS) for allowing clients to access a file system over the internet using a simplified, firewall-friendly protocol.
WebNFS was developed to give Java applets and other internet enabled applications a way of accessing filesystem services over the internet. While NFS provides applications on Unix with full filesystem semantics, not all of these might be needed in a distributed, read-only web environment. Conversely, access restrictions—such as requiring the use of restricted ports for originating requests—normally used in closed environments are not usually applicable in public distributed environments.
WebNFS makes use of a well known port (port 2049 on both UDP and TCP) thus avoiding the overhead and unpredictability of using the ONC RPC portmap protocol. WebNFS adds public filehandles and multicomponent lookups to the NFS protocol.
WebNFS is specified by a number of RFCs:
: WebNFS Client
: WebNFS Server
: NFS URL Scheme
: Security negotiation for WebNFS
In 2007, Sun Microsystems opensourced its WebNFS implementation. The name has since changed to YANFS (Yet Another NFS) to reflect the expanded scope of the project to include a server side implementation.
Legacy
While WebNFS itself did not gain much traction, several important WebNFS features later became part of NFSv4 – such as the usage of port 2049, or the concept of a fixed "root filehandle" (which evolved from WebNFS public filehandles and allows exported filesystems to be accessed without needing the MOUNT protocol to learn their individual root handles first), both together allowing NFSv4 to function without the portmap service. |
https://en.wikipedia.org/wiki/AMA%20Scientific%20Achievement%20Award | The AMA Scientific Achievement Award is awarded by American Medical Association. It may be given to either physicians or non-physician scientists who have contributed significantly to the field of medical science. The award itself consists of a gold medallion.
The recipients are chosen by the AMA's Board of Trustees, and Physician candidates must be AMA members.
Recipients
Source: AMA Awards
1962 – Donald D. Van Slyke, PhD, Upton, New York
1963 – John F. Enders, Phd, Boston, Massachusetts
1964 – René J. Dubos, PhD, New York, New York
1965 – Edward C. Kendall, PhD, Princeton, New Jersey
1966 – Wendell M. Stanley, PhD, Berkeley, California
1967 – Gregory Pincus, ScD, Shrewsbury, Massachusetts
1968 – Arthur Kornberg, MD, Palo Alto, California
1969 – Philip Handler, PhD, Durham, North Carolina
1970 – Choh Hao Li, Phd, Berkeley, California
1971 – Robert B. Woodward, MD, Cambridge, Massachusetts
1972 – William Bennett Kouwenhoven, MD, Baltimore, Maryland
1973 – Edith Hinkley Quimby, ScD, Palo Alto, California
1974 – Philip Abelson, PhD, Washington, District of Columbia
1975 – Rosalyn Yalow, PhD, Bronx, New York; Solomon A. Berson, MD (posthumously)
1976 – Harry Goldblatt, MD, Cleveland, Ohio
1977 – Helen B. Taussig, MD, Baltimore, Maryland
1978 – F. Mason Sones, MD, Cleveland, Ohio
1979 – Orvan W. Hess, MD, New Haven, Connecticut
1980 – Harold E. Kleinert, MD, Louisville, Kentucky
1981 – Hans von Leden, MD, Los Angeles, California
1982 – Willem J. Kolff, PhD, Salt Lake City, Utah
1983 – Maurice R. Hillerman, PhD, West Point, Pennsylvania
1984 – Maurice J. Jurkiewicz, MD, Atlanta, Georgia
1985 – Solomon H. Snyder, MD, Baltimore, Maryland
1986 – George Edward Burch, MD, New Orleans, Louisiana
1987 – Norman E. Shumway, MD, Stanford, California
1988 – Harriet P. Dustan, MD, Birmingham, Alabama
1989 – John G. Morrison, MD, Piedmont, California
1990 – Arthur C. Guyton, MD, Jackson, Mississippi
1991 – Henry Nicholas Wagner, Jr., MD, Baltimore, |
https://en.wikipedia.org/wiki/High%20Performance%20Wireless%20Research%20and%20Education%20Network | The High Performance Wireless Research and Education Network (HPWREN) is a network research program, funded by the National Science Foundation. The program includes the creation, demonstration, and evaluation of a non-commercial, prototype, high-performance, wide-area, wireless network in its Southern California service area.
The HPWREN program is a collaborative, interdisciplinary and multi-institutional cyber-infrastructure for research and education purposes. The program also provides data, and data transmission capabilities, to emergency first responders in its service area.
Network
The program includes the creation, demonstration, and evaluation of a non-commercial, prototype, high-performance, wide-area, wireless network in its service area. Currently, the HPWREN network is used for network analysis research, and it also provides high-speed Internet access to field researchers.
Service area
Southern California, specifically San Diego, Riverside, and Imperial counties.
Backbone nodes
The network includes backbone nodes located at the University of California San Diego (UCSD) and San Diego State University (SDSU) campuses, as well as a number of “hard-to-reach” areas in remote environments.
Operations
The HPWREN backbone itself operates primarily in the licensed spectrum and project researchers use off-the-shelf technology to create a redundant topology. Access links often utilize license-exempt radios.
In 2002, HPWREN researchers conducted an expedition to locate the SEALAB II/III habitat located off the Scripps pier in La Jolla, California. From the MV Kellie Chouest and utilizing a Scorpio ROV to find the site, researchers were able to conduct a live multicast from ship to shore.
Topology
The network spans from the Southern California coast to the inland valleys, on to the high mountains (reaching more than 8700 feet), and out to the remote desert. The network's longest link is 72 miles in distance – reaching from the San Diego Supercomputer Center to |
https://en.wikipedia.org/wiki/Thalassiosira%20pseudonana | Thalassiosira pseudonana is a species of marine centric diatoms. It was chosen as the first eukaryotic marine phytoplankton for whole genome sequencing. T. pseudonana was selected for this study because it is a model for diatom physiology studies, belongs to a genus widely distributed throughout the world's oceans, and has a relatively small genome at 34 mega base pairs. Scientists are researching on diatom light absorption, using the marine diatom of Thalassiosira. The diatom requires a high enough concentration of CO2 in order to utilize C4 metabolism (Clement et al. 2015).
The clone of T. pseudonana that was sequenced is CCMP 1335 and is available from the National Center for Marine Algae and Microbiota at Bigelow Laboratory for Ocean Sciences. This clone was originally collected in 1958 from Moriches Bay (Long Island, New York) and has been maintained continuously in culture.
Morphology
Thalassiosira pseudonana has a radial symmetry. Its biosilica cell wall is divided into two halves, which are joined together by girdle bands, giving them a cylindrical shape or making them appear as a Petri dish. The diameter of their valves ranges from 2 to 9 μm. The valve is made up of silica ribs that radiate from the center with many 18 nm diameter nanopores between them. The face of the valve has 0-1 central fultoportula and a marginal ring of fultoportulae (6-12). The external openings of the central fultoportula appear as rimmed holes, whereas those of the marginal fultoportulae appear as short rimmed tubes, which are sometimes obliquely sectioned at the opening. On the internal face of the valve, two satellite pores surround the central fultoportula, while the marginal fultoportulae are surrounded by three satellite pores. The rimoportula is a rimmed pore located on the valve face, with a size similar to the fultoportula, and positioned between two fultoportulae. The pervalvar axis of T. pseudonana can be either shorter than or equal to the valve diameter. Their cell |
https://en.wikipedia.org/wiki/Nylon%2011 | Nylon 11 or Polyamide 11 (PA 11) is a polyamide, bioplastic and a member of the nylon family of polymers produced by the polymerization of 11-aminoundecanoic acid. It is produced from castor beans by Arkema under the trade name Rilsan.
Nylon 11 is applied in the fields of oil and gas, aerospace, automotive, textiles, electronics and sports equipment, frequently in tubing, wire sheathing, and metal coatings.
History
In 1938, a research director for Thann & Mulhouse, Joseph Zeltner, first conceived the idea of Nylon 11, which was suggested in the works of Wallace Carothers. Thann & Mulhouse had already been involved in processing castor oil for 10-undecenoic-acid, which would eventually be converted into the first amount of 11-aminoundecanoic acid in 1940 with the help of coworkers Michel Genas and Marcel Kastner. In 1944, Kastner sufficiently improved the monomer process and the first patents for Nylon 11 were filed in 1947. The first nylon 11 thread was created in 1950 and full industrial production began with the opening of the Marseilles production facility in 1955, which remains the sole producer of 11-aminoudecanoic acid today.
Currently Arkema polymerizes Nylon 11 in Birdsboro, PA, Changshu, and Serquigny.
Chemistry
The chemical process of creating Nylon 11 begins with ricinoleic acid which makes up 85-90% of castor oil. Ricinoleic acid is first transesterified with methanol creating methyl ricinoleate, which is then cracked to create heptaldehyde and methyl undecylenate. These undergo hydrolysis to create methanol, which is re-used in the initial transesterification of ricinoleic acid, and undecylenic acid that is added on to hydrogen bromide. After hydrolysis, hydrogen bromide then undergoes nucleophilic substitution with ammonia to form 11-aminoundecanoic acid, which is polymerized into nylon 11.
Properties
As seen in the table below, Nylon 11 has lower values of density, flexural and Young's modulus, water absorption, as well as melting and glass |
https://en.wikipedia.org/wiki/Inter-ictal%20spiking | Inter-ictal spiking refers to abnormal neuronal discharges between epileptic seizures. This abnormal activity can originate from one or more cranial lobes, often travels from one lobe to another, and interferes with normal activity from the affected lobe. Patients with severe, intractable forms of epilepsy can experience unremitting abnormal brain activity due to inter-ictal spiking.
This phenomenon has been confirmed through telemetry by electrocorticography in which electrode grids are positioned sub-durally and augmented by intra-cranial depth electrodes. The EEG tracings from these are correlated with external video recordings of physical responses to epileptic seizures to help determine type of seizure and origin of onset.
Neurophysiology |
https://en.wikipedia.org/wiki/Solid%20State%20Communications | Solid State Communications is a peer-reviewed scientific journal of solid-state physics. The journal specializes in short papers on significant developments in the condensed matter science. The journal was established 1963, when the Journal of Physics and Chemistry of Solids split its letters section to create this journal. Elias Burstein served as founding chief editor until 1992, and was succeeded by Manuel Cardona until 2004, when Aron Pinczuk assumed the role. Pinczuk stepped down in 2020.
The journal is published bimonthly by Elsevier and its current editor-in-chief is François Peeters (University of Antwerp).
Abstracting and Indexing
The journal is abstracted and indexing in the following databases:
Cambridge Scientific Abstracts
Chemical Abstracts
Current Contents/Physics, Chemical, & Earth Sciences
Current Contents/SciSearch Database
Current Contents/Social & Behavioral Sciences
MSCI
Engineering Index
INSPEC
PASCAL/CNRS
Research Alert
SSSA/CISA/ECA/ISMEC
Scopus
External links
Bimonthly journals
English-language journals
Hybrid open access journals
Physics journals
Academic journals established in 1963 |
https://en.wikipedia.org/wiki/SPC%20Australia | SPC Global Limited, formerly SPC Ardmona, is an Australian-based global agribusiness that specialises in food manufacturing, in particular large fruit processing and packing and owns and operates a factory in Shepparton and is one of the biggest employers in the Goulburn Valley region.
History
Started in 1917 by a group of fruit growers in Victoria’s Goulburn Valley as a cooperative which they named the Shepparton Fruit Preserving Co. Ltd. The company began operations in February 1918, canning pears, peaches and nectarines under the brand name of SPC. SPC was incorporated as a public listed company in 1912, and Ardmona opened in 1921.
SPC Ardmona was formed in 2002 by the merger of the former Shepparton Preserving Company (SPC) and Ardmona.
SPC Ardmona was bought by Coca-Cola Amatil in 2005 for . In that time, CCA had also invested about in infrastructure, and the Victorian government had provided in co-investment.
It acquired IXL and Taylors brands in 2004,.
SPCA closed its Mooroopna processing plant in 2011 and it's Kyabram factory was sold in 2019.
Coca-Cola Amatil sold SPC in 2019 for to Shepparton Partners Collective, now known as SPC Global Limited who implemented strategies for global growth.
SPC Global carries the SPC, Goulburn Valley, Ardmona, Provital, Pomlife and Helping Humans brands.
Currently, Hussein Rifai is the Chairman of the Company.
In August 2021, SPC Global mandated COVID-19 vaccinations for all staff, contractors, and visitors by November to be allowed entry to all their offices and factory sites. It announced it would provide staff paid vaccination leave and up to two days of special leave to recover if they became unwell after receiving the shot. With that, it became the first Australian company to mandate vaccines for all onsite staff and visitors.
Former Financial problems
SPC Ardmona lost $25 million in 2013, compared to a $70 million profit 8 years earlier. In 2012, SPCA disposed $100 million of fruit bought under co |
https://en.wikipedia.org/wiki/Schr%C3%B6dinger%27s%20Kitten | "Schrödinger's Kitten" is a 1988 novella by American writer George Alec Effinger, which won both a Hugo Award and a Nebula Award, as well as the Japanese Seiun Award.
The story utilizes a form of the many-worlds hypothesis, and is named after the Schrödinger's cat thought experiment. It first appeared in Omni.
Plot summary
The story follows a Middle-Eastern woman, Jehan Fatima Ashûfi, through various realities, ranging from one in which she is raped when still a girl, subsequently abandoned by her family and dies alone, to one in which she is sentenced to death for killing her would-be rapist and being unable to pay the "blood price" to his family, and another in which she becomes a physicist and companion to well-known German scientists ranging from Heisenberg to Schrödinger, and subsequently prevents the Nazis from developing nuclear weapons during World War II by simply forwarding "unintelligible scientific papers" to key politicians looking into the idea.
She is, unusually, aware of the existence of these realities, which she perceives as "visions" and assumes might come to her from Allah. Throughout different points in the story, the adult Jehan of some realities struggles to reconcile her religious upbringing and "visions" with her scientific profession; in the end, however, an aged Jehan finds satisfaction in the explanation of Hugh Everett's theory regarding the possibility of alternate realities, which fits with her personal experiences.
Awards and nominations
"Schrödinger's Kitten" won the Hugo Award for Best Novelette in 1989, as well as a Nebula Award and the Seiun Award.
See also
Many worlds hypothesis
Schrödinger's cat
1988 novels
1988 science fiction novels
Hugo Award for Best Novelette winning works
Nebula Award for Best Novelette-winning works
Physics in fiction
Quantum fiction novels
Works originally published in Omni (magazine)
Theodore Sturgeon Award-winning works |
https://en.wikipedia.org/wiki/RUCAPS | RUCAPS (Really Universal Computer-Aided Production System) was a computer aided design (CAD) system for architects, first developed during the 1970s and 1980s, and today credited as a forerunner of Building Information Modelling (BIM). It ran on minicomputers from Prime Computer and Digital Equipment Corporation (DEC).
Development
The system was initially developed by two graduates of Liverpool University, Dr John Davison and John Watts in the early 1970s. They took their work to architects Gollins Melvin Ward (GMW Architects) in London in the late 1970s, and developed it whilst working on a project for Riyadh University. It became the Really Universal Computer Aided Production System (RUCAPS), and from 1977 was sold through GMW Computers Ltd in several countries worldwide. The term 'building model' (in the sense of BIM as used today) was first used in papers in the mid-1980s: in a 1985 paper by Simon Ruffle, and later in a 1986 paper by Robert Aish - then at GMW Computers - referring to the software's use at London's Heathrow Airport.
RUCAPS was a significant milestone in the development of building modellers, selling many hundreds of copies during the early 1980s when CAD was rare and expensive, and introducing thousands of architects to computer aided design. It is regarded as a forerunner to today's BIM software, and is seen by some writers, e.g.: Jerry Laiserin, as the inspiration behind Autodesk's Revit:
While Autodesk Revit may not contain genomic snippets of Reflex code, Revit clearly is spiritual heir to a lineage of BIM "begats" — RUCAPS begat Sonata, Sonata begat Reflex, and Reflex begat Revit.
RUCAPS was superseded in the mid-late 1980s by Sonata, developed by former GMW employee Jonathan Ingram. This was sold to T2 Solutions (renamed from GMW Computers in 1987), which was eventually bought by Alias|Wavefront but then "disappeared in a mysterious, corporate black hole, somewhere in eastern Canada in 1992." Ingram then went on to develop Reflex, bough |
https://en.wikipedia.org/wiki/Collenia | Collenia is genus of fossil cyanobacteria that form a particular type of stromatolites.
Description
Collenia are stromatolites made up of convex layers flattened in the center, forming columnar colonies. The microorganisms involved were likely photosynthetic bacteria expiring oxygen.
Fossil record
Collenia stromatolites were very common in the late Precambrian, about 2.2 to 2.4 billion years ago. |
https://en.wikipedia.org/wiki/Parity-check%20matrix | In coding theory, a parity-check matrix of a linear block code C is a matrix which describes the linear relations that the components of a codeword must satisfy. It can be used to decide whether a particular vector is a codeword and is also used in decoding algorithms.
Definition
Formally, a parity check matrix H of a linear code C is a generator matrix of the dual code, C⊥. This means that a codeword c is in C if and only if the matrix-vector product (some authors would write this in an equivalent form, cH⊤ = 0.)
The rows of a parity check matrix are the coefficients of the parity check equations. That is, they show how linear combinations of certain digits (components) of each codeword equal zero. For example, the parity check matrix
,
compactly represents the parity check equations,
,
that must be satisfied for the vector to be a codeword of C.
From the definition of the parity-check matrix it directly follows the minimum distance of the code is the minimum number d such that every d - 1 columns of a parity-check matrix H are linearly independent while there exist d columns of H that are linearly dependent.
Creating a parity check matrix
The parity check matrix for a given code can be derived from its generator matrix (and vice versa). If the generator matrix for an [n,k]-code is in standard form
,
then the parity check matrix is given by
,
because
.
Negation is performed in the finite field Fq. Note that if the characteristic of the underlying field is 2 (i.e., 1 + 1 = 0 in that field), as in binary codes, then -P = P, so the negation is unnecessary.
For example, if a binary code has the generator matrix
,
then its parity check matrix is
.
It can be verified that G is a matrix, while H is a matrix.
Syndromes
For any (row) vector x of the ambient vector space, s = Hx⊤ is called the syndrome of x. The vector x is a codeword if and only if s = 0. The calculation of syndromes is the basis for the syndrome decoding algorithm.
See also
Hamm |
https://en.wikipedia.org/wiki/Rytov%20number | The Rytov number is a fundamental scaling parameter for laser propagation through atmospheric turbulence. Rytov numbers greater than 0.2 are generally considered to be strong scintillation. A Rytov number of 0 would indicate no turbulence, thus no scintillation of the beam.
Wave mechanics |
https://en.wikipedia.org/wiki/Evolutionary%20mismatch | Evolutionary mismatch (also "mismatch theory" or "evolutionary trap") is the evolutionary biology concept that a previously advantageous trait may become maladaptive due to change in the environment, especially when change is rapid. It is said this can take place in humans as well as other animals.
Environmental change leading to evolutionary mismatch can be broken down into two major categories: temporal (change of the existing environment over time, e.g. a climate change) or spatial (placing organisms into a new environment, e.g. a population migrating). Since environmental change occurs naturally and constantly, there will certainly be examples of evolutionary mismatch over time. However, because large-scale natural environmental change – like a natural disaster – is often rare, it is less often observed. Another more prevalent kind of environmental change is anthropogenic (human-caused). In recent times, humans have had a large, rapid, and trackable impact on the environment, thus creating scenarios where it is easier to observe evolutionary mismatch.
Because of the mechanism of evolution by natural selection, the environment ("nature") determines ("selects") which traits will persist in a population. Therefore, there will be a gradual weeding out of disadvantageous traits over several generations as the population becomes more adapted to its environment. Any significant change in a population's traits that cannot be attributed to other factors (such as genetic drift and mutation) will be responsive to a change in that population's environment; in other words, natural selection is inherently reactive. Shortly following an environmental change, traits that evolved in the previous environment, whether they were advantageous or neutral, are persistent for several generations in the new environment. Because evolution is gradual and environmental changes often occur very quickly on a geological scale, there is always a period of "catching-up" as the population evol |
https://en.wikipedia.org/wiki/Prescriptive%20authority%20for%20psychologists%20movement | The prescriptive authority for psychologists (RxP) movement is a movement in the United States of America among certain psychologists to give prescriptive authority to psychologists with predoctoral or postdoctoral graduate-level training in clinical psychopharmacology; successful passage of a standardized, national examination (Psychopharmacology Examination for Psychologists - Second Edition; PEP-2); supervised clinical experience; or a certificate from the Department of Defense Psychopharmacology Demonstration Project; or a diploma from the Prescribing Psychologists Register (FICPP or FICPPM) to enable them, according to state law, to prescribe psychotropic medications to treat mental disorders. This approach is non-traditional medical training focused on the specialized training to prescribe for mental health disorders by a psychologist. It includes rigorous didactics and supervised clinical experience. Legislation pertaining to prescriptive authority for psychologists has been introduced over 180 times in over half of the United States. It has passed in six states, due largely to substantial lobbying efforts by the American Psychological Association (APA), the largest professional organization of psychologists in the world with over 146,000 members. Prior to RxP legislation and in American states where it has not been passed, this role has been played by psychiatrists, who possess a medical degree and thus the authority to prescribe medication, but more frequently by primary care providers who can prescribe psychotropics, but lack extensive training in psychotropic drugs and in diagnosing and treating psychological disorders. According to the APA, the movement is a reaction to the growing public need for mental health services, particularly in under-resourced areas where patients have little or no access to psychiatrists.
Currently, in states where RxP legislation has been passed, psychologists who seek prescriptive authority must possess a doctoral degree (Ph |
https://en.wikipedia.org/wiki/Asarone | Asarone is chemical compound of the phenylpropanoid class found in certain plants such as Acorus and Asarum. There are two isomers, α (or trans) and β (or cis). As a volatile fragrance oil, it is used in killing pests and bacteria.
Pharmacology
The main clinical symptom of asarone is prolonged vomiting that sometimes lasted more than 15 hours. Asarone is not metabolized to trimethoxyamphetamine as has been claimed by online vendors.
The Council of Europe Committee of Experts on Flavouring Substances concluded that β-asarone is clearly carcinogenic and has proposed limits for its concentration in flavorings such as bitters made from Acorus calamus (sweet flag).
β-Asarone exhibits anti-fungal activity by inhibiting ergosterol biosynthesis in Aspergillus niger. However, the toxicity and carcinogenicity of asarone means that it may be difficult to develop any practical medication based on it.
See also
Elemicin
2,4,5-Trimethoxypropiophenone
Notes and references
Phenylpropenes
O-methylated phenylpropanoids
Plant toxins |
https://en.wikipedia.org/wiki/Domino%20tiling | In geometry, a domino tiling of a region in the Euclidean plane is a tessellation of the region by dominoes, shapes formed by the union of two unit squares meeting edge-to-edge. Equivalently, it is a perfect matching in the grid graph formed by placing a vertex at the center of each square of the region and connecting two vertices when they correspond to adjacent squares.
Height functions
For some classes of tilings on a regular grid in two dimensions, it is possible to define a height function associating an integer to the vertices of the grid. For instance, draw a chessboard, fix a node with height 0, then for any node there is a path from to it. On this path define the height of each node (i.e. corners of the squares) to be the height of the previous node plus one if the square on the right of the path from to is black, and minus one otherwise.
More details can be found in .
Thurston's height condition
describes a test for determining whether a simply-connected region, formed as the union of unit squares in the plane, has a domino tiling. He forms an undirected graph that has as its vertices the points (x,y,z) in the three-dimensional integer lattice, where each such point is connected to four neighbors: if x + y is even, then (x,y,z) is connected to (x + 1,y,z + 1), (x − 1,y,z + 1), (x,y + 1,z − 1), and (x,y − 1,z − 1), while if x + y is odd, then (x,y,z) is connected to (x + 1,y,z − 1), (x − 1,y,z − 1), (x,y + 1,z + 1), and (x,y − 1,z + 1). The boundary of the region, viewed as a sequence of integer points in the (x,y) plane, lifts uniquely (once a starting height is chosen) to a path in this three-dimensional graph. A necessary condition for this region to be tileable is that this path must close up to form a simple closed curve in three dimensions, however, this condition is not sufficient. Using more careful analysis of the boundary path, Thurston gave a criterion for tileability of a region that was sufficient as well as necessary.
Counting tilin |
https://en.wikipedia.org/wiki/Roundup%20%28issue%20tracker%29 | Roundup is an open-source issue or bug tracking system featuring a command-line, web and e-mail interface. It is written in Python and designed to be highly customizable.
History
Roundup was designed by Ka-Ping Yee for the Software Carpentry project and was developed from 2001 to 2016 under the direction of Richard Jones. Since then, it has been developed by the Roundup community. It was the issue tracker for the Python programming language for 17 years before migrating to GitHub. It was once described as "like Bugzilla without the six years of training, or RT without that tedious MySQL rubbish."
Features
The standard configuration of Roundup features:
a web interface for viewing, editing and searching issues
REST and XMLRPC interfaces for remote automation and web applications
a Mail gateway allowing creation and changing of issues
a database abstraction layer, currently supporting (among others) Python's built-in "anydbm" module, PostgreSQL, MySQL and SQLite
issue-specific "nosy lists", used for e-mail notifications and conversation (each issue effectively becoming a mini mailing list)
an authorization system, based on roles (of users), classes and objects
an interactive shell for backup and restore tasks and for manipulation of objects
Roundup supports several web backends. It can be run standalone, as a background daemon process, as a CGI script or as WSGI application.
Concepts
Roundup is customized by changing the contents of the tracker instance directory:
Database schema
The database schema is defined in a Python file in the tracker instance's root directory; it is
re-read whenever the server is started anew. When changes are found (e.g. new attributes), the tables of the underlying RDBS are altered accordingly.
Page templates
Roundup uses the Template Attribute Language (TAL) to create HTML or XHTML output. Version 1.5.0 adds experimental support for alternative template engines, such as Jinja2.
Templates are named after the classes in d |
https://en.wikipedia.org/wiki/Hydrogenase%20mimic | A hydrogenase mimic or bio-mimetic is an enzyme mimic of hydrogenases.
Bio-mimetic compounds inspired in hydrogenases
One of the more interesting applications of hydrogenases is to produce hydrogen, due its capacity to catalyze its redox reaction:
In the field of hydrogen production, the incorporation of chemical compounds in electrochemical devices to produce molecular hydrogen has been a topic of huge interest in the recent years due to the possibility of using hydrogen as a replacement of the fossil fuels as an energetic carrier. This approach of using materials inspired by natural models to do the same function as their natural counterparts is called bio-mimetic approach. Nowadays this approach has received a big impulse due to the availability of high-resolution crystal structures of several hydrogenases obtained with different techniques. The technical details of these hydrogenases are stored in electronic databases at disposition to who may be interested.
This information has allowed to determine the important parts of the enzyme necessary to catalyze the reaction and determine the pathway of the reaction in a very detailed way. Which allow to have a very good comprehension of what is necessary to catalyze the same reaction using artificial components.
Examples of bio-mimetic compounds inspired in hydrogenase
Several studies have demonstrated the possibility to develop chemical cells inspired by biological models to produce molecular hydrogen, for example: Selvaggi et al. explored the possibility to use energy captured by the PSII, developing for that goal, an organic-inorganic hybrid system replacing the PSII protein complex by microspheres of TiO2 a photo-inducible compound. In order to get the hydrogen production, the TiO2 microspheres were covered with hydrogenases extracted from the marine thermophile Pyrococcus furiosus, in that way the energy of the light was captured by the TiO2 microspheres and used to generate protons and electrons from wate |
https://en.wikipedia.org/wiki/Underwater%20acoustics | Underwater acoustics (also known as hydroacoustics) is the study of the propagation of sound in water and the interaction of the mechanical waves that constitute sound with the water, its contents and its boundaries. The water may be in the ocean, a lake, a river or a tank. Typical frequencies associated with underwater acoustics are between 10 Hz and 1 MHz. The propagation of sound in the ocean at frequencies lower than 10 Hz is usually not possible without penetrating deep into the seabed, whereas frequencies above 1 MHz are rarely used because they are absorbed very quickly.
Hydroacoustics, using sonar technology, is most commonly used for monitoring of underwater physical and biological characteristics. Hydroacoustics can be used to detect the depth of a water body (bathymetry), as well as the presence or absence, abundance, distribution, size, and behavior of underwater plants and animals. Hydroacoustic sensing involves "passive acoustics" (listening for sounds) or active acoustics making a sound and listening for the echo, hence the common name for the device, echo sounder or echosounder.
There are a number of different causes of noise from shipping. These can be subdivided into those caused by the propeller, those caused by machinery, and those caused by the movement of the hull through the water. The relative importance of these three different categories will depend, amongst other things, on the ship type
One of the main causes of hydro acoustic noise from fully submerged lifting surfaces is the unsteady separated turbulent flow near the surface's trailing edge that produces pressure fluctuations on the surface and unsteady oscillatory flow in the near wake. The relative motion between the surface and the ocean creates a turbulent boundary layer (TBL) that surrounds the surface. The noise is generated by the fluctuating velocity and pressure fields within this TBL.
The field of underwater acoustics is closely related to a number of other fields of a |
https://en.wikipedia.org/wiki/Ordinal%20numerical%20competence | In human developmental psychology or non-human primate experiments, ordinal numerical competence or ordinal numerical knowledge is the ability to count objects in order and to understand the greater than and less than relationships between numbers. It has been shown that children as young as two can make some ordinal numerical decisions. There are studies indicating that some non-human primates, like chimpanzees and rhesus monkeys have some ordinal numerical competence.
In humans
Prenatal
There is no evidence to support prenatal ordinal numerical competence. Teratogens such as stress can alter prenatal neural development, leading to diminished competence after birth. Physical effects of teratogens are common, but endocrine effects are harder to measure. These are the factors that influence neural development and by extension the development of ordinal numerical competence. Premature birth is also a risk factor for developmental problems including reduced brain activity. Brain activity is measured from outside the body with electroencephalography.
Infants
There have been a vast number of studies done on infants and their knowledge of numbers. Most research confirms that infants do in fact have a profound innate sense of number, both in abstract and finite ways. Infants as young as 49 hours can accurately match up images with a certain number of objects, with sounds that contain the same number ("ra, ra, ra, ra") as the number of objects in the image. Because the sounds are abstract, or visibly there, we can see that infants as young as 49 hours have some abstract numerical sense as well as concrete numerical sense shown by their recognition of the image with the corresponding number of objects. Similarly, infants around the age of 7 months can also match up images of random objects.
Although children as young as 49 hours can match up the number of sounds with the number of objects, they can only do so at certain ratios. When 1:3 ratios were used (4 so |
https://en.wikipedia.org/wiki/State%20of%20charge | State of charge (SoC) is the level of charge of an electric battery relative to its capacity. SoC is usually expressed as percentage (0% = empty; 100% = full). An alternative form of the same measure is the depth of discharge (DoD), calculated as 100 - SoC (100% = empty; 0% = full). SoC is normally used when discussing the current state of a battery in use, while DoD is most often seen when discussing the lifetime of the battery after repeated use.
In electric vehicles
In a battery electric vehicle (BEV), SoC for the battery pack is the equivalent of a fuel gauge.
It is important to mention that state of charge, presented as a gauge or percentage value on any vehicle dashboard, especially in plug-in hybrid vehicles, may not be representative of a real level of charge. In that particular case, some noticeable amount of energy stored in the electric battery is not shown on the dashboard, and is reserved for hybrid-work operations. It permits a vehicle to accelerate with electric motors mainly using battery energy, while the petrol engine serves as a generator and recharges the battery to the minimum level needed for such operation. Examples of such cars are Mitsubishi Outlander PHEV (all versions/years of production), where 0% of the state of charge presented to the driver is a real 20-22% of charge level (assuming zero level as the lowest level of charge permitted by car producer). Another one is BMW i3 REX (Range Extender version), where about 6% of SOC is reserved for PHEV-alike operations. Tesla has stated that their SoC should be less than 95%, with some commentators saying between 30%-80% . There is some data to back this up as well.
The state of charge (SOC) can help to reduce electrical car's owners' anxiety when they are waiting in the line or stay at home since it will reflect the progress of charging and let owners know when it will be ready.
Determining SoC
Usually, SoC cannot be measured directly but it can be estimated from direct measurement varia |
https://en.wikipedia.org/wiki/Metolachlor | Metolachlor is an organic compound that is widely used as an herbicide. It is a derivative of aniline and is a member of the chloroacetanilide family of herbicides. It is highly effective toward grasses.
Agricultural use
Metolachlor was developed by Ciba-Geigy. Its acts by inhibition of elongases and of the geranylgeranyl pyrophosphate (GGPP) cyclases, which are part of the gibberellin pathway. It is used for grass and broadleaf weed control in corn, soybean, peanuts, sorghum, and cotton. It is also used in combination with other herbicides.
Metolachlor is a popular herbicide in the United States. As originally formulated metolachlor was applied as a racemate, a 1:1 mixture of the (S)- and (R)-stereoisomers. The (R)-enantiomer is less active, and modern production methods afford a higher concentration of S-metolachlor, thus current application rates are far lower than original formulations.
Production and basic structure
Metolachlor is produced from 2-ethyl-6-methylaniline (MEA) via condensation with methoxyacetone. The resulting imine is hydrogenated to give primarily the S-stereoisomeric amine. This secondary amine is acetylated with chloroacetylchloride. Because of the steric effects of the 2,6-disubstituted aniline, rotation about the aryl-C to N bond is restricted. Thus, both the (R)- and the (S)-enantiomers exist as atropisomers. Both atropisomers of (S)-metolachlor exhibit the same biological activity.
Safety and ecological effects
Metolachlor has been detected in ground and surface waters in concentrations ranging from 0.08 to 4.5 parts per billion (ppb) throughout the U.S. It is classified as a Category C pesticide by the United States Environmental Protection Agency (US EPA), which indicates limited evidence of carcinogenicity. Evidence of the bioaccumulation of metolachlor in edible species of fish as well as its adverse effect on the growth and development raise concerns on its effects on human health. Though there is no set maximum concen |
https://en.wikipedia.org/wiki/Layer%20cake%20representation | In mathematics, the layer cake representation of a non-negative, real-valued measurable function defined on a measure space is the formula
for all , where denotes the indicator function of a subset and denotes the super-level set
The layer cake representation follows easily from observing that
and then using the formula
The layer cake representation takes its name from the representation of the value as the sum of contributions from the "layers" : "layers"/values below contribute to the integral, while values above do not.
It is a generalization of Cavalieri's principle and is also known under this name.
An important consequence of the layer cake representation is the identity
which follows from it by applying the Fubini-Tonelli theorem.
An important application is that for can be written as follows
which follows immediately from the change of variables in the layer cake representation of .
See also
Symmetric decreasing rearrangement |
https://en.wikipedia.org/wiki/Pr%C3%A9kopa%E2%80%93Leindler%20inequality | In mathematics, the Prékopa–Leindler inequality is an integral inequality closely related to the reverse Young's inequality, the Brunn–Minkowski inequality and a number of other important and classical inequalities in analysis. The result is named after the Hungarian mathematicians András Prékopa and László Leindler.
Statement of the inequality
Let 0 < λ < 1 and let f, g, h : Rn → [0, +∞) be non-negative real-valued measurable functions defined on n-dimensional Euclidean space Rn. Suppose that these functions satisfy
for all x and y in Rn. Then
Essential form of the inequality
Recall that the essential supremum of a measurable function f : Rn → R is defined by
This notation allows the following essential form of the Prékopa–Leindler inequality: let 0 < λ < 1 and let f, g ∈ L1(Rn; [0, +∞)) be non-negative absolutely integrable functions. Let
Then s is measurable and
The essential supremum form was given by Herm Brascamp and Elliott Lieb. Its use can change the left side of the inequality. For example, a function g that takes the value 1 at exactly one point will not usually yield a zero left side in the "non-essential sup" form but it will always yield a zero left side in the "essential sup" form.
Relationship to the Brunn–Minkowski inequality
It can be shown that the usual Prékopa–Leindler inequality implies the Brunn–Minkowski inequality in the following form: if 0 < λ < 1 and A and B are bounded, measurable subsets of Rn such that the Minkowski sum (1 − λ)A + λB is also measurable, then
where μ denotes n-dimensional Lebesgue measure. Hence, the Prékopa–Leindler inequality can also be used to prove the Brunn–Minkowski inequality in its more familiar form: if 0 < λ < 1 and A and B are non-empty, bounded, measurable subsets of Rn such that (1 − λ)A + λB is also measurable, then
Applications in probability and statistics
Log-concave distributions
The Prékopa–Leindler inequality is useful in the theory of log-concave distributions, as it can be used to show t |
https://en.wikipedia.org/wiki/Florin%20sign | The florin sign (ƒ) is a symbol that is used for the currencies named florin, also called guilder. The Dutch name for the currency is gulden. The symbol "ƒ" is the lowercase version of Ƒ of the Latin alphabet. In many serif typefaces, it can often be substituted with a normal italic small-letter f ( ).
It is used in the following current and obsolete currencies (between brackets their ISO 4217 currency codes):
Current:
Aruban florin (AWG)
Netherlands Antillean guilder (ANG)
Obsolete:
Dutch guilder (NLG; until 2002)
Netherlands Indies guilder (until 1954)
Surinamese guilder (SRG; until 2004)
Italian florin (until 1533)
Currency symbols |
https://en.wikipedia.org/wiki/Biological%20network%20inference | Biological network inference is the process of making inferences and predictions about biological networks. By using these networks to analyze patterns in biological systems, such as food-webs, we can visualize the nature and strength of these interactions between species, DNA, proteins, and more.
The analysis of biological networks with respect to diseases has led to the development of the field of network medicine. Recent examples of application of network theory in biology include applications to understanding the cell cycle as well as a quantitative framework for developmental processes. Good network inference requires proper planning and execution of an experiment, thereby ensuring quality data acquisition. Optimal experimental design in principle refers to the use of statistical and or mathematical concepts to plan for data acquisition. This must be done in such a way that the data information content is enriched, and a sufficient amount of data is collected with enough technical and biological replicates where necessary.
Steps
The general cycle to modeling biological networks is as follows:
Prior knowledge
Involves a thorough literature and database search or seeking an expert's opinion.
Model selection
A formalism to model your system, usually an ordinary differential equation, boolean network, or Linear regression models, e.g. Least-angle regression, by Bayesian network or based on Information theory approaches. it can also be done by the application of a correlation-based inference algorithm, as will be discussed below, an approach which is having increased success as the size of the available microarray sets keeps increasing
Hypothesis/assumptions
Experimental design
Data acquisition
Ensure that high quality data is collected with all the required variables being measured
Network inference
This process is mathematical rigorous and computationally costly.
Model refinement
Cross-check how well the results meet the expectations. The process |
https://en.wikipedia.org/wiki/Stem%20mixing%20and%20mastering | Stem-mixing is a method of mixing audio material based on creating groups of audio tracks and processing them separately prior to combining them into a final master mix. Stems are also sometimes referred to as submixes, subgroups, or buses.
The distinction between a stem and a separation is rather unclear. Some consider stem manipulation to be the same as separation mastering, although others consider stems to be sub-mixes to be used along with separation mastering. It depends on how many separate channels of input are available for mixing and/or at which stage they are on the way towards reducing them to a final stereo mix.
The technique originated in the 1960s, with the introduction of mixing boards equipped with the capability to assign individual inputs to sub-group faders and to work with each sub-group (stem mix) independently from the others. The approach is widely used in recording studios to control, process and manipulate entire groups of instruments such as drums, strings, or backup vocals, in order to streamline and simplify the mixing process. Additionally, as each stem-bus usually has its own inserts, sends and returns, the stem-mix (sub-mix) can be routed independently through its own signal processing chain, to achieve a different effect for each group of instruments. A similar method is also utilised with digital audio workstations (DAWs), where separate groups of audio tracks may be digitally processed and manipulated through discrete chains of plugins.
Stem-mastering is a technique derived from stem mixing. Just as in stem-mixing, the individual audio tracks are grouped together, to allow for independent control and signal processing of each stem, and can be manipulated independently from each other. Most of the mastering engineers require music producers to have at least -3db headroom at each individual track before starting stem mastering process. The reason for this is to leave more space in the mix to make the mastered version sound cleaner |
https://en.wikipedia.org/wiki/Calendrical%20calculation | A calendrical calculation is a calculation concerning calendar dates. Calendrical calculations can be considered an area of applied mathematics.
Some examples of calendrical calculations:
Converting a Julian or Gregorian calendar date to its Julian day number and vice versa .
The number of days between two dates, which is simply the difference in their Julian day numbers.
The dates of moveable holidays, like Christian Easter (the calculation is known as Computus) followed up by Ascension Thursday and Pentecost or Advent Sundays, or the Jewish Passover, for a given year.
Converting a date between different calendars. For instance, dates in the Gregorian calendar can be converted to dates in the Islamic calendar with the Kuwaiti algorithm.
Calculating the day of the week.
Calendrical calculation is one of the five major Savant syndrome characteristics.
Examples
Numerical methods were described in the Journal of the Department of Mathematics, Open University, Milton Keynes, Buckinghamshire (M500) in 1997 and 1998. The following algorithm gives the number of days (d) in month m of year y. The value of m is given on the right of the month in the following list:
January 11 February 12 March 1 April 2 May 3 June 4 July 5 August 6 September 7 October 8 November 9 December 10.
The algorithm enables a computer to print calendar and diary pages for past or future sequences of any desired length from the reform of the calendar, which in England was 3/14 September 1752. The article Date of Easter gives algorithms for calculating the date of Easter. Combining the two enables the page headers to show any fixed or movable festival observed on the day, and whether it is a bank holiday.
The algorithm utilises the integral or floor function: thus is that part of the number x which lies to the left of the decimal point. It is only necessary to work through the complete function when calculating the length of February in a year which is divisible by 100 witho |
https://en.wikipedia.org/wiki/List%20of%20Portuguese%20flags | This is a list of flags used in Portugal.
National flag
Autonomous regions
Municipalities
Government flags
Military flags
Army staff
Navy staff
Historical flags
County of Portugal
Kingdom of Portugal
Portuguese Macau
Military Flags
Merchant marine flags
Navy flags
Royal banners
Governmental flags
Chartered trading company flags
Colonial authorities flag
Proposed flags
Overseas provinces
Capitals of former overseas territories
Political flags
Ethnic groups flags
See also
A Portuguesa
Coat of arms of Portugal
List of personal standards of the Kings of Portugal
List of Portuguese municipal flags
List of Macanese flags
Portuguese vexillology
External links
Portugal
Flags of Portugal
Flags |
https://en.wikipedia.org/wiki/K-mer | In bioinformatics, k-mers are substrings of length contained within a biological sequence. Primarily used within the context of computational genomics and sequence analysis, in which k-mers are composed of nucleotides (i.e. A, T, G, and C), k-mers are capitalized upon to assemble DNA sequences, improve heterologous gene expression, identify species in metagenomic samples, and create attenuated vaccines. Usually, the term k-mer refers to all of a sequence's subsequences of length , such that the sequence AGAT would have four monomers (A, G, A, and T), three 2-mers (AG, GA, AT), two 3-mers (AGA and GAT) and one 4-mer (AGAT). More generally, a sequence of length will have k-mers and total possible k-mers, where is number of possible monomers (e.g. four in the case of DNA).
Introduction
k-mers are simply length subsequences. For example, all the possible k-mers of a DNA sequence are shown below:
A method of visualizing k-mers, the k-mer spectrum, shows the multiplicity of each k-mer in a sequence versus the number of k-mers with that multiplicity. The number of modes in a k-mer spectrum for a species's genome varies, with most species having a unimodal distribution. However, all mammals have a multimodal distribution. The number of modes within a k-mer spectrum can vary between regions of genomes as well: humans have unimodal k-mer spectra in 5' UTRs and exons but multimodal spectra in 3' UTRs and introns.
Forces Affecting DNA k-mer Frequency
The frequency of k-mer usage is affected by numerous forces, working at multiple levels, which are often in conflict. It is important to note that k-mers for higher values of k are affected by the forces affecting lower values of k as well. For example, if the 1-mer A does not occur in a sequence, none of the 2-mers containing A (AA, AT, AG, and AC) will occur either, thereby linking the effects of the different forces.
k = 1
When k = 1, there are four DNA k-mers, i.e., A, T, G, and C. At the molecular level, there |
https://en.wikipedia.org/wiki/Supnick%20matrix | A Supnick matrix or Supnick array – named after Fred Supnick of the City College of New York, who introduced the notion in 1957 – is a Monge array which is also a symmetric matrix.
Mathematical definition
A Supnick matrix is a square Monge array that is symmetric around the main diagonal.
An n-by-n matrix is a Supnick matrix if, for all i, j, k, l such that if
and
then
and also
A logically equivalent definition is given by Rudolf & Woeginger who in 1995 proved that
A matrix is a Supnick matrix iff it can be written as the sum of a sum matrix S and a non-negative linear combination of LL-UR block matrices.
The sum matrix is defined in terms of a sequence of n real numbers {αi}:
and an LL-UR block matrix consists of two symmetrically placed rectangles in the lower-left and upper right corners for which aij = 1, with all the rest of the matrix elements equal to zero.
Properties
Adding two Supnick matrices together will result in a new Supnick matrix (Deineko and Woeginger 2006).
Multiplying a Supnick matrix by a non-negative real number produces a new Supnick matrix (Deineko and Woeginger 2006).
If the distance matrix in a traveling salesman problem can be written as a Supnick matrix, that particular instance of the problem admits an easy solution (even though the problem is, in general, NP hard). |
https://en.wikipedia.org/wiki/Digital%20journalism | Digital journalism, also known as netizen journalism or online journalism, is a contemporary form of journalism where editorial content is distributed via the Internet, as opposed to publishing via print or broadcast. What constitutes digital journalism is debated by scholars; however, the primary product of journalism, which is news and features on current affairs, is presented solely or in combination as text, audio, video, or some interactive forms like storytelling stories or newsgames, and disseminated through digital media technology.
Fewer barriers to entry, lowered distribution costs, and diverse computer networking technologies have led to the widespread practice of digital journalism. It has democratized the flow of information that was previously controlled by traditional media including newspapers, magazines, radio, and television.
Some have asserted that a greater degree of creativity can be exercised with digital journalism when compared to traditional journalism and traditional media. The digital aspect may be central to the journalistic message and remains, to some extent, within the creative control of the writer, editor, and/or publisher.
It has been acknowledged that reports of its growth have tended to be exaggerated. In fact, a 2019 Pew survey showed a 16% decline in the time spent on online news sites since 2016.
Overview
Digital journalism flows as journalism flows and is difficult to pinpoint where it is and where it is going. In partnership with digital media, digital journalism uses facets of digital media to perform journalist tasks, for example, using the internet as a tool rather than a singular form of digital media. There is no absolute agreement as to what constitutes digital journalism. Mu Lin argues that, “Web and mobile platforms demand us to adopt a platform-free mindset for an all-inclusive production approach – create the [digital] contents first, then distribute via appropriate platforms." The repurposing of print content |
https://en.wikipedia.org/wiki/65%2C536 | 65536 is the natural number following 65535 and preceding 65537.
65536 is a power of two: (2 to the 16th power).
65536 is the smallest number with exactly 17 divisors.
In mathematics
65536 is , so in tetration notation 65536 is 42.
When expressed using Knuth's up-arrow notation, 65536 is
,
which is equal to
,
which is equivalent to
or
.
65536 is a superperfect number – a number such that σ(σ(n)) = 2n.
A 16-bit number can distinguish 65536 different possibilities. For example, unsigned binary notation exhausts all possible 16-bit codes in uniquely identifying the numbers 0 to 65535. In this scheme, 65536 is the least natural number that can not be represented with 16 bits. Conversely, it is the "first" or smallest positive integer that requires 17 bits.
65536 is the only power of 2 less than 231000 that does not contain the digits 1, 2, 4, or 8 in its decimal representation.
The sum of the unitary divisors of 65536 is prime (1 + 65536 = 65537, which is prime).
65536 is an untouchable number.
In computing
65536 (216) is the number of different values representable in a number of 16 binary digits (or bits), also known as an unsigned short integer in many computer programming systems.
A 65,536-bit integer can represent up to 265536 (2.00352993...) values.
65,536 is the number of characters in the original Unicode, and currently in a Unicode plane.
This number is a limit in many common hardware and software implementations, some examples of which are:
The Motorola 68000 family, x86 architecture, and other computing platforms have a word size of 16 bits, representing 65536 possible values. (32- or 64-bit operations are supported equally or better in modern microprocessors.)
Many modern CPUs allow a memory page size of 64KiB (65536 bytes) to be configured in their memory-management hardware.
The default buffer size of a pipeline (Unix) is 64KiB (65536 bytes).
65536 is the maximum number of spreadsheet rows supported by Excel 97, Excel 2000, Excel 2002 |
https://en.wikipedia.org/wiki/Nun%20Study | The Nun Study of Aging and Alzheimer's Disease is a continuing longitudinal study, begun in 1986, to examine the onset of Alzheimer's disease. David Snowdon, an Epidemiologist and the founding Nun Study investigator, started the Nun Study at the University of Minnesota, later transferring the study to the University of Kentucky in 1990. In 2008, with Snowdon's retirement, the study returned to the University of Minnesota. The Nun Study was very briefly moved from the University of Minnesota to Northwestern University in 2021 under the directorship of Dr. Margaret Flanagan. The Nun Study is currently housed at the University of Texas Health San Antonio in the Bigg's Institute for Alzheimer's and Neurodegenerative diseases under the continued directorship of Neuropathologist, Dr. Margaret Flanagan.
The Sisters' autobiographies written just before they took their vows (ages 19–21) revealed that positivity was closely related to longevity and idea density, which is related to conversation and writing. This research found that higher idea density scores correlated with a higher chance of having sufficient mental capacity in late-life, despite neurological evidence that showed the onset of Alzheimer's disease.
In 1992, researchers at Rush University Medical Center Rush Alzheimer's Disease Center (RADC), building on the success of the Nun Study, proposed the Rush Religious Orders Study. The Religious Orders Study was funded by the National Institute on Aging in 1993, and was ongoing, as of 2012.
Origin and procedure
The Nun Study began in 1986 with funding by the National Institute on Aging. This study was focused on a group of 678 American Roman Catholic Sisters who were members of the School Sisters of Notre Dame. The purpose of the study was to conclude if activities, academics, past experiences, and disposition are correlated to continued cognitive, neurological, and physical ability as individuals got older, as well as overall longevity. The Nun Study participants |
https://en.wikipedia.org/wiki/Maximum%20subarray%20problem | In computer science, the maximum sum subarray problem, also known as the maximum segment sum problem, is the task of finding a contiguous subarray with the largest sum, within a given one-dimensional array A[1...n] of numbers. It can be solved in time and space.
Formally, the task is to find indices and with , such that the sum
is as large as possible. (Some formulations of the problem also allow the empty subarray to be considered; by convention, the sum of all values of the empty subarray is zero.) Each number in the input array A could be positive, negative, or zero.
For example, for the array of values [−2, 1, −3, 4, −1, 2, 1, −5, 4], the contiguous subarray with the largest sum is [4, −1, 2, 1], with sum 6.
Some properties of this problem are:
If the array contains all non-negative numbers, then the problem is trivial; a maximum subarray is the entire array.
If the array contains all non-positive numbers, then a solution is any subarray of size 1 containing the maximal value of the array (or the empty subarray, if it is permitted).
Several different sub-arrays may have the same maximum sum.
Although this problem can be solved using several different algorithmic techniques, including brute force, divide and conquer, dynamic programming, and reduction to shortest paths, a simple single-pass algorithm known as Kadane's algorithm solves it efficiently.
History
The maximum subarray problem was proposed by Ulf Grenander in 1977 as a simplified model for maximum likelihood estimation of patterns in digitized images.
Grenander was looking to find a rectangular subarray with maximum sum, in a two-dimensional array of real numbers. A brute-force algorithm for the two-dimensional problem runs in O(n6) time; because this was prohibitively slow, Grenander proposed the one-dimensional problem to gain insight into its structure. Grenander derived an algorithm that solves the one-dimensional problem in O(n2) time,
improving the brute force running time of O |
https://en.wikipedia.org/wiki/English%20Longitudinal%20Study%20of%20Ageing | The English Longitudinal Study of Ageing (ELSA) is a longitudinal study that collects multidisciplinary data from a representative sample of the English population aged 50 and older to look at all aspects of aging in England.
The study started in 2002 and there are currently nine waves of completed data and a tenth wave is currently being collected. The survey data are designed to be used for the investigation of a broad set of topics relevant to understanding the ageing process. Both objective and subjective data are collected covering themes such as health trajectories, disability and healthy life expectancy, the determinants of economic position in older age; the links between economic position, physical health, cognition and mental health; the nature and timing of retirement and post-retirement, labour market activity; household and family structure, social networks and social supports; patterns, determinants and consequences of social, civic and cultural participation and predictors of well-being.
ELSA is led by Professor Andrew Steptoe and is jointly run by teams at University College London (UCL), the Institute for Fiscal Studies (IFS), National Centre for Social Research, University of Manchester and the University of East Anglia.
Funding
ELSA is funded jointly by the National Institute on Aging in the US and a consortium of UK government departments: Department of Health and Social Care; Department for Work and Pensions; and Department for Transport coordinated by the National Institute for Health and Care Research (NIHR). Funding has also been provided by the Economic and Social Research Council.
Study design and data collection
The first wave of ELSA achieved a sample comprising 11,050 respondents aged 50 and over on 1 March 2002. Sample members are drawn from respondents to the Health Survey For England (HSE) and the initial data collected for that survey are subsequently linked to the ongoing ELSA measurements. For waves 3, 4, 6, 7 and 9 refreshment |
https://en.wikipedia.org/wiki/IC%20layout%20editor | An Integrated circuit layout editor or IC layout editor is an electronic design automation software tool that allows a user to digitize the shapes and patterns that form an integrated circuit. Typically the view will include the components (usually as pcells), metal routing tracks, vias and electrical pins. Software of this type is similar to computer aided drafting software, but is specialized for the task of integrated circuit layout. The typical flow for the layout of analog circuits might be :
1. The layout engineer receives the schematic from the designer in electrical form
2. Either the tool or the layout engineer creates a physical view of the circuit including all of the required components, wires, layers and pads.
3. The layout engineer positions the components to minimize both the area required and the negative effects of layout parasitics upon the circuit performance and also to allow efficient routing to components.
4. The layout engineer uses metal routing and other layers at times to connect all of the components, again taking care to avoid unwanted layout parasitics.
5. The layout engineer uses DRC and LVS checks to ensure that the circuit is both manufacturable and functional. Other tools in include field solver verification to check for important specs such as device resistance and sources of problems such as electromigration or too thin wires resulting in burn up of wires causing shorts or open circuits.
6. Other checks include ESD, XOR, EOS and verification with the foundry called Mebes check to ensure the boolean algorithms that generate the mask layers are done as intended. Boolean generation is quite often done in the layout editor.
Layout used to be done on sticks and yards of strings for very basic components. The advent of computers particularly mainframes and mini computers helped bring layout to the digital world of computers. In the 80's and 90's quite a bit of layout editing was done on personal pc's using such tools as IC Editors |
https://en.wikipedia.org/wiki/Niko%20Bellic | Niko Bellic is a fictional character and the playable protagonist of Rockstar North's 2008 video game Grand Theft Auto IV, the sixth main instalment in Rockstar Games' Grand Theft Auto series. He also makes non-playable appearances in the game's episodic content The Lost and Damned and The Ballad of Gay Tony, both released in 2009. Michael Hollick provided the character's voice and motion capture.
Within the game's storyline, Niko is an ex-soldier from Eastern Europe, who was shaped by his experiences in an unidentified war he fought in, developing a very cynical view on life. After becoming involved with a Russian crime syndicate, and discovering that his unit was sold out to enemy forces, he decides to move to Liberty City to pursue the American Dream, inspired by his cousin Roman’s personal tales of luxury and riches that he had experienced while living there for the past decade. However, upon his arrival, he discovers that those stories were greatly exaggerated, and attempts to improve his and Roman's financial situation by becoming involved with the local criminal underworld. As the game's story progresses, Niko works for various prominent fictional crime figures, in the hopes of finding the traitor who betrayed his unit during the war, while slowly learning to let go of his past and quest for revenge, and attempting to leave the criminal life that comes with major risks.
Niko's character received critical acclaim for his maturity, moral ambiguity, and personal growth, and has been called one of the best protagonists in the series. For his role, Hollick won Best Performance by a Human Male at the 2008 Spike Video Game Awards.
Nationality
Niko's nationality is unspecified in the game and is subject to debate. It was believed by some that he was Russian, Serbian, or Croatian. Executive producer Sam Houser spoke on the matter, saying that Niko is "from that grey part of broken-down Eastern Europe", suggesting that Niko's nationality was left intentionally vague |
https://en.wikipedia.org/wiki/Iridomyrmex | Iridomyrmex is a genus of ants called rainbow ants (referring to their blue-green iridescent sheen) first described by Austrian entomologist Gustav Mayr in 1862. He placed the genus in the subfamily Dolichoderinae of the family Formicidae. It has 79 described species and five fossil species. Most of these ants are native to Australia; others are found in Asia and Oceania, and they have been introduced to Brazil, New Zealand, and the United Arab Emirates. Fossil species are known from China, France, and the United States.
These ants are known to be an ecologically dominant and important group of ants, but they are sometimes regarded as pests because they disturb soil and enter human houses. Farmers in rural Australia place animal carcasses on meat ant (I. purpureus) mounds as a method of disposing of them; meat ants consume the carcass and reduce it to bones in a matter of weeks. Meat ants also engage in ritualised fighting, which helps prevent casualties and solve territorial disputes between neighbouring colonies. The largest members of this genus are those of the I. purpureus species group, measuring .
After their nuptial flight, queen ants may establish a colony by themselves, by budding, or cooperatively, where a subset of the colony migrates to a new location or when multiple queens help find a suitable nesting spot, but they display intolerance to each other when workers are present. The eggs take 44 to 61 days to fully develop into adults. Ants of the genus live in a wide variety of habitats and nest in soil in numbers that range from a few hundred individuals to over 300,000 in a single colony. Depending on the species, nests are large mounds covered in pebbles with multiple entrances, while others live above ground in twig nests. In some cases, ants dwell in several nest sites connected by paths; some of these nests can extend to in length. Some species associate with caterpillars and butterflies that provide the ants with secretions and honeydew, and |
https://en.wikipedia.org/wiki/SitePal | SitePal is a speaking avatar platform for small and medium-sized businesses developed by Oddcast.
SitePal allows users to deploy "virtual employees" on websites that can welcome visitors, guide them around the site and answer questions.
The use of SitePal on commercial websites has been controversial because many visitors report finding them annoying. Some research has shown that they can increase sales in comparison to using static photographs.
Development
The technology used was the result of more than 4 years of research at Stanford University. The research was based on a literature review and other previous work in the field of artificial intelligence research.
The SitePal AI option uses the AIML programming language, which is partially editable by users. This allows web designers to simulate normal human conversation by using keywords or key phrases that the bot can respond to.
Features
The company provides web designers with options to customize the chosen avatar. A large selection of faces, clothing, hair, backgrounds, voices and other details are available. If a web designer wants to use a particular face, Sitepal can create one from a photo. Thus, a mascot or a known face can be simulated.
Speech
Sitepal avatars talk through text-to-speech (tts) software.
A short paragraph can be written (up to 900 characters) and the text-to-speech engine will compile the actual speech, which can be reproduced and edited. The tts engine is not perfect, but it comes close to actual speech and is easy to understand. Tts can be further enhanced by some commands, like /laugh and /loud which make the avatar laugh or talk loud. Even pronunciation is possible.
The web designer can record and upload his or her own audio messages. Alternatively Sitepal offers professional voice acting service at extra cost.
User interaction
The company provides 5 options for visitor interaction:
No interaction. The avatar simply says a pre-fixed message.
FAQ mode. Questions can be confi |
https://en.wikipedia.org/wiki/Rapid%20prototyping | Rapid prototyping is a group of techniques used to quickly fabricate a scale model of a physical part or assembly using three-dimensional computer aided design (CAD) data.
Construction of the part or assembly is usually done using 3D printing or "additive layer manufacturing" technology.
The first methods for rapid prototyping became available in mid 1987 and were used to produce models and prototype parts. Today, they are used for a wide range of applications and are used to manufacture production-quality parts in relatively small numbers if desired without the typical unfavorable short-run economics. This economy has encouraged online service bureaus. Historical surveys of RP technology start with discussions of simulacra production techniques used by 19th-century sculptors. Some modern sculptors use the progeny technology to produce exhibitions and various objects. The ability to reproduce designs from a dataset has given rise to issues of rights, as it is now possible to interpolate volumetric data from 2D images.
As with CNC subtractive methods, the computer-aided-design – computer-aided manufacturing CAD -CAM workflow in the traditional rapid prototyping process starts with the creation of geometric data, either as a 3D solid using a CAD workstation, or 2D slices using a scanning device. For rapid prototyping this data must represent a valid geometric model; namely, one whose boundary surfaces enclose a finite volume, contain no holes exposing the interior, and do not fold back on themselves. In other words, the object must have an "inside". The model is valid if for each point in 3D space the computer can determine uniquely whether that point lies inside, on, or outside the boundary surface of the model. CAD post-processors will approximate the application vendors' internal CAD geometric forms (e.g., B-splines) with a simplified mathematical form, which in turn is expressed in a specified data format which is a common feature in additive manufacturing: STL |
https://en.wikipedia.org/wiki/Navigation%20function | Navigation function usually refers to a function of position, velocity, acceleration and time which is used to plan robot trajectories through the environment. Generally, the goal of a navigation function is to create feasible, safe paths that avoid obstacles while allowing a robot to move from its starting configuration to its goal configuration.
Potential functions as navigation functions
Potential functions assume that the environment or work space is known. Obstacles are assigned a high potential value, and the goal position is assigned a low potential. To reach the goal position, a robot only needs to follow the negative gradient of the surface.
We can formalize this concept mathematically as following: Let be the state space of all possible configurations of a robot. Let denote the goal region of the state space.
Then a potential function is called a (feasible) navigation function if
if and only if no point in is reachable from .
For every reachable state, , the local operator produces a state for which .
Probabilistic navigation function
Probabilistic navigation function is an extension of the classical navigation function for static stochastic scenarios. The function is defined by permitted collision probability, which limits the risk during motion. The Minkowski sum used for in the classical definition is replaced with a convolution of the geometries and the Probability Density Functionss of locations. Denoting the target position by , the Probabilistic navigation function is defined as:
where is a predefined constant like in the classical navigation function, which ensures the Morse nature of the function. is the distance to the target position , and takes into account all obstacles, defined as
where is based on the probability for a collision at location . The probability for a collision is limited by a predetermined value , meaning:
and,
where is the probability to collide with the i-th obstacle.
A map is said to be a |
https://en.wikipedia.org/wiki/Clascal | Clascal is an object-oriented programming language (and associated discontinued compiler) developed in 1983 by the Personal Office Systems (POS) division (later renamed The Lisa Division, then later The 32-Bit Systems Division) of Apple Computer. Clascal was used to program applications for the Lisa Office System, the operating environment of the Lisa.
Developed as an extension of Lisa Pascal, which in turn harked back to the UCSD Pascal model originally implemented on the Apple II, the language was strongly influenced by the Xerox Palo Alto Research Center (PARC) release of Smalltalk-80, v1 (which had been formerly ported to the Lisa), and by Modula. According to Larry Tesler, Clascal was developed as a replacement for Apple's version of Smalltalk, which was "too slow" and because the experience offered by the Smalltalk syntax was too unfamiliar for most people.
Clascal was the basis for Object Pascal on the Apple Macintosh in 1985. With the demise of the Lisa in 1986, Pascal and Object Pascal continued to be used in the Macintosh Programmer's Workshop for systems and application development for several more years, until it was finally supplanted by the languages C and C++. The MacApp application framework was based on Toolkit originally written in Clascal.
Object Pascal, in turn, served as the basis for Borland's Delphi. |
https://en.wikipedia.org/wiki/Bionic%20architecture | Bionic architecture is a contemporary movement that studies the physiological, behavioural, and structural adaptions of biological organisms as a source of inspiration for designing and constructing expressive buildings. These structures are designed to be self-sufficient, being able to structurally modify themselves in response to the fluctuating internal and external forces such as changes in weather and temperature.
Although this style of architecture has existed since the early 18th century period, the movement only began to mature in the early 21st century, following society's growing concerns over climate change and global warming. These influences led to bionic architecture being used to draw society away from its anthropocentric environment, by creating landscapes that allow for the harmonious relationship between nature and society. This is achieved through having an in-depth understanding of the complex interactions between form, material, and structure in order to ensure that the building's design supports a more sustainable environment. As a result, architects will rely upon the use of high-tech, artificial materials and techniques in order to conserve energy and materials, lower the consumption of construction and increase the practicality and reliability of their building structures.
History and theoretical framework
The word ‘bionic architecture’ is derived from the Greek word ‘bios’ (life) as well as the English word ‘technics’ (to study). The term was originally used to describe the scientific trend of ‘transferring technologies into life-forms’. The term ‘bionic’ was first used in 1958 by U.S army colonel, Jack E. Steele and Soviet scientist, Otto Schmitt during an astronomer project that focused on research surrounding the field of robotics. In their project, both researchers initially recognised the concept of bionics as ‘the science of systems based on living creatures’. The idea was then expanded upon in 1997 by Janine Benyus, who coined the |
https://en.wikipedia.org/wiki/Carbon%E2%80%93nitrogen%20bond | A carbon–nitrogen bond is a covalent bond between carbon and nitrogen and is one of the most abundant bonds in organic chemistry and biochemistry.
Nitrogen has five valence electrons and in simple amines it is trivalent, with the two remaining electrons forming a lone pair. Through that pair, nitrogen can form an additional bond to hydrogen making it tetravalent and with a positive charge in ammonium salts. Many nitrogen compounds can thus be potentially basic but its degree depends on the configuration: the nitrogen atom in amides is not basic due to delocalization of the lone pair into a double bond and in pyrrole the lone pair is part of an aromatic sextet.
Similar to carbon–carbon bonds, these bonds can form stable double bonds, as in imines; and triple bonds, such as nitriles. Bond lengths range from 147.9 pm for simple amines to 147.5 pm for C-N= compounds such as nitromethane to 135.2 pm for partial double bonds in pyridine to 115.8 pm for triple bonds as in nitriles.
A CN bond is strongly polarized towards nitrogen (the electronegativities of C and N are 2.55 and 3.04, respectively) and subsequently molecular dipole moments can be high: cyanamide 4.27 D, diazomethane 1.5 D, methyl azide 2.17, pyridine 2.19. For this reason many compounds containing CN bonds are water-soluble. N-philes are group of radical molecules which are specifically attracted to the C=N bonds.
Carbon-nitrogen bond can be analyzed by X-ray photoelectron spectroscopy (XPS). Depending on the bonding states the peak positions differ in N1s XPS spectra.
Nitrogen functional groups
See also
Cyanide
Other carbon bonds with group 15 elements: carbon–nitrogen bonds, carbon–phosphorus bonds
Other carbon bonds with period 2 elements: carbon–lithium bonds, carbon–beryllium bonds, carbon–boron bonds, carbon–carbon bonds, carbon–nitrogen bonds, carbon–oxygen bonds, carbon–fluorine bonds
Carbon–hydrogen bond |
https://en.wikipedia.org/wiki/State%E2%80%93action%E2%80%93reward%E2%80%93state%E2%80%93action | State–action–reward–state–action (SARSA) is an algorithm for learning a Markov decision process policy, used in the reinforcement learning area of machine learning. It was proposed by Rummery and Niranjan in a technical note with the name "Modified Connectionist Q-Learning" (MCQ-L). The alternative name SARSA, proposed by Rich Sutton, was only mentioned as a footnote.
This name reflects the fact that the main function for updating the Q-value depends on the current state of the agent "S1", the action the agent chooses "A1", the reward "R" the agent gets for choosing this action, the state "S2" that the agent enters after taking that action, and finally the next action "A2" the agent chooses in its new state. The acronym for the quintuple (st, at, rt, st+1, at+1) is SARSA. Some authors use a slightly different convention and write the quintuple (st, at, rt+1, st+1, at+1), depending on which time step the reward is formally assigned. The rest of the article uses the former convention.
Algorithm
A SARSA agent interacts with the environment and updates the policy based on actions taken, hence this is known as an on-policy learning algorithm. The Q value for a state-action is updated by an error, adjusted by the learning rate alpha. Q values represent the possible reward received in the next time step for taking action a in state s, plus the discounted future reward received from the next state-action observation.
Watkin's Q-learning updates an estimate of the optimal state-action value function based on the maximum reward of available actions. While SARSA learns the Q values associated with taking the policy it follows itself, Watkin's Q-learning learns the Q values associated with taking the optimal policy while following an exploration/exploitation policy.
Some optimizations of Watkin's Q-learning may be applied to SARSA.
Hyperparameters
Learning rate (alpha)
The learning rate determines to what extent newly acquired information overrides old information. A |
https://en.wikipedia.org/wiki/Optical%20properties%20of%20water%20and%20ice | The refractive index of water at 20 °C for visible light is 1.33. The refractive index of normal ice is 1.31 (from List of refractive indices). In general, an index of refraction is a complex number with real and imaginary parts, where the latter indicates the strength of absorption loss at a particular wavelength. In the visible part of the electromagnetic spectrum, the imaginary part of the refractive index is very small. However, water and ice absorb in infrared and close the infrared atmospheric window thereby contributing to the greenhouse effect ...
The absorption spectrum of pure water is used in numerous applications, including light scattering and absorption by ice crystals and cloud water droplets, theories of the rainbow, determination of the single-scattering albedo, ocean color, and many others.
Quantitative description of the refraction index
Over the wavelengths from 0.2 μm to 1.2 μm, and over temperatures from −12 °C to 500 °C, the real part of the index of refraction of water can be calculated by the following empirical expression:
Where:
,
, and
and the appropriate constants are
= 0.244257733, = 0.00974634476, = −0.00373234996, = 0.000268678472, = 0.0015892057, = 0.00245934259, = 0.90070492, = −0.0166626219, = 273.15 K, = 1000 kg/m3, = 589 nm, = 5.432937, and = 0.229202.
In the above expression, T is the absolute temperature of water (in K), is the wavelength of light in nm, is the density of the water in kg/m3, and n is the real part of the index of refraction of water.
Volumic mass of water
In the above formula, the density of water also varies with temperature and is defined by:
with:
= −3.983035 °C
= 301.797 °C
= 522528.9 °C2
= 69.34881 °C
= 999.974950 kg / m3
Refractive index (real and imaginary parts) for liquid water
The total refractive index of water is given as m = n + ik. The absorption coefficient α' is used in the Beer–Lambert law with the prime here signifying base e convention. Values are for water |
https://en.wikipedia.org/wiki/2007%20pet%20food%20recalls | Beginning in March 2007, there was a widespread recall of many brands of cat and dog foods due to contamination with melamine and cyanuric acid. The recalls in North America, Europe, and South Africa came in response to reports of kidney failure in pets. Initially, the recalls were associated with the consumption of mostly wet pet foods made with wheat gluten from a single Chinese company.
After more than three weeks of complaints from consumers, the recall began voluntarily with the Canadian company Menu Foods on 16 March 2007, when a company test showed sickness and death in some of the test animals. In the following weeks, several other companies that had received the contaminated wheat gluten also voluntarily recalled dozens of pet food brands. One month after the initial recall, contaminated rice protein from a different source in China was also identified as being associated with kidney failure in pets in the United States, while contaminated corn gluten was associated with kidney failure with pets in South Africa. As a result of investigating the 2007 pet food recalls, a broader Chinese protein export contamination investigation unfolded, raising concerns about the safety of the human food supply.
By the end of March, veterinary organizations reported more than 100 pet deaths among nearly 500 cases of kidney failure. However, many sources speculate that the actual number of affected pets may never be known, and experts think that the actual death toll could potentially reach into the thousands. In the United States there was extensive media coverage of the recall, with calls for greater government regulation. Reports of widespread and possibly intentional adulteration of Chinese animal feed with melamine raised the issue of melamine contamination in the human food supply, both in China and abroad.
Research has focused on the combination of melamine and cyanuric acid in causing kidney failure. Reports that cyanuric acid may be an independently and potential |
https://en.wikipedia.org/wiki/Origin%20of%20avian%20flight | Around 350 BCE, Aristotle and other philosophers of the time attempted to explain the aerodynamics of avian flight. Even after the discovery of the ancestral bird Archaeopteryx which lived over 150 million years ago, debates still persist regarding the evolution of flight. There are three leading hypotheses pertaining to avian flight: Pouncing Proavis model, Cursorial model, and Arboreal model.
In March 2018, scientists reported that Archaeopteryx was likely capable of flight, but in a manner substantially different from that of modern birds.
Flight characteristics
For flight to occur, four physical forces (thrust and drag, lift and weight) must be favorably combined. In order for birds to balance these forces, certain physical characteristics are required. Asymmetrical wing feathers, found on all flying birds with the exception of hummingbirds, help in the production of thrust and lift. Anything that moves through the air produces drag due to friction. The aerodynamic body of a bird can reduce drag, but when stopping or slowing down a bird will use its tail and feet to increase drag. Weight is the largest obstacle birds must overcome in order to fly. An animal can more easily attain flight by reducing its absolute weight. Birds evolved from other theropod dinosaurs that had already gone through a phase of size reduction during the Middle Jurassic, combined with rapid evolutionary changes. Flying birds during their evolution further reduced relative weight through several characteristics such as the loss of teeth, shrinkage of the gonads out of mating season, and fusion of bones. Teeth were replaced by a lightweight bill made of keratin, the food being processed by the bird's gizzard. Other advanced physical characteristics evolved for flight are a keel for the attachment of flight muscles and an enlarged cerebellum for fine motor coordination. These were gradual changes, though, and not strict conditions for flight: the first birds had teeth, at best a small keel |
https://en.wikipedia.org/wiki/Ecological%20release | Ecological release refers to a population increase or population explosion that occurs when a species is freed from limiting factors in its environment. Sometimes this may occur when a plant or animal species is introduced, for example, to an island or to a new territory or environment other than its native habitat. When this happens, the new arrivals may find themselves suddenly free from the competitors, diseases, or predatory species, etc. in their previous environment, allowing their population numbers to increase beyond their previous limitations. Another common example of ecological release can occur if a disease or a competitor or a keystone species, such as a top predator, is removed from a community or ecosystem. Classical examples of this latter dynamics include population explosions of sea urchins in California's offshore kelp beds, for example, when human hunters began to kill too many sea otters, and/or sudden population explosions of jackrabbits if hunters or ranchers kill too many coyotes.
The foreign species either flourishes into a local population or dies out. Not all released species will become invasive; most released species that don't immediately die out tend to find a small niche in the local ecosystem.
Ecological release also occurs when a species expands its niche within its own habitat or into a new habitat.
Origin
The term ecological release first appeared in the scientific literature in 1972 in the American Zoologist journal discussing the effects of the introduction of a sea snail on an isolated ecosystem, Easter Island. One of the first studies that linked niche shifts to the presence and absence of competitors was by Lack and Southern where habitat broadness of song birds was positively correlated to the absence of a related species.
Common example
Invasive species are an excellent example of successful ecological release because low levels of biodiversity, an abundance of resources, and particular life history traits allow their nu |
https://en.wikipedia.org/wiki/Inspec | Inspec is a major indexing database of scientific and technical literature, published by the Institution of Engineering and Technology (IET), and formerly by the Institution of Electrical Engineers (IEE), one of the IET's forerunners.
Inspec coverage is extensive in the fields of physics, computing, control, and engineering. Its subject coverage includes astronomy, electronics, communications, computers and computing, computer science, control engineering, electrical engineering, information technology, physics, manufacturing, production and mechanical engineering. Now, due to emerging concept of technology for business, Inspec also includes information technology for business in its portfolio. Inspec indexed few journals publishing high quality research by integrating technology into management, economics and social sciences domains. The sample journals include Annual Review of Financial Economics, Aslib Journal of Information Management, Australian Journal of Management and, International Journal of Management, Economics and Social Sciences.
Inspec was started in 1967 as an outgrowth of the Science Abstracts service. The electronic records were distributed on magnetic tape. In the 1980s, it was available in the U.S. through the Knowledge Index, a low-priced dial-up version of the Dialog service for individual users, which made it popular. For nearly 50 years, the IET has employed scientists to manually review items to be included in Inspec, hand-indexing the literature using their own expertise of the subject area and make a judgement call about which terms and classification codes should be applied. Thanks to this work, a significant thesaurus has been developed which enables content to be indexed far more accurately and in context, which in turn helps end-users discover relevant literature that may otherwise have remained hidden from typical search queries, making Inspec an essential tool for prior art, patentability searches and patent drafting.
Access to |
https://en.wikipedia.org/wiki/Axiom%20of%20reducibility | The axiom of reducibility was introduced by Bertrand Russell in the early 20th century as part of his ramified theory of types. Russell devised and introduced the axiom in an attempt to manage the contradictions he had discovered in his analysis of set theory.
History
With Russell's discovery (1901, 1902) of a paradox in Gottlob Frege's 1879 Begriffsschrift and Frege's acknowledgment of the same (1902), Russell tentatively introduced his solution as "Appendix B: Doctrine of Types" in his 1903 The Principles of Mathematics. This contradiction can be stated as "the class of all classes that do not contain themselves as elements". At the end of this appendix Russell asserts that his "doctrine" would solve the immediate problem posed by Frege, but "there is at least one closely analogous contradiction which is probably not soluble by this doctrine. The totality of all logical objects, or of all propositions, involves, it would seem a fundamental logical difficulty. What the complete solution of the difficulty may be, I have not succeeded in discovering; but as it affects the very foundations of reasoning..."
By the time of his 1908 Mathematical logic as based on the theory of types Russell had studied "the contradictions" (among them the Epimenides paradox, the Burali-Forti paradox, and Richard's paradox) and concluded that "In all the contradictions there is a common characteristic, which we may describe as self-reference or reflexiveness".
In 1903, Russell defined predicative functions as those whose order is one more than the highest-order function occurring in the expression of the function. While these were fine for the situation, impredicative functions had to be disallowed:
He repeats this definition in a slightly different way later in the paper (together with a subtle prohibition that they would express more clearly in 1913):
This usage carries over to Alfred North Whitehead and Russell's 1913 Principia Mathematica wherein the authors devote an entire subs |
https://en.wikipedia.org/wiki/Vegv%C3%ADsir | A (Icelandic for "wayfinder", ) is an Icelandic magical stave intended to help the bearer find their way through rough weather. The symbol is attested in the Huld Manuscript, collected in Iceland by Geir Vigfusson in Akureyri in 1860, and does not have any earlier attestations.
A leaf of the manuscript provides an image of the , gives its name, and, in prose, declares that "if this sign is carried, one will never lose one's way in storms or bad weather, even when the way is not known".
It has been claimed that it also features in the Galdrabók, a magical grimoire. although this latter location is denied and contested by Jackson Crawford. Stephen E. Flowers lists the Vegvisir in his translation of the Galdrabók, but in a later publication cites it in “Isländische Zauberzeichen und Zauberbücher” by Ólafur Davíðsson rather than the Galdrabók. It is also only claimed to be in the Huld manuscript by Daniel McCoy. Tomáš Vlasatý claims that it is not only in the Huld manuscript but also in two other Icelandic grimoires, Galdrakver (designated Lbs 2917 a 4to and Lbs 4627 8vo) and has Christian roots.
The is often confused to be a Viking symbol. There is however no evidence of this, and the Huld Manuscript, where it is mentioned, was collected eight centuries after the end of the Viking Age.
Etymology
is a compound word formed from the two Icelandic words, and . means 'way, road, path' (), and , inflection form of , 'to show, to let know, to guide' ().
is derived from the Old Norse , Proto-Germanic , or the Proto-Indo-European .
is derived from the Old Norse meaning 'to show, point out, indicate', or the Proto-Germanic or , meaning 'to visit'.
('way') + ('pointer') derives its meaning from the same word as the English wise. It points someone the right way.
See also
Helm of Awe
Notes
Bibliography
Flowers, Stephen (1989). The Galdrabók: An Icelandic Grimoire. Samuel Weiser, Inc.
Justin Foster Huld Manuscript of Galdrastafir Witchcraft Magic Symbols and |
https://en.wikipedia.org/wiki/Positive%20form | In complex geometry, the term positive form refers to several classes of real differential forms of Hodge type (p, p).
(1,1)-forms
Real (p,p)-forms on a complex manifold M are forms which are of type (p,p) and real, that is, lie in the intersection A real (1,1)-form is called semi-positive (sometimes just positive), respectively, positive (or positive definite) if any of the following equivalent conditions holds:
is the imaginary part of a positive semidefinite (respectively, positive definite) Hermitian form.
For some basis in the space of (1,0)-forms, can be written diagonally, as with real and non-negative (respectively, positive).
For any (1,0)-tangent vector , (respectively, ).
For any real tangent vector , (respectively, ), where is the complex structure operator.
Positive line bundles
In algebraic geometry, positive definite (1,1)-forms arise as curvature forms of ample line bundles (also known as positive line bundles). Let L be a holomorphic Hermitian line bundle on a complex manifold,
its complex structure operator. Then L is equipped with a unique connection preserving the Hermitian structure and satisfying
.
This connection is called the Chern connection.
The curvature of the Chern connection is always a
purely imaginary (1,1)-form. A line bundle L is called positive if is a positive (1,1)-form. (Note that the de Rham cohomology class of is times the first Chern class of L.) The Kodaira embedding theorem claims that a positive line bundle is ample, and conversely, any ample line bundle admits a Hermitian metric with positive.
Positivity for (p, p)-forms
Semi-positive (1,1)-forms on M form a convex cone. When M is a compact complex surface, , this cone is self-dual, with respect to the Poincaré pairing :
For (p, p)-forms, where , there are two different notions of positivity. A form is called
strongly positive if it is a linear combination of products of semi-positive forms, with positive real coefficients. A real (p, p)-form |
https://en.wikipedia.org/wiki/Therefore%20sign | In logical argument and mathematical proof, the therefore sign, , is generally used before a logical consequence, such as the conclusion of a syllogism. The symbol consists of three dots placed in an upright triangle and is read therefore. While it is not generally used in formal writing, it is used in mathematics and shorthand.
History
According to Cajori, A History of Mathematical Notations, Johann Rahn used both the therefore and because signs to mean "therefore"; in the German edition of Teutsche Algebra (1659) the therefore sign was prevalent with the modern meaning, but in the 1668 English edition Rahn used the because sign more often to mean "therefore". Other authors in the 18th century also used three dots in a triangle shape to signify "therefore", but as with Rahn, there wasn't much in the way of consistency as to how the triangle was oriented; because with its current meaning appears to have originated in the 19th century. In the 20th century, the three-dot notation for 'therefore' became very rare in continental Europe, but it remains popular in Anglophone countries.
Example of use
Used in a syllogism:
All gods are immortal.
Zeus is a god.
∴ Zeus is immortal.
and in mathematics
Other uses
In meteorology, the 'therefore' sign is used to indicate 'moderate rain' on a station model; the similar typographic symbol asterism (⁂, three asterisks) indicates moderate snow.
Freemasonry
In Freemasonry traditions, the symbol is used to indicate a Masonic abbreviation (rather than the period mark used conventionally with some abbreviations). For example, "R∴W∴ John Smith" is an abbreviation for "Right Worshipful John Smith" (the term Right Worshipful is an honorific and indicates that Brother Smith is a Grand Lodge officer).
Unicode
The symbol has a Unicode code point at . See Unicode input for keyboard-entering methods.
Similar signs
The inverted form, , known as the because sign, is sometimes used as a shorthand form of "because".
The character (visarg |
https://en.wikipedia.org/wiki/Catalogue%20of%20Life | The Catalogue of Life is an online database that provides an index of known species of animals, plants, fungi, and microorganisms. It was created in 2001 as a partnership between the global Species 2000 and the American Integrated Taxonomic Information System. The Catalogue is used by research scientists, citizen scientists, educators, and policy makers. The Catalogue is also used by the Biodiversity Heritage Library, the Barcode of Life Data System, Encyclopedia of Life, and the Global Biodiversity Information Facility. The Catalogue currently compiles data from 165 peer-reviewed taxonomic databases that are maintained by specialist institutions around the world. the COL Checklist lists 2,067,951 of the world's 2.2m extant species known to taxonomists on the planet at present time.
Structure
The Catalogue of Life employs a simple data structure to provide information on synonymy, grouping within a taxonomic hierarchy, common names, distribution and ecological environment. It provides a dynamic edition, which is updated monthly (and in which data can change without tracking of those changes) and an Annual Checklist, which provides a dated, verifiable reference for the usage of names and associated data. Development of the Catalogue of Life was funded through the Species 2000 europa (EuroCat), 4d4Life, i4Life projects in 2003–2013, and later by the Naturalis Biodiversity Center, Leiden, the Netherlands and Species Files group at Illinois Natural History Survey in Champaign-Urbana.
Current people governing the CoL, contributors, and other relevant information which changes over time, are listed on the CoL Web site.
Usage
Much of the use of the Catalogue is to provide a backbone taxonomy for other global data portals and biological collections. Through the i4Life project, it has formal partnerships with Global Biodiversity Information Facility, European Nucleotide Archive, Encyclopedia of Life, European Consortium for the Barcode of Life, IUCN Red List, and Life W |
https://en.wikipedia.org/wiki/Somatotopic%20arrangement | Somatotopy is the point-for-point correspondence of an area of the body to a specific point on the central nervous system. Typically, the area of the body corresponds to a point on the primary somatosensory cortex (postcentral gyrus). This cortex is typically represented as a sensory homunculus which orients the specific body parts and their respective locations upon the homunculus. Areas such as the appendages, digits, penis, and face can draw their sensory locations upon the somatosensory cortex. The areas which are finely controlled (e.g., the digits) have larger portions of the somatosensory cortex whereas areas which are coarsely controlled (e.g., the trunk) have smaller portions. Areas such as the viscera do not have sensory locations on the post central gyrus.
Macaques, a kind of monkey, already exhibit somatotopy in their somatosensory and motor systems at birth.
Sensorimotor mapping of the human cerebellum
Functional magnetic resonance imaging (fMRI) was employed to determine areas of activation in the cerebellar cortex in humans during a series of motor tasks. The activation areas for movements of lips, tongue, hands, and feet were determined and found to be sharply confined to lobules and sublobules and their sagittal zones in the rostral and caudal spinocerebellar cortex. The activation mapped as two distinct homunculoid representations. One, a more extended representation, was located upside down in the superior cerebellum, and a second one, doubled and smaller, in the inferior cerebellum.
See also
Retinotopy
Topographic map (neuroanatomy)
Cortical homunculus |
https://en.wikipedia.org/wiki/Identity%20score | An identity score is a system for detecting identity theft. Identity scores are increasingly being adopted as a means to prevent fraud in business and as a tool to verify and correct public records.
Identity scores incorporate, a broad set of consumer data that gauges a person's legitimacy. Identity score components can include (but are not limited to) personal identifiers, public records, Internet data, government records, corporate data, predicted behavior patterns based on empirical data, self-assessed behavior patterns, and credit records.
Business and consumer identity scores
Identity scoring was originally developed for use by financial services firms to measure the fraud risk for new customers opening accounts. Typical external credit and fraud checks often fail to detect erroneous background information.
Identity scoring is also being tested as a means for financial institutions to comply with criminal investigations and antiterrorism measures, such as the Bank Secrecy Act (BSA) and the USA PATRIOT Act. Usage of fraud verification tools and third-party authentication systems to verify identities and “red flag” suspicious activity is greatly enhanced by identity scoring.
Public records, private records, and credit records
Identity scores are built from collecting information from a variety of sources and analyzing discernible patterns from the total information. These records can generally be broken down into three categories: Public records, private records, and credit records.
Public records can include (but are not limited to) any of the following sources:
Federal, state and local government records
Financial records like bankruptcies, liens and judgments
Property ownership records
Registered Voter Records
Law enforcement records for felony and misdemeanor convictions
Private (non-credit) records can include (but are not limited to) any of the following sources:
Bill and utility payments
Collected personal information from marketers or affiliates
|
https://en.wikipedia.org/wiki/Illegal%20number | An illegal number is a number that represents information which is illegal to possess, utter, propagate, or otherwise transmit in some legal jurisdiction. Any piece of digital information is representable as a number; consequently, if communicating a specific set of information is illegal in some way, then the number may be illegal as well.
Background
A number may represent some type of classified information or trade secret, legal to possess only by certain authorized persons. An AACS encryption key (09 F9 11 02 9D 74 E3 5B D8 41 56 C5 63 56 88 C0) that came to prominence in May 2007 is an example of a number claimed to be a secret, and whose publication or inappropriate possession is claimed to be illegal in the United States. It allegedly assists in the decryption of any HD DVD or Blu-ray Disc released before this date. The issuers of a series of cease-and-desist letters claim that the key itself is therefore a copyright circumvention device, and that publishing the key violates Title 1 of the US Digital Millennium Copyright Act.
In part of the DeCSS court order and in the AACS legal notices, the claimed protection for these numbers is based on their mere possession and the value or potential use of the numbers. This makes their status and legal issues surrounding their distribution quite distinct from that of copyright infringement.
Any image file or an executable program can be regarded as simply a very large binary number. In certain jurisdictions, there are images that are illegal to possess, due to obscenity or secrecy/classified status, so the corresponding numbers could be illegal.
In 2011 Sony sued George Hotz and members of fail0verflow for jailbreaking the PlayStation 3. Part of the lawsuit complaint was that they had published PS3 keys. Sony also threatened to sue anyone who distributed the keys. Sony later accidentally retweeted an older dongle key through its fictional Kevin Butler character.
Flags and steganography
As a protest of the DeCSS c |
https://en.wikipedia.org/wiki/Transmissometer | A transmissometer or transmissiometer is an instrument for measuring the extinction coefficient of the atmosphere and sea water, and for the determination of visual range. It operates by sending a narrow, collimated beam of energy (usually a laser) through the propagation medium. A narrow field of view receiver at the designated measurement distance determines how much energy is arriving at the detector, and determines the path transmission and/or extinction coefficient. In a transmissometer the extinction coefficient is determined by measuring direct light transmissivity, and the extinction coefficient is then used to calculate visibility range.
Atmospheric extinction is a wavelength dependent phenomenon, but the most common wavelength in use for transmissometers is 550 nm, which is in the middle of the visible waveband, and allows a good approximation of visual range.
Transmissometers are also referred to as telephotometers, transmittance meters, or hazemeters.
Transmissometers are also used by oceanographers and limnologists to measure the optical properties of natural water. In this context, a transmissometer measures the transmittance or attenuation of incident radiation from a light source with a wavelength of around 660 nm, generally through a shorter distance than in air, as water has a smaller maximum visibility distance.
EMOR - Extended MOR Technology
Latest generation transmissometer technology makes use of a co-located forward scatter visibility sensor on the transmitter unit to allow for higher accuracies over an Extended Meteorological Optical Range or EMOR. After 10,000 meters the accuracy of transmissometer technology diminishes, and at higher visibilities forward scatter visibility sensor technology is more accurate. The co-location of the two sensors allows for the most accurate technology to be used when reporting current visibility. The forward scatter sensor also enables auto-alignment and auto-calibration of the transmissometer device. He |
https://en.wikipedia.org/wiki/Citizen%20Weather%20Observer%20Program | The Citizen Weather Observer Program (CWOP) is a network of privately owned electronic weather stations concentrated in the United States but also located in over 150 countries. Network participation allows volunteers with computerized weather stations to send automated surface weather observations to the National Weather Service (NWS) by way of the Meteorological Assimilation Data Ingest System (MADIS). This data is then used by the Rapid Refresh (RAP) forecast model to produce short term forecasts (3 to 12 hours into the future) of conditions across the contiguous United States. Observations are also redistributed to the public.
Origin
The CWOP was originally set up by amateur radio operators experimenting with packet radio, but now includes Internet-only connected stations, as well as amateur radio Automatic Packet Reporting System (APRS) stations. , more than 13,000 stations worldwide report regularly to the network.
Description
The Citizen Weather Observer Program is a program to collect surface weather observations from thousands of privately operated weather stations, into the FindU database, and forward it to the Meteorological Assimilation Data Ingest System (MADIS), operated by the National Oceanic and Atmospheric Administration (NOAA).
FindU
The FindU database is a set of privately operated Internet servers, run by Steve Dimse, (amateur radio callsign K4HG). Numerous IGates (Internet Gateways) receive broadcast amateur radio Automatic Packet Reporting System (APRS) packets containing position and short messages (including telemetry such weather observations), and forward the data to the FindU servers via APRS-IS on the Internet. Weather observations may be polled directly from FindU, and the data is forwarded to MADIS for ingest. APRS messages may also originate directly from computers on the Internet without being broadcast on the radio waves.
MADIS
The Meteorological Assimilation Data Ingest System (MADIS) integrates weather observations from numero |
https://en.wikipedia.org/wiki/Frederick%20Rossini | Frederick Dominic Rossini (July 18, 1899 – October 12, 1990) was an American thermodynamicist noted for his work in chemical thermodynamics.
In 1920, at the age of twenty-one, Rossini entered Carnegie-Mellon University in Pittsburgh, and soon was awarded a full-time teaching scholarship. He graduated with a B.S. in chemical engineering in 1925, followed by an M.S. degree in science in physical chemistry in 1926.
As a result of reading Lewis and Randall's classical 1923 textbook Thermodynamics and the Free Energy of Chemical Substances he wrote to Gilbert N. Lewis and as a result he was offered a teaching fellowship at the University of California at Berkeley. Among his teachers were Gilbert Lewis and William Giauque. Rossini's doctoral dissertation on the heat capacities of strong electrolytes in aqueous solution was supervised by Merle Randall. His Ph.D. degree was awarded in 1928, after only 21 months of graduate work, even though he continued to serve as a teaching fellow throughout this entire period. He worked at the National Bureau of Standards (Washington, DC) from 1928 to 1950.
In 1932, Frederick Rossini, Edward W. Washburn, and Mikkel Frandsen authored "The Calorimetric Determination of the Intrinsic Energy of Gases as a Function of the Pressure." This experiment resulted in the development of the Washburn Correction for bomb calorimetry, a decrease or correction of the results of a calorimetric procedure to normal states.
In 1950, he published his popular textbook Chemical Thermodynamics. In that year he also moved to the Carnegie Institute of Technology (Pittsburgh), where he remained until 1960. He served as dean of the Notre Dame College of Science from 1960 to 1967.
In 1973, Dr. Rossini spent the spring academic quarter at Baldwin-Wallace College, in Berea Ohio, as the first distinguished professor to occupy the Charles J. Strosacker Chair of Science. The Baldwin-Wallace College student union was named after "the late Dr. strosacker, who was vic |
https://en.wikipedia.org/wiki/Community%20%28ecology%29 | In ecology, a community is a group or association of populations of two or more different species occupying the same geographical area at the same time, also known as a biocoenosis, biotic community, biological community, ecological community, or life assemblage. The term community has a variety of uses. In its simplest form it refers to groups of organisms in a specific place or time, for example, "the fish community of Lake Ontario before industrialization".
Community ecology or synecology is the study of the interactions between species in communities on many spatial and temporal scales, including the distribution, structure, abundance, demography, and interactions between coexisting populations. The primary focus of community ecology is on the interactions between populations as determined by specific genotypic and phenotypic characteristics. It is important to understand the origin, maintenance, and consequences of species diversity when evaluating community ecology.
Community ecology also takes into account abiotic factors that influence species distributions or interactions (e.g. annual temperature or soil pH). For example, the plant communities inhabiting deserts are very different from those found in tropical rainforests due to differences in annual precipitation. Humans can also affect community structure through habitat disturbance, such as the introduction of invasive species.
On a deeper level the meaning and value of the community concept in ecology is up for debate. Communities have traditionally been understood on a fine scale in terms of local processes constructing (or destructing) an assemblage of species, such as the way climate change is likely to affect the make-up of grass communities. Recently this local community focus has been criticized. Robert Ricklefs, a professor of biology at the University of Missouri and author of Disintegration of the Ecological Community, has argued that it is more useful to think of communities on a regional sc |
https://en.wikipedia.org/wiki/I%20Love%20a%20Rainy%20Night | "I Love a Rainy Night" is a song co-written and recorded by American country music artist Eddie Rabbitt. It was released in November 1980 as the second single from his album Horizon. It reached number one on the Hot Country Singles, Billboard Hot 100, and Adult Contemporary Singles charts in early 1981. It was written by Rabbitt, Even Stevens, and David Malloy.
Song history
According to music historian Fred Bronson, "I Love a Rainy Night" was 12 years in the making. Rabbitt had a collection of old tapes he kept in the basement of his home. While rummaging through the tapes one day in 1980, he heard a fragment of a song he had recorded one rainy night in the late 1960s.
"It brought back the memory of sitting in a small apartment, staring out the window at one o'clock in the morning, watching the rain come down," wrote Bronson in The Billboard Book of Number One Hits. "He sang into his tape recorder, 'I love a rainy night, I love a rainy night.'"
Upon rediscovery of the old lyrics, Rabbitt completed the song (with help from frequent songwriting partners Even Stevens and David Malloy) and recorded it.
The result included vivid descriptions of a man's fondness for thunderstorms and the peace it brings him ("I love to hear the thunder/watch the lightnin' when it lights up the sky/you know it makes me feel good") and a renewed sense of hope the storms bring ("Showers wash all my cares away/I wake up to a sunny day").
The song's other distinctive feature is its rhythmic pattern of alternating finger snaps and hand claps, which was included with the help of percussionist Farrell Morris, who, according to The Billboard Book of Number One Country Hits, mixed two tracks of each to complete the record.
Chart performance
On February 28, the song succeeded Dolly Parton's hit film theme song "9 to 5" in the number-one position on the Billboard Hot 100 pop singles chart. On March 14, Parton's song returned to the top spot – the last time that the pop chart featured back-to-ba |
https://en.wikipedia.org/wiki/The%20Principles%20and%20Practice%20of%20Medicine | The Principles and Practice of Medicine: Designed for the Use of Practitioners and Students of Medicine is a medical textbook by Sir William Osler. It was first published in 1892 by D. Appleton & Company, while Osler was professor of Medicine at Johns Hopkins University. The book established Osler as the world's leading authority in the teaching of modern medicine.
The text was translated into French, German, Russian, Portuguese, Spanish and Chinese, and for over 40 years it was the world's most significant medical textbook.
First edition
Osler dedicated the book to his teachers; William Arthur Johnson, James Bovell and Robert Palmer Howard. There are 11 sections, preceded by a list of charts and illustrations.
Later years
After 1927, its popularity was succeeded by Cecil Textbook of Medicine.
A revised eleventh edition appeared in 1932. |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.