source stringlengths 31 227 | text stringlengths 9 2k |
|---|---|
https://en.wikipedia.org/wiki/Demihypercube | In geometry, demihypercubes (also called n-demicubes, n-hemicubes, and half measure polytopes) are a class of n-polytopes constructed from alternation of an n-hypercube, labeled as hγn for being half of the hypercube family, γn. Half of the vertices are deleted and new facets are formed. The 2n facets become 2n (n−1)-demicubes, and 2n (n−1)-simplex facets are formed in place of the deleted vertices.
They have been named with a demi- prefix to each hypercube name: demicube, demitesseract, etc. The demicube is identical to the regular tetrahedron, and the demitesseract is identical to the regular 16-cell. The demipenteract is considered semiregular for having only regular facets. Higher forms do not have all regular facets but are all uniform polytopes.
The vertices and edges of a demihypercube form two copies of the halved cube graph.
An n-demicube has inversion symmetry if n is even.
Discovery
Thorold Gosset described the demipenteract in his 1900 publication listing all of the regular and semiregular figures in n-dimensions above three. He called it a 5-ic semi-regular. It also exists within the semiregular k21 polytope family.
The demihypercubes can be represented by extended Schläfli symbols of the form h{4,3,...,3} as half the vertices of {4,3,...,3}. The vertex figures of demihypercubes are rectified n-simplexes.
Constructions
They are represented by Coxeter-Dynkin diagrams of three constructive forms:
... (As an alternated orthotope) s{21,1,...,1}
... (As an alternated hypercube) h{4,3n−1}
.... (As a demihypercube) {31,n−3,1}
H.S.M. Coxeter also labeled the third bifurcating diagrams as 1k1 representing the lengths of the three branches and led by the ringed branch.
An n-demicube, n greater than 2, has n(n−1)/2 edges meeting at each vertex. The graphs below show less edges at each vertex due to overlapping edges in the symmetry projection.
In general, a demicube's elements can be determined from the original n-cube: (with Cn,m = mth-face count i |
https://en.wikipedia.org/wiki/Infanticide%20in%20primates | Infanticide in non-human primates occurs when an individual kills its own or another individual's dependent young. Five hypotheses have been proposed to explain infanticide in non-human primates: exploitation, resource competition, parental manipulation, sexual selection, and social pathology.
Hypotheses for infanticide
Exploitation
Infanticide in non-human primates occurs as a result of exploitation when the individuals performing the infanticide directly benefit from consumption or use of their victim. The individual can become a resource: food (cannibalism); a protective buffer against aggression, or a prop to obtain maternal experience.
The form of exploitation in non-human primates most attributable to adult females is when non-lactating females take an infant from its mother (allomothering) and forcibly retain it until starvation. This behavior is known as the "aunting to death" phenomenon; these non-lactating female primates gain mothering-like experience, yet lack the resources to feed the infant. This behaviour has been seen in captive bonobos, but not wild ones. It is not clear if it is a natural bonobo trait or the result of living in captivity. Male orangutans have not been directly observed practicing infanticide as a reproductive strategy, but recorded case of a male abducting an infant almost resulting in said infant dying from dehydration was observed. Additionally, a possible case of infanticide has been inferred, in which a mother orangutan had lost an infant and received a serious injury on her foot shortly after a new male had been introduced nearby. Although not directly observed, it is inferred this male attacked the female and killed her infant.
Resource competition
Resource competition results when there are too few resources in a particular area to support the existing population. In primates, resource competition is a prime motivator for infanticide. Infanticide motivated by resource competition can occur both outside of and within f |
https://en.wikipedia.org/wiki/Radiobiology | Radiobiology (also known as radiation biology, and uncommonly as actinobiology) is a field of clinical and basic medical sciences that involves the study of the effects of ionizing radiation on living things, in particular health effects of radiation. Ionizing radiation is generally harmful and potentially lethal to living things but can have health benefits in radiation therapy for the treatment of cancer and thyrotoxicosis. Its most common impact is the induction of cancer with a latent period of years or decades after exposure. High doses can cause visually dramatic radiation burns, and/or rapid fatality through acute radiation syndrome. Controlled doses are used for medical imaging and radiotherapy.
Health effects
In general, ionizing radiation is harmful and potentially lethal to living beings but can have health benefits in radiation therapy for the treatment of cancer and thyrotoxicosis.
Most adverse health effects of radiation exposure may be grouped in two general categories:
deterministic effects (harmful tissue reactions) due in large part to the killing or malfunction of cells following high doses; and
stochastic effects, i.e., cancer and heritable effects involving either cancer development in exposed individuals owing to mutation of somatic cells or heritable disease in their offspring owing to mutation of reproductive (germ) cells.
Stochastic
Some effects of ionizing radiation on human health are stochastic, meaning that their probability of occurrence increases with dose, while the severity is independent of dose. Radiation-induced cancer, teratogenesis, cognitive decline, and heart disease are all stochastic effects induced by ionizing radiation.
Its most common impact is the stochastic induction of cancer with a latent period of years or decades after exposure. The mechanism by which this occurs is well understood, but quantitative models predicting the level of risk remain controversial. The most widely accepted model posits that the incidence |
https://en.wikipedia.org/wiki/Fibroepithelial%20neoplasm | A fibroepithelial neoplasm (or tumor) is a biphasic tumor. They consist of epithelial tissue, and stromal or mesenchymal tissue. They may be benign or malignant.
Examples include:
Brenner tumor of the ovary
Fibroadenoma of the breast
Phyllodes tumor of the breast |
https://en.wikipedia.org/wiki/DBTSS | DBTSS, the DataBase of Transcriptional Start Sites, is a database hosted by the Human Genome Center at the University of Tokyo. It contains the exact positions of transcriptional start sites in the genomes of various organisms.
See also
Transcription |
https://en.wikipedia.org/wiki/Blockchain | A blockchain is a distributed ledger with growing lists of records (blocks) that are securely linked together via cryptographic hashes. Each block contains a cryptographic hash of the previous block, a timestamp, and transaction data (generally represented as a Merkle tree, where data nodes are represented by leaves). Since each block contains information about the previous block, they effectively form a chain (compare linked list data structure), with each additional block linking to the ones before it. Consequently, blockchain transactions are irreversible in that, once they are recorded, the data in any given block cannot be altered retroactively without altering all subsequent blocks.
Blockchains are typically managed by a peer-to-peer (P2P) computer network for use as a public distributed ledger, where nodes collectively adhere to a consensus algorithm protocol to add and validate new transaction blocks. Although blockchain records are not unalterable, since blockchain forks are possible, blockchains may be considered secure by design and exemplify a distributed computing system with high Byzantine fault tolerance.
A blockchain was created by a person (or group of people) using the name (or pseudonym) Satoshi Nakamoto in 2008 to serve as the public distributed ledger for bitcoin cryptocurrency transactions, based on previous work by Stuart Haber, W. Scott Stornetta, and Dave Bayer. The implementation of the blockchain within bitcoin made it the first digital currency to solve the double-spending problem without the need of a trusted authority or central server. The bitcoin design has inspired other applications and blockchains that are readable by the public and are widely used by cryptocurrencies. The blockchain may be considered a type of payment rail.
Private blockchains have been proposed for business use. Computerworld called the marketing of such privatized blockchains without a proper security model "snake oil"; however, others have argued that permis |
https://en.wikipedia.org/wiki/693%20%28number%29 | 693 (six hundred [and] ninety-three) is the natural number following 692 and preceding 694.
In mathematics
693 has twelve divisors: 1, 3, 7, 9, 11, 21, 33, 63, 77, 99, 231, and 693. Thus, 693 is tied with 315 for the highest number of divisors for any odd natural number below 900. The smallest positive odd integer with more divisors is 945, which has 16 divisors. Consequently, 945 is also the smallest odd abundant number, having an abundancy index of 1920/945 ≈ 2.03175.
693 appears as the first three digits after the decimal point in the decimal form for the natural logarithm of 2. To 10 digits, this number is 0.6931471805. As a result, if an event has a constant probability of 0.1% of occurring, 693 is the smallest number of trials that must be performed for there to be at least a 50% chance that the event occurs at least once. More generally, for any probability p, the probability that the event occurs at least once in a sample of n items, assuming the items are independent, is given by the following formula:
1 − (1 − p)n
For p = 10−3 = 0.001, plugging in n = 692 gives, to four decimal places, 0.4996, while n = 693 yields 0.5001.
693 is the lowest common multiple of 7, 9, and 11. Multiplying 693 by 5 gives 3465, the smallest positive integer divisible by 3, 5, 7, 9, and 11.
693 is a palindrome in bases 32, 62, 76, 98, 230, and 692. It is also a palindrome in binary: 1010110101.
The reciprocal of 693 has a period of six: = 0..
693 is a triangular matchstick number. |
https://en.wikipedia.org/wiki/Nonlinear%20algebra | Nonlinear algebra is the nonlinear analogue to linear algebra, generalizing notions of spaces and transformations coming from the linear setting. Algebraic geometry is one of the main areas of mathematical research supporting nonlinear algebra, while major components coming from computational mathematics support the development of the area into maturity.
The topological setting for nonlinear algebra is typically the Zariski topology, where closed sets are the algebraic sets. Related areas in mathematics are tropical geometry, commutative algebra, and optimization.
Algebraic geometry
Nonlinear algebra is closely related to algebraic geometry, where the main objects of study include algebraic equations, algebraic varieties, and schemes.
Computational nonlinear algebra
Current methods in computational nonlinear algebra can be broadly broken into two domains: symbolic and numerical. Symbolic methods often rely on the computation of Gröbner bases and resultants. On the other hand, numerical methods typically use algebraically founded homotopy continuation, with a base field of the complex numbers. |
https://en.wikipedia.org/wiki/H%C5%8Dch%C5%8Dd%C5%8D | is a traditional Japanese culinary art form of filleting a fish or fowl without touching it with one's hands.
It is also known as or , and survives to the present day, with occasional demonstrations, particularly in Kyoto.
Ritual origin
It is a Shinto ritual, properly an offering to the gods, and originates in the court cuisine of yūsoku ryōri, dating to the Heian period.
Technique
The filleting is done using only a and a pair of , without touching the fish with one's hands. The chef is dressed in Heian period clothing, most notably an hat and robe. The hitatare features long sleeves and a drawstring, which is used to tie up the sleeves during the ceremony.
Schools
The oldest school is , which originated with in the early Heian period (9th century). He was also known as , due to the mansion he built at the intersection of Shijō Street and Ōmiya street (current Ōmiya Station), hence the name of the style.
The main surviving school is the . The current head (29th generation) is (art name ), of restaurant in Kyoto's Nishijin neighborhood. This style originated in the early Kamakura period (late 12th century), in warrior households that had been bestowed the "Ikama" name by the emperor. In this school the art is called , hence this name is commonly used today.
Demonstrations
The ritual is occasionally done as an offering at shrines, with irregular schedule, and private displays are available by appointment with practitioners. The main event featuring hōchōdō is a demonstration by many practitioners at the , held annually in Kyoto in December.
Notes |
https://en.wikipedia.org/wiki/Department%20of%20Physics%2C%20Quaid-e-Azam%20University | The Department of Physics, Quaid-e-Azam University () (founded as the Institute of Physics, QAU), is an academic and research department of the Quaid-e-Azam University (Qau), Pakistan. It is also referred to as the Institute of Theoretical Physics. In 2018, it was officially renamed as al-Khazini Department, named after al-Khazini.
Established in 1966 with efforts led by Abdus Salam, the institute was located in Rawalpindi, Punjab Province and offered research in mathematics and theoretical physics. Professor Riazuddin served its first and founding director of the institute after shifting to the present QAU campus in 1972. The institute is considered a birthplace of school of theoretical physics of Pakistan. It is also a birthplace of "Theoretical Physics Group" (TPG), now shifted to PINSTECH, and invited scientists from all over the world. The department has collaborated with institutes including Kahuta Research Laboratories, National Centre for Physics, Government College University, the International Centre for Theoretical Physics (ICTP) and CERN.
Foundation and history
Between 1966 and 1968, Abdus Salam's doctoral students had returned to Pakistan after earning their doctorates and experience. These physicists had been under Abdus Salam's influence and were eager to establish an institute of physics. Abdus Salam enabled them to engage their research in theoretical physics and to establish the country's first school of theoretical physics. Abdus Salam provided his support to establish the institute and gave the founding directorship of the institute to his pupil student, Riazuddin.
Abdus Salam and Riazuddin founded the first "Theoretical Physics Group (TPG)" in 1968 that gave birth to Pakistan's school of theoretical physics. The infrastructure was built by the Pakistan Atomic Energy Commission (PAEC) after Raziuddin Siddiqui also persuaded the PAEC to provide its full support to establish this institute as a birthplace of theoretical physics. The new institu |
https://en.wikipedia.org/wiki/Schur%20test | In mathematical analysis, the Schur test, named after German mathematician Issai Schur, is a bound on the operator norm of an integral operator in terms of its Schwartz kernel (see Schwartz kernel theorem).
Here is one version. Let be two measurable spaces (such as ). Let be an integral operator with the non-negative Schwartz kernel , , :
If there exist real functions and and numbers such that
for almost all and
for almost all , then extends to a continuous operator with the operator norm
Such functions , are called the Schur test functions.
In the original version, is a matrix and .
Common usage and Young's inequality
A common usage of the Schur test is to take Then we get:
This inequality is valid no matter whether the Schwartz kernel is non-negative or not.
A similar statement about operator norms is known as Young's inequality for integral operators:
if
where satisfies , for some , then the operator extends to a continuous operator , with
Proof
Using the Cauchy–Schwarz inequality and inequality (1), we get:
Integrating the above relation in , using Fubini's Theorem, and applying inequality (2), we get:
It follows that for any .
See also
Hardy–Littlewood–Sobolev inequality |
https://en.wikipedia.org/wiki/Acid1 | Acid1, originally called the Box Acid Test, is a test page for web browsers. It was developed in October 1998 and was important in establishing baseline interoperability between early web browsers, especially for the Cascading Style Sheets 1.0 specification. As with acid tests for gold which produce a quick and obvious assessment of the quality of a piece of metal, the web acid tests were designed to produce a clear indication of a browser's compliance to web standards.
History
Acid1 tests many features on one page against a reference image. Acid1 was developed by Todd Fahrner, who was frustrated with the lack of stringent tests to improve browser interoperability. After looking at tests developed by Braden McDaniel that used reference renderings to clarify the intended result, Fahrner developed a comprehensive test that resulted in a quirky-looking graphic. In 1999, the test was incorporated into the CSS1 test suite. The text used in Acid1 is an allusion to T. S. Eliot's poem The Hollow Men. Acid1 is included as an offline Easter egg in Internet Explorer 5 for Mac OS, accessible by typing 'about:tasman', with the text replaced by the names of the developers.
By early 2008, all major browsers passed the Acid1 test.
Acid1 has served as inspiration for Acid2 and Acid3.
See also
Comparison of layout engines |
https://en.wikipedia.org/wiki/PPSSPP | PPSSPP (an acronym for "PlayStation Portable Simulator Suitable for Playing Portably") is a free and open-source PSP emulator for Windows, macOS, Linux, iOS, Android, Nintendo WiiU, Nintendo Switch, BlackBerry 10, MeeGo, Pandora, Xbox Series X/S and Symbian with an increased focus on speed and portability. It was first released to the public on November 1, 2012, licensed under the GNU GPLv2 or later. The PPSSPP project was created by Henrik Rydgård, one of the co-founders of the Dolphin emulator.
Features and development
PPSSPP supports save states, dynamic recompilation (JIT) and has rudimentary support of ad hoc wireless networking. To decode PSP multimedia data PPSSPP uses the FFmpeg software library, which was enhanced to enable it to handle Sony's proprietary ATRAC3plus audio format as used by the PSP. PPSSPP offers graphical features that are enhancements over the PSP's capabilities, such as higher screen resolutions, antialiasing, image scaling, support for shaders, and linear and anisotropic filtering.
The ports of PPSSPP for mobile devices offer additional features specific to each platform, such as 'immersive mode' for Android devices, support of the multimedia buttons within Symbian devices and screen stretching on BlackBerry 10 devices to support square screens. All ports of PPSSPP for mobile devices support the use of accelerometers, keyboards and gamepads as input devices.
PPSSPP also supports the Vulkan API, which was added in v1.5.4 release and is intended to provide a substantial performance boost on supported devices.
Portability
Since its inception, PPSSPP has had a focus on portability with support for multiple architectures and operating systems. While initially only supporting Microsoft Windows and Android, this quickly grew to include Blackberry 10, Symbian, macOS, Linux and later iOS. The source code also unofficially supports a wide variety of operating systems and platforms, including Raspberry Pi, Loongson, Maemo, Universal Windows |
https://en.wikipedia.org/wiki/Peanut%20Corporation%20of%20America | Peanut Corporation of America (PCA) was a peanut-processing business which is now defunct as a result of one of the most massive and lethal food-borne contamination events in U.S. history.
PCA was founded in 1977 and initially run by Hugh Parnell, father of Stewart Parnell, with him and two other sons. The company was sold in 1994–1995 with the senior Parnell retiring, and with Stewart Parnell and others remaining with the new company as consultants. In 2000, PCA returned to Stewart Parnell via a private sale. Over this history, PCA came to operate processing facilities in Blakely, Georgia, Suffolk, Virginia, and Plainview, Texas, providing peanut and peanut butter products primarily to the "institutional food" market (schools, prisons and nursing homes), to food manufacturers for use in cookies, snacks, ice cream, and dog treats, and to other markets.
By 2007, prior to closing its doors, PCA had grown to 90 employees and was doing $25 million in annual sales. It has been estimated to have been manufacturing roughly 2.5% of processed peanuts in the U.S. at that time.
PCA permanently halted its operations after it was found to be the source of a massive Salmonella outbreak in the U.S., during late 2008 and early 2009. The 2008 contamination followed a long history of food quality issues. There had been concerns about sanitation at the company since at least the mid-1980s, when the company was run by its founder, Stewart Parnell's father, Hugh Parnell. In addition, in the years just prior to its sale and Hugh Parnell's retirement, PCA was sued: by American Candy Company in 1990, and by Zachary Confections Inc. of Frankfort, Indiana in 1991, after discovery that PCA's peanut products exceeded the FDA tolerance level for aflatoxin, a mold-derived toxin common to peanuts. Moreover, as a result of the coming contamination event, investigations would show that some PCA processing was being done without FDA knowledge and oversight, and other food handling and processing |
https://en.wikipedia.org/wiki/C1orf185 | Chromosome 1 open reading frame 185, also known as C1orf185, is a protein that in humans is encoded by the C1orf185 gene. In humans, C1orf185 is a lowly expressed protein that has been found to be occasionally expressed in the circulatory system.
Gene
C1orf185 is located on chromosome 1 in humans on the positive strand between bases 51,102,221 and 51,148,086. There are 5 exons in the main splice isoform, however the number and selection of exons varies based on the isoform
mRNA and Protein Isoforms
C1orf185 has 5 different splice isoforms in humans.
Protein
C1orf185 is a member of the pfam15842 protein family, containing a domain of unknown function, DUF4718. This family of proteins is between 130 and 224 amino acids long, and is found only in eukaryotes..
The main splice isoform of C1orf185 has a molecular weight of 22.4 kDa and an isoelectric point of 7.67. It contains a transmembrane domain spanning from positions 15 to 37. There is also a conserved serine-rich region from S123 to S142, which could possibly indicate function as a "splicing activator".
C1orf185 contains 3 primary subcellular domains: an extracellular domain which spans the amino acids from positions 1 to 14, a transmembrane domain from positions 15–37, and a large intracellular domain from positions 38–199.
Below are predicted secondary and tertiary structures of C1orf185, modeled using the Chou-Fasman secondary structure prediction tool and the I-TASSER protein structure and function prediction tool. Chou-Fasman predicts a mixture of α-helices, β-sheets, and other structural turns and coils, which can be seen modeled on the I-TASSER prediction.
Regulation of Expression
Gene Level Regulation
Below is a diagram showing the locations of predicted transcription factor binding sites in the C1orf185 promoter, along with a table describing the attributes of each individual binding site. The transcription factors were found and analyzed using the ElDorado tool from Genomatix.
Matrix similar |
https://en.wikipedia.org/wiki/Application%20performance%20engineering | Application performance engineering is a method to develop and test application performance in various settings, including mobile computing, the cloud, and conventional information technology (IT).
Methodology
According to the American National Institute of Standards and Technology, nearly four out of every five dollars spent on the total cost of ownership of an application is directly attributable to finding and fixing issues post-deployment. A full one-third of this cost could be avoided with better software testing.
Application performance engineering attempts to test software before it is published. While practices vary among organizations, the method attempts to emulate the real-world conditions that software in development will confront, including network deployment and access by mobile devices. Techniques include network virtualization.
See also
Network virtualization
Performance engineering
Service virtualization
Software performance testing |
https://en.wikipedia.org/wiki/Haloarchaeobius | Haloarchaeobius (common abbreviation Hab.) is a genus of halophilic archaea in the family of Halobacteriaceae.
Phylogeny
The currently accepted taxonomy is based on the List of Prokaryotic names with Standing in Nomenclature (LPSN) and National Center for Biotechnology Information (NCBI).
See also
List of Archaea genera |
https://en.wikipedia.org/wiki/Elive | Elive is a non-commercial Linux distribution based on Debian. It uses the Enlightenment desktop environment, offering a live DVD and a persistent USB image for 32- and 64-bit computers with Intel or AMD x86 processors.
History
Elive was established in early 2005 as a customized Knoppix Live DVD running the Enlightenment desktop. The first version to appear publicly was called Elive and referenced multiple times by DistroWatch's Ladislav Bodnar and Susan Linton. It was also distributed by UK-based LinuxFormat magazine in 2007, as well as being offered on their cover-disc.
In July 2007, Susan Linton wrote for Distrowatch, "I love Elive and version 1.0 is a wonderful first full release. Just about everything works and works well." At the time, the review was somewhat mixed, with certain criticisms pertaining to laptop usage, such as (at that time) lack of CPU throttling, or WEP when connecting to WiFi.
Controversy regarding payment model
In 2010, version 2.0 was released with improvements like upgrade mode, the "nurse" and more. Linux Magazine and Linux Journal especially touted the tight integration of the E17 window manager in their reviews. However, this version required a payment for installation to hard disk which seriously impacted the initial popularity and was subsequently changed to a voluntary donation. A March 2010 article by Koen Vervloesem of LWN.net criticized Elive 2.0 for requiring payment partway through installation to a hard drive.
Eight years after 2.0, version 3.0 was released. Elive was no longer pay-to-install, but its prolonged development cycle and certain other factors lead to mixed reviews of the project.
Releases
Three separate versions are currently available, as of 1 July 2023:
Stable (version 3.0.6), based on Debian Wheezy and E17. It has a 32-bit release only.
Beta (now at version 3.8.30), which offers a 32- and 64-bit release. It is based on Debian Bullseye and uses the E16 desktop environment.
Retrowave Stable based on 3.8 |
https://en.wikipedia.org/wiki/The%20Llura%20Liggett%20Gund%20Award | The Llura Liggett Gund Award honors researchers for career achievements that have significantly advanced the research and development of preventions, treatments and cures for eye disease.
The award is the highest tribute presented by the Foundation Fighting Blindness, a nonprofit organization that funds research and clinical trials on eye disease. It is named after Llura Liggett Gund, a national trustee of the Foundation and wife of Gordon Gund, the Foundation’s cofounder and chairman. Awardees receive a custom-designed Steuben crystal sculpture.
According to its citation, the award recognizes “an individual whose outstanding dedication and commitment to retinal science for 15 years or more has resulted in highly significant advancements or breakthroughs in retinal degenerative disease research.”
Recipients
The award is presented as warranted.
2016 - Richard Weleber, Casey Eye Institute, Oregon Health & Science University - for career achievement in the field of retinal-disease therapies.
2015 - José-Alain Sahel, Institut de la Vision in Paris - for his work on vision-saving treatments and cures.
2013 - William W. Hauswirth, Ph.D., the Maida and Morris Rybaczki Professor of Ophthalmology at the University of Florida College of Medicine. He received the award for using non-harmful adeno-associated viruses (AAVs) to transport healthy DNA into retinal cells at the back of the eye, showing that gene therapy could restore vision in animal models, including Briard dogs. His team’s approach was then used in human clinical trials where children and young adults virtually blind from Leber congenital amaurosis (LCA) have had significant vision restored.
2012 - Edwin Stone, M.D., Ph.D., the Seamans-Hauser Chair of Molecular Ophthalmology at the University of Iowa. Dr. Stone’s lab has been a leader in discovering and characterizing genes that cause retinal degenerations such as retinitis pigmentosa, Leber congenital amaurosis, and Stargardt disease, among other conditions. |
https://en.wikipedia.org/wiki/International%20Movement%20Writing%20Alphabet | The International Movement Writing Alphabet (IMWA) is a set of symbols that can be used to describe and record movement. Its creator, Valerie Sutton, also invented MovementWriting, a writing system which employs IMWA. It in turn has several application areas within which it is specialised.
Application areas
Sign language transcription
Sutton SignWriting is optimised for sign languages and has the most development so far.
Dance notation
DanceWriting is a form of dance notation.
Mimestry notation
MimeWriting is for classic mimestry.
Kinesiology
SportsWriting is for the kinesiology of ice skating and gymnastics.
Identification numbers
The IMWA has more than 27,000 elements that are represented by unique identification numbers. Each identification number specifies six attributes——as dash-separated values. The symbol is specified with a three-digit value whereas all other attributes use a two-digit value (e.g., 01-01-001-01-01-01).
There are eight categories: hand, movement, face, head, upper body, full body, space, and punctuation.
There are 40 groups. The are based on the 40 groups.
History
The IMWA was originally designed for describing sign language and consequently was named Sutton's Sign Symbol Sequence (SSS) by its inventor, Valerie Sutton. The original symbol set, SSS-95, was limited in size due to memory constraints in personal computers at the time. The SSS-99 symbol set expanded the number of symbols, and the SSS-2002 set was the first to use the current identification numbering system. The final version, SSS-2004, was renamed International Movement Writing Alphabet (SSS-IMWA) to reflect its usefulness in applications beyond sign language.
External links
MovementWriting
IMWA Design Documents
IMWA Keyboard Design
writing systems
constructed scripts |
https://en.wikipedia.org/wiki/Invisibility | Invisibility is the state of an object that cannot be seen. An object in this state is said to be invisible (literally, "not visible"). The phenomenon is studied by physics and perceptual psychology.
Since objects can be seen by light in the visible spectrum from a source reflecting off their surfaces and hitting the viewer's eye, the most natural form of invisibility (whether real or fictional) is an object that neither reflects nor absorbs light (that is, it allows light to pass through it). This is known as
transparency, and is seen in many naturally occurring materials (although no naturally occurring material is 100% transparent).
Invisibility perception depends on several optical and visual factors. For example, invisibility depends on the eyes of the observer and/or the instruments used. Thus an object can be classified as "invisible to" a person, animal, instrument, etc. In research on sensorial perception it has been shown that invisibility is perceived in cycles.
Invisibility is often considered to be the supreme form of camouflage, as it does not reveal to the viewer any kind of vital signs, visual effects, or any frequencies of the electromagnetic spectrum detectable to the human eye, instead making use of radio, infrared or ultraviolet wavelengths.
In illusion optics, invisibility is a special case of illusion effects: the illusion of free space.
The term is often used in fantasy and science fiction, where objects cannot be seen by means of magic or hypothetical technology.
Practical efforts
Technology can be used theoretically or practically to render real-world objects invisible.
Making use of a real-time image displayed on a wearable display, it is possible to create a see-through effect. This is known as active camouflage. Though stealth technology is declared to be invisible to radar, all officially disclosed applications of the technology can only reduce the size and/or clarity of the signature detected by radar.
In 2003 the Chilean s |
https://en.wikipedia.org/wiki/Colored%20music%20notation | Colored music notation is a technique used to facilitate enhanced learning in young music students by adding visual color to written musical notation. It is based upon the concept that color can affect the observer in various ways, and combines this with standard learning of basic notation.
Basis
Viewing color has been widely shown to change an individual's emotional state and stimulate neurons. The Lüscher color test observes from experiments that when individuals are required to contemplate pure red for varying lengths of time, [the experiments] have shown that this color decidedly has a stimulating effect on the nervous system; blood pressure increases, and respiration rate and heart rate both increase. Pure blue, on the other hand, has the reverse effect; observers experience a decline in blood pressure, heart rate, and breathing. Given these findings, it has been suggested that the influence of colored musical notation would be similar.
Music education
In music education, color is typically used in method books to highlight new material. Stimuli received through several senses excite more neurons in several localized areas of the cortex, thereby reinforcing the learning process and improving retention. This information has been proven by other researchers; Chute (1978) reported that "elementary students who viewed a colored version of an instructional film scored significantly higher on both immediate and delayed tests than did students who viewed a monochrome version".
Color studies
Effect on achievement
A researcher in this field, George L. Rogers is the Director of Music Education at Westfield State College. He is also the author of 25 articles in publications that include the Music Educators Journal, The Instrumentalist, and the Journal of Research in Music Education. In 1991, George L. Rogers did a study that researched the effect of color-coded notation on music achievement of elementary instrumental students. Rogers states that the color-co |
https://en.wikipedia.org/wiki/Stereotype%20%28UML%29 | A stereotype is one of three types of extensibility mechanisms in the Unified Modeling Language (UML), the other two being tags and constraints. They allow designers to extend the vocabulary of UML in order to create new model elements, derived from existing ones, but that have specific properties that are suitable for a particular domain or otherwise specialized usage. The nomenclature is derived from the original meaning of stereotype, used in printing. For example, when modeling a network you might need to have symbols for representing routers and hubs. By using stereotyped nodes you can make these things appear as primitive building blocks.
Graphically, a stereotype is rendered as a name enclosed by guillemets (« » or, if guillemets proper are unavailable, << >>) and placed above the name of another element. In addition or alternatively it may be indicated by a specific icon. The icon image may even replace the entire UML symbol. For instance, in a class diagram stereotypes can be used to
classify method behavior, e.g. with «constructor» and «getter» and
refine the classifier itself, e.g. with «interface».
One alternative to stereotypes, suggested by Peter Coad in his book Java Modeling in Color with UML: Enterprise Components and Process is the use of colored archetypes. The archetypes indicated by different-colored UML boxes can be used in combination with stereotypes. This added definition of meaning indicates the role that the UML object plays within the larger software system.
Stereotype attributes
From version 2.0 the previously independent tagged value is considered to be a stereotype attribute. The name tagged value is still kept. Each stereotype has zero or more tag definitions, and all stereotyped UML elements have the corresponding number of tagged values.
UML-defined stereotypes
Become
In UML, become is a keyword for a specific UML stereotype, and applies to a dependency (modeled as a dashed arrow). Become shows that the source modeling el |
https://en.wikipedia.org/wiki/PCB%20reverse%20engineering | Reverse engineering of Printed circuit boards (sometimes called “cloning”, or PCB RE) is the process of generating fabrication and design data for an existing circuit board, either closely or exactly replicating its functionality.
Obtaining circuit board design data is not by necessity malicious or aimed at intellectual property theft. The data generated in the reverse engineering process can be used for troubleshooting, repair, redesign and re-manufacturing, or even testing the security of a device to be used in a restricted environment.
Uses
Legacy product support
Legacy systems need maintenance and replacement parts to operate past their intended life cycle. Demand for parts that are no longer being manufactured can lead to material shortages of parts, called DMS/DMSMS.
There is much demand that entire government divisions have been created to regulate and plan the obsolescence of those systems and parts. Areas commonly affected by technical obsolescence include power station controls, ATC and aviation controls, medical imaging systems, and many aspects of military technology.
There are many legacy systems developed in the 70s, 80s or 90s whose original manufacturer is no longer in business or no longer has the original design data, but whose original equipment is still in use. In many cases exact Form, fit and function is required, either that so parts can “handshake” properly with the existing framework, or to avoid requirements of time-consuming and costly testing.
For industries with highly regulated electronics, (like military or aerospace) this approach can vastly reduce the time required to fabricate replacement parts for system repairs, since the new part's specifications match the original design exactly and therefore do not need to undergo the same level of rigorous re-certification and testing that would be required of a newly designed or revised circuit board.
For example, a power company in Florida was forced to shut down due to the failure of |
https://en.wikipedia.org/wiki/Carboxylate%20transporter | A carboxylate transporter is a membrane transport protein that transports carboxylate.
They are responsible for the reabsorption of filtered carboxylate in renal physiology, resulting in a 100% reabsorption in the proximal tubule.
In proximal tubule
In the renal proximal tubule, there are several kinds of carboxylate transporters in the apical membrane and the basolateral membrane.
Apical
Na-monocarboxylate cotransporter
3Na-dicarboxylate cotransporter
3Na-tricarboxylate cotransporter
Basolateral
H-monocarboxylate cotransporter
organic anion-dicarboxylate exchanger |
https://en.wikipedia.org/wiki/1st%20Word | 1st Word is a word processor program for the Atari ST developed by GST Computer Systems and published in 1985. It was given away with all ST systems from December 1985 for the next two years. Although it was relatively well received, it was a very simple program, lacking most power features and was very slow when working in large documents. In spite of any limitations, its wide availability made the program's .DOC file format became a de facto standard for the platform and was widely supported by other programs like desktop publishing systems.
1st Word Plus was a greatly improved version released by GST when the bundling deal ended in 1987. This addressed performance issues and added dozens of features that made it one of the faster and most feature-packed word processors on the platform. Among its more notable additions were a spell checker, mail merge, and support for footnotes and similar long-document editing features. This became one of the best selling programs on the ST and saw a number of revisions over its lifetime. Plus was later ported to the Acorn Archimedes and IBM PC under GEM as First Word Plus.
In 1990, an entirely unrelated program known as 1st Word Plus 4.0 was released by Compo Software. This used a new file format, and while it could load files from the original 1st Word, the older versions could not read the newer files.
History
Atari released the ST in the summer of 1985, and to ensure there was some useful software at release, they bundled it with the ST Writer word processor. This was a purely text-mode program that had been ported from the best-selling Atari 8-bit program, AtariWriter. The company made it clear ST Writer was being offered only as a stop-gap solution while a graphical user interface (GUI) program was being developed, known as GEMwrite.
For reasons unknown, the company contracted GST Computer Systems of Cambridge, UK to port their Sinclair QL products to the ST. In December 1985, Atari began bundling their 1st Word with th |
https://en.wikipedia.org/wiki/Belgian%20aircraft%20registration%20and%20serials | Belgian owned and operated aircraft are identified by either registration letters or serial numbers for military aircraft.
Civil aircraft
An aircraft registration is a unique alphanumeric string that identifies a civil aircraft, in similar fashion to a licence plate on an automobile. In accordance with the Convention on International Civil Aviation all aircraft must be registered with a national aviation authority and they must carry proof of this registration in the form of a legal document called a Certificate of Registration at all times when in operation.
1913 allocation
The first use of aircraft registrations was based on the radio callsigns allocated at the London International Radiotelegraphic Conference in 1913. The format was a single letter prefix followed by four other letters (like A-BCDE). The major nations operating aircraft were allocated a single letter prefix but minor countries had to share a single letter prefix but were allocated exclusive use of the first letter of the suffix. Belgium was not considered a major operator of aircraft and was allocated the prefix and first letter suffix O-B. When the conference allocated the same prefix it made sure that they were in different parts of the world, the other user of the O prefix was Peru and they were allocated O-P. The first allocation was O-BEBE to a Fokker D.VII on 1 March 1920.
1928 allocation
Following the allocation of radio callsigns to Belgium of ON, OO, OP, OQ, OR, OS and OT at the 1927 International Radio-Telegraph Conference. The callsign allocation did not align with those allocated for aircraft registrations and in 1928 the International Convention of Air Navigation re-allocated the aircraft registration prefix to align with the callsigns. Belgium could use all or any letter groups that had been allocated as radio callsigns and in 1929 the prefix OO was selected. The first allocation was OO-AJT to a Stampe et Vertongen RSV 26 in March 1929 and the format was still in use in 2011.
Use |
https://en.wikipedia.org/wiki/Mobile%20device%20forensics | Mobile device forensics is a branch of digital forensics relating to recovery of digital evidence or data from a mobile device under forensically sound conditions. The phrase mobile device usually refers to mobile phones; however, it can also relate to any digital device that has both internal memory and communication ability, including PDA devices, GPS devices and tablet computers.
Some of the mobile companies had tried to duplicate the model of the phones which is illegal. So, We see so many new models arriving every year which is the forward step to the further generations. The Process of cloning the mobile phones/devices in crime was widely recognised for some years, but the forensic study of mobile devices is a relatively new field, dating from the late 1990s and early 2000s. A proliferation of phones (particularly smartphones) and other digital devices on the consumer market caused a demand for forensic examination of the devices, which could not be met by existing computer forensics techniques.
Mobile devices can be used to save several types of personal information such as contacts, photos, calendars and notes, SMS and MMS messages. Smartphones may additionally contain video, email, web browsing information, location information, and social networking messages and contacts.
There is growing need for mobile forensics due to several reasons and some of the prominent reasons are:
Use of mobile phones to store and transmit personal and corporate information
Use of mobile phones in online transactions
Law enforcement, criminals and mobile phone devices
Mobile device forensics can be particularly challenging on a number of levels:
Evidential and technical challenges exist. For example, cell site analysis following from the use of a mobile phone usage coverage, is not an exact science. Consequently, whilst it is possible to determine roughly the cell site zone from which a call was made or received, it is not yet possible to say with any degree of certain |
https://en.wikipedia.org/wiki/Cadaeic%20Cadenza | "Cadaeic Cadenza" is a 1996 short story by Mike Keith. It is an example of constrained writing, a story with restrictions on how it can be written. It is also one of the most prodigious examples of piphilology, being written in "pilish". The word "cadaeic" is the alphabetical equivalent of 3.141593, the first six decimal digits of pi when rounded, where each digit is replaced by the Nth letter of the alphabet; a cadenza is a solo passage in music.
In addition to the main restriction, the author attempts to mimic portions, or entire works, of different types and pieces of literature ("The Raven", "Jabberwocky", the lyrics of Yes, "The Love Song of J. Alfred Prufrock", Rubaiyat, Hamlet, and Carl Sandburg's Grass) in story, structure, and rhyme.
The limitations
When the number of letters in each word is written out they form the first 3835 digits of pi.
{|style="border: none; text-align: center;"
|-
|One||/||A||Poem||/||A||Raven||/||Midnights||so||dreary,||tired||and||weary,
|-
|3 ||.||1||4 || ||1||5 || ||9 ||2 ||6 ||5 ||3 ||5
|}
While in this example each word is the same number of letters as the next digit of pi (and ten letters for the digit 0), some sections use words of more than ten letters as a one followed by another digit:
{|style="border: none; text-align: center;"
|-
|And||fear||overcame||my||being||–||the||fear||of||"forevermore".
|-
|3 ||4 ||8 ||2 ||5 || ||3 ||4 ||2 ||11
|}
where 11 represents two consecutive digit "1"s in pi.
Taking "A" as 1, "B" as 2, "C" as 3, etc., the name of the piece itself is based on pi, as "Cadaeic" is the first 7 digits of pi, when rounded to that number of significant digits.
C a d a e i c
3.1 4 1 5 9 3
Near a Raven
The first part of Cadaeic Cadenza is slightly changed from an earlier version, "Near a Raven", which was a retelling of Edgar Allan Poe's "The Raven".
See also
Six nines in pi (handled at the start of chapter 2, "Change") |
https://en.wikipedia.org/wiki/Chromatin%20bridge | Chromatin bridge is a mitotic occurrence that forms when telomeres of sister chromatids fuse together and fail to completely segregate into their respective daughter cells. Because this event is most prevalent during anaphase, the term anaphase bridge is often used as a substitute. After the formation of individual daughter cells, the DNA bridge connecting homologous chromosomes remains fixed. As the daughter cells exit mitosis and re-enter interphase, the chromatin bridge becomes known as an interphase bridge. These phenomena are usually visualized using the laboratory techniques of staining and fluorescence microscopy.
Background
The faithful inheritance of genetic information from one cellular generation to the next heavily relies on the duplication of deoxyribonucleic acid (DNA), as well as the formation of two identical daughter cells. This complicated cellular process, known as mitosis, depends on a multitude of cellular checkpoints, signals, interactions and signal cascades for accurate and faithful functioning. Cancer, characterized by uncontrollable cell growth mechanisms and high tendencies for proliferation and metastasis, is highly prone to mitotic mistakes. As a result, several forms of chromosomal aberrations occur, including, but not limited to, binucleated cells, multipolar spindles and micronuclei. Chromatin bridges may serve as a marker of cancer activity.
Process of formation
Chromatin bridges may form by any number of processes wherein chromosomes remain topologically entangled during mitosis. One way in which this may occur is the failure to resolve joint molecules formed during homologous recombination mediated DNA repair, a process that ensures that replicated chromosomes are intact before chromosomes are segregated during cell division. In particular, genetic studies have demonstrated that the loss of the enzymes BLM (Bloom's Syndrome Helicase) or FANCM each result in a dramatic increase in the number of chromatin bridges. This occurs beca |
https://en.wikipedia.org/wiki/High-energy%20phosphate | High-energy phosphate can mean one of two things:
The phosphate-phosphate (phosphoanhydride/phosphoric anhydride/macroergic/phosphagen) bonds formed when compounds such as adenosine diphosphate (ADP) and adenosine triphosphate (ATP) are created.
The compounds that contain these bonds, which include the nucleoside diphosphates and nucleoside triphosphates, and the high-energy storage compounds of the muscle, the phosphagens. When people speak of a high-energy phosphate pool, they speak of the total concentration of these compounds with these high-energy bonds.
High-energy phosphate bonds are usually pyrophosphate bonds, acid anhydride linkages formed by taking phosphoric acid derivatives and dehydrating them. As a consequence, the hydrolysis of these bonds is exergonic under physiological conditions, releasing Gibbs free energy.
Except for PPi → 2 Pi, these reactions are, in general, not allowed to go uncontrolled in the human cell but are instead coupled to other processes needing energy to drive them to completion. Thus, high-energy phosphate reactions can:
provide energy to cellular processes, allowing them to run
couple processes to a particular nucleoside, allowing for regulatory control of the process
drive a reaction out of equilibrium (drive it to the right) by promoting one direction of the reaction faster than the equilibrium can relax.
The one exception is of value because it allows a single hydrolysis, ATP + H2O → AMP + PPi, to effectively supply the energy of hydrolysis of two high-energy bonds, with the hydrolysis of PPi being allowed to go to completion in a separate reaction. The AMP is regenerated to ATP in two steps, with the equilibrium reaction ATP + AMP ↔ 2ADP, followed by regeneration of ATP by the usual means, oxidative phosphorylation or other energy-producing pathways such as glycolysis.
Often, high-energy phosphate bonds are denoted by the character '~'. In this "squiggle" notation, ATP becomes A-P~P~P. The squiggle notation was i |
https://en.wikipedia.org/wiki/Cap%27n%20Crunch | Cap'n Crunch is a corn and oat breakfast cereal manufactured since 1963 by Quaker Oats Company, a subsidiary of PepsiCo since 2001. Since the original product introduction, marketed simply as Cap'n Crunch, Quaker Oats has since introduced numerous flavors and seasonal variations, some for a limited time—and currently offers a Cap'n Crunch product line.
The original Cap'n Crunch cereal was developed to recall a recipe of brown sugar and butter over rice. It was one of the first cereals to use an oil coating to deliver its flavoring, which required an innovative baking process. The taste has been described as similar to the UK and Ireland cereal Golden Nuggets.
Product history
Pamela Low, a flavorist at Arthur D. Little, developed the original Cap'n Crunch flavor in 1963—recalling a recipe of brown sugar and butter her grandmother Luella Low served over rice at her home in Derry, New Hampshire.
Low created the flavor coating for Cap'n Crunch, describing it as giving the cereal a quality she called "want-more-ishness". After her death in 2007, The Boston Globe called Low "the mother of Cap'n Crunch". At Arthur D. Little, Low had also worked on the flavors for Heath, Mounds and Almond Joy candy bars.
In 1965, the Quaker Oats Company awarded the Fredus N. Peters Award to Robert Rountree Reinhart, Sr. for his leadership in directing the development team of Cap'n Crunch. Reinhart developed a technique in the manufacture of Cap'n Crunch, using oil in its recipe as a flavor delivery mechanism—which initially made the cereal difficult to bake properly.
Marketing
The character was created by Allan Burns, who became known for co-creating The Munsters and The Mary Tyler Moore Show. The commercials themselves were originally produced by Jay Ward Productions. Quaker Oats had a marketing plan for Cap'n Crunch, before it had developed the cereal. The product line is heralded by a cartoon mascot named Cap'n Crunch. Cap'n Crunch is depicted as a late 18th-century naval captain |
https://en.wikipedia.org/wiki/Projections%20onto%20convex%20sets | In mathematics, projections onto convex sets (POCS), sometimes known as the alternating projection method, is a method to find a point in the intersection of two closed convex sets. It is a very simple algorithm and has been rediscovered many times. The simplest case, when the sets are affine spaces, was analyzed by John von Neumann. The case when the sets are affine spaces is special, since the iterates not only converge to a point in the intersection (assuming the intersection is non-empty) but to the orthogonal projection of the point onto the intersection. For general closed convex sets, the limit point need not be the projection. Classical work on the case of two closed convex sets shows that the rate of convergence of the iterates is linear.
There are now extensions that consider cases when there are more than two sets, or when the sets are not convex, or that give faster convergence rates. Analysis of POCS and related methods attempt to show that the algorithm converges (and if so, find the rate of convergence), and whether it converges to the projection of the original point. These questions are largely known for simple cases, but a topic of active research for the extensions. There are also variants of the algorithm, such as Dykstra's projection algorithm. See the references in the further reading section for an overview of the variants, extensions and applications of the POCS method; a good historical background can be found in section III of.
Algorithm
The POCS algorithm solves the following problem:
where C and D are closed convex sets.
To use the POCS algorithm, one must know how to project onto the sets C and D separately.
The algorithm starts with an arbitrary value for and then generates the sequence
The simplicity of the algorithm explains some of its popularity. If the intersection of C and D is non-empty, then the sequence generated by the algorithm will converge to some point in this intersection.
Unlike Dykstra's projection algor |
https://en.wikipedia.org/wiki/Social%20profiling | Social profiling is the process of constructing a social media user's profile using his or her social data. In general, profiling refers to the data science process of generating a person's profile with computerized algorithms and technology. There are various platforms for sharing this information with the proliferation of growing popular social networks, including but not limited to LinkedIn, Google+, Facebook and Twitter.
Social profile and social data
A person's social data refers to the personal data that they generate either online or offline (for more information, see social data revolution). A large amount of these data, including one's language, location and interest, is shared through social media and social network. Users join multiple social media platforms and their profiles across these platforms can be linked using different methods to obtain their interests, locations, content, and friend list. Altogether, this information can be used to construct a person's social profile.
Meeting the user's satisfaction level for information collection is becoming more challenging. This is because of too much "noise" generated, which affects the process of information collection due to explosively increasing online data. Social profiling is an emerging approach to overcome the challenges faced in meeting user's demands by introducing the concept of personalized search while keeping in consideration user profiles generated using social network data. A study reviews and classifies research inferring users social profile attributes from social media data as individual and group profiling. The existing techniques along with utilized data sources, the limitations, and challenges were highlighted.
The prominent approaches adopted include machine learning,
ontology, and fuzzy logic. Social media data from Twitter and Facebook have been used by most of the studies to infer the social attributes of users. The literature showed that user social attributes, including age, |
https://en.wikipedia.org/wiki/Pharmacogenomics | Pharmacogenomics is the study of the role of the genome in drug response. Its name (pharmaco- + genomics) reflects its combining of pharmacology and genomics. Pharmacogenomics analyzes how the genetic makeup of a patient affects their response to drugs. It deals with the influence of acquired and inherited genetic variation on drug response, by correlating DNA mutations (including single-nucleotide polymorphisms, copy number variations, and insertions/deletions) with pharmacokinetic (drug absorption, distribution, metabolism, and elimination), pharmacodynamic (effects mediated through a drug's biological targets), and/or immunogenic endpoints.
Pharmacogenomics aims to develop rational means to optimize drug therapy, with regard to the patients' genotype, to achieve maximum efficiency with minimal adverse effects. It is hoped that by using pharmacogenomics, pharmaceutical drug treatments can deviate from what is dubbed as the "one-dose-fits-all" approach. Pharmacogenomics also attempts to eliminate trial-and-error in prescribing, allowing physicians to take into consideration their patient's genes, the functionality of these genes, and how this may affect the efficacy of the patient's current or future treatments (and where applicable, provide an explanation for the failure of past treatments). Such approaches promise the advent of precision medicine and even personalized medicine, in which drugs and drug combinations are optimized for narrow subsets of patients or even for each individual's unique genetic makeup.
Whether used to explain a patient's response (or lack of it) to a treatment, or to act as a predictive tool, it hopes to achieve better treatment outcomes and greater efficacy, and reduce drug toxicities and adverse drug reactions (ADRs). For patients who do not respond to a treatment, alternative therapies can be prescribed that would best suit their requirements. In order to provide pharmacogenomic recommendations for a given drug, two possible types of |
https://en.wikipedia.org/wiki/Acoustical%20Society%20of%20America | The Acoustical Society of America (ASA) is an international scientific society founded in 1929 dedicated to generating, disseminating and promoting the knowledge of acoustics and its practical applications. The Society is primarily a voluntary organization of about 7500 members and attracts the interest, commitment, and service of many professionals.
History
In the summer of 1928, Floyd R. Watson and Wallace Waterfall (1900–1974), a former doctoral student of Watson, were invited by UCLA's Vern Oliver Knudsen to an evening dinner at Knudsen's beach club in Santa Monica. The three physicists decided to form a society of acoustical engineers interested in architectural acoustics. In the early part of December 1928, Wallace Waterfall sent letters to sixteen people inquiring about the possibility of organizing such a society. Harvey Fletcher offered the use of the Bell Telephone Laboratories at 463 West Street in Manhattan as a meeting place for an organizational, initial meeting to be held on December 27, 1928. The meeting was attended by forty scientists and engineers who started the Acoustical Society of America (ASA). Temporary officers were elected: Harvey Fletcher as president, V. O. Knudsen as vice-president, Wallace Waterfall as secretary, and Charles Fuller Stoddard (1876–1958) as treasurer. A constitution and by-laws were drafted. The first issue of the Journal of the Acoustical Society of America was published in October 1929.
Technical committees
The Society has 13 technical committees that represent specialized interests in the field of acoustics. The committees organize technical sessions at conferences and are responsible for the representation of their sub-field in ASA publications. The committees include:
Acoustical oceanography
Animal bioacoustics
Architectural acoustics
Biomedical acoustics
Computational acoustics (Technical Specialty Group)
Acoustical engineering
Musical acoustics
Noise
Physical acoustics
Psychoacoustics
Signal processing in acous |
https://en.wikipedia.org/wiki/Institute%20of%20Solid%20State%20Physics%20%28China%29 | The Institute of Solid State Physics (ISSP, ) is one of over a dozen divisions comprising the Hefei Institutes of Physical Science, Chinese Academy of Sciences (CAS) situated in Hefei, Anhui. The ISSP was founded in March 1982 by the renowned scientist couple Ke T'ing-sui and He Yizhen; it now employs over 200 staff members, has combined facilities of over 25,000 square meters of laboratory space, and is a training base to over 200 graduate students in the fields of condensed matter physics, materials physics and chemistry.
The ISSP is research intensive in the areas of nano-materials technology, novel functional materials, computational physics, internal friction and defects of solids, materials physics under extreme environments, environmental and energy nanomaterials, nuclear engineering and special metallic materials. Most of its research is funded by the National Natural Science Foundation of China, the Chinese Academy of Sciences, the Natural Science Foundation of Anhui Province and the private sector. The ISSP also has close ties with the University of Science and Technology of China in both R&D and training of graduate students.
Research Laboratories
Laboratory of Internal Friction and Defects in Solids
Laboratory for Computational Materials Sciences
Functional Materials Laboratory
Laboratory of Nanomaterials & Nanostructures
Applied Technology Laboratory of Materials
Applied Technology Laboratory of Materials Nanomaterials
Laboratory of Structure Research
Center for Energy Matter in Extreme Environments
Center for Environmental and Energy Nanomaterials
Notable people
T. S. Kê - Pioneered the study of the dynamics of dislocations in solids and founder of the ISSP.
He Yizhen - Pioneer of studies in amorphous physics and metal glasses in China and co-founder of the ISSP.
Kong Qingping () - Recipient of the 2014 Zener Medal |
https://en.wikipedia.org/wiki/Eriodictyol | Eriodictyol is a bitter-masking flavanone, a flavonoid extracted from yerba santa (Eriodictyon californicum), a plant native to North America. Eriodictyol is one of the four flavanones identified in this plant as having taste-modifying properties, the other three being homoeriodictyol, its sodium salt, and sterubin.
Eriodictyol was also found in the twigs of Millettia duchesnei, in Eupatorium arnottianum, and its glycosides (eriocitrin) in lemons and rose hips (Rosa canina). |
https://en.wikipedia.org/wiki/Whitewater%20Interactive%20System%20Development%20with%20Object%20Models | Wisdom (Whitewater Interactive System Development with Object Models) is a software development process and method to design software-intensive interactive systems. It is based on object modelling, and focuses human-computer interaction (HCI) in order to model the software architecture of the system i.e. it is architecture-centric. The focus on HCI while being architecture-centric places Wisdom as a pioneer method within human-centered software engineering. Wisdom was conceived by Nuno Nunes and first published in the years 1999-2000 in order to close the gaps of existing software engineering methods regarding the user interface design.
Notably, the Wisdom method identifies for each use case the tasks of the user, the interaction spaces of the user interface, and the system responsibilities that support that user activity, which are complemented with the data entities used in each case, completing a usable software architecture, an MVC model. The Wisdom model clarifies the relation between the human and the computer-based system, allows rationalization over the software artifacts that must be implemented, therefore facilitating effort affection for a software development team.
From Wisdom, other relevant contributions were derived targeting the enhancement of software development based on the Wisdom model, such as: CanonSketch, Hydra Framework Cruz's
Another relevant contribution is related to effort estimation of software development, the iUCP method, which is based in traditional UCP method leveling the estimation based on the predicted user interface design. A comparison study was carried out using both methods, revealing that there is positive effect in the usage of iUCP when compared to UCP when considering the user interface design, a recurrent situation in nowadays software systems development. |
https://en.wikipedia.org/wiki/Forward%20anonymity | Forward anonymity is a property of a cryptographic system which prevents an attacker who has recorded past encrypted communications from discovering its contents and participants in the future. This property is analogous to forward secrecy.
An example of a system which uses forward anonymity is a public key cryptography system, where the public key is well-known and used to encrypt a message, and an unknown private key is used to decrypt it. In this system, one of the keys is always said to be compromised, but messages and their participants are still unknown by anyone without the corresponding private key.
In contrast, an example of a system which satisfies the perfect forward secrecy property is one in which a compromise of one key by an attacker (and consequent decryption of messages encrypted with that key) does not undermine the security of previously used keys. Forward secrecy does not refer to protecting the content of the message, but rather to the protection of keys used to decrypt messages.
History
Originally introduced by Whitfield Diffie, Paul van Oorschot, and Michael James Wiener to describe a property of STS (station-to-station protocol) involving a long term secret, either a private key or a shared password.
Public Key Cryptography
Public Key Cryptography is a common form of a forward anonymous system. It is used to pass encrypted messages, preventing any information about the message from being discovered if the message is intercepted by an attacker. It uses two keys, a public key and a private key. The public key is published, and is used by anyone to encrypt a plaintext message. The Private key is not well known, and is used to decrypt cyphertext. Public key cryptography is known as an asymmetric decryption algorithm because of different keys being used to perform opposing functions. Public key cryptography is popular because, while it is computationally easy to create a pair of keys, it is extremely difficult to determine the private key kno |
https://en.wikipedia.org/wiki/Freshwater%20phytoplankton | Freshwater phytoplankton is the phytoplankton occurring in freshwater ecosystems. It can be distinguished between limnoplankton (lake phytoplankton), heleoplankton (phytoplankton in ponds), and potamoplankton (river phytoplankton). They differ in size as the environment around them changes. They are affected negatively by the change in salinity in the water.
Ecology
Temperature
Temperature correlates with various characteristics of phytoplankton community: in a study of American freshwater plankton by Huisman et al, higher temperatures correlates with increased species richness, increased overall biomass, and decreased cell size. This may have serious implications under global warming, as given lower phytoplankton biomass with the same cell density would provide less food to the grazers and higher trophic levels, and smaller size may favour consumers that are able to access them at the expense of those that prefer larger algae.
Chlorophyll α
Chlorophyll α is the core photosynthetic pigment that all phytoplankton possess. Concentrations of this pigment, which can be measured remotely, is used as a proxy for phytoplankton biomass in a given location general, the more chlorophyll a, the more phytoplankton biomass, although the CHL a to C ratio May vary between species, and even within the species.
Water body characteristics
Freshwater phytoplankton shows a strong correlation with both surface area and depth of the water body they inhabit. Species richness increases in larger lakes with surface area and decreases in deeper lakes. Decreases due to depth are associated with lower chlorophyll α concentrations.
Salinity
Almost all freshwater phytoplankton die when salinity levels exceed 8%. Between 0% and 8% however, some species may grow preferentially with some amount of salt available. This may be due to the presence of the salt itself, or the hydrodynamic processes that occur with water stratified due to unequal salinity.
Light Utilization
Cyanobacteria are |
https://en.wikipedia.org/wiki/Carlo%20Miranda | Carlo Miranda (15 August 1912 – 28 May 1982) was an Italian mathematician, working on mathematical analysis, theory of elliptic partial differential equations and complex analysis: he is known for giving the first proof of the Poincaré–Miranda theorem, for Miranda's theorem in complex analysis, and for writing an influential monograph in the theory of elliptic partial differential equations.
Selected works
Scientific works
Articles
, available at Gallica.
, available at NUMDAM.
.
Books
.
.
.
.
(and for the second volume).
: two volumes collecting his most important mathematical papers in their original language and typographical form, and a full list of Miranda's publications.
Commemorative, historical, and survey works
. This work completes a survey of with the same title, by elucidating the role of some scientists and adding a further bibliography.
.
.
See also
Luigi Amerio
Schottky's theorem
Singular integral
Notes |
https://en.wikipedia.org/wiki/Structure-based%20assignment | Structure-Based Assignment (SBA) is a technique to accelerate the resonance assignment which is a key bottleneck of NMR (Nuclear magnetic resonance) structural biology. A homologous (similar) protein is used as a template to the target protein in SBA. This template protein provides prior structural information about the target protein and leads to faster resonance assignment . By analogy, in X-ray Crystallography, the molecular replacement technique allows solution of the crystallographic phase problem when a homologous structural model is
known, thereby facilitating rapid structure determination. Some of the SBA algorithms
are CAP which is an RNA assignment algorithm which performs an exhaustive search over
all permutations, MARS which is a program for robust automatic backbone assignment and Nuclear Vector Replacement (NVR) which is a molecular replacement like approach for SBA of resonances and
sparse Nuclear Overhauser Effect (NOE)'s. |
https://en.wikipedia.org/wiki/VPIM | Voice Profile for Internet Mail (VPIMv2) is defined in RFC 3801, an Internet standards track protocol.
VPIM defines a subset of the Internet multimedia messaging protocols (MIME) for use between voice processing server platforms.
This and related protocols define the familiar myEmailAddress@myDomain.com type of e-mail addressing everyone now uses on the internet, extending it from regular e-mail to voice mail and fax systems.
RFC 3801 refers to other RFCs for details on addressing [RFC 3804] and [RFC 3191], FAX [RFC 3192], and GSTN Address Extensions [RFC 2846]. These RFCs are mostly concerned with e-mail address formats, dividing addresses into left hand and right hand sides (LHS, before the @, and RHS, after the @), how protocols are extended to support voice messaging, faxes, and AMIS addressing (which supports phone number dialing), and how to register new values of GSTN service selectors VPIM, VOICE, AMIS, and SYSNUM with IANA, the Internet Assigned Numbers Authority.
The RFCs define VPIM gateways to "translate between dissimilar environments." Both "onramps" (gateways which connects from another voice mail networking protocol to VPIM) and "offramps" (gateways which connect from VPIM to another voice mail networking protocol) are defined, casting VPIM as the inter-system "highway" that connects the different voice messaging systems together.
Messaging and voice messaging systems such as Microsoft's Exchange, Cisco System's Unity and Unity Connection, and Callware's Callegra.UC product lines provide VPIM gateways with on and off ramp gateways to allow integration of similar and dissimilar messaging systems. |
https://en.wikipedia.org/wiki/Bernard%20H.%20Lavenda | Bernard Howard Lavenda (born September 18, 1945) is a retired professor of chemical physics at the University of Camerino and expert on irreversible thermodynamics. He has contributed to many areas of physics, including that of Brownian motion, and in the establishment of the statistical basis of thermodynamics, and non-Euclidean geometrical theories of relativity. He was the scientific coordinator of the "European Thermodynamics Network" in the European Commission Program of Human Capital and Mobility. He was also a proponent for the establishment of, and scientific director of, a National (Italian) Centre for Thermodynamics, and has acted as scientific consultant to companies such as the ENI Group, where he helped to found TEMA, a consulting firm for SNAM Progetti, ENEA (Italian National Agency for New Technologies, Energy and the Environment), and the Solar Energy Research Institute in Golden, Colorado. He has had over 130 scientific papers published in international journals, some critical of the new fashions and modes in theoretical physics.
Professor Lavenda currently lives in Trevignano Romano near Rome, is married with two adult children and two grandchildren, for whom his textbook "A New Perspective on Thermodynamics" is dedicated.
Biography
Early years
Bernard H. Lavenda was born in New York City. After completing secondary school in North Adams, Massachusetts, he attended Clark University where he graduated cum laude in 1966 with a B.Sc in chemistry. Having passed the entrance examination for the doctoral program at the Weizmann Institute of Science, he began experimental work on enzymes under the direction of Ephraim Katzir, who was later to become the President of Israel. Realizing that he was not made out for experimental work, he came under the influence of Ephraim's brother, Aaron, after reading his book Nonequilibrium Thermodynamics in Biophysics, coauthored with Peter Curran.
After the Six Days War, Aaron Katchalsky helped him secure a students |
https://en.wikipedia.org/wiki/Spacecraft%20call%20signs | Spacecraft call signs are radio call signs used for communication in crewed spaceflight. These are not formalized or regulated to the same degree as other equivalent forms of transportation, like aircraft. The three nations currently launching crewed space missions use different methods to identify the ground and space radio stations; the United States uses either the names given to the space vehicles or else the project name and mission number. Russia traditionally assigns code names as call signs to individual cosmonauts, more in the manner of aviator call signs, rather than to the spacecraft.
The only continuity in call signs for spacecraft has been the issuance of "ISS"-suffixed (or "-1SS", for its visual similarity) call signs by various countries in the Amateur Radio service as a citizen of their country has been assigned there. The first Amateur Radio call sign assigned to the International Space Station was NA1SS by the United States. OR4ISS (Belgium), GB1SS (UK), DP0ISS (Germany), and RS0ISS (Russia) are examples of others, but are not all-inclusive of others also issued.
United States
In America's first crewed space program Project Mercury, the astronauts named their individual spacecraft. These names each consisted of a significant word followed by the number 7 (representing the seven original astronauts) and were used as the call signs by the capsule communicators (CAPCOMs).
In Project Gemini, the astronauts were not officially permitted to name their two-man spacecraft, which was identified by "Gemini" followed by the mission number (3 through 12). A notable exception was that Gus Grissom named his Gemini 3 spacecraft Molly Brown after the Titanic survivor, as a joke based on his experience with his Liberty Bell 7 capsule sinking. This name was used as a call sign by CAPCOM L. Gordon Cooper, without NASA's approval.
Starting with the second Gemini flight, Gemini 4, NASA used the Lyndon B. Johnson Space Center (JSC) to house the flight control ce |
https://en.wikipedia.org/wiki/Microsemi | Microsemi Corporation was an Aliso Viejo, California-based provider of semiconductor and system solutions for aerospace & defense, communications, data center and industrial markets.
In February 2018, it was announced that Chandler, Arizona-based Microchip Technology was acquiring the company for over US$10 billion, pending regulatory approval. In May 2018, it was announced that Microchip had completed its acquisition of Microsemi. In August 2018, Microchip discovered that Microsemi shipped large orders to distributors on discount before the closing of the acquisition and had a culture of excessive extravagance, casting some doubt on the future prospect of the acquisition.
History
1959 to 1970: Early years
Microsemi was founded in February 1959 in Culver City, California as MicroSemiconductor. It incorporated in Delaware on September 27, 1960. A trade catalog and price lists from this early period can be found at the Smithsonian Institution.
In March 1963, A. Feldon represented Microsemiconductor Corp in the Advisory Panel of Suppliers at the IEEE session on "Components for Miniaturized Electronic Assemblies".
According to the Missiles and Rockets magazine, in 1964 "Molecular packaging of integrated circuits has been suggested by Microsemiconductor Corp. This would involve the same process the company uses for diodes it supplies to the Improved Minuteman program."
Microsemiconductor is cited in 1965 as supplier of miniature silicon diodes that are "ideal to use as beam profile detectors or to measure depth-dose distribution of small collimated beams." by University of California's Ernest O. Lawrence Radiation Laboratory.
In 1966, two of the company's engineers, E.S. Resnond and M.W. Stillwell, published a Production Engineering Measures Study, on High Voltage Oscillators.
As of 1969, Microsemiconductor's address was still 11250 Plaza Court, Culver City, CA 90230. On August 13, 1969, Standard Resources Corporation, from Copiague, New York, a close-end managemen |
https://en.wikipedia.org/wiki/Black%20hole%20stability%20conjecture | The black hole stability conjecture is the conjecture that a perturbed Kerr black hole will settled back down to a stable state. This has been an open problem in general relativity for some time.
A 2016 paper proved the stability of slowly rotating Kerr black holes in de Sitter space.
A limited stability result for Kerr black holes in Schwarzschild space-time was published by Klainerman et. al. in 2017.
Culminating in 2022, a series of papers was published by Klainerman et. al. which present a proof of the conjecture for slowly rotating Kerr black holes in Minkowski space-time.
See also
Final state conjecture |
https://en.wikipedia.org/wiki/Lamprin | In molecular biology, the lamprin family of proteins consists of several lamprin proteins from the Sea lamprey Petromyzon marinus. Lamprin, an insoluble non-collagen, non-elastin protein, is the major connective tissue component of the fibrillar extracellular matrix of lamprey annular cartilage.
Although not generally homologous to any other protein, soluble lamprins contain a tandemly repeated peptide sequence (GGLGY), which is present in both silk moth chorion proteins and spider dragline silk. Strong homologies to this repeat sequence are also present in several mammalian and avian elastins. It is thought that these proteins share a structural motif which promotes self-aggregation and fibril formation in proteins through interdigitation of hydrophobic side chains in beta-sheet/beta-turn structures, a motif that has been preserved in recognisable form over several hundred million years of evolution. |
https://en.wikipedia.org/wiki/Electron%20configuration | In atomic physics and quantum chemistry, the electron configuration is the distribution of electrons of an atom or molecule (or other physical structure) in atomic or molecular orbitals. For example, the electron configuration of the neon atom is , meaning that the 1s, 2s and 2p subshells are occupied by 2, 2 and 6 electrons respectively.
Electronic configurations describe each electron as moving independently in an orbital, in an average field created by all other orbitals. Mathematically, configurations are described by Slater determinants or configuration state functions.
According to the laws of quantum mechanics, for systems with only one electron, a level of energy is associated with each electron configuration and in certain conditions, electrons are able to move from one configuration to another by the emission or absorption of a quantum of energy, in the form of a photon.
Knowledge of the electron configuration of different atoms is useful in understanding the structure of the periodic table of elements. This is also useful for describing the chemical bonds that hold atoms together, and for understanding the chemical formulas of compounds and the geometries of molecules. In bulk materials, this same idea helps explain the peculiar properties of lasers and semiconductors.
Shells and subshells
Electron configuration was first conceived under the Bohr model of the atom, and it is still common to speak of shells and subshells despite the advances in understanding of the quantum-mechanical nature of electrons.
An electron shell is the set of allowed states that share the same principal quantum number, n (the number before the letter in the orbital label), that electrons may occupy. An atom's nth electron shell can accommodate 2n2 electrons. For example, the first shell can accommodate 2 electrons, the second shell 8 electrons, the third shell 18 electrons and so on. The factor of two arises because the allowed states are doubled due to electron spin—each |
https://en.wikipedia.org/wiki/Hardy%20hierarchy | In computability theory, computational complexity theory and proof theory, the Hardy hierarchy, named after G. H. Hardy, is a hierarchy of sets of numerical functions generated from an ordinal-indexed family of functions hα: N → N (where N is the set of natural numbers, {0, 1, ...}) called Hardy functions. It is related to the fast-growing hierarchy and slow-growing hierarchy.
Hardy hierarchy is introduced by Stanley S. Wainer in 1972, but the idea of its definition comes from Hardy's 1904 paper, in which Hardy exhibits a set of reals with cardinality .
Definition
Let μ be a large countable ordinal such that a fundamental sequence is assigned to every limit ordinal less than μ. The Hardy functions hα: N → N, for α < μ, is then defined as follows:
if α is a limit ordinal.
Here α[n] denotes the nth element of the fundamental sequence assigned to the limit ordinal α. A standardized choice of fundamental sequence for all α ≤ ε0 is described in the article on the fast-growing hierarchy.
The Hardy hierarchy is a family of numerical functions. For each ordinal , a set is defined as the smallest class of functions containing , zero, successor and projection functions, and closed under limited primitive recursion and limited substitution (similar to Grzegorczyk hierarchy).
defines a modified Hardy hierarchy of functions by using the standard fundamental sequences, but with α[n+1] (instead of α[n]) in the third line of the above definition.
Relation to fast-growing hierarchy
The Wainer hierarchy of functions fα and the Hardy hierarchy of functions hα are related by fα = hωα for all α < ε0. Thus, for any α < ε0, hα grows much more slowly than does fα. However, the Hardy hierarchy "catches up" to the Wainer hierarchy at α = ε0, such that fε0 and hε0 have the same growth rate, in the sense that fε0(n-1) ≤ hε0(n) ≤ fε0(n+1) for all n ≥ 1.
Notes |
https://en.wikipedia.org/wiki/Eye%20pinning | Eye pinning, also known as eye flashing or eye blazing, is a form of body language used by parrots. The term that refers to the rapid and very conspicuous dilation and constriction of the pupils of the bird's eyes in response to an external stimulus. Unlike humans, parrots are able to control this reflex and use it as a form of nonverbal communication. It is a common behavior in amazons, macaws, Poicephalus species and the African grey parrots.
It can be an indication that the parrot is feeling excited, angry, afraid or is interested in something. In some circumstances, it may also be a warning that the parrot is currently in a state of being where it will bite if touched. Male budgerigars will perform eye pinning as part of their courtship behavior, pinning the eyes while singing, fluffing and head bobbing, while amazon parrots may pin their eyes to show excitement during play, alongside a fanned tail and raised head and neck feathers.
In the Panama amazon, eye pinning has been noted during vocal communication with humans. The parrot's eyes were observed to noticeably contract when talking or mimicking other sounds of human origin, video recordings indicating that the pupils began contracting several milliseconds prior to the utterance, perhaps suggesting some sort of 'internal rehearsal' process. This was not seen to be the case in the experiment with innate parrot sounds such as short squawks, neither was this behavior observed in a non-talking blue-fronted amazon. In certain circumstances, eye pinning would also occur when the bird heard specific favored words or sounds from humans.
This behavior has also been observed in domestic chickens and woodpigeons. |
https://en.wikipedia.org/wiki/Coherent%20backscattering | In physics, coherent backscattering is observed when coherent radiation (such as a laser beam) propagates through a medium which has a large number of scattering centers (such as milk or a thick cloud) of size comparable to the wavelength of the radiation.
The waves are scattered many times while traveling through the medium. Even for incoherent radiation, the scattering typically reaches a local maximum in the direction of backscattering. For coherent radiation, however, the peak is two times higher.
Coherent backscattering is very difficult to detect and measure for two reasons. The first is fairly obvious, that it is difficult to measure the direct backscatter without blocking the beam, but there are methods for overcoming this problem. The second is that the peak is usually extremely sharp around the backward direction, so that a very high level of angular resolution is needed for the detector to see the peak without averaging its intensity out over the surrounding angles where the intensity can undergo large dips. At angles other than the backscatter direction, the light intensity is subject to numerous essentially random fluctuations called speckles.
This is one of the most robust interference phenomena that survives multiple scattering, and it is regarded as an aspect of a quantum mechanical phenomenon known as weak localization (Akkermans et al. 1986). In weak localization, interference of the direct and reverse paths leads to a net reduction of light transport in the forward direction. This phenomenon is typical of any coherent wave which is multiple scattered. It is typically discussed for light waves, for which it is similar to the weak localization phenomenon for electrons in disordered semi-conductors and often seen as the precursor to Anderson (or strong) localization of light. Weak localization of light can be detected since it is manifested as an enhancement of light intensity in the backscattering direction. This substantial enhancement is called |
https://en.wikipedia.org/wiki/Monarch%20butterfly%20migration | Monarch butterfly migration is the phenomenon, mainly across North America, where the subspecies Danaus plexippus plexippus migrates each summer and autumn to and from overwintering sites on the West Coast of California or mountainous sites in Central Mexico. Other subspecies perform minor migrations or none at all. This massive movement of butterflies has been called "one of the most spectacular natural phenomena in the world".
The monarchs begin their southern migration from September to October. Eastern and northeastern populations, up to 500,000 monarch butterflies, migrate at this time. Originating in southern Canada and the United States, they travel to overwintering sites in central Mexico. The butterflies arrive at their roosting sites in November. They remain in their roosts during the winter months and then begin their northern migration in March. No individual butterfly completes the entire round trip. Female monarchs lay eggs for a subsequent generation during the northward migration. Four generations are involved in the annual cycle and the generation undertaking the southbound migration live eight times longer than their parents and grandparents.
Similarly, the western populations migrate annually from regions west of the Rocky Mountains to overwintering sites on the coast of California.
Not all monarch populations make major migrations. Monarchs migrate short distances in Australia and New Zealand. There are some populations, for instance in Florida and the Caribbean, that do not migrate, as well as another subspecies distributed in the Caribbean, Central America and northern South America. Additional overwintering sites have been identified in Arizona and northern Florida.
Historical accounts
As late as 1951, monarchs were mistakenly thought to overwinter as adults or pupae. Roosts of thousands were observed in southern regions of North America.
Migrating western populations of Danaus plexippus and their overwintering sites were known long befo |
https://en.wikipedia.org/wiki/Structural%20induction | Structural induction is a proof method that is used in mathematical logic (e.g., in the proof of Łoś' theorem), computer science, graph theory, and some other mathematical fields. It is a generalization of mathematical induction over natural numbers and can be further generalized to arbitrary Noetherian induction. Structural recursion is a recursion method bearing the same relationship to structural induction as ordinary recursion bears to ordinary mathematical induction.
Structural induction is used to prove that some proposition holds for all of some sort of recursively defined structure, such as
formulas, lists, or trees. A well-founded partial order is defined on the structures ("subformula" for formulas, "sublist" for lists, and "subtree" for trees). The structural induction proof is a proof that the proposition holds for all the minimal structures and that if it holds for the immediate substructures of a certain structure , then it must hold for also. (Formally speaking, this then satisfies the premises of an axiom of well-founded induction, which asserts that these two conditions are sufficient for the proposition to hold for all .)
A structurally recursive function uses the same idea to define a recursive function: "base cases" handle each minimal structure and a rule for recursion. Structural recursion is usually proved correct by structural induction; in particularly easy cases, the inductive step is often left out. The length and ++ functions in the example below are structurally recursive.
For example, if the structures are lists, one usually introduces the partial order "<", in which whenever list is the tail of list . Under this ordering, the empty list is the unique minimal element. A structural induction proof of some proposition then consists of two parts: A proof that is true and a proof that if is true for some list , and if is the tail of list , then must also be true.
Eventually, there may exist more than one base case |
https://en.wikipedia.org/wiki/Pople%20notation | The Pople notation is named after the Nobel laureate John Pople and is a simple method of presenting second-order spin coupling systems in NMR.
The notation labels each (NMR active) nucleus with a letter of the alphabet. The difference in chemical shift, δ, relative to the J-coupling between nuclei mirrors the separation of the letter labels in the Latin alphabet. The letters used tend to be limited to A,B,M,N,X,Y.
For example, AB indicates two nuclei which have similar chemical shifts (Δδ similar to or smaller than J), whereas AX indicates two which lie further apart on the spectrum (Δδ significantly larger than J). A2B would similarly indicate a spin system containing two equivalent nuclei (A) and a third, inequivalent one (B). Nuclei which are in equivalent chemical environments (that is, symmetry-related), but inequivalent magnetic environments are distinguished with a prime; e.g. AA'. This key aspect of the notation, i.e., using a prime to differentiate between chemical equivalence only compared to full magnetic equivalence, was introduced by Richards and Schaefer in 1958.
The notation can be used to represent systems of more than two nuclei, for example AMX represents three nuclei, each moderately separated from the others, and ABX represents two nuclei whose peaks are closely spaced and one other nucleus which is more distant.
Examples:
PHCl2 is an AX system whereas
CH3CH2F is an A3M2X system, |
https://en.wikipedia.org/wiki/Anthesis | Anthesis is the period during which a flower is fully open and functional. It may also refer to the onset of that period.
The onset of anthesis is spectacular in some species. In Banksia species, for example, anthesis involves the extension of the style far beyond the upper perianth parts. Anthesis of flowers is sequential within an inflorescence, so when the style and perianth are different colours, the result is a striking colour change that gradually sweeps along the inflorescence.
Flowers with diurnal anthesis generally are brightly colored in order to attract diurnal insects, such as butterflies.
Flowers with nocturnal anthesis generally are white or less colorful, and as such, they contrast more strongly with the night. These flowers typically attract nocturnal insects including many moth species. |
https://en.wikipedia.org/wiki/Design%20rationale | A design rationale is an explicit documentation of the reasons behind decisions made when designing a system or artifact. As initially developed by W.R. Kunz and Horst Rittel, design rationale seeks to provide argumentation-based structure to the political, collaborative process of addressing wicked problems.
Overview
A design rationale is the explicit listing of decisions made during a design process, and the reasons why those decisions were made. Its primary goal is to support designers by providing a means to record and communicate the argumentation and reasoning behind the design process.
It should therefore include:
the reasons behind a design decision,
the justification for it,
the other alternatives considered,
the trade offs evaluated, and
the argumentation that led to the decision.
Several science areas are involved in the study of design rationales, such as computer science cognitive science, artificial intelligence, and knowledge management. For supporting design rationale, various frameworks have been proposed, such as QOC, DRCS, IBIS, and DRL.
History
While argumentation formats can be traced back to Stephen Toulmin's work in the 1950s datums, claims, warrants, backings and rebuttals, the origin of design rationale can be traced back to W.R. Kunz and Horst Rittel's development of the Issue-Based Information System (IBIS) notation in 1970. Several variants on IBIS have since been proposed.
The first was Procedural Hierarchy of Issues (PHI), first described in Ray McCall's PhD Dissertation although not named at the time.
IBIS was also modified, in this case to support Software Engineering, by Potts & Bruns. The Potts & Bruns approach was then extended by the Decision Representation Language (DRL). which itself was extended by RATSpeak.
Questions Options and Criteria (QOC), also known as Design Space Analysis is an alternative representation for argumentation-based rationale, as are Win-Win and the Decision Recommendation and Intent Model |
https://en.wikipedia.org/wiki/Cgroups | cgroups (abbreviated from control groups) is a Linux kernel feature that limits, accounts for, and isolates the resource usage (CPU, memory, disk I/O, etc.) of a collection of processes.
Engineers at Google started the work on this feature in 2006 under the name "process containers". In late 2007, the nomenclature changed to "control groups" to avoid confusion caused by multiple meanings of the term "container" in the Linux kernel context, and the control groups functionality was merged into the Linux kernel mainline in kernel version 2.6.24, which was released in January 2008. Since then, developers have added many new features and controllers, such as support for kernfs in 2014, firewalling, and unified hierarchy. cgroup v2 was merged in Linux kernel 4.5 with significant changes to the interface and internal functionality.
Versions
There are two versions of cgroups.
Cgroups was originally written by Paul Menage and Rohit Seth, and merged into the mainline Linux kernel in 2007. Afterwards this is called cgroups version 1.
Development and maintenance of cgroups was then taken over by Tejun Heo. Tejun Heo redesigned and rewrote cgroups. This rewrite is now called version 2, the documentation of cgroup-v2 first appeared in Linux kernel 4.5 released on 14 March 2016.
Unlike v1, cgroup v2 has only a single process hierarchy and discriminates between processes, not threads.
Features
One of the design goals of cgroups is to provide a unified interface to many different use cases, from controlling single processes (by using nice, for example) to full operating system-level virtualization (as provided by OpenVZ, Linux-VServer or LXC, for example). Cgroups provides:
Resource limiting
groups can be set to not exceed a configured memory limit, which also includes the file system cache, I/O bandwidth limit, CPU quota limit, or CPU set limit.
Prioritization
some groups may get a larger share of CPU utilization or disk I/O throughput
Accounting
measures a group's |
https://en.wikipedia.org/wiki/NDUFB2 | NADH dehydrogenase [ubiquinone] 1 beta subcomplex subunit 2, mitochondrial is an enzyme that in humans is encoded by the NDUFB2 gene. NADH dehydrogenase (ubiquinone) 1 beta subcomplex, 2, 8kDa is an accessory subunit of the NADH dehydrogenase (ubiquinone) complex, located in the mitochondrial inner membrane. It is also known as Complex I and is the largest of the five complexes of the electron transport chain.
Structure
The NDUFB2 gene, located on the q arm of chromosome 7 in position 34, is 9,966 base pairs long and is composed of 4 exons. The NDUFB2 protein weighs 12 kDa and is composed of 105 amino acids. NDUFB2 is a subunit of the enzyme NADH dehydrogenase (ubiquinone), the largest of the respiratory complexes. The structure is L-shaped with a long, hydrophobic transmembrane domain and a hydrophilic domain for the peripheral arm that includes all the known redox centers and the NADH binding site. NDUFB3 is one of about 31 hydrophobic subunits that form the transmembrane region of Complex I. It has been noted that the N-terminal hydrophobic domain has the potential to be folded into an alpha helix spanning the inner mitochondrial membrane with a C-terminal hydrophilic domain interacting with globular subunits of Complex I. The highly conserved two-domain structure suggests that this feature is critical for the protein function and that the hydrophobic domain acts as an anchor for the NADH dehydrogenase (ubiquinone) complex at the inner mitochondrial membrane. Hydropathy analysis revealed that this subunit and 4 other subunits have an overall hydrophilic pattern, even though they are found within the hydrophobic protein (HP) fraction of complex I.
Function
The human NDUFB2 gene codes for a subunit of Complex I of the respiratory chain, which transfers electrons from NADH to ubiquinone. However, NDUFB2 is an accessory subunit of the complex that is believed not to be involved in catalysis. Initially, NADH binds to Complex I and transfers two electrons to the isoa |
https://en.wikipedia.org/wiki/Mixed-function%20oxidase | Mixed-function oxidase is the name of a family of oxidase enzymes that catalyze a reaction in which each of the two atoms of oxygen in O2 is used for a different function in the reaction.
Oxidase is a general name for enzymes that catalyze oxidations in which molecular oxygen is the electron acceptor but oxygen atoms do not appear in the oxidized product. Often, oxygen is reduced to either water (cytochrome oxidase of the mitochondrial electron transfer chain) or hydrogen peroxide (dehydrogenation of fatty acyl-CoA in peroxisomes). Most of the oxidases are flavoproteins.
The name "mixed-function oxidase" indicates that the enzyme oxidizes two different substrate simultaneously. Desaturation of fatty acyl-CoA in vertebrates is an example of the mixed-function oxidase reaction. In the process, saturated fatty acyl-CoA and NADPH are oxidized by molecular oxygen (O2) to produce monounsaturated fatty acyl-CoA, NADP+ and 2 molecules of water.
Reaction
The mixed-function oxidase reaction proceeds as follows:
AH + BH2 + O2 --> AOH + B + H2O (H2O as catalyst.)
Medical significance
High levels of mixed-function oxidase activity has been studied for their activation effects in human colon carcinoma cell lines, to study the susceptibility to certain cancers. The research has been successful in mice but remains inconclusive in humans. |
https://en.wikipedia.org/wiki/Carrier%20generation%20and%20recombination | In the solid-state physics of semiconductors, carrier generation and carrier recombination are processes by which mobile charge carriers (electrons and electron holes) are created and eliminated. Carrier generation and recombination processes are fundamental to the operation of many optoelectronic semiconductor devices, such as photodiodes, light-emitting diodes and laser diodes. They are also critical to a full analysis of p-n junction devices such as bipolar junction transistors and p-n junction diodes.
The electron–hole pair is the fundamental unit of generation and recombination in inorganic semiconductors, corresponding to an electron transitioning between the valence band and the conduction band where generation of electron is a transition from the valence band to the conduction band and recombination leads to a reverse transition.
Overview
Like other solids, semiconductor materials have an electronic band structure determined by the crystal properties of the material. Energy distribution among electrons is described by the Fermi level and the temperature of the electrons. At absolute zero temperature, all of the electrons have energy below the Fermi level; but at non-zero temperatures the energy levels are filled following a Fermi-Dirac distribution.
In undoped semiconductors the Fermi level lies in the middle of a forbidden band or band gap between two allowed bands called the valence band and the conduction band. The valence band, immediately below the forbidden band, is normally very nearly completely occupied. The conduction band, above the Fermi level, is normally nearly completely empty. Because the valence band is so nearly full, its electrons are not mobile, and cannot flow as electric current.
However, if an electron in the valence band acquires enough energy to reach the conduction band (as a result of interaction with other electrons, holes, photons, or the vibrating crystal lattice itself), it can flow freely among the nearly empty conduction |
https://en.wikipedia.org/wiki/Pseudo-differential%20operator | In mathematical analysis a pseudo-differential operator is an extension of the concept of differential operator. Pseudo-differential operators are used extensively in the theory of partial differential equations and quantum field theory, e.g. in mathematical models that include ultrametric pseudo-differential equations in a non-Archimedean space.
History
The study of pseudo-differential operators began in the mid 1960s with the work of Kohn, Nirenberg, Hörmander, Unterberger and Bokobza.
They played an influential role in the second proof of the Atiyah–Singer index theorem via K-theory. Atiyah and Singer thanked Hörmander for assistance with understanding the theory of pseudo-differential operators.
Motivation
Linear differential operators with constant coefficients
Consider a linear differential operator with constant coefficients,
which acts on smooth functions with compact support in Rn.
This operator can be written as a composition of a Fourier transform, a simple multiplication by the
polynomial function (called the symbol)
and an inverse Fourier transform, in the form:
Here, is a multi-index, are complex numbers, and
is an iterated partial derivative, where ∂j means differentiation with respect to the j-th variable. We introduce the constants to facilitate the calculation of Fourier transforms.
Derivation of formula ()
The Fourier transform of a smooth function u, compactly supported in Rn, is
and Fourier's inversion formula gives
By applying P(D) to this representation of u and using
one obtains formula ().
Representation of solutions to partial differential equations
To solve the partial differential equation
we (formally) apply the Fourier transform on both sides and obtain the algebraic equation
If the symbol P(ξ) is never zero when ξ ∈ Rn, then it is possible to divide by P(ξ):
By Fourier's inversion formula, a solution is
Here it is assumed that:
P(D) is a linear differential operator with constant coefficients,
its symbol P(ξ) |
https://en.wikipedia.org/wiki/Space-based%20solar%20power | Space-based solar power (SBSP, SSP) is the concept of collecting solar power in outer space with solar power satellites (SPS) and distributing it to Earth. Its advantages include a higher collection of energy due to the lack of reflection and absorption by the atmosphere, the possibility of very little night, and a better ability to orient to face the Sun. Space-based solar power systems convert sunlight to some other form of energy (such as microwaves) which can be transmitted through the atmosphere to receivers on the Earth's surface.
Various SBSP proposals have been researched since the early 1970s, but none is economically viable with present-day space launch costs. Some technologists speculate that this may change in the distant future with space manufacturing from asteroids or lunar material, or with radical new space launch technologies other than rocketry.
Besides cost, SBSP also introduces several technological hurdles, including the problem of transmitting energy from orbit. Since wires extending from Earth's surface to an orbiting satellite are not feasible with current technology, SBSP designs generally include the wireless power transmission with its concomitant conversion inefficiencies, as well as land use concerns for antenna stations to receive the energy at Earth's surface. The collecting satellite would convert solar energy into electrical energy, power a microwave transmitter or laser emitter, and transmit this energy to a collector (or microwave rectenna) on Earth's surface. Contrary to appearances in fiction, most designs propose beam energy densities that are not harmful if human beings were to be inadvertently exposed, such as if a transmitting satellite's beam were to wander off-course. But the necessarily vast size of the receiving antennas would still require large blocks of land near the end users. The service life of space-based collectors in the face of long-term exposure to the space environment, including degradation from radiation |
https://en.wikipedia.org/wiki/BLAKE%20%28hash%20function%29 | BLAKE is a cryptographic hash function based on Daniel J. Bernstein's ChaCha stream cipher, but a permuted copy of the input block, XORed with round constants, is added before each ChaCha round. Like SHA-2, there are two variants differing in the word size. ChaCha operates on a 4×4 array of words. BLAKE repeatedly combines an 8-word hash value with 16 message words, truncating the ChaCha result to obtain the next hash value. BLAKE-256 and BLAKE-224 use 32-bit words and produce digest sizes of 256 bits and 224 bits, respectively, while BLAKE-512 and BLAKE-384 use 64-bit words and produce digest sizes of 512 bits and 384 bits, respectively.
The BLAKE2 hash function, based on BLAKE, was announced in 2012. The BLAKE3 hash function, based on BLAKE2, was announced in 2020.
History
BLAKE was submitted to the NIST hash function competition by Jean-Philippe Aumasson, Luca Henzen, Willi Meier, and Raphael C.-W. Phan. In 2008, there were 51 entries. BLAKE made it to the final round consisting of five candidates but lost to Keccak in 2012, which was selected for the SHA-3 algorithm.
Algorithm
Like SHA-2, BLAKE comes in two variants: one that uses 32-bit words, used for computing hashes up to 256 bits long, and one that uses 64-bit words, used for computing hashes up to 512 bits long. The core block transformation combines 16 words of input with 16 working variables, but only 8 words (256 or 512 bits) are preserved between blocks.
It uses a table of 16 constant words (the leading 512 or 1024 bits of the fractional part of π), and a table of 10 16-element permutations:
σ[0] = 0 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15
σ[1] = 14 10 4 8 9 15 13 6 1 12 0 2 11 7 5 3
σ[2] = 11 8 12 0 5 2 15 13 10 14 3 6 7 1 9 4
σ[3] = 7 9 3 1 13 12 11 14 2 6 5 10 4 0 15 8
σ[4] = 9 0 5 7 2 4 10 15 14 1 11 12 6 8 3 13
σ[5] = 2 12 6 10 0 11 8 3 4 13 7 5 15 14 1 9
σ[6] = 12 5 1 15 14 13 4 10 0 7 6 3 9 2 8 11
σ[7] = 13 11 7 14 12 |
https://en.wikipedia.org/wiki/Riemannian%20manifold | In differential geometry, a Riemannian manifold or Riemannian space , so called after the German mathematician Bernhard Riemann, is a real, smooth manifold M equipped with a positive-definite inner product gp on the tangent space TpM at each point p.
The family gp of inner products is called a Riemannian metric (or Riemannian metric tensor). Riemannian geometry is the study of Riemannian manifolds.
A common convention is to take g to be smooth, which means that for any smooth coordinate chart on M, the n2 functions
are smooth functions. These functions are commonly designated as .
With further restrictions on the , one could also consider Lipschitz Riemannian metrics or measurable Riemannian metrics, among many other possibilities.
A Riemannian metric (tensor) makes it possible to define several geometric notions on a Riemannian manifold, such as angle at an intersection, length of a curve, area of a surface and higher-dimensional analogues (volume, etc.), extrinsic curvature of submanifolds, and intrinsic curvature of the manifold itself.
Introduction
In 1828, Carl Friedrich Gauss proved his Theorema Egregium ("remarkable theorem" in Latin), establishing an important property of surfaces. Informally, the theorem says that the curvature of a surface can be determined entirely by measuring distances along paths on the surface. That is, curvature does not depend on how the surface might be embedded in 3-dimensional space. See Differential geometry of surfaces. Bernhard Riemann extended Gauss's theory to higher-dimensional spaces called manifolds in a way that also allows distances and angles to be measured and the notion of curvature to be defined, again in a way that is intrinsic to the manifold and not dependent upon its embedding in higher-dimensional spaces. Albert Einstein used the theory of pseudo-Riemannian manifolds (a generalization of Riemannian manifolds) to develop his general theory of relativity. In particular, his equations for gravitation are c |
https://en.wikipedia.org/wiki/March%20of%20Dimes%20Prize%20in%20Developmental%20Biology | The March of Dimes Prize in Developmental Biology is awarded once a year by the March of Dimes. It carries a $150,000 award "to scientific leaders who have pioneered research to advance our understanding of prenatal development and pregnancy".
Laureates
Source: March of Dimes
2022-23 Patrica Hunt
2021 Alan W. Flake
2020 Susan Fisher
2019 Myriam Hemberger
2018 Allan C. Spradling
2017 Charles David Allis
2016 Victor R. Ambros and Gary B. Ruvkun
2015 Rudolf Jaenisch
2014 Huda Y. Zoghbi
2013 Eric N. Olson
2012 Elaine Fuchs and Howard Green
2011 Patricia Ann Jacobs and David C. Page
2010 Shinya Yamanaka
2009 Kevin P. Campbell and Louis M. Kunkel
2008 Philip A. Beachy and Clifford Tabin
2007 Anne McLaren and Janet Rossant
2006 Alexander Varshavsky
2005 Mario R. Capecchi and Oliver Smithies
2004 Mary F. Lyon
2003 Pierre Chambon and Ronald M. Evans
2002 Seymour Benzer and Sydney Brenner
2001 and Thomas M. Jessell
2000 H. Robert Horvitz
1999 Martin J. Evans and Richard L. Gardner
1998 Davor Solter
1997 Walter J. Gehring and David S. Hogness
1996 Beatrice Mintz and Ralph L. Brinster
1993 Alex Jones and Vinny Testaverde
1986 Joseph Warkany
See also
List of biology awards
List of medicine awards |
https://en.wikipedia.org/wiki/Infrastructure%20%28number%20theory%29 | In mathematics, an infrastructure is a group-like structure appearing in global fields.
Historic development
In 1972, D. Shanks first discovered the infrastructure of a real quadratic number field and applied his baby-step giant-step algorithm to compute the regulator of such a field in binary operations (for every ), where is the discriminant of the quadratic field; previous methods required binary operations. Ten years later, H. W. Lenstra published a mathematical framework describing the infrastructure of a real quadratic number field in terms of "circular groups". It was also described by R. Schoof and H. C. Williams, and later extended by H. C. Williams, G. W. Dueck and B. K. Schmid to certain cubic number fields of unit rank one and by J. Buchmann and H. C. Williams to all number fields of unit rank one. In his habilitation thesis, J. Buchmann presented a baby-step giant-step algorithm to compute the regulator of a number field of arbitrary unit rank. The first description of infrastructures in number fields of arbitrary unit rank was given by R. Schoof using Arakelov divisors in 2008.
The infrastructure was also described for other global fields, namely for algebraic function fields over finite fields. This was done first by A. Stein and H. G. Zimmer in the case of real hyperelliptic function fields. It was extended to certain cubic function fields of unit rank one by R. Scheidler and A. Stein. In 1999, S. Paulus and H.-G. Rück related the infrastructure of a real quadratic function field to the divisor class group. This connection can be generalized to arbitrary function fields and, combining with R. Schoof's results, to all global fields.
One-dimensional case
Abstract definition
A one-dimensional (abstract) infrastructure consists of a real number , a finite set together with an injective map . The map is often called the distance map.
By interpreting as a circle of circumference and by identifying with , one can see a one-dimensional infr |
https://en.wikipedia.org/wiki/Woodhull%20Freedom%20Foundation | The Woodhull Freedom Foundation, also known as Woodhull Sexual Freedom Alliance, is an American non-profit organization founded in 2003 that advocates for sexual freedom as a fundamental human right. The organization is based in Washington, D.C., United States. Named after an influential member of the American woman's suffrage movement, Victoria Woodhull, its focus includes analyzing groups and individuals that seek to perpetuate a culture of sexual repression.
Sexual Freedom Day, officially recognized in 2011 in Washington, DC, and held every September 23, celebrates the birthday of Victoria Woodhull. The Woodhull Freedom Foundation (WFF) has held the Sexual Freedom Summit annually since 2010. Organization members have included LGBT activist Jeffrey Montgomery, former chairwoman of the United States Commission on Civil Rights Mary Frances Berry, writer Eric Rofes, lawyer Lawrence G. Walters, and activist Dan Massey.
In the furtherance of activities relating to its goals, the organization has allied itself with groups including the Foundation for Individual Rights and Expression, the National Gay and Lesbian Task Force, the American Civil Liberties Union, the Gay and Lesbian Activists Alliance, the Tully Center for Free Speech at Syracuse University, National Coalition Against Censorship, the Heartland Institute, National Association of Scholars, American Booksellers Foundation for Free Expression, Accuracy in Academia, and the American Council of Trustees and Alumni. An academic paper in the Journal of Homosexuality characterized the organization as one "that addresses both international and national sexual freedom issues as well as a host of other health and human rights issues."
History
Foundation: 2003
The organization was founded in 2003 with the name Woodhull Freedom Foundation. It began with a focus on global and domestic human rights, specifically looking at sexual freedom. It is named for Victoria Woodhull (18381927), the first woman to own a company o |
https://en.wikipedia.org/wiki/Percent-encoding | URL encoding, officially known as percent-encoding, is a method to encode arbitrary data in a Uniform Resource Identifier (URI) using only the limited US-ASCII characters legal within a URI. Although it is known as URL encoding, it is also used more generally within the main Uniform Resource Identifier (URI) set, which includes both Uniform Resource Locator (URL) and Uniform Resource Name (URN). As such, it is also used in the preparation of data of the application/x-www-form-urlencoded media type, as is often used in the submission of HTML form data in HTTP requests.
Percent-encoding in a URI
Types of URI characters
The characters allowed in a URI are either reserved or unreserved (or a percent character as part of a percent-encoding). Reserved characters are those characters that sometimes have special meaning. For example, forward slash characters are used to separate different parts of a URL (or more generally, a URI). Unreserved characters have no such meanings. Using percent-encoding, reserved characters are represented using special character sequences. The sets of reserved and unreserved characters and the circumstances under which certain reserved characters have special meaning have changed slightly with each revision of specifications that govern URIs and URI schemes.
Other characters in a URI must be percent-encoded.
Reserved characters
When a character from the reserved set (a "reserved character") has a special meaning (a "reserved purpose") in a certain context, and a URI scheme says that it is necessary to use that character for some other purpose, then the character must be percent-encoded. Percent-encoding a reserved character involves converting the character to its corresponding byte value in ASCII and then representing that value as a pair of hexadecimal digits (if there is a single hex digit, a leading zero is added). The digits, preceded by a percent sign (%) as an escape character, are then used in the URI in place of the reserved chara |
https://en.wikipedia.org/wiki/Toile | Toile (French for "canvas") is a textile fabric comparable to fine batiste with a cloth weave. Natural silk or chemical fiber filaments are usually used as materials. The word toile can refer to the fabric itself or to a test garment sewn from calico. The French term toile entered the English language around the 12th century, was used in the middle ages and meanwhile has disappeared.
Etymology
Middle English toile, from French toile ("cloth"), from Old French teile, from Latin tela ("web"), from Proto-Indo-European *(s)teg ("to cover") (see List of Proto-Indo-European roots in Wiktionary). In Australian and British terminology, a toile is a version of a garment made by a fashion designer or dressmaker to test a pattern. They are usually made of calico, as multiple toiles may be made in the process of perfecting a design. In the United States toiles are sometimes referred to as muslins, because during the Middle Ages they were made from the cheap, unbleached muslin-fabric available in different weights.
Toile de Jouy
The French "Toile de Jouy" simply means "cloth from Jouy" in English and describes a type of fabric printing.
"Toile de Jouy", sometimes abbreviated to simply "toile", is a type of decorating pattern consisting of a white or off-white background on which is a repeated pattern depicting a fairly complex scene, generally of a pastoral theme such as a couple having a picnic by a lake or an arrangement of flowers. The pattern portion consists of a single color, most often black, dark red, or blue. Greens, browns, and magenta toile patterns are less common, but not unheard of. Toile is most associated with fabrics (curtains and upholstery in particular, especially chintz), though toile wallpaper is also popular. Toile can also be used on teapots, beddings, clothing, etc. In upper-class (primarily American, but also northern European) society, toile is often seen on dresses or aprons used at such events as country-themed garden parties or tea parties.
To |
https://en.wikipedia.org/wiki/Delandistrogene%20moxeparvovec | Delandistrogene moxeparvovec, sold under the brand name Elevidys, is a recombinant gene therapy used for the treatment of Duchenne muscular dystrophy. It is designed to deliver into the body a gene that leads to production of Elevidys micro-dystrophin that contains selected domains of the dystrophin protein present in normal muscle cells. It is an adeno-associated virus vector-based gene therapy that is given by injection into a vein.
The most commonly reported side effects include vomiting, nausea, acute liver injury, pyrexia (fever), and thrombocytopenia (abnormally low platelet count in the blood).
Delandistrogene moxeparvovec was approved for medical use in the United States in June 2023. It was developed by Sarepta Therapeutics and is manufactured by Catalent.
Medical uses
Delandistrogene moxeparvovec is indicated for the treatment of ambulatory children aged four through five years of age with Duchenne muscular dystrophy with a confirmed mutation in the DMD gene.
Delandistrogene moxeparvovec is designed to deliver into the body a gene that leads to production of Elevidys micro-dystrophin, a shortened protein (138 kDa, compared to the 427 kDa dystrophin protein of normal muscle cells) that contains selected domains of the dystrophin protein present in normal muscle cells. FDA states that the conditional approval is based on detection of successful gene expression; evidence of clinical improvement is still pending.
History
The accelerated US Food and Drug Administration (FDA) approval of delandistrogene moxeparvovec was based on data from a randomized clinical trial that established that delandistrogene moxeparvovec increased the expression of the Elevidys micro-dystrophin protein observed in delandistrogene moxeparvovec-treated individuals aged four to five years with Duchenne muscular dystrophy.
Society and culture
Economics
Initial pricing was announced at for a single treatment which is expected to last lifetime. |
https://en.wikipedia.org/wiki/Sonar%20signal%20processing | Sonar systems are generally used underwater for range finding and detection. Active sonar emits an acoustic signal, or pulse of sound, into the water. The sound bounces off the target object and returns an “echo” to the sonar transducer. Unlike active sonar, passive sonar does not emit its own signal, which is an advantage for military vessels. But passive sonar cannot measure the range of an object unless it is used in conjunction with other passive listening devices. Multiple passive sonar devices must be used for triangulation of a sound source. No matter whether active sonar or passive sonar, the information included in the reflected signal can not be used without technical signal processing. To extract the useful information from the mixed signal, some steps are taken to transfer the raw acoustic data.
Active Sonar
For active sonar, six steps are needed during the signal processing system.
Signal generation
To generate a signal pulse typical analog implementations are oscillators and voltage controlled oscillators (VCO) which are followed by modulators. Amplitude modulation is used to weight the pulse envelopes and to translate the signal spectrum up to some suitable carrier frequency for transmission.
First, in sonar system, the acoustic pressure field can be represented as . The field function includes four variables: time and spatial coordinate . Thus, according to the Fourier transform, in frequency domain
In the formula is temporal frequency and is spatial frequency.
We often define as elemental signal, for the reason that any 4-D can be generated by taking a linear combination of elemental signals.
Obviously, the direction of gives the direction of propagation of waves, and the speed of the waves is
The wavelength is
Temporal sampling
In modern world, digital computers do contribute a lot to higher speed and efficiency in data analysis. Thus, it is necessary to convert an analog signal into a digital signal by sample the signal in time domai |
https://en.wikipedia.org/wiki/FAM110A | Protein FAM110A, also known as protein family with sequence similarity 110, A, C20orf55 or BA371L19.3 is encoded by the FAM110A gene. FAM110A is located on chromosome 20 and is a part of the greater FAM110 gene family, consisting of FAM110A, FAM110B, and FAM110C.
Gene
Overview
In humans, FAM110A is located on the plus strand at 20p13. The gene transcript is found from base pairs 833,715 to 846,279, with a total transcript length of 12,564 base pairs. The FAM110A mRNA transcript is predicted to contain two exons. An upstream promoter region for FAM110A is predicted to be 1,111 base pairs long. Six different mRNA transcripts of FAM110A are predicted, all differing in their 5' untranslated regions.
Homology
219 organisms have been reported to have orthologs with the human FAM110A gene.
Regulation
Protein
Overview
All human FAM110A transcript variants encode the same protein, which is 295 amino acids in length. The human FAM110A protein is projected to weigh 31.3 kiladaltons and have an isoelectric point of 10.5.
Human FAM110A is predicted to contain one standard deviation less than average frequencies of methionine, asparagine, and isoleucine residues, while containing one standard deviation higher frequencies of serine and proline residues. Human FAM110A is also predicted to contain a frequency of arginine residues two standard deviations higher than average. The presence of a high frequency of arginine residues is also apparent in the FAM110A chimpanzee, mouse, chicken and zebrafish orthologs, indicating that it may play a vital role to the function of the gene due to its high conservation.
FAM110A is predicted to be hydrophilic and soluble.
The tertiary structure of FAM110A is predicted to be 80% disordered.
Post-translational modification
The N-terminal glycine residue FAM110A is not predicted to be myristolated (confidence: 0.97), indicating that FAM110A is not membrane-associated.
It is predicted that FAM110A contains no sulfation of tyrosine residu |
https://en.wikipedia.org/wiki/CD%20player | A CD player is an electronic device that plays audio compact discs, which are a digital optical disc data storage format. CD players were first sold to consumers in 1982. CDs typically contain recordings of audio material such as music or audiobooks. CD players may be part of home stereo systems, car audio systems, personal computers, or portable CD players such as CD boomboxes. Most CD players produce an output signal via a headphone jack or RCA jacks. To use a CD player in a home stereo system, the user connects an RCA cable from the RCA jacks to a hi-fi (or other amplifier) and loudspeakers for listening to music. To listen to music using a CD player with a headphone output jack, the user plugs headphones or earphones into the headphone jack.
Modern units can play audio formats other than the original CD PCM audio coding, such as MP3, AAC and WMA. DJs playing dance music at clubs often use specialized players with an adjustable playback speed to alter the pitch and tempo of the music. Audio engineers using CD players to play music for an event through a sound reinforcement system use professional audio-grade CD players. CD playback functionality is also available on CD-ROM/DVD-ROM drive equipped computers as well as on DVD players and most optical disc-based home video game consoles.
History
American inventor James T. Russell is known for inventing the first system to record digital video information on an optical transparent foil that is lit from behind by a high-power halogen lamp. Russell's patent application was first filed in 1966, and he was granted a patent in 1970. Following litigation, Sony and Philips licensed Russell's recording patents (then held by a Canadian company, Optical Recording Corp.) in the 1980s.
The compact disc is not based on Russell's invention, it is an evolution of LaserDisc technology, where a focused laser beam is used that enables the high information density required for high-quality digital audio signals.
Prototypes were deve |
https://en.wikipedia.org/wiki/Armorial%20of%20Montenegro | This is a list of coats of arms of Montenegro. Most municipalities of Montenegro have their own coat of arms. Many Montenegrin military units and other public agencies and some private families have coats of arms. There are also many historical Montenegrin coat of arms throughout history.
Montenegro
Historical coat of arms
Medieval Period
17th–20th Century
Late-20th Century
Municipalities of Montenegro |
https://en.wikipedia.org/wiki/Making%20Mathematics%20Count | Making Mathematics Count is the title of a report on mathematics education in the United Kingdom (U.K.).
The report was written by Adrian Smith as leader of an "Inquiry into Post–14 Mathematics Education", which was commissioned by the UK Government in 2002. The report recommended an increase in mathematics schooling; the report recommended that statistics be taught as part of the natural sciences rather than as part of the mathematics curriculum.
Inquiry and report
Making Mathematics Count is the title of a report on mathematics education in the United Kingdom (U.K.). The report was written by Adrian Smith as leader of an "Inquiry into Post–14 Mathematics Education", which was commissioned by the UK Government in 2002. The purpose of the Inquiry was:
"To make recommendations on changes to the curriculum, qualifications and
pedagogy for those aged 14 and over in schools, colleges and higher
education institutions to enable those students to acquire the mathematical
knowledge and skills necessary to meet the requirements of employers and
of further and higher education."
Publication of the report was followed two years later by a conference of 241 delegates, who included mathematics teachers, college lecturers, as well as university mathematicians, head teachers, local authority consultants and advisers, and other mathematics professionals. There is a report of the conclusions of this conference, which was intended to bring together policymakers and practitioners to share information and discuss ways in which changes in mathematics education could be implemented to benefit schools, teachers and students.
Influence
The Smith report has influenced debate on U.K. educational policy. A particular concern of the report was where and how statistics should be taught: the report recommended that statistics should be embedded in application subjects and taught by teachers of those subjects where it is applied. The government decision was that statistics teaching should re |
https://en.wikipedia.org/wiki/Plant%20rights | Plant rights are rights to which plants may be entitled. Such issues are often raised in connection with discussions about human rights, animal rights, biocentrism, or sentiocentrism.
Philosophy
Samuel Butler's Erewhon contains a chapter, "The Views of an Erewhonian Philosopher Concerning the Rights of Vegetables".
On the question of whether animal rights can be extended to plants, animal rights philosopher Tom Regan argues that animals acquire rights due to being aware, what he calls "subjects-of-a-life". He argues that this does not apply to plants, and that even if plants did have rights, abstaining from eating meat would still be moral due to the use of plants to rear animals.
According to philosopher Michael Marder, the idea that plants should have rights derives from "plant subjectivity", which is distinct from human personhood. Paul W. Taylor holds that all life has inherent worth and argues for respect for plants, but does not assign them rights. Christopher D. Stone, the son of investigative journalist I. F. Stone, proposed in a 1972 paper titled "Should Trees Have Standing?" that, if corporations are assigned rights, so should natural objects such as trees. Citing the broadening of rights of blacks, Jews, women, and fetuses as examples, Stone explains that, throughout history, societies have been conferring rights to new "entities" which, at the time, people thought to be "unthinkable".
Whilst not appealing directly to "rights", Matthew Hall has argued that plants should be included within the realm of human moral consideration. His Plants as Persons: A Philosophical Botany discusses the moral background of plants in western philosophy and contrasts this with other traditions, including indigenous cultures, which recognise plants as persons—active, intelligent beings that are appropriate recipients of respect and care. Hall backs up his call for the ethical consideration of plants with arguments based on plant neurobiology, which says that plants are a |
https://en.wikipedia.org/wiki/Wirmer%20Flag | The Wirmer Flag (German: Wirmer-Flagge), also known commercially as the flag of German Resistance 20 July or the Stauffenberg flag, is a design by Josef Wirmer. Wirmer was a resistance fighter against the Nazi Regime and part of the 20 July plot. According to his idea, the flag was to become the new flag of Germany after the successful assassination attempt against Hitler and the transfer of power to the conspirators. First discussed by the Parlamentarischer Rat ("Parliamentary Council") in 1948/49 as the federal flag, the design served in modified form as the party flag of the Christian Democratic Union of Germany (CDU) from 1953 until around 1970 and as the model for the Free Democratic Party's (FDP) party symbol. The flag then disappeared from public perception.
In 1999, Reinhold Oberlercher, the thought leader of the Neo-Nazi German College ("Deutsches Kolleg"), declared the flag to be that of the Fourth Reich, aspired to by the college in his revised draft constitution of 9 November 1999, it became the college's flag. The German college also included Horst Mahler, who published the proclamation of the Reichsbürger movement on 14 December 2003. As a result, the flag became popular in the movement, which at the time received little media and social attention and was thus increasingly used by right-wing extremist and populist groups, which were met with criticism from various sides. Due to its high presence at the Dresden Pegida demonstrations, the flag was also referred to as the "Pegida flag" in some media reports.
Origin
Ottfried Neubecker's design
In the booklet Die Reichseinheitsflagge ("The Unity Flag of the Reich"), the vexillologist Ottfried Neubecker proposed in 1926 a flag in the color sequence black-gold-red as a compromise in the dispute over the German national flag ("Flaggenstreit"). He made two proposals for this: firstly, a tricolour in this color sequence, and secondly, a red Nordic Cross, outlined in gold, on a black background. This design a |
https://en.wikipedia.org/wiki/Loop%20perforation | Loop perforation is an approximate computing technique that allows to regularly skip some iterations of a loop.
It relies on one parameter: the perforation rate. The perforation rate can be interpreted as the number of iteration to skip each time or the number of iterations to perform before skipping one.
Variants of loop perforation include those that skip iterations deterministically at regular intervals, those that skip iterations at the beginning or the end of the loop, and those that skip a random sample of iterations. The compiler may select the perforation variant at the compile-time, or include instrumentation that allows the runtime system to adaptively adjust the perforation strategy and perforation rate to satisfy the end-to-end accuracy goal.
Loop perforation techniques were first developed by MIT senior researchers Martin C. Rinard and Stelios Sidiroglou.
Code examples
The examples that follows provide the result of loop perforation applied on this C-like source code
for (int i = 0; i < N; i++) {
// do things
}
Skip n iterations each time
for (int i = 0; i < N; i++) {
// do things
i = i + skip_factor;
}
Skip one iteration after n
int count = 0;
for (int i = 0; i < N; i++) {
if (count == skip_factor) {
count = 0;
} else {
// do things
count++;
}
}
See also
Approximate computing
Task skipping
Memoization
Software optimization |
https://en.wikipedia.org/wiki/Manin%E2%80%93Drinfeld%20theorem | In mathematics, the Manin–Drinfeld theorem, proved by and , states that the difference of two cusps of a modular curve has finite order in the Jacobian variety. |
https://en.wikipedia.org/wiki/Paytm | Paytm (acronym for "pay through mobile") is an Indian multinational financial technology company, that specializes in digital payments and financial services, based in Noida. It was founded in 2010 by Vijay Shekhar Sharma under One97 Communications. The company offers mobile payment services to consumers and enables merchants to receive payments through its QR code, Payment Soundbox, Android based-point of sale machine and online payment gateway offerings. In partnership with financial institutions, Paytm offers financial services such as microloans and buy now, pay later to its consumers and merchants. Apart from bill payments and money transfer, the company also provides ticketing services, retail brokerage products and online games.
Paytm's parent company, One97 Communications, got listed on the Indian stock exchanges on 18 November 2021 after an initial public offering, which was the largest in India at the time. For the fiscal year 2022–23, Paytm's gross merchandise value (GMV) was reported to be .
History
Paytm was founded in August 2010 with an initial investment of by its founder Vijay Shekhar Sharma in Noida, Delhi NCR. It started off as a prepaid mobile and DTH recharge platform, and later added debit card, postpaid mobile and landline bill payments in 2013.
In October 2011, Sapphire Ventures (fka SAP Ventures) invested $10 million in One97 Communications Ltd. By January 2014, the company had launched the Paytm Wallet, which the Indian Railways and Uber added as a payment option. It launched into e-commerce with online deals and bus ticketing. In 2015, it added education fees, metro recharges, electricity, gas, and water bill payments. Paytm's registered user base grew from 1.18 crore in August 2014 to 10.4 crore in August 2015. Its travel business crossed $500 million in annualised GMV run rate, with 20 lakh tickets booked per month.
In March 2015, Paytm received its huge stake from Chinese e-commerce company Alibaba Group, after Ant Financial Servic |
https://en.wikipedia.org/wiki/Gravity%20of%20Earth | The gravity of Earth, denoted by , is the net acceleration that is imparted to objects due to the combined effect of gravitation (from mass distribution within Earth) and the centrifugal force (from the Earth's rotation).
It is a vector quantity, whose direction coincides with a plumb bob and strength or magnitude is given by the norm .
In SI units this acceleration is expressed in metres per second squared (in symbols, m/s2 or m·s−2) or equivalently in newtons per kilogram (N/kg or N·kg−1). Near Earth's surface, the acceleration due to gravity, accurate to 2 significant figures, is . This means that, ignoring the effects of air resistance, the speed of an object falling freely will increase by about per second every second. This quantity is sometimes referred to informally as little (in contrast, the gravitational constant is referred to as big ).
The precise strength of Earth's gravity varies with location. The agreed upon value for is by definition. This quantity is denoted variously as , (though this sometimes means the normal gravity at the equator, ), , or simply (which is also used for the variable local value).
The weight of an object on Earth's surface is the downwards force on that object, given by Newton's second law of motion, or (). Gravitational acceleration contributes to the total gravity acceleration, but other factors, such as the rotation of Earth, also contribute, and, therefore, affect the weight of the object. Gravity does not normally include the gravitational pull of the Moon and Sun, which are accounted for in terms of tidal effects.
Variation in magnitude
A non-rotating perfect sphere of uniform mass density, or whose density varies solely with distance from the centre (spherical symmetry), would produce a gravitational field of uniform magnitude at all points on its surface. The Earth is rotating and is also not spherically symmetric; rather, it is slightly flatter at the poles while bulging at the Equator: an oblate spheroid. |
https://en.wikipedia.org/wiki/Lipid%20A%20acylase | Antimicrobial peptide resistance and lipid A acylation protein PagP
is a family of several bacterial antimicrobial peptide resistance and lipid A acylation (PagP) proteins. The bacterial outer membrane enzyme PagP transfers a palmitate chain from a phospholipid to lipid A. In a number of pathogenic Gram-negative bacteria, PagP confers resistance to certain cationic antimicrobial peptides produced during the host innate immune response. |
https://en.wikipedia.org/wiki/Hanany%E2%80%93Witten%20transition | In theoretical physics the Hanany–Witten transition, also called the Hanany–Witten effect, refers to any process in a superstring theory in which two p-branes cross resulting in the creation or destruction of a third p-brane. A special case of this process was first discovered by Amihay Hanany and Edward Witten in 1996. All other known cases of Hanany–Witten transitions are related to the original case via combinations of S-dualities and T-dualities. This effect can be expanded to string theory, 2 strings cross together resulting in the creation or destruction of a third string.
The original effect
The original Hanany–Witten transition was discovered in type IIB superstring theory in flat, 10-dimensional Minkowski space. They considered a configuration of NS5-branes, D5-branes and D3-branes which today is called a Hanany–Witten brane cartoon. They demonstrated that a subsector of the corresponding open string theory is described by a 3-dimensional Yang–Mills gauge theory. However they found that the string theory space of solutions, called the moduli space, only agreed with the known Yang-Mills moduli space if whenever an NS5-brane and a D5-brane cross, a D3-brane stretched between them is created or destroyed.
They also presented various other arguments in support of their effect, such as a derivation from the worldvolume Wess–Zumino terms. This proof uses the fact that the flux from each brane renders the action of the other brane ill-defined if one does not include the D3-brane.
The S-rule
Furthermore, they discovered the S-rule, which states that in a supersymmetric configuration the number of D3-branes stretched between a D5-brane and an NS5-brane may only be equal to 0 or 1. Then the Hanany-Witten effect implies that after the D5-brane and the NS5-brane cross, if there was a single D3-brane stretched between them it will be destroyed, and if there was not one then one will be created. In other words, there cannot be more than one D3 brane that stre |
https://en.wikipedia.org/wiki/Winemaking%20cooperative | A winemaking cooperative is an agricultural cooperative which is involved in winemaking, and which in a similar way to other cooperatives is owned by its members. The members in a winemaking cooperative are usually vineyard owners, who deliver grapes to the cooperative, which is involved in production of wine from the grapes and the subsequent marketing activities.
Winemaking cooperatives are responsible for a significant proportion of the total wine production in many major wine-producing countries, including most of the classical European wine countries, but their importance varies much between different wine regions within these countries. Cooperatives tend to be more important in regions where the wine’s selling price is relatively low and average size of vineyard holdings is small.
While some winemaking cooperatives were established in the 19th century, the majority were established in the early 1930s following the Great Depression.
Advantages to members
The advantage to members of a cooperative, in comparison to pursuing winemaking and marketing on their own, consists in pooling resources and sharing costs for winemaking and marketing, which call for costly equipment and technical expertise. There are also other financial advantages, including certain European Union subsidies for cooperatives located in EU countries.
Wines from cooperatives are often allowed to be described as producer-bottled according to the wine laws of the country in question, which is sometimes an advantage in marketing. The French term corresponding to this is mis(e) en bouteille à la propriété, while the German is Erzeugerabfüllung.
Alternatives available to wine-growers
Producing and marketing wine on their own is usually not a realistic possibility for many vineyard owners with small holdings. However, being a member of a winemaking cooperative is not the only option available. Selling grapes on the open market, entering into long-term contracts with negociants or other winem |
https://en.wikipedia.org/wiki/Borel%27s%20lemma | In mathematics, Borel's lemma, named after Émile Borel, is an important result used in the theory of asymptotic expansions and partial differential equations.
Statement
Suppose U is an open set in the Euclidean space Rn, and suppose that f0, f1, ... is a sequence of smooth functions on U.
If I is any open interval in R containing 0 (possibly I = R), then there exists a smooth function F(t, x) defined on I×U, such that
for k ≥ 0 and x in U.
Proof
Proofs of Borel's lemma can be found in many text books on analysis, including and , from which the proof below is taken.
Note that it suffices to prove the result for a small interval I = (−ε,ε), since if ψ(t) is a smooth bump function with compact support in (−ε,ε) equal identically to 1 near 0, then ψ(t) ⋅ F(t, x) gives a solution on R × U. Similarly using a smooth partition of unity on Rn subordinate to a covering by open balls with centres at δ⋅Zn, it can be assumed that all the fm have compact support in some fixed closed ball C. For each m, let
where εm is chosen sufficiently small that
for |α| < m. These estimates imply that each sum
is uniformly convergent and hence that
is a smooth function with
By construction
Note: Exactly the same construction can be applied, without the auxiliary space U, to produce a smooth function on the interval I for which the derivatives at 0 form an arbitrary sequence.
See also |
https://en.wikipedia.org/wiki/Equinalysis | Equinalysis is a computer software program designed to capture and analyse equine locomotion by visually tracking and quantifying biomechanical data. The system was developed in 2004 by consultant farrier, Haydn Price with the intent of allowing veterinarians, farriers, horse trainers and physiotherapists to highlight subtle changes in a horse's locomotion and provide a video record of how a horse's movements change during the course of its working life. This then allows the user to improve the horse's performance with various techniques and treatment plans, such as appropriate shoeing regimes.
Operation
For the analysis, polystyrene markers are placed at specific points on the horse's limbs, mainly over the joints. Then the horse is walked and trotted in-hand, and filmed with a video camera from all angles on a hard, flat surface. The information is then collated and downloaded on to a CD or DVD, which is analyzed on a computer by an accredited individual. The specialist software program records the movement of the markers and produces data that can be used to quantify stride length, body symmetry, joint flexion and extension, and soundness. The resulting baseline of facts - which is presented in a hard-copy portfolio of information for future reference - then provides the horse owner with a valuable 'baseline measurement' of movement and soundness.
Reliability
A 2011 study published in the Journal of Equine Veterinary Science found the system did not produce repeatable data from day to day, and was therefore not sufficiently reliable for use in clinical evaluations of equine lameness.
See also
Skeletal system of the horse |
https://en.wikipedia.org/wiki/Proof%20of%20authority | Proof of authority (PoA) is an algorithm used with blockchains that delivers comparatively fast transactions through a consensus mechanism based on identity as a stake. The most notable platforms using PoA are VeChain, Bitgert, Palm Network and Xodex.
Proof-of-authority
In PoA-based networks, transactions and blocks are validated by approved accounts, known as validators. Validators run software allowing them to put transactions in blocks. The process is automated and does not require validators to be constantly monitoring their computers. It, however, does require maintaining the computer (the authority node) uncompromised. The term was coined by Gavin Wood, co-founder of Ethereum and Parity Technologies.
With PoA, individuals earn the right to become validators, so there is an incentive to retain the position that they have gained. By attaching a reputation to identity, validators are incentivized to uphold the transaction process, as they do not wish to have their identities attached to a negative reputation. This is considered more robust than PoS (proof-of-stake) - PoS, while a stake between two parties may be even, it does not take into account each party’s total holdings. This means that incentives can be unbalanced.
On the other hand, PoA only allows non-consecutive block approval from any one validator, meaning that the risk of serious damage is centralized to the authority node.
PoA is suited for both private networks and public networks, like POA Network or Eurus, where trust is distributed. |
https://en.wikipedia.org/wiki/Alachlor | Alachlor is an herbicide from the chloroacetanilide family. It is an odorless, white solid. The greatest use of alachlor is for control of annual grasses and broadleaf weeds in crops. Use of alachlor is illegal in the European Union and no products containing alachlor are currently registered in the United States.
Its mode of action is elongase inhibition, and inhibition of geranylgeranyl pyrophosphate (GGPP) cyclisation enzymes, part of the gibberellin pathway. It is marketed under the trade names Alanex, Bronco, Cannon, Crop Star, Intrro, Lariat, Lasso, Micro-Tech and Partner.
Uses
The largest use of alachlor is as a herbicide for control of annual grasses and broadleaf weeds in crops, primarily on corn, sorghum, and soybeans.
Application details
Alachlor mixes well with other herbicides. It is marketed in mixed formulations with atrazine, glyphosate, trifluralin and imazaquin. It is a selective, systemic herbicide, absorbed by germinating shoots and by roots. Its mode of action is elongase inhibition, and inhibition of geranylgeranyl pyrophosphate (GGPP) cyclisation enzymes, part of the gibberellin pathway. Stated more simply, it works by interfering with a plant's ability to produce protein and by interfering with root growth.
It is most commonly available as microgranules containing 15% active ingredients (AI), or emulsifiable concentrate containing 480 g/ litre of AI. Homologuation in Europe requires a maximum dose of 2,400 g per hectare of AI, or 5 litres/hectare of emulsifiable concentrate or 17 kg/ha of microgranules. The products are applied as either pre-drilling, soil incorporated or pre-emergence.
Safety
The United States Environmental Protection Agency (EPA) classifies the herbicide as toxicity class III - slightly toxic. The Maximum Contaminant Level Goal (MCLG) for Alachlor is zero, to prevent long-term effects. The Maximum Contaminant Level (MCL) for drinking water is two parts per billion (2 ppb).
The EPA cited the following long-term ef |
https://en.wikipedia.org/wiki/S5%20%28modal%20logic%29 | In logic and philosophy, S5 is one of five systems of modal logic proposed by
Clarence Irving Lewis and Cooper Harold Langford in their 1932 book Symbolic Logic.
It is a normal modal logic, and one of the oldest systems of modal logic of any kind. It is formed with propositional calculus formulas and tautologies, and inference apparatus with substitution and modus ponens, but extending the syntax with the modal operator necessarily and its dual possibly .
The axioms of S5
The following makes use of the modal operators ("necessarily") and ("possibly").
S5 is characterized by the axioms:
K: ;
T: ,
and either:
5: ;
or both of the following:
4: , and
B: .
The (5) axiom restricts the accessibility relation of the Kripke frame to be Euclidean, i.e. .
Kripke semantics
In terms of Kripke semantics, S5 is characterized by frames where the accessibility relation is an equivalence relation: it is reflexive, transitive, and symmetric.
Determining the satisfiability of an S5 formula is an NP-complete problem. The hardness proof is trivial, as S5 includes the propositional logic. Membership is proved by showing that any satisfiable formula has a Kripke model where the number of worlds is at most linear in the size of the formula.
Applications
S5 is useful because it avoids superfluous iteration of qualifiers of different kinds. For example, under S5, if X is necessarily, possibly, necessarily, possibly true, then X is possibly true. Unbolded qualifiers before the final "possibly" are pruned in S5. While this is useful for keeping propositions reasonably short, it also might appear counter-intuitive in that, under S5, if something is possibly necessary, then it is necessary.
Alvin Plantinga has argued that this feature of S5 is not, in fact, counter-intuitive. To justify, he reasons that if X is possibly necessary, it is necessary in at least one possible world; hence it is necessary in all possible worlds and thus is true in all possible worlds. Such reasoning |
https://en.wikipedia.org/wiki/Internal%20carotid%20artery | The internal carotid artery (Latin: arteria carotis interna) is an artery in the neck which supplies the anterior circulation of the brain.
In human anatomy, the internal and external carotids arise from the common carotid arteries, where these bifurcate at cervical vertebrae C3 or C4. The internal carotid artery supplies the brain, including the eyes, while the external carotid nourishes other portions of the head, such as the face, scalp, skull, and meninges.
Classification
Terminologia Anatomica in 1998 subdivided the artery into four parts: "cervical", "petrous", "cavernous", and "cerebral". However, in clinical settings, the classification system of the internal carotid artery usually follows the 1996 recommendations by Bouthillier, describing seven anatomical segments of the internal carotid artery, each with a corresponding alphanumeric identifier—C1 cervical, C2 petrous, C3 lacerum, C4 cavernous, C5 clinoid, C6 ophthalmic, and C7 communicating. The Bouthillier nomenclature remains in widespread use by neurosurgeons, neuroradiologists and neurologists.
The segments are subdivided based on anatomical and microsurgical landmarks and surrounding anatomy, more than angiographic appearance of the artery. An alternative embryologic classification system proposed by Pierre Lasjaunias and colleagues is invaluable when it comes to explanation of many internal carotid artery variants. An older clinical classification, based on pioneering work by Fischer, is mainly of historical significance.
The segments of the internal carotid artery are as follows:
Cervical segment, or C1, identical to the commonly used cervical portion
Petrous segment, or C2
Lacerum segment, or C3
C2 and C3 compose the commonly termed petrous portion
Cavernous segment, or C4, almost identical to the commonly used cavernous portion
Clinoid segment, or C5. This segment is not identified in some earlier classifications and lies between the commonly used cavernous portion and cerebral or su |
https://en.wikipedia.org/wiki/Genetic%20ecology | Genetic ecology is the study of the stability and expression of varying genetic material within abiotic mediums. Typically, genetic data is not thought of outside of any organism save for criminal forensics. However, genetic material has the ability to be taken up by various organisms that exist within an abiotic medium through natural transformations that may occur. Thus, this field of study focuses on interaction, exchange, and expression of genetic material that may not be shared by species had they not been in the same environment.
History
E.B. Ford was the first geneticist to begin work in this field of study. E.B. Ford worked mostly during the 1950s and is most noted for his work with Maniola jurtina and published a book entitled Ecological Genetics in 1975. This type of evolutionary biological study was only possible after gel electrophoresis had been designed in 1937. Prior to this, a high throughput method for DNA analysis did not exist. This field of study began to become more popular following the 1980s with the development of polymerase chain reaction (PCR 1985) and poly-acrylamide gel electrophoresis (p. 1967). With this technology, segments of DNA could be sequenced, amplified, and proteins produced using bacterial transformations. The genetic material along with the proteins could be analyzed and more correct phylogenetic trees could be created.
Since E.B. Ford's research, multiple other genetic ecologists have continued study within the field of genetic ecology such as PT Hanford Alina von Thaden, and many others.
Gene transfer
Genetic information may transfer throughout an ecosystem in multiple ways. The first of which, on the smallest scale, being bacterial gene transfer (see bacterial transformations). Bacteria have the ability to exchange DNA. This DNA exchange, or horizontal gene transfer, may provide various species of bacteria with the genetic information they need to survive in an environment. This can help many bacterial species surviv |
https://en.wikipedia.org/wiki/Inverterization | Inverterization is the process of using an inverter-based motor drive intended to operate an electric motor at a variable speed for the sake of improved efficiency. The main job of the inverter is to control the speed and torque of the motor to match work requirements.
Principle
A fixed speed drive can run in two states: on or off. The energy consumption of such a system is higher because it runs at full power even when that is not needed. Fixed speed drives are designed to manage starting at full speed, which does not maximize efficiency. An inverter represents one of the basic components of a variable speed drive (VSD) that includes other components such as a rectifier, an intermediate circuit and a control unit.
The basic work of an inverter within a VSD is simple. A microcontroller integrated with the inverter manages the speed of a motor according to demands. For example, in a refrigerator data such as temperature, humidity and motor speed are gathered through sensors to keep the motor running efficiently. Based on such data, a microcontroller optimizes performance. Motor speed is optimized and keeps the refrigerator’s temperature constant.
Applications
Inverterization has multiple applications in home appliances and industrial drives. Most major home appliances, including refrigerators, washing machines, fans, vacuum cleaners and pumps, are inverterized.
Advantages
Inverterization offers advantages in multiple domains:
Higher durability through smoothly starting engines improves efficiency through an increase in power factor, and improves process control.
Reducing motor speed saves energy. For example, slowing down a fan motor from 100% to 80% when less airflow is needed can save as much as 50% ofn energy use.
Inverterization reduces noise and vibration.
Inverterization has become popular due to the low cost volume production of Intelligent Power Modules (IPMs) and the availability of easy-to-use designs. It is popular to use in electrical app |
https://en.wikipedia.org/wiki/WDDX | WDDX (Web Distributed Data eXchange) is a programming language-, platform- and transport-neutral data interchange mechanism designed to pass data between different environments and different computers.
History
WDDX was created by Simeon Simeonov of Allaire Corporation in 1998, initially for the ColdFusion server environment.
WDDX was open-sourced later that year.
Usage
WDDX is functionally comparable to XML-RPC and WIDL. The specification supports simple data types such as number, string, boolean, etc., and complex aggregates of these in forms such as structures, arrays and recordsets (row/column data, typically coming from database queries).
The data is encoded into XML using an XML 1.0 DTD, producing a platform-independent but relatively bulky representation. The XML-encoded data can then be sent to another computer using HTTP, FTP, or other transmission mechanism. The receiving computer must have WDDX-aware software to translate the encoded data into the receiver's native data representation. WDDX can also be used to serialize data structures to storage (file system or database). Many applications use WDDX to pass complex data to browsers where it can be manipulated with JavaScript, as an alternative to JSON.
Example from php.net:
<wddxPacket version='1.0'>
<header comment='PHP'/>
<data>
<struct>
<var name='pi'>
<number>3.1415926</number>
</var>
<var name='cities'>
<array length='3'>
<string>Austin</string>
<string>Novato</string>
<string>Seattle</string>
</array>
</var>
</struct>
</data>
</wddxPacket>
Adoption
WDDX is mainly used by ColdFusion and, as February 2022, still supported by Adobe.
Outside ColdFusion, libraries exist to read or write this format, Ruby, Python, PHP, Java, C++, .NET, Actionscript, lisp, Haskell, Perl.
PHP used to offer a comprehensive support for WDDX, which could be used as a format to store session information until the version 7.4. |
https://en.wikipedia.org/wiki/Large%20countable%20ordinal | In the mathematical discipline of set theory, there are many ways of describing specific countable ordinals. The smallest ones can be usefully and non-circularly expressed in terms of their Cantor normal forms. Beyond that, many ordinals of relevance to proof theory still have computable ordinal notations (see ordinal analysis). However, it is not possible to decide effectively whether a given putative ordinal notation is a notation or not (for reasons somewhat analogous to the unsolvability of the halting problem); various more-concrete ways of defining ordinals that definitely have notations are available.
Since there are only countably many notations, all ordinals with notations are exhausted well below the first uncountable ordinal ω1; their supremum is called Church–Kleene ω1 or ω (not to be confused with the first uncountable ordinal, ω1), described below. Ordinal numbers below ω are the recursive ordinals (see below). Countable ordinals larger than this may still be defined, but do not have notations.
Due to the focus on countable ordinals, ordinal arithmetic is used throughout, except where otherwise noted. The ordinals described here are not as large as the ones described in large cardinals, but they are large among those that have constructive notations (descriptions). Larger and larger ordinals can be defined, but they become more and more difficult to describe.
Generalities on recursive ordinals
Ordinal notations
Recursive ordinals (or computable ordinals) are certain countable ordinals: loosely speaking those represented by a computable function. There are several equivalent definitions of this: the simplest is to say that a computable ordinal is the order-type of some recursive (i.e., computable) well-ordering of the natural numbers; so, essentially, an ordinal is recursive when we can present the set of smaller ordinals in such a way that a computer (Turing machine, say) can manipulate them (and, essentially, compare them).
A different defin |
https://en.wikipedia.org/wiki/ISO/IEC%207816 | ISO/IEC 7816 is an international standard related to electronic identification cards with contacts, especially smart cards, and more recently, contactless mobile devices, managed jointly by the International Organization for Standardization (ISO) and the International Electrotechnical Commission (IEC).
It is developed by ISO/IEC JTC 1 (Joint Technical Committee 1) / SC 17 (Subcommittee 17).
The following describes the different parts of this standard.
Note: abstracts and dates, when present, are mere quotations from the ISO website, and are neither guaranteed at the time of edition nor in the future.
Parts
ISO/IEC 7816-1:2011 Part 1: Cards with contacts—Physical characteristics
ISO/IEC 7816-2:2007 Part 2: Cards with contacts—Dimensions and location of the contacts
ISO/IEC 7816-3:2006 Part 3: Cards with contacts—Electrical interface and transmission protocols
ISO/IEC 7816-4:2013 Part 4: Organization, security and commands for interchange
ISO/IEC 7816-5:2004 Part 5: Registration of application providers
ISO/IEC 7816-6:2016 Part 6: Interindustry data elements for interchange
ISO/IEC 7816-7:1999 Part 7: Interindustry commands for Structured Card Query Language (SCQL)
ISO/IEC 7816-8:2016 Part 8: Commands and mechanisms for security operations
ISO/IEC 7816-9:2017 Part 9: Commands for card management
ISO/IEC 7816-10:1999 Part 10: Electronic signals and answer to reset for synchronous cards
ISO/IEC 7816-11:2017 Part 11: Personal verification through biometric methods
ISO/IEC 7816-12:2005 Part 12: Cards with contacts—USB electrical interface and operating procedures
ISO/IEC 7816-13:2007 Part 13: Commands for application management in a multi-application environment
ISO/IEC 7816-15:2016 Part 15: Cryptographic information application
7816-1: Cards with contacts — Physical characteristics
Created in 1987, updated in 1998, amended in 2003, updated in 2011.
This part describes the physical characteristics of the card, primarily by reference to ISO/IEC 7810 I |
https://en.wikipedia.org/wiki/Splint%20%28programming%20tool%29 | Splint, short for Secure Programming Lint, is a programming tool for statically checking C programs for security vulnerabilities and coding mistakes. Formerly called LCLint, it is a modern version of the Unix lint tool.
Splint has the ability to interpret special annotations to the source code, which gives it stronger checking than is possible just by looking at the source alone. Splint is used by gpsd as part of an effort to design for zero defects.
Splint is free software released under the terms of the GNU General Public License.
Main development activity on Splint stopped in 2010. According to the CVS at SourceForge, as of September 2012 the most recent change in the repository was in November 2010. A Git repository at GitHub has more recent changes, starting in July 2019.
Example
#include <stdio.h>
int main()
{
char c;
while (c != 'x');
{
c = getchar();
if (c = 'x')
return 0;
switch (c) {
case '\n':
case '\r':
printf("Newline\n");
default:
printf("%c",c);
}
}
return 0;
}
Splint's output:
<nowiki>
Variable c used before definition
Suspected infinite loop. No value used in loop test (c) is modified by test or loop body.
Assignment of int to char: c = getchar()
Test expression for if is assignment expression: c = 'x'
Test expression for if not boolean, type char: c = 'x'
Fall through case (no preceding break)
</nowiki>
Fixed source:
#include <stdio.h>
int main()
{
int c = 0; // Added an initial assignment definition.
while (c != 'x') {
c = getchar(); // Corrected type of c to int
if (c == 'x') // Fixed the assignment error to make it a comparison operator.
return 0;
switch (c) {
case '\n':
case '\r':
printf("Newline\n");
break; // Added break statement to prevent fall-through.
default:
printf("%c",c);
break; //Added break stateme |
https://en.wikipedia.org/wiki/Vectorial%20addition%20chain | In mathematics, for positive integers k and s, a vectorial addition chain is a sequence V of k-dimensional vectors of nonnegative integers vi for −k + 1 ≤ i ≤ s together with a sequence w,
such that
⋮
⋮
vi =vj+vr for all 1≤i≤s with -k+1≤j, r≤i-1
vs = [n0,...,nk-1]
w = (w1,...ws), wi=(j,r).
For example, a vectorial addition chain for [22,18,3] is
V=([1,0,0],[0,1,0],[0,0,1],[1,1,0],[2,2,0],[4,4,0],[5,4,0],[10,8,0],[11,9,0],[11,9,1],[22,18,2],[22,18,3])
w=((-2,-1),(1,1),(2,2),(-2,3),(4,4),(1,5),(0,6),(7,7),(0,8))
Vectorial addition chains are well suited to perform multi-exponentiation:
Input: Elements x0,...,xk-1 of an abelian group G and a vectorial addition chain of dimension k computing [n0,...,nk-1]
Output:The element x0n0...xk-1nr-1
for i =-k+1 to 0 do yi → xi+k-1
for i = 1 to s do yi →yj×yr
return ys
Addition sequence
An addition sequence for the set of integer S ={n0, ..., nr-1} is an addition chain v that contains every element of S.
For example, an addition sequence computing
{47,117,343,499}
is
(1,2,4,8,10,11,18,36,47,55,91,109,117,226,343,434,489,499).
It's possible to find addition sequence from vectorial addition chains and vice versa, so they are in a sense dual.
See also
Addition chain
Addition-chain exponentiation
Exponentiation by squaring
Non-adjacent form |
https://en.wikipedia.org/wiki/Body%20image%20%28neuroscience%29 | Body image is a complex construct, often used in the clinical context of describing a patient's cognitive perception of their own body. The medical concept began with the work of the Austrian neuropsychiatrist and psychoanalyst Paul Schilder, described in his book The Image and Appearance of the Human Body first published in 1935. The term “body image” was officially introduced by Schilder himself and his widely used definition is: “body image is the picture of our own body we form in our mind, that is to say the way in which the body appears to ourselves”. In research with the term “body image” we currently refer to a conscious mental representation of one’s own body, which involves affects, attitudes, perceptual components and cognition. On the contrary the term “body schema” was initially used to describe an unconscious body mental representation fundamental for action. Keizer and colleagues (2013) suggest the following definition: “[body schema is] an unconscious, sensorimotor, representation of the body that is invoked in action. In light of recent scientific developments regarding the multisensory integration of body sensations, the distinction between body image and body schema appears simplistic and probably no longer useful for scientific research and clinical purposes.
Clinical significance
In the clinical setting, body image disturbances are relatively frequent and involve both psychiatric and neurological disorders. Disturbances in the perception of one's body are present in psychiatric disorders such as:
anorexia nervosa
bulimia nervosa
binge eating disorder
psychotic spectrum disorders
body dysmorphic disorder
body integrity dysphoria (not included in DSM-5).
Cotard delusion
Body image disorders are common in eating disorders and are referred to as "body image disturbance.
Disturbances in the body image are also present in neurological conditions such as:
somatoparaphrenia
unilateral neglect
Alice in Wonderland syndrome
See also
Bo |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.