source
stringlengths
31
227
text
stringlengths
9
2k
https://en.wikipedia.org/wiki/RS-422
RS-422, also known as TIA/EIA-422, is a technical standard originated by the Electronic Industries Alliance, first issued in 1975, that specifies electrical characteristics of a digital signaling circuit. It was meant to be the foundation of a suite of standards that would replace the older RS-232C standard with standards that offered much higher speed, better immunity from noise, and longer cable lengths. RS-422 systems can transmit data at rates as high as 10 Mbit/s, or may be sent on cables as long as at lower rates. It is closely related to RS-423, which uses the same signaling systems but on a different wiring arrangement. RS-422 specifies differential signaling, with every data line paired with a dedicated return line. It is the voltage difference between these two lines that define the mark and space, rather than, as in RS-232, the difference in voltage between a data line and a local ground. As the ground voltage can differ at either end of the cable, this required RS-232 to use signals with voltage magnitudes greater than 5 volts. Moving to dedicated return lines and always defining ground in reference to the sender allows RS-422 to use 0.4 V, allowing it to run at much higher speeds. RS-423 differs primarily in that it has a single return pin instead of one for each data pin. Standard scope RS-422 is the common short form title of American National Standards Institute (ANSI) standard ANSI/TIA/EIA-422-B Electrical Characteristics of Balanced Voltage Differential Interface Circuits and its international equivalent ITU-T Recommendation T-REC-V.11, also known as X.27. These technical standards specify the electrical characteristics of the balanced voltage digital interface circuit. RS-422 provides for data transmission, using balanced, or differential, signaling, with unidirectional/non-reversible, terminated or non-terminated transmission lines, point to point, or multi-drop. In contrast to EIA-485, RS-422/V.11 does not allow multiple drivers but only m
https://en.wikipedia.org/wiki/Fatou%E2%80%93Bieberbach%20domain
In mathematics, a Fatou–Bieberbach domain is a proper subdomain of , biholomorphically equivalent to . That is, an open set is called a Fatou–Bieberbach domain if there exists a bijective holomorphic function whose inverse function is holomorphic. It is well-known that the inverse can not be polynomial. History As a consequence of the Riemann mapping theorem, there are no Fatou–Bieberbach domains in the case n = 1. Pierre Fatou and Ludwig Bieberbach first explored such domains in higher dimensions in the 1920s, hence the name given to them later. Since the 1980s, Fatou–Bieberbach domains have again become the subject of mathematical research.
https://en.wikipedia.org/wiki/Dry%20needling
Dry needling, also known as trigger point dry needling and intramuscular stimulation, is a treatment technique used by various healthcare practitioners, including physical therapists, physicians, and chiropractors, among others. Acupuncturists usually maintain that dry needling is adapted from acupuncture, but others consider dry needling as a variation of trigger point injections. It involves the use of either solid filiform needles or hollow-core hypodermic needles for therapy of muscle pain, including pain related to myofascial pain syndrome. Dry needling is mainly used to treat myofascial trigger points, but it is also used to target connective tissue, neural ailments, and muscular ailments. The American Physical Therapy Association defines dry needling as a technique used to treat dysfunction of skeletal muscle and connective tissue, minimize pain, and improve or regulate structural or functional damage. There is conflicting evidence regarding the effectiveness of dry needling. Some results suggest that it is an effective treatment for certain kinds of muscle pain, while other studies have shown no benefit compared to a placebo; however, not enough high-quality, long-term, and large-scale studies have been done on the technique to draw clear conclusions about its efficacy. Currently, dry needling is being practiced in the United States, Canada, Europe, Australia, and other parts of the world. Origin Etymology and terminology The origin of the term dry needling is attributed to Janet G. Travell. In her 1983 book, Myofascial Pain and Dysfunction: Trigger Point Manual, Travell uses the term dry needling to differentiate between two hypodermic needle techniques when performing trigger point therapy. However, Travell did not elaborate on the details on the techniques of dry needling; the current techniques of dry needling were based on the traditional and western medical acupuncture. Initial techniques The two techniques Travell described are the injection of a
https://en.wikipedia.org/wiki/Oliver%20E.%20Buckley%20Prize
The Oliver E. Buckley Condensed Matter Prize is an annual award given by the American Physical Society "to recognize and encourage outstanding theoretical or experimental contributions to condensed matter physics." It was endowed by AT&T Bell Laboratories as a means of recognizing outstanding scientific work. The prize is named in honor of Oliver Ellsworth Buckley, a former president of Bell Labs. Before 1982, it was known as the Oliver E. Buckley Solid State Prize. It is one of the most prestigious awards in the field of condensed matter physics. The prize is normally awarded to one person but may be shared if multiple recipients contributed to the same accomplishments. Nominations are active for three years. The prize was endowed in 1952 and first awarded in 1953. Since 2012, the prize has been co-sponsored by HTC-VIA Group. Recipients See also List of physics awards
https://en.wikipedia.org/wiki/Evolutionarily%20stable%20state
A population can be described as being in an evolutionarily stable state when that population's "genetic composition is restored by selection after a disturbance, provided the disturbance is not too large" (Maynard Smith, 1982). This population as a whole can be either monomorphic or polymorphic. This is now referred to as convergent stability. History & connection to evolutionary stable strategy While related to the concept of an evolutionarily stable strategy (ESS), evolutionarily stable states are not identical and the two terms cannot be used interchangeably. An ESS is a strategy that, if adopted by all individuals within a population, cannot be invaded by alternative or mutant strategies. This strategy becomes fixed in the population because alternatives provide no fitness benefit that would be selected for. In comparison, an evolutionarily stable state describes a population that returns as a whole to its previous composition even after being disturbed. In short: the ESS refers to the strategy itself, uninterrupted and supported through natural selection, while the evolutionarily stable state refers more broadly to a population-wide balance of one or more strategies that may be subjected to temporary change. The term ESS was first used by John Maynard Smith in an essay from the 1972 book On Evolution. Maynard Smith developed the ESS drawing in part from game theory and Hamilton’s work on the evolution of sex ratio. The ESS was later expanded upon in his book Evolution and the Theory of Games in 1982, which also discussed the evolutionarily stable state. Mixed v. single strategies There has been variation in how the term is used and exploration of under what conditions an evolutionarily stable state might exist. In 1984, Benhard Thomas compared “discrete” models in which all individuals use only one strategy to “continuous” models in which individuals employ mixed strategies. While Maynard Smith had originally defined an ESS as being a single “uninvadable
https://en.wikipedia.org/wiki/Berkeley%20SETI%20Research%20Center
The Berkeley SETI Research Center (BSRC) conducts experiments searching for optical and electromagnetic transmissions from intelligent extraterrestrial civilizations. The center is based at the University of California, Berkeley. The Berkeley SETI Research Center has several SETI searches operating at various wavelengths, from radio, through infrared, to visible light. These include SERENDIP, SEVENDIP, NIROSETI, Breakthrough Listen, and SETI@home. The research center is also involved in the development of new telescopes and instrumentation. The Berkeley SETI Research Center is independent of, but collaborates with, researchers at the SETI Institute. No unambiguous signals from extraterrestrial intelligence have been found. Breakthrough Listen The Berkeley SETI Research Center also hosts the Breakthrough Listen program, which is a ten-year initiative with $100 million funding begun in July 2015 to actively search for intelligent extraterrestrial communications in the universe, in a substantially expanded way, using resources that had not previously been extensively used for the purpose. It has been described as the most comprehensive search for alien communications to date. Announced in July 2015, the project is observing for thousands of hours every year on two major radio telescopes, the Green Bank Observatory in West Virginia, the Parkes Observatory in Australia, and the Automated Planet Finder telescope. SETI@home The center also created the SETI@home, an Internet-based public volunteer computing project employing the BOINC software platform, hosted by their Space Sciences Laboratory. Its purpose is to analyze radio data from radio telescopes for signs of extraterrestrial intelligence. SERENDIP The SERENDIP program takes advantage of ongoing "mainstream" radio telescope observations and analyzes deep space radio telescope data that it obtains while other astronomers are using the telescope. SERENDIP observations have been conducted at frequencies bet
https://en.wikipedia.org/wiki/KCNG4
Potassium voltage-gated channel subfamily G member 4 is a protein that in humans is encoded by the KCNG4 gene. The protein encoded by this gene is a voltage-gated potassium channel subunit.
https://en.wikipedia.org/wiki/Localized%20disease
A localized disease is an infectious or neoplastic process that originates in and is confined to one organ system or general area in the body, such as a sprained ankle, a boil on the hand, an abscess of finger. A localized cancer that has not extended beyond the margins of the organ involved can also be described as localized disease, while cancers that extend into other tissues are described as invasive. Tumors that are non-hematologic in origin but extend into the bloodstream or lymphatic system are known as metastatic. Localized diseases are contrasted with disseminated diseases and systemic diseases. Some diseases are capable of changing from local to disseminated diseases. Pneumonia, for example, is generally confined to one or both lungs but can become disseminated through sepsis, in which the microorganism responsible for the pneumonia "seeds" the bloodstream or lymphatic system and is transported to distant sites in the body. When that occurs, the process is no longer described as a localized disease, but rather as a disseminated disease. See also Disease Nosology
https://en.wikipedia.org/wiki/Explicit%20Congestion%20Notification
Explicit Congestion Notification (ECN) is an extension to the Internet Protocol and to the Transmission Control Protocol and is defined in RFC 3168 (2001). ECN allows end-to-end notification of network congestion without dropping packets. ECN is an optional feature that may be used between two ECN-enabled endpoints when the underlying network infrastructure also supports it. Conventionally, TCP/IP networks signal congestion by dropping packets. When ECN is successfully negotiated, an ECN-aware router may set a mark in the IP header instead of dropping a packet in order to signal impending congestion. The receiver of the packet echoes the congestion indication to the sender, which reduces its transmission rate as if it detected a dropped packet. Rather than responding properly or ignoring the bits, some outdated or faulty network equipment has historically dropped or mangled packets that have ECN bits set. , measurements suggested that the fraction of web servers on the public Internet for which setting ECN prevents network connections had been reduced to less than 1%. Passive support has existed in Ubuntu Linux since 12.04 and in Windows Server since 2012. Passive support in the most popular websites has increased from 8.5% in 2012 to over 70% in May 2017. Adoption across the Internet now requires clients to actively request ECN. In June 2015, Apple announced that ECN will be enabled by default on its supported and future products, to help drive the adoption of ECN signaling industry-wide. Operation ECN requires specific support at both the Internet layer and the transport layer for the following reasons: In TCP/IP, routers operate within the Internet layer, while the transmission rate is handled by the endpoints at the transport layer. Congestion may be handled only by the transmitter, but since it is known to have happened only after a packet was sent, there must be an echo of the congestion indication by the receiver to the transmitter. Without ECN, cong
https://en.wikipedia.org/wiki/Pseudo%20Jacobi%20polynomials
In mathematics, the term Pseudo Jacobi polynomials was introduced by Lesky for one of three finite sequences of orthogonal polynomials y. Since they form an orthogonal subset of Routh polynomials it seems consistent to refer to them as Romanovski-Routh polynomials, by analogy with the terms Romanovski-Bessel and Romanovski-Jacobi used by Lesky. As shown by Askey for two other sequencesth is finite sequence orthogonal polynomials of can be expressed in terms of Jacobi polynomials of imaginary argument. In following Raposo et al. they are often referred to simply as Romanovski polynomials.
https://en.wikipedia.org/wiki/Retrial%20queue
In queueing theory, a discipline within the mathematical theory of probability, a retrial queue is a model of a system with finite capacity, where jobs which arrive and find the system busy wait for some time before trying again to enter the system. Examples of such systems include making restaurant reservations and packet switching networks.
https://en.wikipedia.org/wiki/Nuclear%20magnetic%20moment
The nuclear magnetic moment is the magnetic moment of an atomic nucleus and arises from the spin of the protons and neutrons. It is mainly a magnetic dipole moment; the quadrupole moment does cause some small shifts in the hyperfine structure as well. All nuclei that have nonzero spin also possess a nonzero magnetic moment and vice versa, although the connection between the two quantities is not straightforward or easy to calculate. The nuclear magnetic moment varies from isotope to isotope of an element. For a nucleus of which the numbers of protons and of neutrons are both even in its ground state (i.e. lowest energy state), the nuclear spin and magnetic moment are both always zero. In cases with odd numbers of either or both protons and neutrons, the nucleus often has nonzero spin and magnetic moment. The nuclear magnetic moment is not sum of nucleon magnetic moments, this property being assigned to the tensorial character of the nuclear force, such as in the case of the most simple nucleus where both proton and neutron appear, namely deuterium nucleus, deuteron. Measurement methods The methods for measuring nuclear magnetic moments can be divided into two broad groups in regard to the interaction with internal or external applied fields. Generally the methods based on external fields are more accurate. Different experimental techniques are designed in order to measure nuclear magnetic moments of a specific nuclear state. For instance, the following techniques are aimed to measure magnetic moments of an associated nuclear state in a range of life-times τ: Nuclear Magnetic Resonance (NMR) ms. Time Differential Perturbed Angular Distribution (TDPAD) s. Perturbed Angular Correlation (PAC) ns. Time Differential Recoil Into Vacuum (TDRIV) ps. Recoil Into Vacuum (RIV) ns. Transient Field (TF) ns. Techniques as Transient Field have allowed measuring the g factor in nuclear states with life-times of few ps or less. Shell model According to the shell mode
https://en.wikipedia.org/wiki/Nebulette
Nebulette is a cardiac-specific isoform belonging to the nebulin family of proteins. It is encoded by the NEBL gene. This family is composed of 5 members: nebulette, nebulin, N-RAP, LASP-1 and LASP-2. Nebulette localizes to Z-discs of cardiac muscle and appears to regulate the length of actin thin filaments. Structure Nebulette is a 116.4 kDa protein composed of 1014 amino acids. As a member of the nebulin family of proteins, nebulette is characterized by 35 amino acid stretches of ‘‘nebulin repeats’’, which are actin binding domains containing a conserved SDxxYK motif. Like nebulin, nebulette has an acidic region with unknown structure at its N-terminus, and a serine-rich region adjacent to an SH3 domain at its C-terminus. Though nebulette shares structural features with nebulin, nebulin is expressed preferentially in skeletal muscle and has an enormous size (600-900 kDa), while nebulette is expressed in cardiac muscle at Z-disc regions and is significantly smaller (roughly 1/6 of the size). Nebulette interacts with actin, tropomyosin, alpha-actinin. Xin, and XIRP2. Function Nebulette was identified in 1995 by Moncman and Wang using primary cultures of chicken embryonic cardiomyocytes by immunoprecipitations with certain anti-nebulin monoclonal antibodies. Normal expression of nebulette is essential for the assembly and contractile function of myofibrils. Specifically, nebulette appears to regulate the stability and length of actin thin filaments, as well as beating frequencies of cardiomyocytes; reduction of full-length nebulette protein in cardiomyocytes resulted in reduced thin filament lengths, depressed beating frequencies and loss of thin filament regulatory proteins troponin I and tropomyosin. Clinical significance Mutations in the NEBL gene have been associated with dilated cardiomyopathy. Studies in transgenic mice have supported their causative role in endocardial fibroelastosis and dilated cardiomyopathy. Further reading
https://en.wikipedia.org/wiki/Food%20and%20Drug%20Administration%20%28Myanmar%29
The Department of Food and Drug Administration (; abbreviated FDA) is Burma's food safety regulatory body, which oversees the safety and quality of food, drugs, medical devices and cosmetics. FDA was established in 1995. The agency was established under the 1992 National Drug Law, which also established the Food and Drug Board of Authority, which regulates and controls the manufacture, import, export, storage, distribution and sale of food and drugs, in the interests of public safety. See also Ministry of Health (Burma)
https://en.wikipedia.org/wiki/Doxing
Doxing or doxxing is the act of publicly providing personally identifiable information about an individual or organization, usually via the Internet. Historically, the term has been used interchangeably to refer to both the aggregation of this information from public databases and social media websites (like Facebook), as well as the publication of previously private information obtained through criminal or otherwise fraudulent means (such as hacking and social engineering). The aggregation and provision of previously published material is generally a legal practice, though it may be subject to laws concerning stalking and intimidation. Doxing may be carried out for reasons such as online shaming, extortion, and vigilante aid to law enforcement. It also may be associated with hacktivism. Etymology "Doxing" is a neologism. It originates from a spelling alteration of the abbreviation "docs", for "documents", and refers to "compiling and releasing a dossier of personal information on someone". Essentially, doxing is revealing and publicizing the records of an individual, which were previously private or difficult to obtain. The term dox derives from the slang "dropping dox", which, according to a contributor to Wired, Mat Honan, was "an old-school revenge tactic that emerged from hacker culture in 1990s". Hackers operating outside the law in that era used the breach of an opponent's anonymity as a means to expose opponents to harassment or legal repercussions. Consequently, doxing often comes with a negative connotation because it can be a means of revenge via the violation of privacy. History The practice of publishing personal information about individuals as a form of vigilantism predates the Internet, via physical media such as newspapers and pamphlets. For example, in response to the Stamp Act 1765 in the Thirteen Colonies, radical groups such as the Sons of Liberty harassed tax collectors and those who did not comply with boycotts on British goods, by publi
https://en.wikipedia.org/wiki/Bone%20morphogenetic%20protein%204
Bone morphogenetic protein 4 is a protein that in humans is encoded by BMP4 gene. BMP4 is found on chromosome 14q22-q23. BMP4 is a member of the bone morphogenetic protein family which is part of the transforming growth factor-beta superfamily. The superfamily includes large families of growth and differentiation factors. BMP4 is highly conserved evolutionarily. BMP4 is found in early embryonic development in the ventral marginal zone and in the eye, heart blood and otic vesicle. Discovery Bone morphogenetic proteins were originally identified by an ability of demineralized bone extract to induce endochondral osteogenesis in vivo in an extraskeletal site. Function BMP4 is a polypeptide belonging to the TGF-β superfamily of proteins. It, like other bone morphogenetic proteins, is involved in bone and cartilage development, specifically tooth and limb development and fracture repair. This particular family member plays an important role in the onset of endochondral bone formation in humans. It has been shown to be involved in muscle development, bone mineralization, and ureteric bud development. BMP4 stimulates differentiation of overlying ectodermal tissue. Bone morphogenetic proteins are known to stimulate bone formation in adult animals. This is thought that inducing osteoblastic commitment and differentiation of stem cells such as mesenchymal stem cells.BMPs are known to play a large role in embryonic development. In the embryo BMP4 helps establish dorsal-ventral axis formation in Xenopus frog through inducing ventral mesoderm. In mice targeted inactivation of BMP4 disrupts mesoderm from forming. As well establishes dorsal-ventral patterning of the developing neural tube with the help of BMP7, and inducing dorsal characters. BMP4 also limits the extent to which neural differentiation in Xenopus embryos occurs by inducing epidermis formation rather than neural tissue. They can aid in inducing the lateral characteristics in somites. Somites are required fo
https://en.wikipedia.org/wiki/Hemoglobin%20Hopkins-2
Hemoglobin Hopkins-2 (Hb Hop-2) is a mutation of the protein hemoglobin, which is responsible for the transportation of oxygen through the blood from the lungs to the musculature of the body in vertebrates. The specific mutation in Hemoglobin Hopkins-2 results in two abnormal α chains (human hemoglobin consist of 2 α and 2 β polypeptides usually termed chains). The mutation is the result of histidine 112 being replaced with aspartic acid in the protein's polypeptide sequence. Additionally, within one of the mutated alpha chains, there are substitutes at 114 and 118, two points on the amino acid chain. This mutation can cause sickle cell anemia. Following the initial discovery of hemoglobin, two researchers working at Johns Hopkins Hospital in the mid-twentieth century, Ernest W. Smith and J.V. Torbert, discovered the Hopkins-2 mutation of hemoglobin. Work by Harvey A. Itano and Elizabeth A. Robinson in 1960 confirmed Smith's and Torbert's finding and emphasized the importance of the alpha loci in the mutation. Later in the twentieth century, Samuel Charache, another Hopkins affiliated scientist and doctor, studied the physiological impacts of the variant on health. His findings suggest that the variant plays no effect clinically. History In the mid-1900s, many factors prompted hemoglobin research in Baltimore, Maryland and ultimately led to the discovery of Hemoglobin Hopkins-2. First, the development of new technology, including x-ray crystallography and protein chemistry, that could be utilized in molecular biology studies catalyzed research. Furthermore, the large presence of thalassemia (Hb H), a disorder in which the alpha gene is dysfunctional, in Southeast Asia and southern China further concerned researchers as, if left untreated, the mutation could result in bone deformities, swelling of the spleen, slowed growth rate, or cardiac dysfunction. Additionally, Max Perutz's, a Cambridge researcher, discovery of hemoglobin's basic tertiary structure in
https://en.wikipedia.org/wiki/F-ratio%20%28oceanography%29
In oceanic biogeochemistry, the f-ratio is the fraction of total primary production fuelled by nitrate (as opposed to that fuelled by other nitrogen compounds such as ammonium). The ratio was originally defined by Richard Eppley and Bruce Peterson in one of the first papers estimating global oceanic production. This fraction was originally believed significant because it appeared to directly relate to the sinking (export) flux of organic marine snow from the surface ocean by the biological pump. However, this interpretation relied on the assumption of a strong depth-partitioning of a parallel process, nitrification, that more recent measurements has questioned. Overview Gravitational sinking of organisms (or the remains of organisms) transfers particulate organic carbon from the surface waters of the ocean to its deep interior. This process is known as the biological pump, and quantifying it is of interest to scientists because it is an important aspect of the Earth's carbon cycle. Essentially, this is because carbon transported to the deep ocean is isolated from the atmosphere, allowing the ocean to act as a reservoir of carbon. This biological mechanism is accompanied by a physico-chemical mechanism known as the solubility pump which also acts to transfer carbon to the ocean's deep interior. Measuring the flux of sinking material (so-called marine snow) is usually done by deploying sediment traps which intercept and store material as it sinks down the water column. However, this is a relatively difficult process, since traps can be awkward to deploy or recover, and they must be left in situ over a long period to integrate the sinking flux. Furthermore, they are known to experience biases and to integrate horizontal as well as vertical fluxes because of water currents. For this reason, scientists are interested in ocean properties that can be more easily measured, and that act as a proxy for the sinking flux. The f-ratio is one such proxy. "New" and "rege
https://en.wikipedia.org/wiki/Fundamental%20theorem%20of%20asset%20pricing
The fundamental theorems of asset pricing (also: of arbitrage, of finance), in both financial economics and mathematical finance, provide necessary and sufficient conditions for a market to be arbitrage-free, and for a market to be complete. An arbitrage opportunity is a way of making money with no initial investment without any possibility of loss. Though arbitrage opportunities do exist briefly in real life, it has been said that any sensible market model must avoid this type of profit. The first theorem is important in that it ensures a fundamental property of market models. Completeness is a common property of market models (for instance the Black–Scholes model). A complete market is one in which every contingent claim can be replicated. Though this property is common in models, it is not always considered desirable or realistic. Discrete markets In a discrete (i.e. finite state) market, the following hold: The First Fundamental Theorem of Asset Pricing: A discrete market on a discrete probability space is arbitrage-free if, and only if, there exists at least one risk neutral probability measure that is equivalent to the original probability measure, P. The Second Fundamental Theorem of Asset Pricing: An arbitrage-free market (S,B) consisting of a collection of stocks S and a risk-free bond B is complete if and only if there exists a unique risk-neutral measure that is equivalent to P and has numeraire B. In more general markets When stock price returns follow a single Brownian motion, there is a unique risk neutral measure. When the stock price process is assumed to follow a more general sigma-martingale or semimartingale, then the concept of arbitrage is too narrow, and a stronger concept such as no free lunch with vanishing risk (NFLVR) must be used to describe these opportunities in an infinite dimensional setting. In continuous time, a version of the fundamental theorems of asset pricing read: Let be a d-dimensional semimartingale market (a collectio
https://en.wikipedia.org/wiki/Acoustic%20phase%20conjugation
Acoustic phase conjugation is a set of techniques meant to perform phase conjugation on acoustic waves. Techniques Acoustic phase conjugation can appear in a solid when the sound velocity is modulated by an electromagnetic field. The generation of the conjugate wave can be seen as the decay of a photon into two phonons, as seen on the diagram. The two phonons have opposite wave vectors k and -k (they will propagate in opposite directions) and a frequency two time smaller than that of the photon. Parametric pumping techniques can be performed in several media: In piezoelectric crystals, a nonlinear piezoelectric effect will produce a modulation of fractions of a percent. In magnetic crystals, a modulation of dozens of percent can be reached through the magneto-acoustic coupling, which can be improved by combining the magnetostriction and spin reorientation transition effects or using the magnetoacoustic resonance. A "supercritical" or "giant" amplification, up to 80 dB, can be obtained beyond the threshold of instability of phonons in magnetoacoustic media. In semiconductors, parametric interaction between phonons and plasmon can be generated by an alternative electric field or a modulated optical pump. Applications The auto-compensation of phase distortion and auto-focusing properties of the conjugate wave are used in non-destructive testing techniques. In medical therapy, they can be combined with giant amplification for tumor destruction, like lithotripsy and hyperthermia therapy. Acoustic imaging can be improved by applying selective phase conjugation on some harmonics of the incident wave. This narrows the focal distribution of those harmonics and reduces the sidelobes and reverberation noise, thus increasing the image resolution. Selective acoustic phase conjugation can be used to detect isoechogenic objects whose nonlinear parameters differ from that of the medium. The linear acoustic properties of such objects are close to that of the medium which
https://en.wikipedia.org/wiki/Phenamidine
Phenamidine is an antiprotozoal drug of the amidine class used in veterinary medicine. It is used to treat Babesia infection (babesiosis) dogs, horses, and cattle. Because the drug causes frequent allergic reactions, it is usually combined with an antihistamine.
https://en.wikipedia.org/wiki/Potassium%20propanoate
Potassium propanoate or potassium propionate has formula K(C2H5COO). Its melting point is 410 °C. It is the potassium salt of propanoic acid. Use It is used as a food preservative and is represented by the food labeling E number E283 in Europe and by the INS number 283 in Australia and New Zealand.
https://en.wikipedia.org/wiki/Physiological%20Reviews
Physiological Reviews is a journal published quarterly by the American Physiological Society which has been published since 1921. The editor in chief of the journal is Sadis Matalon (University of Alabama at Birmingham). The journal's first managing editor, who served to his death in 1946, was Dr. Donald R. Hooker. Among the cadre of editors at the journal's inception were William Henry Howell, Lafayette Mendel, and John Macleod. From 1932 to 1950, the chairman of the board of editors of the journal was Anton J. Carlson. Other notable people who have served on the journal's editorial board include John Jacob Abel (c.1935), Ernest William Goodpasture (c.1938). According to the Journal Citation Reports, the journal has a 2020 impact factor of 37.312 Internet Archive page set (values in table linked to specific pages)
https://en.wikipedia.org/wiki/JArchitect
JArchitect is a static analysis tool for Java code. This tool supports a large number of code metrics, allows for visualization of dependencies using directed graphs and dependency matrix. The tools also performs code base snapshots comparison, and validation of architectural and quality rules. User-defined rules can be written using LINQ queries. This possibility is named CQLinq. The tool also comes with a large number of predefined CQLinq code rules. Features The main features of JArchitect are: Dependency visualization (using dependency graphs, and dependency matrix) Software metrics (JArchitect currently supports 82 code metrics: cyclomatic complexity; afferent and efferent coupling; relational cohesion; percentage of code covered by tests, etc.) Declarative code rule over LINQ query (CQLinq) JArchitect can tell you what has been changed between 2 builds Code rule through LINQ Query (CQLinq) The tool proposes live code query and code rule through LINQ query. This is one of the innovations of JArchitect. For example: - Classes inherit from a particular class: // classes inherit from a particular class from t in Types where t.IsClass && t.DeriveFrom ("CBase") select t - The 10 most complex methods (Source Code Cyclomatic complexity) // The 10 most complex methods (from m in Methods orderby select new { m, m.CyclomaticComplexity }).Take(10) In addition, the tool proposes a live CQLinq query editor with code completion and embedded documentation. See also Design Structure Matrix Software visualization External links JArchitect Blog PCWorld Reviews InfoQ press release Java Code Geek review Heise.de Press Release Static program analysis tools Software metrics Java development tools
https://en.wikipedia.org/wiki/PANDIT%20%28database%29
PANDIT is a database of multiple sequence alignments and phylogenetic trees covering many common protein domains. See also Pfam: database of protein domains Phylogeny Sequence alignment
https://en.wikipedia.org/wiki/Crenobacter%20cavernea
Crenobacter cavernea Cave-375 is a gram negative bacterium that is closely related to a previously discovered Crenobacter cavernae strain K1W11S-77ͭ. C. cavernea Cave-375 has not directly been described morphologically, however the related strain K1W11S-77ͭ is a "rod-shaped, motile, and strictly aerobic novel bacteria". Its metabolism has not yet been determined. C. cavernea Cave-375 was first identified from a water sample coming from a dripping stalactite. This stalactite was located in the Algar do Pena cave in the karst Estremadura Limestone Massif in central western Portugal. C. cavernea Cave-375 was first isolated and "grown on nutrient agar at 25 degrees Celsius". Its ecology is not yet known. With the sequencing of the genome of C. cavernea Cave-375, the ecological impact should be able to be identified. Diversity C. cavernea Cave-375 belongs in the Proteobacteria phylum, Neisseriaceae family, and Crenobacter cavernea species. By comparing the 16s rRNA of the CAVE-375 stain to Crenobacter cavernea species, a 99% similarity value was calculated. When comparing DNA-DNA hybridization using a Genome-to-Genome Distance Calculator, a 62.66% hybridization percentage was found. Genome "Genomic DNA was extracted from C. cavernea Cave-375 using an NZY microbial gDNA isolation kit (NZYTech, Portugal)". The whole genome was then sequenced using whole genome shotgun sequencing method. With this, "17,325,372 high-quality raw sequences were assembled into 15 contigs with an N50 value of 323,281 and a total genome size of 2,273,143 base pairs (2.9 Mb)". NCBI Prokaryotic Genome Annotation Pipeline was able to identify a 65.9% GC content and sequencing coding for proteins and tRNA. "2,779 protein coding sequences and 63 tRNA sequences" were identified using this method.
https://en.wikipedia.org/wiki/Mir-444%20microRNA%20precursor%20family
In molecular biology mir-444 microRNA is a short RNA molecule. MicroRNAs function to regulate the expression levels of other genes by several mechanisms. See also MicroRNA
https://en.wikipedia.org/wiki/Certified%20broadcast%20networking%20engineer
Certified Broadcast Networking Engineer (CBNE) is a title granted to an individual that passes the exam requirements of the certification. The certification is regulated by the Society of Broadcast Engineers (SBE) and shows competence in computer network equipment and their specialized applications in radio and television broadcast stations. The CBNE title is protected by copyright laws. Individuals who use the title without consent from the Society of Broadcast Engineers could face legal action. The SBE certifications were created to recognize individuals who practice in career fields which are not regulated by state licensing or Professional Engineering programs. External links Certified Broadcast Networking Engineer (CBNE) Requirements & Application SBE Official Website See also List of post-nominal letters Broadcast engineering Professional certification in engineering
https://en.wikipedia.org/wiki/Gaetano%20Fichera
Gaetano Fichera (8 February 1922 – 1 June 1996) was an Italian mathematician, working in mathematical analysis, linear elasticity, partial differential equations and several complex variables. He was born in Acireale, and died in Rome. Biography He was born in Acireale, a town near Catania in Sicily, the elder of the four sons of Giuseppe Fichera and Marianna Abate. His father Giuseppe was a professor of mathematics and influenced the young Gaetano starting his lifelong passion. In his young years he was a talented football player. On 1 February 1943 he was in the Italian Army and during the events of September 1943 he was taken prisoner by the Nazist troops, kept imprisoned in Teramo and then sent to Verona: he succeeded in escaping from there and reached the Italian region of Emilia-Romagna, spending with partisans the last year of war. After the war he was first in Rome and then in Trieste, where he met Matelda Colautti, who became his wife in 1952. Education and academic career After graduating from the liceo classico in only two years, he entered the University of Catania at the age of 16, being there from 1937 to 1939 and studying under Pia Nalli. Then he went to the university of Rome, where in 1941 he earned his laurea with magna cum laude under the direction of Mauro Picone, when he was only 19. He was immediately appointed by Picone as an assistant professor to his chair and as a researcher at the Istituto Nazionale per le Applicazioni del Calcolo, becoming his pupil. After the war he went back to Rome working with Mauro Picone: in 1948 he became "Libero Docente" (free professor) of mathematical analysis and in 1949 he was appointed as full professor at the University of Trieste. As he remembers in , in both cases one of the members of the judging commission was Renato Caccioppoli, which become a close friend of him. From 1956 onward he was full professor at the University of Rome in the chair of mathematical analysis and then at the Istituto Nazionale
https://en.wikipedia.org/wiki/Cambium%20Networks
Cambium Networks is a wireless infrastructure provider that offers fixed wireless and Wi-Fi to broadband service providers and enterprises to provide Internet access. An American telecommunications infrastructure company, it provides wireless technology, including Enterprise WiFi, switching solutions, Internet of Things, and fixed wireless broadband and Wi-Fi for enterprises. Publicly traded on the NASDAQ stock exchange, it spun out of Motorola in October 2011. Products Cambium Networks manufactures point-to-point backhaul, point-to-multipoint communication wide area network (WAN), Wi-Fi indoor and outdoor access, and cloud-based network management systems. In 2020, the company collaborated with Facebook to add mesh networking technology Terragraph that allows high-speed internet connections where laying fiber optic cable is not viable. As of 2021 the company has shipped 10 million radios. Products are available in point-to-point and point-to-multipoint configurations. Its cnWave fixed wireless solution provides multi-gigabit throughputs. It includes both the original Motorola-designed products using the Canopy protocol and the PtP backhauls that were rebranded from Orthogon Systems, which Motorola acquired in 2006. Cambium Networks’ solutions are used by broadband service providers and managed service providers to connect business and residential locations in dense urban, suburban, rural and remote locations, including education and healthcare. Enterprise Wi-Fi and Switching Cambium Networks also manufactures Wireless LAN (WLAN) Wi-Fi access points including Wi-Fi 6E and intelligent switches along with cloud=management systems. In 2022, Spectralink added interoperability with Cambium Networks access points and Wi-Fi phones and handsets as part of its enterprise wireless certification program. History Cambium Networks was created when Motorola Solutions sold the Canopy and Orthogon businesses in 2011. Cambium evolved the platform and expanded it to thr
https://en.wikipedia.org/wiki/NASA%20360
NASA 360 is a half-hour vodcast developed by NASA in partnership with the National Institute of Aerospace. The show premiered in August 2008. It has aired on more than 450 TV stations across the country, is available on air and cruise lines, and is consistently one of the top-downloaded programs on the NASA.gov website. It is currently in its tenth season. Description NASA 360 is one of four programs in NASA's award-winning eClips suite of web-based shows designed to encourage careers in science, technology, engineering, and mathematics. NASA 360 is written, produced, and edited by Timothy J. Allen, Tom Shortridge, and Scott Bednar; Rebecca Jaramillo is the Senior Educator and Project Coordinator for NASA 360, and Harla Sherwood the Principal Investigator - all of the National Institute of Aerospace. NASA 360 shows how NASA has changed and continues to change life on Earth by examining how technologies developed by or for NASA are being used in everything from space exploration to everyday consumer products. These include lithium ion batteries, medical innovations, sporting equipment, and automotive and aircraft safety and efficiency, among many more. NASA 360 is shot on-location at NASA centers across the country, as well as at other relevant sites across the globe. The fifth season marked the revitalization of NASA 360 and features new hosts Caleb Kinchlow and Molly McKinney, B-roll, animations, and interviews conducted with NASA researchers, engineers, and astronauts, as well as with outside sources with expertise relevant to the topics being discussed. The show is produced for a young adult audience, and stylistically this is accomplished through the use of hand-held cameras, quick edits, and numerous transitions, effects, and filters used in post-production. NASA 360 was originally created in 2006 by producers Kevin Krigsvold and Michael Bibbo. It was hosted by Johnny Alonso and Jennifer Pulley through 2012. Twenty-three episodes were produced during this
https://en.wikipedia.org/wiki/Desert%20locust
The desert locust (Schistocerca gregaria) is a species of locust, a periodically swarming, short-horned grasshopper in the family Acrididae. They are found primarily in the deserts and dry areas of northern and eastern Africa, Arabia, and southwest Asia. During population surge years, they may extend north into parts of western Spain and southern Italy, south into Eastern Africa, and east in northern India. The desert locust shows periodic changes in its body form and can change in response to environmental conditions, over several generations, from a solitary, shorter-winged, highly fecund, non-migratory form to a gregarious, long-winged, and migratory phase in which they may travel long distances into new areas. In some years, they may thus form locust plagues, invading new areas, where they may consume all vegetation including crops, and at other times, they may live unnoticed in small numbers. During plague years, desert locusts can cause widespread damage to crops, as they are highly mobile and feed on large quantities of any kind of green vegetation, including crops, pasture, and fodder. A typical swarm can be made up of and fly in the direction of the prevailing wind, up to in one day. Even a very small, locust swarm can eat the same amount of food in a day as about 35,000 people. As an international transboundary pest that threatens agricultural production and livelihoods in many countries in Africa, the Near East, and southwest Asia, their populations have been routinely monitored through a collaborative effort between countries and the United Nations Food and Agriculture Organization (FAO) Desert Locust Information Service (DLIS), which provides global and national assessments, forecasts, and early warning to affected countries and the international community. The desert locust's migratory nature and capacity for rapid population growth present major challenges for control, particularly in remote semiarid areas, which characterize much of their range
https://en.wikipedia.org/wiki/KIF6
Kinesin family member 6 is a protein that in humans is encoded by the KIF6 gene. This gene encodes a member of the kinesin family of proteins. Members of this family are part of a multisubunit complex that functions as a microtubule motor in intracellular organelle transport.
https://en.wikipedia.org/wiki/UEFI%20Forum
UEFI Forum, Inc. is an alliance between technology companies to coordinate the development of the UEFI specifications. The board of directors includes representatives from twelve promoter companies: AMD, American Megatrends, ARM, Apple, Dell, Hewlett Packard Enterprise, HP Inc., Insyde Software, Intel, Lenovo, Microsoft, and Phoenix Technologies. Overview The non-profit corporation has assumed responsibility for the management and promotion of the Unified Extensible Firmware Interface (UEFI) specification, a bootloader and runtime interface between platform firmware and an operating system. The original EFI specification was developed by Intel and was used as the starting point from which the UEFI version(s) were developed. The goal of the organization is to replace the aging PC BIOS. In addition to the UEFI specification, the forum is responsible for a UEFI Platform Initialization (PI) specification, which addresses the firmware internal architecture as well as firmware-to-hardware interfaces. The forum also is responsible for Self-Certification Test suites, which defines conformance to the specifications that it defines. In October 2013, the Advanced Configuration and Power Interface (ACPI) assets have also been transferred into the forum. The forum is responsible for the management and promotion of future ACPI specifications, which provides static tables at boot time and dynamic control methods as the primary runtime interfaces between the OS and system firmware for system configuration, power management and RAS (Reliability, Availability and Supportability) features. ACPI "Revision 5.0" is used as the starting point from which future ACPI version(s) will be developed. Published specifications UEFI Specification version 2.8, published March, 2019 UEFI Shell Specification version 2.2, published January 26, 2016 UEFI Platform Initialization Specification version 1.7, published January, 2019 UEFI Platform Initialization Distribution Packaging Specification
https://en.wikipedia.org/wiki/Causes%20of%20Parkinson%27s%20disease
Parkinson's disease (PD) is a degenerative disorder of the central nervous system. Most people with PD have idiopathic Parkinson's disease (having no specific known cause). A small proportion of cases, however, can be attributed to known genetic factors. Other factors such as environmental toxins, herbicides, pesticides, and fungicides, have been associated with the risk of developing PD, but no causal relationships have been proven. Genetic factors Traditionally, Parkinson's disease has been considered a non-genetic disorder. However, around 15% of individuals with PD have a first-degree relative who has the disease. At least 5–15% of cases are known to occur because of a mutation in one of several specific genes, transmitted in either an autosomal-dominant or autosomal-recessive pattern. Mutations in specific genes have been conclusively shown to cause PD. A large number of these genes are linked to translation. Genes which have been implicated in autosomal-dominant PD include PARK1 and PARK4, PARK5, PARK8, PARK11 and GIGYF2 and PARK13 which code for alpha-synuclein (SNCA), UCHL1, leucine-rich repeat kinase 2 (LRRK2 or dardarin) (LRRK2 and Htra2 respectively Genes such as PARK2, PARK6, PARK7 and PARK9 which code for parkin (PRKN), PTEN-induced putative kinase 1 (PINK1), DJ-1 and ATP13A2 respectively have been implicated in the development of autosomal-recessive PD Furthermore, mutations in genes including those that code for SNCA, LRRK2 and glucocerebrosidase (GBA) have been found to be risk factors for sporadic PD In most cases, people with these mutations will develop PD. With the exception of LRRK2, however, they account for only a small minority of cases of PD. The most extensively studied PD-related genes are SNCA and LRRK2. SNCA gene The role of the SNCA gene is significant in PD because the alpha-synuclein protein is the main component of Lewy bodies, which appear as a primary biomarker in the disease. Missense mutations of the gene (in which a si
https://en.wikipedia.org/wiki/Z-transform
In mathematics and signal processing, the Z-transform converts a discrete-time signal, which is a sequence of real or complex numbers, into a complex frequency-domain (the z-domain or z-plane) representation. It can be considered a discrete-time equivalent of the Laplace transform (the s-domain or s-plane). This similarity is explored in the theory of time-scale calculus. While the continuous-time Fourier transform is evaluated on the s-domain's vertical axis (the imaginary axis), the discrete-time Fourier transform is evaluated along the z-domain's unit circle. The s-domain's left half-plane maps to the area inside the z-domain's unit circle, while the s-domain's right half-plane maps to the area outside of the z-domain's unit circle. One of the means of designing digital filters is to take analog designs, subject them to a bilinear transform which maps them from the s-domain to the z-domain, and then produce the digital filter by inspection, manipulation, or numerical approximation. Such methods tend not to be accurate except in the vicinity of the complex unity, i.e. at low frequencies. History The basic idea now known as the Z-transform was known to Laplace, and it was re-introduced in 1947 by W. Hurewicz and others as a way to treat sampled-data control systems used with radar. It gives a tractable way to solve linear, constant-coefficient difference equations. It was later dubbed "the z-transform" by Ragazzini and Zadeh in the sampled-data control group at Columbia University in 1952. The modified or advanced Z-transform was later developed and popularized by E. I. Jury. The idea contained within the Z-transform is also known in mathematical literature as the method of generating functions which can be traced back as early as 1730 when it was introduced by de Moivre in conjunction with probability theory. From a mathematical view the Z-transform can also be viewed as a Laurent series where one views the sequence of numbers under consideration as the (La
https://en.wikipedia.org/wiki/Cyclic%20and%20separating%20vector
In mathematics, the notion of a cyclic and separating vector is important in the theory of von Neumann algebras, and in particular in Tomita–Takesaki theory. A related notion is that of a vector which is cyclic for a given operator. The existence of cyclic vectors is guaranteed by the Gelfand–Naimark–Segal (GNS) construction. Definitions Given a Hilbert space H and a linear space A of bounded linear operators in H, an element Ω of H is said to be cyclic for A if the linear space AΩ = {aΩ: a ∈ A} is norm-dense in H. The element Ω is said to be separating if aΩ = 0 with a in A implies a = 0. Any element Ω of H defines a semi-norm p on A by p(a) = ||aΩ||. Saying that Ω is separating is equivalent with saying that p is actually a norm. If Ω is cyclic for A then it is separating for the commutant A′, which is the von Neumann algebra of all bounded operators in H which commute with all operators of A. Indeed, if a belongs to A′ and satisfies aΩ = 0 then one has for all b in A that 0 = baΩ = abΩ. Because the set of bΩ with b in A is dense in H this implies that a vanishes on a dense subspace of H. By continuity this implies that a vanishes everywhere. Hence, Ω is separating for A′. The following stronger result holds if A is a *-algebra (an algebra which is closed under taking adjoints) and contains the identity operator 1. For a proof, see Proposition 5 of Part I, Chapter 1 of. Proposition If A is a *-algebra of bounded linear operators in H and 1 belongs to A then Ω is cyclic for A if and only if it is separating for the commutant A′. A special case occurs when A is a von Neumann algebra. Then a vector Ω which is cyclic and separating for A is also cyclic and separating for the commutant A′ Positive linear functionals A positive linear functional ω on a *-algebra A is said to be faithful if ω(a) = 0, where a is a positive element of A, implies a = 0. Every element Ω of H defines a positive linear functional ωΩ on a *-algebra A of bounded linear operators i
https://en.wikipedia.org/wiki/HOCPCA
HOCPCA (3-hydroxycyclopent-1-enecarboxylic acid) is a compound with an affinity for the GHB receptor 39 times greater than that of GHB itself. See also T-HCA
https://en.wikipedia.org/wiki/DOME%20MicroDataCenter
A microDataCenter contains compute, storage, power, cooling and networking in a very small volume, sometimes also called a "DataCenter-in-a-box". The term has been used to describe various incarnations of this idea over the past 20 years. Late 2017 a very tightly integrated version was shown at SuperComputing conference 2017: the DOME microDataCenter. Key features are its hot-watercooling, fully solid-state and being built with commodity components and standards only. DOME project DOME is a Dutch government-funded project between IBM and ASTRON in form of a public–private partnership to develop technology roadmaps targeting the Square Kilometer Array (SKA), the world's largest planned radio telescope. It will be built in Australia and South Africa during the late 2010s and early 2020s. One of the 7 DOME projects is MicroDataCenter (previously called Microservers) that are small, inexpensive and computationally efficient. The goal for the MicroDataCenter is the capability to be used both near the SKA antennas to do early processing of the data, and inside much larger supercomputers that will do the big data analysis. These servers can be deployed in very large numbers and in environmentally extreme locations such as in deserts where the antennas will be located and not in only in cooled datacenters. A common misconception is that microservers offer only low performance. This is caused by the first microservers being based on Atoms or early 32bit ARM cores. The aim of the DOME MicroDataCenter project is to deliver high performance at low cost and low power. A key characteristic of a MicroDataCenter is its packaging: very small form factor that allows short communication distances. This is based on using Microservers, eliminating all unnecessary components by integrating as much as possible from the traditional compute server into a single SoC (Server on a chip). A microserver will not deliver the highest possible single-thread performance, instead, it offers an e
https://en.wikipedia.org/wiki/Builder%27s%20plate
A builder's plate is usually a metal plate that is attached to railway locomotives and rolling stock, bogies, construction equipment, trucks, automobiles, large household appliances, bridges, ships and more. It gives such information as the name of the manufacturer, the place and country of manufacture, the model number, the serial number, as well as the date of manufacture or date of fabrication of the item or unit. Gallery See also Engine number Head badge Nameplate Railroadiana VRP VIN
https://en.wikipedia.org/wiki/Ambient%20isotopy
In the mathematical subject of topology, an ambient isotopy, also called an h-isotopy, is a kind of continuous distortion of an ambient space, for example a manifold, taking a submanifold to another submanifold. For example in knot theory, one considers two knots the same if one can distort one knot into the other without breaking it. Such a distortion is an example of an ambient isotopy. More precisely, let and be manifolds and and be embeddings of in . A continuous map is defined to be an ambient isotopy taking to if is the identity map, each map is a homeomorphism from to itself, and . This implies that the orientation must be preserved by ambient isotopies. For example, two knots that are mirror images of each other are, in general, not equivalent. See also Isotopy Regular homotopy Regular isotopy
https://en.wikipedia.org/wiki/Internet%20Cat%20Video%20Festival
The Internet Cat Video Festival was a national competition that celebrates cat videos on the internet. Many of these festivals include appearances by special guests and celebricats (such as Grumpy Cat and the creator of Nyan Cat), live music, costume contests, art projects, and booths hosting local animal resource nonprofits. The 2013 Minneapolis show featured a cat sculpture made out of butter. From Conception - 2015 The idea for the first Internet Cat Video Festival came from Katie Hill, a program associate at the Walker Art Center, who suggested it early in 2012 as a sort of joke. The Festival quickly became a viral phenomenon, drawing a crowd of more than 10,000 fans to the Walker's Open Field. The event caught the attention of local, national, and international audiences and media including the New York Times. Walker Art Center was joined by community partners Feline Rescue, Animal Humane Society, and The Wildcat Sanctuary. After the inaugural event, the Internet Cat Video Festival went on tour. Stops included UMass, Boston, the Museum of Photographic Arts in San Diego, and Memphis' Brooks Museum. The festival was even included in the Vienna Independent Shorts in Vienna, Austria and the Jerusalem Film Festival. The next big Fest took place in Oakland, CA on May 11, 2013. All profits benefitted the East Bay SPCA. On June 21 and 22, 2013, the Cat Video Fest was at Portland's Hollywood Theatre in partnership with the Theatre's "Future So Bright" project. On Saturday, October 19, 2013, the Cat Video Festival was held in Chicago, IL, supported by the Chicago Cat Rescue and Tree House. Moving into the second year, catvidfest was booked for 15 tour dates, including festivals in San Francisco, Chicago, Brooklyn, and New York, plus experimental versions at ArtPrize in Grand Rapids, Michigan, and at CultureTECH in Derry Northern Ireland. In 2013, the Internet Cat Video Festival returned to Minneapolis and screens at the Great Minnesota Get-Together: the Sta
https://en.wikipedia.org/wiki/Spontaneous%20generation
Spontaneous generation is a superseded scientific theory that held that living creatures could arise from nonliving matter and that such processes were commonplace and regular. It was hypothesized that certain forms, such as fleas, could arise from inanimate matter such as dust, or that maggots could arise from dead flesh. The doctrine of spontaneous generation was coherently synthesized by the Greek philosopher and naturalist Aristotle, who compiled and expanded the work of earlier natural philosophers and the various ancient explanations for the appearance of organisms. Spontaneous generation was taken as scientific fact for two millennia. Though challenged in the 17th and 18th centuries by the experiments of the Italian biologists Francesco Redi and Lazzaro Spallanzani, it was not discredited until the work of the French chemist Louis Pasteur and the Irish physicist John Tyndall in the mid-19th century. Rejection of spontaneous generation is no longer controversial among biologists. By the middle of the 19th century, experiments by Pasteur and others were considered to have disproven the traditional theory of spontaneous generation. Attention has turned instead to the origin of life, since all life seems to have evolved from a single form around four billion years ago. Description "Spontaneous generation" means both the supposed processes by which different types of life might repeatedly emerge from specific sources other than seeds, eggs, or parents, and the theoretical principles presented in support of any such phenomena. Crucial to this doctrine are the ideas that life comes from non-life and that no causal agent, such as a parent, is needed. Supposed examples included the seasonal generation of mice and other animals from the mud of the Nile, the emergence of fleas from inanimate matter such as dust, or the appearance of maggots in dead flesh. Such ideas have something in common with the modern hypothesis of the origin of life, which asserts that life em
https://en.wikipedia.org/wiki/Videobombing
Videobombing is the phenomenon of an unexpected appearance in a video of individuals who were not intended to be there by the operators of the camera or the individuals being filmed. The act of inserting oneself into someone else's video is often done in order to play a practical joke on the cameraman or his subjects, and sometimes in a deliberate attempt to create a video that could go viral. For example, individuals wishing to videobomb someone may insert themselves into a video of a famous celebrity being interviewed, or a live news broadcast being filmed in the streets or in another area where outsiders are present. The individual will usually attempt to amuse the home audience by performing humorous and strange actions in the background, and may even distract and annoy the news reporter. Individuals aware that a news broadcast will occur nearby them may even prepare beforehand and turn up to the live broadcast adorning a silly costume. Some individuals such as Chris Bosh and Paul Yarrow, do this repeatedly. A related term is photobombing, which applies the concept to a still photograph instead of video footage. 1970s-1980s: John 3:16 Rainbow Man Rollen Frederick Stewart (born February 23, 1944), also known as "Rock 'n' Rollen" and "Rainbow Man", is a pioneer of videobombing who established himself as celebrity in American sports culture by being best known for wearing a rainbow-colored afro-style wig and, later, holding up signs reading "John 3:16" at stadium sporting events around the United States and overseas in the 1970s and 1980s. Stewart became a born-again Christian, and was determined to "get the message out" via television. His first major appearance was at the 1977 NBA Finals; by the time of the 1979 MLB All-Star Game, broadcasters actively tried to avoid showing him. He "appeared behind NFL goal posts, near Olympic medal stands, and even at the Augusta National Golf Club." At the 1982 Indianapolis 500, he was behind the pits of race winner Gordo
https://en.wikipedia.org/wiki/Shifted%20force%20method
The net electrostatic force acting on a charged particle with index contained within a collection of particles is given as: where is the spatial coordinate, is a particle index, is the separation distance between particles and , is the unit vector from particle to particle , is the force magnitude, and and are the charges of particles and , respectively. With the electrostatic force being proportional to , individual particle-particle interactions are long-range in nature, presenting a challenging computational problem in the simulation of particulate systems. To determine the net forces acting on particles, the Ewald or Lekner summation methods are generally employed. One alternative and usually computationally faster technique based on the notion that interactions over large distances (e.g. > 1 nm) are insignificant to the net forces acting in certain systems is the method of spherical truncation. The equations for basic truncation are: where is the cutoff distance. Simply applying this cutoff method introduces a discontinuity in the force at that results in particles experiencing sudden impulses when other particles cross the boundary of their respective interaction spheres. In the particular case of electrostatic forces, as the force magnitude is large at the boundary, this unphysical feature can compromise simulation accuracy. A way to correct this problem is to shift the force to zero at , thus removing the discontinuity. This can be accomplished with a variety of functions, but the most simple/computationally efficient approach is to simply subtract the value of the electrostatic force magnitude at the cutoff distance as such: As mentioned before, the shifted force (SF) method is generally suited for systems that do not have net electrostatic interactions that are long-range in nature. This is the case for condensed systems that show electric-field screening effects. Note that anisotropic systems (e.g. interfaces) may not be accurately simulat
https://en.wikipedia.org/wiki/LM358
The LM358 is a low-power dual operational amplifier integrated circuit, originally introduced by National Semiconductor. It supports an operating voltage of +3 to +32 volts (single power supply) or ±1.5 to ±16 volts (dual power supplies). Input voltage can range from −0.3 to +32 volts with single power supply. Small negative input voltages below ground are acceptable because the bipolar junction transistors at the input stage are configured such that their base-emitter junction voltage provides just enough voltage differential between the collector and base for the transistors to function.
https://en.wikipedia.org/wiki/Trisomic%20rescue
Trisomic rescue (also known as trisomy rescue or trisomy zygote rescue) is a genetic phenomenon in which a fertilized ovum containing three copies of a chromosome loses one of these chromosomes (anaphase lag) to form a diploid chromosome complement. If both of the retained chromosomes come from the same parent, then uniparental disomy results. If the retained chromosomes come from different parents (that is, one copy from each) then there are no phenotypic or genotypic anomalies. The mechanism of trisomic rescue has been well confirmed in vivo, and alternative mechanisms that occur in trisomies are rare in comparison. Many trisomic conditions result in stillborn infants (major exceptions include trisomies 13, 18, 21). Trisomic rescue may be a natural means to keep a fetus as viable as possible (though uniparental disomy may occur and result in syndromes such as Prader-Willi and Angelman syndromes due to genetic imprinting). Indeed, spontaneous trisomic rescue has been observed in vitro. Similarly, monosomic rescue may also be a natural means to keep fetal viability via restoration of a disomic zygote.
https://en.wikipedia.org/wiki/Poincar%C3%A9%20metric
In mathematics, the Poincaré metric, named after Henri Poincaré, is the metric tensor describing a two-dimensional surface of constant negative curvature. It is the natural metric commonly used in a variety of calculations in hyperbolic geometry or Riemann surfaces. There are three equivalent representations commonly used in two-dimensional hyperbolic geometry. One is the Poincaré half-plane model, defining a model of hyperbolic space on the upper half-plane. The Poincaré disk model defines a model for hyperbolic space on the unit disk. The disk and the upper half plane are related by a conformal map, and isometries are given by Möbius transformations. A third representation is on the punctured disk, where relations for q-analogues are sometimes expressed. These various forms are reviewed below. Overview of metrics on Riemann surfaces A metric on the complex plane may be generally expressed in the form where λ is a real, positive function of and . The length of a curve γ in the complex plane is thus given by The area of a subset of the complex plane is given by where is the exterior product used to construct the volume form. The determinant of the metric is equal to , so the square root of the determinant is . The Euclidean volume form on the plane is and so one has A function is said to be the potential of the metric if The Laplace–Beltrami operator is given by The Gaussian curvature of the metric is given by This curvature is one-half of the Ricci scalar curvature. Isometries preserve angles and arc-lengths. On Riemann surfaces, isometries are identical to changes of coordinate: that is, both the Laplace–Beltrami operator and the curvature are invariant under isometries. Thus, for example, let S be a Riemann surface with metric and T be a Riemann surface with metric . Then a map with is an isometry if and only if it is conformal and if . Here, the requirement that the map is conformal is nothing more than the statement that is, Metric a
https://en.wikipedia.org/wiki/Russula%20decolorans
Russula decolorans, commonly known as the graying russula, is an edible Russula mushroom found in groups in coniferous forests. Description The cap is convex, with a depressed centre when old, often brick-red and slippery when young. The cap grows up to 10 cm. The flesh is white and turns grey when old. It has a mild taste. The spores are pale ochre. Uses The edible mushroom is commonly harvested for food in Finland. However, it is not recommended to eat it, as it has several lookalikes with unknown edibility. See also List of Russula species
https://en.wikipedia.org/wiki/Moshi%20Monsters
Moshi Monsters was a British web browser massively multiplayer online role-playing game (MMORPG) aimed at children aged 6–12, with over 80 million registered users in 150 territories worldwide. Users could choose from one of six virtual pet monsters (Katsuma, Poppet, Furi, Zommer, Luvli, and Diavlo) they could create, name and nurture. Once their pet had been customized, players could navigate their way around Monstro City, take daily puzzle challenges to earn 'Rox' (a virtual currency), play games, personalize their room and communicate with other users in a safe environment, although this has been disputed. The servers for the game were officially closed on December 13, 2019, due to the game requiring Adobe Flash Player, which ended support on 31 December 2020. Mind Candy continues to actively use the "Moshi" intellectual property in the mobile app Moshi, an app featuring bedtime stories, games and guided mindfulness meditations aimed at children. History The game was created in late 2007 by Michael Acton Smith, and developed in 2008 by entertainment company Mind Candy and officially launched in April 2008. As of December 2009, there were at least 10 million players registered. In March 2010, Mind Candy announced that there were 15 million users and by September 2010, that number had surpassed 25 million. In June 2011, it was announced that there were 50 million users. On 13 November 2019, it was announced that Moshi Monsters would be closing down on 13 December 2019, due to the end of support for Adobe Flash Player. Gameplay Moshlings The Monsters (in-game pets) keep their own pets, called "Moshlings". They came in a variety of themed sets, including Arties, Beasties, Kitties, and Spookies. Those who were not paying members could keep two "Moshlings" in their room, whilst paying members could keep up to six and visit other pets in the zoo. Other Merchandise Since its digital popularity, Moshi Monsters has grown commercially to include physical products, i
https://en.wikipedia.org/wiki/K.%20C.%20Nag
Keshab Chandra Nag or K.C. Nag () (10 July 1893 – 6 February 1987), was an Indian Bengali mathematician, author of various mathematics textbooks and educator. Early life K. C. Nag was born in Nagpara, Gurap, Hooghly, Bengal, British India (present-day West Bengal, India) on the holy day of Rath Yatra, 10 July 1893. His Father was Raghunath Nag and Mother Khiroda Sundari Debi. He lost his father at an early age of three. He was only cared for by his mother. Education Keshab Chandra Nag started his education in a Bengali Medium School at his village in Gurap. At that time that was the only school at Gurap. From Class VII he changed his school to Bhastara Yojneshshar Uccha Vidyalaya (Yojneshshar High School), 3 miles from his village. He would start walking early in the morning to reach his school and came back home at evening every day. In Class IX he got admitted to Kishenganj High School. In 1912, he passed the entrance examination with a First Class and joined Ripon College (now Surendranath College), Kolkata, in Science. In 1914, he passed the I.Sc examination with a First Class. After this due to severe financial crisis he had to discontinue his education and start earning money. Working life He started his career as Third Master in Bhastara Yojgeshshar Uccha Vidyalaya. He also did private tuitions when teaching there. His family was dependent on him but he resigned from his job to pursue higher studies. In 1917 he passed B.A with Mathematics and Sanskrit. He then received a job offer from Kishenganj High School as a Mathematics Teacher. He taught for some time in that school, after which he got another offer from Baharampur Krishnanath Collegiate School and joined the school as a mathematics teacher. In 1919 he got Diksha from Ma Sarada Devi. During that time the Maharaja of Kasimbazar (Cossimbazar) Manindra Chandra Nandi was a great admirer of Keshab Chandra. Maharaj allowed Keshab Chandra to use his vast library. In that library, he studied extensively abou
https://en.wikipedia.org/wiki/Thue%27s%20lemma
In modular arithmetic, Thue's lemma roughly states that every modular integer may be represented by a "modular fraction" such that the numerator and the denominator have absolute values not greater than the square root of the modulus. More precisely, for every pair of integers with , given two positive integers and such that , there are two integers and such that and Usually, one takes and equal to the smallest integer greater than the square root of , but the general form is sometimes useful, and makes the uniqueness theorem (below) easier to state. The first known proof is attributed to who used a pigeonhole argument. It can be used to prove Fermat's theorem on sums of two squares by taking m to be a prime p that is congruent to 1 modulo 4 and taking a to satisfy a2 + 1 = 0 mod p. (Such an "a" is guaranteed for "p" by Wilson's theorem.) Uniqueness In general, the solution whose existence is asserted by Thue's lemma is not unique. For example, when there are usually several solutions , provided that and are not too small. Therefore, one may only hope for uniqueness for the rational number , to which is congruent modulo if y and m are coprime. Nevertheless, this rational number need not be unique; for example, if , and , one has the two solutions . However, for and small enough, if a solution exists, it is unique. More precisely, with above notation, if and , with and then This result is the basis for rational reconstruction, which allows using modular arithmetic for computing rational numbers for which one knows bounds for numerators and denominators. The proof is rather easy: by multiplying each congruence by the other and subtracting, one gets The hypotheses imply that each term has an absolute value lower than , and thus that the absolute value of their difference is lower than . This implies that , hence the result. Computing solutions The original proof of Thue's lemma is not efficient, in the sense that it does not provi
https://en.wikipedia.org/wiki/Black-Body%20Theory%20and%20the%20Quantum%20Discontinuity%2C%201894%E2%80%931912
Black-Body Theory and the Quantum Discontinuity, 1894–1912 (1978; second edition 1987) is a book by the philosopher Thomas Kuhn, in which the author surveys the development of quantum mechanics. The second edition has a new afterword. Summary Kuhn surveys the development of quantum mechanics by Max Planck at the end of the 19th century. He argues that Planck misread his own earlier work. Reception Alexander Bird describes Kuhn's book as "masterly", writing that it "differs from traditional history of science less in the kind of explanation offered and more in the vast erudition and scholarly attention to detail displayed." According to philosopher Tim Maudlin, Planck and the Black Body Discontinuity (sic) "is a mixed bag: some good historiography and some poor analysis."
https://en.wikipedia.org/wiki/Barlow%27s%20law
Barlow's law is an incorrect physical law proposed by Peter Barlow in 1825 to describe the ability of wires to conduct electricity. It says that the strength of the effect of electricity passing through a wire varies inversely with the square root of its length and directly with the square root of its cross-sectional area, or, in modern terminology: where I is electric current, A is the cross-sectional area of the wire, and L is the length of the wire. Barlow formulated his law in terms of the diameter d of a cylindrical wire. Since A is proportional to the square of d the law becomes for cylindrical wires. Barlow undertook his experiments with the aim of determining whether long-distance telegraphy was feasible and believed that he proved that it was not. The publication of Barlow's law delayed research into telegraphy for several years, until 1831 when Joseph Henry and Philip Ten Eyck constructed a circuit 1,060 feet long, which used a large battery to activate an electromagnet. Barlow did not investigate the dependence of the current strength on electric tension (that is, voltage). He endeavoured to keep this constant, but admitted there was some variation. Barlow was not entirely certain that he had found the correct law, writing "the discrepancies are rather too great to enable us to say, with confidence, that such is the law in question." In 1827, Georg Ohm published a different law, in which current varies inversely with the wire's length, not its square root; that is, where is a constant dependent on the circuit setup. Ohm's law is now considered the correct law, and Barlow's false. The law Barlow proposed was not in error due to poor measurement; in fact, it fits Barlow's careful measurements quite well. Heinrich Lenz pointed out that Ohm took into account "all the conducting resistances … of the circuit", whereas Barlow did not. Ohm explicitly included a term for what we would now call the internal resistance of the battery. Barlow did not hav
https://en.wikipedia.org/wiki/Bamboo%20%28elephant%29
Bamboo was an Asian elephant who resided at the Woodland Park Zoo in Seattle, Washington prior to being moved to the Oklahoma City Zoo in 2015. She was at the center of controversy for several years. Her reputation as a troubled elephant, allegedly due to past abuse and an inadequate life at the Zoo, caused animal rights activists such as Friends of Woodland Park Zoo Elephants to fight for the release of the Zoo's elephants to a sanctuary. According to the Zoo's website, Bamboo weighed 8,800 pounds and was the most inquisitive of its three elephants. Early years Bamboo was born in Thailand in November 1966 and was captured from the wild as a very young calf. She was imported into the United States and arrived at the Woodland Park Zoo in Seattle, Washington, on June 1, 1967. She lived at the Children's Zoo in the Family Farm before moving to the old Elephant House. Former Director of the Zoo, David Hancocks, recalled that Bamboo was once a playful, trustworthy, and cooperative elephant who could be walked on Zoo grounds and use the entire park as her playground. According to Hancocks, who left the Zoo in 1984, harsh management methods such as discipline and overnight chaining were reintroduced after his departure. As a result, Bamboo earned a reputation as a dangerous and "difficult" elephant. A "difficult" elephant On November 3, 2000, another Asian elephant named Chai gave birth to a female calf named Hansa. The new addition changed Bamboo's daily routine, and she was not initially welcoming toward Hansa and had to be housed separately. Spending much of her time alone in a small barn stall, she displayed what is believed to be stereotypical behavior for elephants - pacing and bobbing her head continuously - caused by inadequate environmental and social conditions and not displayed by elephants living in the wild. A YouTube video taken in May 2001 shows Bamboo exhibiting this stereotypical behavior, pacing around her small space repeatedly in circles. Bamboo was
https://en.wikipedia.org/wiki/Raw%20bar
A raw bar is a small restaurant or a bar within a restaurant where live shellfish are shucked and served. Raw bars typically offer a variety of raw and cooked seafood and shellfish that is served cold. Seafood-based dishes may also be offered, and additional, non-seafood foods may also be part of the fare. Raw bars may offer alcoholic beverages such as oyster shooters, as well as wine and sake that is paired with various foods. Additional accompaniments may include condiments, sauces and foods such as lemon and lime. Several restaurants in the United States offer raw bars, some of which are seasonal. Fare Raw seafood Raw bars may serve a selection of raw oysters, clams, quahogs (hard clams), scallops and mussels. Varieties of hard clam may include littlenecks, which are less than 1.5 inches (3.8 cm) in size, and cherrystones, which are up to 2 inches (5 cm). Various types of oysters may be served. Some raw bars may offer oyster shooters, a type of cocktail prepared with raw oyster. Some also offer ceviche, a dish prepared with raw seafood that is cured with citrus juices, particularly lime. Thinly sliced octopus (octopus carpaccio) is another raw bar item. Cooked seafood Raw bars sometimes supplement the menu with cooked versions of the same and additional seafoods and shellfish that are typically served cold, such as clam chowder, oyster stew, poached shrimp, shrimp cocktail, cooked or seared scallops, mussels, crab legs, lobster, cured salmon, sea urchin and steamers (steamed clams). Other cooked foods Sometimes lightly cooked liver or foie gras is a raw bar item. Dishes The plateau de fruits de mer is a seafood dish sometimes offered by raw bars that is prepared with raw and cooked shellfish and cold on a platter, usually on a bed of ice. Accompaniments and condiments Raw bars may offer wine or sake to accompany and be paired with the various foods. Condiments, such as cocktail sauce and lemon, may be available, which are typically served with raw oysters.
https://en.wikipedia.org/wiki/Onavo
Onavo, Inc. was an Israeli mobile web analytics company owned by Meta Platforms. The company primarily performed its activities via consumer mobile apps, including the virtual private network (VPN) service Onavo Protect, which analysed web traffic sent through the VPN to provide statistics on the usage of other apps. Guy Rosen and Roi Tiger founded Onavo in 2010. In October 2013, Onavo was acquired by Facebook, which used Onavo's analytics platform to monitor competitors. This influenced Facebook to make various business decisions, including its 2014 acquisition of WhatsApp. Since the acquisition, Onavo was frequently classified as being spyware, as the VPN was used to monetize application usage data collected within an allegedly privacy-focused environment. In August 2018, Facebook pulled Onavo Protect from the iOS App Store due to violations of Apple's policy forbidding apps from collecting data on the usage of other apps. In February 2019, in response to criticism over a Facebook market research program employing similar techniques (including, in particular, being targeted towards teens), Onavo announced that it would close the Android version of Protect as well. History Onavo was founded in 2010 by Roi Tiger and Guy Rosen. Onavo had two rounds of funding: the first was a Series A investment for $3 million from Magma Venture Partners and Sequoia Capital in May 2011. The second was a Series B investment of $13 million from Magma Ventures, Sequoia Capital, and Horizons Ventures. Onavo's sale to Facebook is one of the top exits for Magma Venture Partners and other Israeli venture capital firms. On October 13, 2013, Facebook bought Onavo for approximately $120 million. The Australian Competition & Consumer Commission (ACCC) initiated legal proceedings against Facebook on December 16, 2020, alleging that Facebook engaged in "false, misleading or deceptive conduct" by using personal data collected from Onavo "for its own commercial purposes" contrary to Onavo's
https://en.wikipedia.org/wiki/Flammulina%20filiformis
Flammulina filiformis is a species of agaric (gilled mushroom) in the family Physalacriaceae. It is widely cultivated in East Asia, and well known for its role in Japanese and Chinese cuisine. Until recently, the species was considered to be conspecific with the European Flammulina velutipes, but DNA sequencing has shown that the two are distinct. Description Basidiocarps are agaricoid and grow in clusters. Individual fruit bodies are up to tall, the cap convex at first, becoming flat when expanded, up to across. The cap surface is smooth, viscid when damp, ochraceous yellow to yellow-brown. The lamellae (gills) are cream to yellowish white. The stipe (stem) is smooth, pale yellow at the apex, yellow-brown to dark brown towards the base, and lacking a ring. The spore print is white, the spores (under a microscope) smooth, inamyloid, ellipsoid to cylindrical, c. 5 to 7 by 3 to 3.5μm. There is a significant difference in appearance between wild and cultivated basidiocarps. Cultivated enokitake are not exposed to light, resulting in white or pallid fruit bodies with long stipes and small caps. Taxonomy Flammulina filiformis was originally described from China in 2015 as a variety of F. velutipes, based on internal transcribed spacer sequences. Further molecular research using a combination of different sequences has shown that F. filiformis and F. velutipes are distinct and should be recognized as separate species. Etymology The names enokitake (榎茸、エノキタケ, ), enokidake (榎茸、エノキダケ) and enoki (榎、エノキ) are derived from the Japanese language. In Mandarin Chinese, the mushroom is called (, "gold needle mushroom") or 金菇 (jīngū, "gold mushroom"). In India it is called futu, in Korean, it is called paengi beoseot (팽이버섯) which means "mushroom planted near catalpa", and nấm kim châm in Vietnamese. Distribution and habitat The fungus is found on dead wood of Betula platyphylla, Broussonetia papyrifera, Dipentodon sinicus, Neolitsea sp., Salix spp, and other broad-leaved
https://en.wikipedia.org/wiki/Cell%20proliferation
Cell proliferation is the process by which a cell grows and divides to produce two daughter cells. Cell proliferation leads to an exponential increase in cell number and is therefore a rapid mechanism of tissue growth. Cell proliferation requires both cell growth and cell division to occur at the same time, such that the average size of cells remains constant in the population. Cell division can occur without cell growth, producing many progressively smaller cells (as in cleavage of the zygote), while cell growth can occur without cell division to produce a single larger cell (as in growth of neurons). Thus, cell proliferation is not synonymous with either cell growth or cell division, despite these terms sometimes being used interchangeably. Stem cells undergo cell proliferation to produce proliferating "transit amplifying" daughter cells that later differentiate to construct tissues during normal development and tissue growth, during tissue regeneration after damage, or in cancer. The total number of cells in a population is determined by the rate of cell proliferation minus the rate of cell death. Cell size depends on both cell growth and cell division, with a disproportionate increase in the rate of cell growth leading to production of larger cells and a disproportionate increase in the rate of cell division leading to production of many smaller cells. Cell proliferation typically involves balanced cell growth and cell division rates that maintain a roughly constant cell size in the exponentially proliferating population of cells. Cell proliferation occurs by combining cell growth with regular "G1-S-M-G2" cell cycles to produce many diploid cell progeny. In single-celled organisms, cell proliferation is largely responsive to the availability of nutrients in the environment (or laboratory growth medium). In multicellular organisms, the process of cell proliferation is tightly controlled by gene regulatory networks encoded in the genome and executed mainly
https://en.wikipedia.org/wiki/Thermal%20paste
Thermal paste (also called thermal compound, thermal grease, thermal interface material (TIM), thermal gel, heat paste, heat sink compound, heat sink paste or CPU grease) is a thermally conductive (but usually electrically insulating) chemical compound, which is commonly used as an interface between heat sinks and heat sources such as high-power semiconductor devices. The main role of thermal paste is to eliminate air gaps or spaces (which act as thermal insulation) from the interface area in order to maximize heat transfer and dissipation. Thermal paste is an example of a thermal interface material. As opposed to thermal adhesive, thermal paste does not add mechanical strength to the bond between heat source and heat sink. It has to be coupled with a fastener such as screws to hold the heat sink in place and to apply pressure, spreading and thinning the thermal paste. Composition Thermal paste consists of a polymerizable liquid matrix and large volume fractions of electrically insulating, but thermally conductive filler. Typical matrix materials are epoxies, silicones (silicone grease), urethanes, and acrylates; solvent-based systems, hot-melt adhesives, and pressure-sensitive adhesive tapes are also available. Aluminum oxide, boron nitride, zinc oxide, and increasingly aluminum nitride are used as fillers for these types of adhesives. The filler loading can be as high as 70–80% by mass, and raises the thermal conductivity of the base matrix from 0.17–0.3 W/(m·K) (watts per meter-kelvin) up to about 4 W/(m·K), according to a 2008 paper. Silver thermal compounds may have a conductivity of 3 to 8 W/(m·K) or more, and consist of micronized silver particles suspended in a silicone/ceramic medium. However, metal-based thermal paste can be electrically conductive and capacitive; if some flows onto the circuits, it can lead to malfunction and damage. The most effective (and most expensive) pastes consist almost entirely of liquid metal, usually a variation of the all
https://en.wikipedia.org/wiki/Articularis%20genus%20muscle
The articularis genus (also known as the subcrureus muscle) is a small skeletal muscle located anteriorly on the thigh just above the knee. Structure It arises from the anterior surface of the lower part of the body of the femur, deep to the vastus intermedius, close to the knee and from the deep fibers of the vastus intermedius. Its insertion is on the synovial membrane of the knee-joint. Blood supply It is supplied by the lateral femoral circumflex artery. Innervation It is innervated by branches of the femoral nerve (L2-L4). Variation Flat, wispy and highly variable, sometimes consisting of several separate muscular bundles, this muscle is without a distinct investing fascia and ranges 1.5–3 cm in width. It is usually distinct from the vastus intermedius, but occasionally blended with it. Function Articularis genus pulls the suprapatellar bursa superiorly during extension of the knee, and prevents impingement of the synovial membrane between the patella and the femur.
https://en.wikipedia.org/wiki/Order%20%28journal%29
Order (subtitled A Journal on the Theory of Ordered Sets and its Applications) is a quarterly peer-reviewed academic journal on order theory and its applications, published by Springer Science+Business Media. It was established in 1984 by Ivan Rival (University of Calgary). From 2010 to 2018, its editor-in-chief was Dwight Duffus (Emory University). He was succeeded in 2019 by Ryan R. Martin (Iowa State University). Abstracting and indexing The journal is abstracted and indexed in: According to the Journal Citation Reports, the journal has a 2017 impact factor of 0.353.
https://en.wikipedia.org/wiki/John%20Pell%20%28mathematician%29
John Pell (1 March 1611 – 12 December 1685) was an English mathematician and political agent abroad. Early life He was born at Southwick in Sussex. His father, also named John Pell, was from Southwick, and his mother was Mary Holland, from Halden in Kent. The second of two sons, Pell's older brother was Thomas Pell. By the time he was six, they were orphans, their father dying in 1616 and their mother the following year. John Pell the elder had a fine library, which proved valuable to the young Pell as he grew up. He was educated at Steyning Grammar School and entered Trinity College, Cambridge, at the age of 13. During his university career he became an accomplished linguist; even before taking a B.A. degree in 1629, he corresponded with Henry Briggs and other mathematicians. He was promoted by seniority to M.A. in 1630 and taught in the short-lived Chichester Academy set up by Samuel Hartlib. On 3 July 1632 he married Ithamaria Reginald (also rendered as Ithamara or Ithumaria, with the surname Reginolles), sister of the writer and polymath Bathsua Makin. They had four sons and four daughters. Ithumaria died in 1661. Some time before 1669 Pell remarried. Pell spent much of the 1630s working under Hartlib's influence, on topics in the area of pedagogy, encyclopedism and pansophy, combinatorics, and the legacy of Trithemius. By 1638 he had formulated a proposal for a universal language. In mathematics, he concentrated on expanding the scope of algebra in the theory of equations, and on mathematical tables. As part of a joint lobbying effort with Hartlib to find himself support to continue as a researcher, he had his short Idea of Mathematics printed in October 1638. It brought interested responses from Johann Moriaen and Marin Mersenne. Academic and diplomat His reputation and the influence of Sir William Boswell, the English resident, with the States-General procured his election in 1644 to the chair of mathematics in Amsterdam, after an earlier attempt immediate
https://en.wikipedia.org/wiki/Syntrophus
Syntrophus is a Gram negative bacterial genus from the family of Syntrophaceae. Phylogeny The currently accepted taxonomy is based on the List of Prokaryotic names with Standing in Nomenclature (LPSN) and National Center for Biotechnology Information (NCBI) See also List of bacterial orders List of bacteria genera
https://en.wikipedia.org/wiki/Shotgun%20surgery
Shotgun surgery is an antipattern in software development and occurs where a developer adds features to an application codebase which span a multiplicity of implementors or implementations in a single change. This is common practice in many programming scenarios, as a great amount of programming effort is usually expended on adding new features to increase the value of programming assets. As a consequence, these new features may require adding code in several places simultaneously where the code itself looks very similar and may only have slight variations. Owing to the fast-paced nature of commercial software development, there may not be sufficient time to remodel (or refactor) a system to support the new features trivially. As a consequence, the practice of copy-and-paste programming is prevalent; the code is written in a single place then simply copied to all other places where that implementation is required (with any required changes applied in-place). This practice is generally frowned on by the refactoring community as a direct violation of the Once and Only Once principle – ultimately any change to the new functionality may require widespread changes. Further, any potential software bug in this new feature will be replicated many-fold and can make bug fixing particularly difficult and tedious. Even in the absence of copied code, the implementations are guaranteed to be very similar and just as prone to requirements change or bug fixing. This form of software development tends to favour short-term improvement (in the form of additional features) at the cost of long-term maintainability and stability. Example The canonical example of this practice is logging which generally adds prologue code to many functions simultaneously, for example: void Func() { ... } void Func2() { ... } ... void FuncN() { ... } Could be transformed to: void Func() { printf("Entering Func\n"); ... } void Func2() { printf("Entering Func2\n"); ... } ... void Fun
https://en.wikipedia.org/wiki/Extinct%20radionuclide
An extinct radionuclide is a radionuclide that was formed by nucleosynthesis before the formation of the Solar System, about 4.6 billion years ago, but has since decayed to virtually zero abundance and is no longer detectable as a primordial nuclide. Extinct radionuclides were generated by various processes in the early Solar system, and became part of the composition of meteorites and protoplanets. All widely documented extinct radionuclides have half-lives shorter than 100 million years. Short-lived radioisotopes that are found in nature are continuously generated or replenished by natural processes, such as cosmic rays (cosmogenic nuclides), background radiation, or the decay chain or spontaneous fission of other radionuclides. Short-lived isotopes that are not generated or replenished by natural processes are not found in nature, so they are known as extinct radionuclides. Their former existence is inferred from a superabundance of their stable or nearly stable decay products. Examples of extinct radionuclides include iodine-129 (the first to be noted in 1960, inferred from excess xenon-129 concentrations in meteorites, in the xenon-iodine dating system), aluminium-26 (inferred from extra magnesium-26 found in meteorites), and iron-60. The Solar System and Earth are formed from primordial nuclides and extinct nuclides. Extinct nuclides have decayed away, but primordial nuclides still exist in their original state (undecayed). There are 251 stable primordial nuclides, and remainders of 35 primordial radionuclides that have very long half-lives. List of extinct radionuclides A partial list of radionuclides not found on Earth, but for which decay products are present: Plutonium-244 and samarium-146 have half-lives long enough to still be present on Earth, but they have not been confirmed experimentally to be present. Notable isotopes with shorter lives still being produced on Earth include: Manganese-53 and beryllium-10 are produced by cosmic ray spallation
https://en.wikipedia.org/wiki/Pseudoalgebra
In algebra, given a 2-monad T in a 2-category, a pseudoalgebra for T is a 2-category-version of algebra for T, that satisfies the laws up to coherent isomorphisms. See also Operad
https://en.wikipedia.org/wiki/Counting%20problem%20%28complexity%29
In computational complexity theory and computability theory, a counting problem is a type of computational problem. If R is a search problem then is the corresponding counting function and denotes the corresponding decision problem. Note that cR is a search problem while #R is a decision problem, however cR can be C Cook-reduced to #R (for appropriate C) using a binary search (the reason #R is defined the way it is, rather than being the graph of cR, is to make this binary search possible). Counting complexity class If NX is a complexity class associated with non-deterministic machines then #X = {#R | R ∈ NX} is the set of counting problems associated with each search problem in NX. In particular, #P is the class of counting problems associated with NP search problems. Just as NP has NP-complete problems via many-one reductions, #P has complete problems via parsimonious reductions, problem transformations that preserve the number of solutions. See also GapP External links Computational problems
https://en.wikipedia.org/wiki/Recent%20African%20origin%20of%20modern%20humans
In paleoanthropology, the recent African origin of modern humans or the "Out of Africa" theory (OOA) is the most widely accepted model of the geographic origin and early migration of anatomically modern humans (Homo sapiens). It follows the early expansions of hominins out of Africa, accomplished by Homo erectus and then Homo neanderthalensis. The model proposes a "single origin" of Homo sapiens in the taxonomic sense, precluding parallel evolution in other regions of traits considered anatomically modern, but not precluding multiple admixture between H. sapiens and archaic humans in Europe and Asia. H. sapiens most likely developed in the Horn of Africa between 300,000 and 200,000 years ago, although an alternative hypothesis argues that diverse morphological features of H. sapiens appeared locally in different parts of Africa and converged due to gene flow between different populations within the same period. The "recent African origin" model proposes that all modern non-African populations are substantially descended from populations of H. sapiens that left Africa after that time. There were at least several "out-of-Africa" dispersals of modern humans, possibly beginning as early as 270,000 years ago, including 215,000 years ago to at least Greece, and certainly via northern Africa and the Arabian Peninsula about 130,000 to 115,000 years ago. There is evidence that modern humans had reached China around 80,000 years ago. Practically all of these early waves seem to have gone extinct or retreated back, and present-day humans outside Africa descend mainly from a single expansion out 70,000–50,000 years ago. The most significant "recent" wave out of Africa took place about 70,000–50,000 years ago, via the so-called "Southern Route", spreading rapidly along the coast of Asia and reaching Australia by around 65,000–50,000 years ago, (though some researchers question the earlier Australian dates and place the arrival of humans there at 50,000 years ago at earliest,
https://en.wikipedia.org/wiki/Unavailable%20name
In zoological nomenclature, an unavailable name is a name that does not conform to the rules of the International Code of Zoological Nomenclature and that therefore is not available for use as a valid name for a taxon. Such a name does not fulfil the requirements in Articles 10 through 20 of the Code, or is excluded under Article 1.3. Unavailable names include names that have not been published, such as "Oryzomys hypenemus" and "Ubirajara jubatus", names without an accompanying description (nomina nuda), such as the subgeneric name Micronectomys proposed for the Nicaraguan rice rat, names proposed with a rank below that of subspecies (infrasubspecific names), such as Sorex isodon princeps montanus for a form of the taiga shrew, and various other categories. Despite the frequent confusion caused by common sense, an unavailable name is not necessarily a nomen nudum. A good examplification of this is the case of the unavailable dinosaur name "Ubirajara jubatus", which was assumed by common sense to be a nomen nudum before a detailed analysis of its nomenclatural status. Notes
https://en.wikipedia.org/wiki/CLCC1
Chloride channel CLIC-like 1 also known as CLCC1 is a human gene. The protein encoded by this gene is a chloride channel which is related in sequence to the S. cerevisiae MID-1 stretch-activated channel. CLCC1 is located in the membranes of intracellular compartments including endoplasmic reticulum and the Golgi apparatus. It is highly expressed in the testis and moderately in the spleen, liver, kidney, heart, brain, and lung.
https://en.wikipedia.org/wiki/Data%20manipulation%20language
A data manipulation language (DML) is a computer programming language used for adding (inserting), deleting, and modifying (updating) data in a database. A DML is often a sublanguage of a broader database language such as SQL, with the DML comprising some of the operators in the language. Read-only selecting of data is sometimes distinguished as being part of a separate data query language (DQL), but it is closely related and sometimes also considered a component of a DML; some operators may perform both selecting (reading) and writing. A popular data manipulation language is that of Structured Query Language (SQL), which is used to retrieve and manipulate data in a relational database. Other forms of DML are those used by IMS/DLI, CODASYL databases, such as IDMS and others. SQL In SQL, the data manipulation language comprises the SQL-data change statements, which modify stored data but not the schema or database objects. Manipulation of persistent database objects, e.g., tables or stored procedures, via the SQL schema statements, rather than the data stored within them, is considered to be part of a separate data definition language (DDL). In SQL these two categories are similar in their detailed syntax, data types, expressions etc., but distinct in their overall function. The SQL-data change statements are a subset of the SQL-data statements; this also contains the SELECT query statement, which strictly speaking is part of the DQL, not the DML. In common practice though, this distinction is not made and SELECT is widely considered to be part of DML, so the DML consists of all SQL-data statements, not only the SQL-data change statements. The SELECT ... INTO ... form combines both selection and manipulation, and thus is strictly considered to be DML because it manipulates (i.e. modifies) data. Data manipulation languages have their functional capability organized by the initial word in a statement, which is almost always a verb. In the case of SQL, these verbs
https://en.wikipedia.org/wiki/Carnegie%20collection
The Carnegie Collection was a series of authentic replicas based on dinosaurs and other extinct prehistoric creatures, using fossils featured at the Carnegie Museum of Natural History as references. They were produced by Florida-based company Safari Ltd., known for their hand-painted replicas, from 1988 to 2015, and became known as "the world’s premier line of scale model dinosaur figures." Description 65 models representing 53 species of dinosaurs and other prehistoric animals were produced for the line. Each of the models was hand-painted, ensuring that no two copies of the same model are identical. Each model was sculpted by artist Forest Rogers and authenticated by paleontologists associated with the Carnegie Museum, such as Matt Lamanna, as well as various species-specific experts. Most of the animals were sculpted at a 1:40 scale (where one inch on the model represents 40 inches on the real creature), although some models representing smaller creatures were produced at a larger scale. Models in the collection range greatly in size from 24 inches long (original Diplodocus) to only three inches long (original Dimetrodon) with all shapes and sizes represented in-between. On the underside of each model is information detailing its name, year of initial production, and copyright information. The models feature an informational hang tag providing scientific details about the animal represented by the replica. In some cases, the dinosaurs were packaged in cardboard display boxes, in which case a small booklet featuring information on each dinosaur featured in the collection was included in lieu of the hang tags. In some instances, two or three models would be packaged together in a box. Examples include Dimetrodon and Deinonychus, Protoceratops and Euoplocephalus, Apatosaurus and Apatosaurus Baby, Elasmosaurus and Mosasaurus, and Australopithecus Male/Female pair and Smilodon. The boxes are not often seen today, and most of the time the dinosaurs are found fre
https://en.wikipedia.org/wiki/Purity%20test
A purity test is a self-graded survey that assesses the participants' supposed degree of innocence in worldly matters (sex, drugs, deceit, and other activities assumed to be vices), generally on a percentage scale with 100% being the most and 0% being the least pure. Online purity tests were among the earliest of Internet memes, popular on Usenet beginning in the early 1980s. However, similar types of tests circulated under various names long before the existence of the Internet. Historical examples The Columbia University humor magazine, The Jester, reported in its October 1935 issue on a campus wide "purity test" conducted at Barnard College in 1935. The issue of The Jester was briefly censored, with distribution curtailed until the director of activities at the university could review the article. According to the editor-in-chief of The Jester, "We printed the survey to clear up some of the misconceptions that Columbia and the outside world have about Barnard girls," he said. "The results seem to establish that Barnard girls are quite regular. I fail to see anything off-color in the story. It's a sociological study." In 1936, The Indian Express reported that students at Toronto University were "under-going a 'purity test', which took "the form of twenty very personal questions, designed to determine the state of their morals and their 'purity ratio'. For example, so many marks are lost for smoking, drinking, and every time the sinner kisses a girl or boy. Then, after truthfully answering all the questions, the total number of bad marks are added up and subtracted from a hundred. What is left, if any, is the 'purity ratio'. The test is unofficial and just what it will prove when completed nobody knows." Alan Dundes, a professor of anthropology and folklore at the University of California, Berkeley, and Carl R. Pagter included examples of purity tests in their 1975 book Work Hard and You Shall Be Rewarded: Urban Folklore from the Paperwork Empire. They noted, "
https://en.wikipedia.org/wiki/Wildlife
Wildlife refers to undomesticated animal species, but has come to include all organisms that grow or live wild in an area without being introduced by humans. Wildlife was also synonymous to game: those birds and mammals that were hunted for sport. Wildlife can be found in all ecosystems. Deserts, plains, grasslands, woodlands, forests, and other areas, including the most developed urban areas, all have distinct forms of wildlife. While the term in popular culture usually refers to animals that are untouched by human factors, most scientists agree that much wildlife is affected by human activities. Some wildlife threaten human safety, health, property, and quality of life. However, many wild animals, even the dangerous ones, have value to human beings. This value might be economic, educational, or emotional in nature. Humans have historically tended to separate civilization from wildlife in a number of ways, including the legal, social, and moral senses. Some animals, however, have adapted to suburban environments. This includes such animals as feral cats, dogs, mice, and rats. Some religions declare certain animals to be sacred, and in modern times, concern for the natural environment has provoked activists to protest against the exploitation of wildlife for human benefit or entertainment. Global wildlife populations have decreased by 68% since 1970 as a result of human activity, particularly overconsumption, population growth, and intensive farming, according to a 2020 World Wildlife Fund's Living Planet Report and the Zoological Society of London's Living Planet Index measure, which is further evidence that humans have unleashed a sixth mass extinction event. According to CITES, it has been estimated that annually the international wildlife trade amounts to billions of dollars and it affects hundreds of millions of animal and plant specimen. Interactions with humans Trade For food Stone Age people and hunter-gatherers relied on wildlife, both plants and anim
https://en.wikipedia.org/wiki/Shamba%20%28agroforestry%20system%29
Shamba (Swahili for 'plantation'; pl. Mashamba) is an agroforestry system practiced in East Africa, particularly Kenya. In these lands various crops are combined: bananas, beans, yams and corn, to which are added timber resources, beekeeping, medicinal herbs, mushrooms, forest fruits, fodder for livestock, etc. Thanks to this polyculture, farmers obtain a higher share of income, food and employment. Furthermore, it is more sustainable and generates a smaller ecological footprint than monocultures. The system was created in Kenya in the mid-19th century to extend land cultivation and meet the demand for firewood. A 2009 study estimated that in Kenya, where the state owns these plantations, there are some 160,000 hectares of shambas. However, the management of the shambas has been tainted by corruption, which Kenyan farmers' associations have denounced on many occasions. Large wooded areas were allowed to be converted to farmland. Moreover, the introduction of non-native species such as eucalyptus or cypress has been a problem. Wangari Maathai called for "not sacrificing native forests at the expense of exotic plantations". See also Kuojtakiloyan, agroforestry system in Mexico Dehesa, agroforestry system in Iberia Taungya, agroforestry system in Myanmar
https://en.wikipedia.org/wiki/Strange%20matter
Strange matter (or strange quark matter) is quark matter containing strange quarks. In extreme environments, strange matter is hypothesized to occur in the core of neutron stars, or, more speculatively, as isolated droplets that may vary in size from femtometers (strangelets) to kilometers, as in the hypothetical strange stars. At high enough density, strange matter is expected to be color superconducting. Ordinary matter, also referred to as atomic matter, is composed of atoms, with nearly all matter concentrated in the atomic nuclei. Nuclear matter is a liquid composed of neutrons and protons, and they are themselves composed of up and down quarks. Quark matter is a condensed form of matter composed entirely of quarks. When quark matter does not contain strange quarks, it is sometimes referred to as non-strange quark matter. Context In particle physics and astrophysics, the term 'strange matter' is used in two different contexts, one broader and the other more specific and hypothetical: In the broader context, our current understanding of the laws of nature predicts that strange matter could be created when nuclear matter (made of protons and neutrons) is compressed beyond a critical density. At this critical pressure and density, the protons and neutrons dissociate into quarks, yielding quark matter and potentially strange matter. A more specific hypothesis is that quark matter is the true ground state of all matter, and thus more stable than ordinary nuclear matter. This idea is known as the "strange matter hypothesis", or the Bodmer–Witten assumption. Under this hypothesis, the nuclei of the atoms we see around us are only metastable, even when the external critical pressure is zero, and given enough time (or the right stimulus) the nuclei would decay into stable droplets of strange matter. Droplets of strange matter are also referred to as strangelets. Stability of strange matter only at high pressure In the general context, strange matter might occur
https://en.wikipedia.org/wiki/Hyper-IgM%20syndrome%20type%205
The fifth type of hyper-IgM syndrome has been characterized in three patients from France and Japan. The symptoms are similar to hyper IgM syndrome type 2, but the AICDA gene is intact. These three patients instead had mutations in the catalytic domain of uracil-DNA glycosylase, an enzyme that removes uracil from DNA. In hyper-IgM syndromes, patients are deficient in the immunoglobulins, IgG, IgE and IgA types since the antibody producing B cells can not carry out the gene recombination steps necessary to class switch from immunoglobulin M (IgM) to the other three immunoglobulins types. Hyper IgM syndromes Hyper IgM syndromes is a group of primary immune deficiency disorders characterized by defective CD40 signaling; via B cells affecting class switch recombination (CSR) and somatic hypermutation. Immunoglobulin (Ig) class switch recombination deficiencies are characterized by elevated serum IgM levels and a considerable deficiency in Immunoglobulins G (IgG), A (IgA) and E (IgE). As a consequence, people with HIGM have an increased susceptibility to infections. Signs and symptoms Hyper IgM syndrome can have the following syndromes: Infection/Pneumocystis pneumonia (PCP), which is common in infants with hyper IgM syndrome, is a serious illness. PCP is one of the most frequent and severe opportunistic infections in people with weakened immune systems. Hepatitis (Hepatitis C) Chronic diarrhea Hypothyroidism Neutropenia Arthritis Encephalopathy (degenerative) Cause Different genetic defects cause HIgM syndrome, the vast majority are inherited as an X-linked recessive genetic trait and most with the condition are male. IgM is the form of antibody that all B cells produce initially before they undergo class switching. Healthy B cells efficiently switch to other types of antibodies as needed to attack invading bacteria, viruses, and other pathogens. In people with hyper IgM syndromes, the B cells keep making IgM antibodies because can not switch to a differe
https://en.wikipedia.org/wiki/Pendulum%20%28mechanics%29
A pendulum is a body suspended from a fixed support so that it swings freely back and forth under the influence of gravity. When a pendulum is displaced sideways from its resting, equilibrium position, it is subject to a restoring force due to gravity that will accelerate it back toward the equilibrium position. When released, the restoring force acting on the pendulum's mass causes it to oscillate about the equilibrium position, swinging it back and forth. The mathematics of pendulums are in general quite complicated. Simplifying assumptions can be made, which in the case of a simple pendulum allow the equations of motion to be solved analytically for small-angle oscillations. Simple gravity pendulum A simple gravity pendulum is an idealized mathematical model of a real pendulum. This is a weight (or bob) on the end of a massless cord suspended from a pivot, without friction. Since in this model there is no frictional energy loss, when given an initial displacement it will swing back and forth at a constant amplitude. The model is based on these assumptions: The rod or cord on which the bob swings is massless, inextensible and always remains taut. The bob is a point mass. Motion occurs only in two dimensions, i.e. the bob does not trace an ellipse but an arc. The motion does not lose energy to friction or air resistance. The gravitational field is uniform. The support does not move. The differential equation which represents the motion of a simple pendulum is where is the magnitude of the gravitational field, is the length of the rod or cord, and is the angle from the vertical to the pendulum. Small-angle approximation The differential equation given above is not easily solved, and there is no solution that can be written in terms of elementary functions. However, adding a restriction to the size of the oscillation's amplitude gives a form whose solution can be easily obtained. If it is assumed that the angle is much less than 1 radian (often cite
https://en.wikipedia.org/wiki/Stephen%20L.%20Adler
Stephen Louis Adler (born November 30, 1939) is an American physicist specializing in elementary particles and field theory. He is currently professor emeritus in the school of natural sciences at the Institute for Advanced Study in Princeton, New Jersey. Biography Adler was born in New York City. He received an A.B. degree at Harvard University in 1961, where he was a Putnam Fellow in 1959, and a Ph.D. from Princeton University in 1964. Adler completed his doctoral dissertation, titled High energy neutrino reactions and conservations hypotheses, under the supervision of Sam Treiman. He is the son of Irving Adler, an American author, teacher, mathematician, scientist and political activist, and Ruth Adler and older brother of Peggy Adler. Adler became a member of the Institute for Advanced Study in 1966, becoming a full professor of theoretical physics in 1969, and was named "New Jersey Albert Einstein Professor" at the institute in 1979. He was elected a member of the American Academy of Arts and Sciences in 1974, and a member of the National Academy of Sciences in 1975. He has won the J. J. Sakurai Prize from the American Physical Society in 1988, and the Dirac Medal of the International Centre for Theoretical Physics in 1998, among other awards. Adler's seminal papers on high energy neutrino processes, current algebra, soft pion theorems, sum rules, and perturbation theory anomalies helped lay the foundations for the current standard model of elementary particle physics. In 2012, Adler contributed to a family venture when he wrote the foreword for his then 99-year-old father's 87th book, Solving the Riddle of Phyllotaxis: Why the Fibonacci Numbers and the Golden Ratio Occur on Plants. The book's diagrams are by his sister Peggy. Trace dynamics In his book Quantum Theory as an Emergent Phenomenon, published 2004, Adler presented his trace dynamics, a framework in which quantum field theory emerges from a matrix theory. In this matrix theory, particles are
https://en.wikipedia.org/wiki/Petzval%20field%20curvature
Petzval field curvature, named for Joseph Petzval, describes the optical aberration in which a flat object normal to the optical axis (or a non-flat object past the hyperfocal distance) cannot be brought properly into focus on a flat image plane. Field curvature can be corrected with the use of a field flattener, designs can also incorporate a curved focal plane like in the case of the human eye in order to improve image quality at the focal surface. Analysis Consider an "ideal" single-element lens system for which all planar wave fronts are focused to a point at distance f from the lens. Placing this lens the distance f from a flat image sensor, image points near the optical axis will be in perfect focus, but rays off axis will come into focus before the image sensor, dropping off by the cosine of the angle they make with the optical axis. This is less of a problem when the imaging surface is spherical, as in the human eye. Most current photographic lenses are designed to minimize field curvature, and so effectively have a focal length that increases with ray angle. Lenses of short focal lengths (ultra wide, wide and normal) below 50 mm typically suffer more from field curvature. Telephoto lenses typically have very little or no visible field curvature. The Petzval lens is one design which has significant field curvature; images taken with the lens are very sharp in the centre, but at greater angles the image is out of focus. Film cameras may be able to bend their image planes to compensate, particularly when the lens is fixed and known. This also includes plate film, which could still be bent slightly. Digital sensors are difficult to bend, although experimental products have been produced. By 2016 the only consumer cameras featuring curved sensors were "selfie" Sony Cybershot KW-1 and KW-11. Large mosaics of sensors (necessary anyway due to limited chip sizes) can be shaped to simulate a bend over larger scales. The Petzval field curvature is equal to the
https://en.wikipedia.org/wiki/Order%20convergence
In mathematics, specifically in order theory and functional analysis, a filter in an order complete vector lattice is order convergent if it contains an order bounded subset (that is, is contained in an interval of the form ) and if where is the set of all order bounded subsets of X, in which case this common value is called the order limit of in Order convergence plays an important role in the theory of vector lattices because the definition of order convergence does not depend on any topology. Definition A net in a vector lattice is said to decrease to if implies and in A net in a vector lattice is said to order-converge to if there is a net in that decreases to and satisfies for all . Order continuity A linear map between vector lattices is said to be order continuous if whenever is a net in that order-converges to in then the net order-converges to in is said to be sequentially order continuous if whenever is a sequence in that order-converges to in then the sequence order-converges to in Related results In an order complete vector lattice whose order is regular, is of minimal type if and only if every order convergent filter in converges when is endowed with the order topology. See also
https://en.wikipedia.org/wiki/Analogue%20electronics
Analogue electronics () are electronic systems with a continuously variable signal, in contrast to digital electronics where signals usually take only two levels. The term "analogue" describes the proportional relationship between a signal and a voltage or current that represents the signal. The word analogue is derived from the Greek word meaning "proportional". Analogue signals An analogue signal uses some attribute of the medium to convey the signal's information. For example, an aneroid barometer uses the angular position of a needle on top of a contracting and expanding box as the signal to convey the information of changes in atmospheric pressure. Electrical signals may represent information by changing their voltage, current, frequency, or total charge. Information is converted from some other physical form (such as sound, light, temperature, pressure, position) to an electrical signal by a transducer which converts one type of energy into another (e.g. a microphone). The signals take any value from a given range, and each unique signal value represents different information. Any change in the signal is meaningful, and each level of the signal represents a different level of the phenomenon that it represents. For example, suppose the signal is being used to represent temperature, with one volt representing one degree Celsius. In such a system, 10 volts would represent 10 degrees, and 10.1 volts would represent 10.1 degrees. Another method of conveying an analogue signal is to use modulation. In this, some base carrier signal has one of its properties altered: amplitude modulation (AM) involves altering the amplitude of a sinusoidal voltage waveform by the source information, frequency modulation (FM) changes the frequency. Other techniques, such as phase modulation or changing the phase of the carrier signal, are also used. In an analogue sound recording, the variation in pressure of a sound striking a microphone creates a corresponding variation in t
https://en.wikipedia.org/wiki/Tetrad%20%28geometry%20puzzle%29
In geometry, a tetrad is a set of four simply connected disjoint planar regions in the plane, each pair sharing a finite portion of common boundary. It was named by Michael R. W. Buckley in 1975 in the Journal of Recreational Mathematics. A further question was proposed that became a puzzle, whether the 4 regions could be congruent, with or without holes, other enclosed regions. Fewest sides and vertices The solutions with four congruent tiles include some with five sides. However, their placement surrounds an uncovered hole in the plane. Among solutions without holes, the ones with the fewest possible sides are given by a hexagon identified by Scott Kim as a student at Stanford University. It is not known whether five-sided solutions without holes are possible. Kim's solution has 16 vertices, while some of the pentagon solutions have as few as 11 vertices. It is not known whether fewer vertices are possible. Congruent polyform solutions Gardner offered a number of polyform (polyomino, polyiamond, and polyhex) solutions, with no holes.
https://en.wikipedia.org/wiki/List%20of%20common%20microcontrollers
This is a list of common microcontrollers listed by brand. Altera In 2015, Altera was acquired by Intel. Nios II 32-bit configurable soft microprocessor Nios 16-bit configurable soft processor Analog Devices Blackfin Super Harvard Architecture Single-Chip Computer (SHARC) TigerSHARC ADSP-21xx digital signal processor MicroConverter Family – ARM7 and 8051 cores ARM While Arm is a fabless semiconductor company (it does not manufacture or sell its own chips), it licenses the ARM architecture family design to a variety of companies. Those companies in turn sell billions of ARM-based chips per year—12 billion ARM-based chips shipped in 2014, about 24 billion ARM-based chips shipped in 2020, some of those are popular chips in their own right. Atmel In 2016, Atmel was sold to Microchip Technology. AT89 series (Intel 8051 architecture) AT90, ATtiny, ATmega, ATxmega series (AVR architecture) (Atmel Norway design) AT91SAM (ARM architecture) AVR32 (32-bit AVR architecture) (Atmel Norway design) MARC4 Cypress Semiconductor In 2020, Cypress Semiconductor was acquired by Infineon Technologies. CY8C2xxxx (PSoC1), M8C CY8C3xxxx (PSoC3), 8051 CY8C4xxxx (PSoC4), ARM Cortex-M0 CY8C5xxxx (PSoC5), ARM Cortex-M3 PSoC (Programmable System on Chip) ELAN Microelectronics Corp. ELAN Microelectronics Corporation is an IC designer and provider of 8-bit microcontrollers and PC Peripheral ICs. Headquartered in Hsinchu Science Park, the Silicon Valley of Taiwan, ELAN's microcontroller product range includes the following: EM78PXXX Low Pin-Count MCU Family EM78PXXX GPIO Type MCU Family EM78PXXXN ADC Type MCU Family These are clones of the 12- and 14-bit Microchip PIC line of processors, but with a 13-bit instruction word. EPSON Semiconductor 4-bit S1C6x family 8-bit S1C88 family 16-bit S1C17 family 32-bit S1C33 family Espressif Systems Espressif Systems, a company with headquarters in Shanghai, China made its debut in the microcontroller scene with their
https://en.wikipedia.org/wiki/Hyperlexia
Hyperlexia is a syndrome characterized by a child's precocious ability to read. It was initially identified by Norman E. Silberberg and Margaret C. Silberberg (1967), who defined it as the precocious ability to read words without prior training in learning to read, typically before the age of five. They indicated that children with hyperlexia have a significantly higher word-decoding ability than their reading comprehension levels. Children with hyperlexia also present with an intense fascination for written material at a very early age. Hyperlexic children are characterized by word-reading ability well above what would be expected given their age. First named and scientifically described in 1967 (Silverberg and Silverberg), it can be viewed as a superability in which word recognition ability goes far above expected levels of skill. Some hyperlexics, however, have trouble understanding speech. Some experts believe that most children with hyperlexia, or perhaps even all of them, lie on the autism spectrum. However, one expert, Darold Treffert, proposes that hyperlexia has subtypes, only some of which overlap with autism. Between five and twenty percent of autistic children have been estimated to be hyperlexic. Hyperlexic children are often fascinated by letters or numbers. They are extremely good at decoding language and thus often become very early readers. Some English-speaking hyperlexic children learn to spell long words (such as elephant) before they are two years old and learn to read whole sentences before they turn three. Etymology The word hyperlexia is derived from the Greek terms 'over, beyond, overmuch, above measure' and 'word'. Development Although hyperlexic children usually learn to read in a non-communicative way, several studies have shown that they can acquire reading comprehension and communicative language after the onset of hyperlexia. They follow a different developmental trajectory relative to neurotypical individuals, with milestones b
https://en.wikipedia.org/wiki/Caldisericum
Caldisericum exile is a species of bacteria sufficiently distinct from other bacteria to be placed in its own family, order, class and phylum. It is the first member of the thermophilic candidate phylum OP5 to be cultured and described.
https://en.wikipedia.org/wiki/Kleo%20Bare%20Metal%20Backup
Kleo Bare Metal Backup for Servers is a free file system cloning program created by Carroll-Net Inc. Kleo creates an all-inclusive backup of the operating system, device drivers and applications. Kleo is bundled with a LiveCD which includes tools for server recovery. Operation Kleo is a graphical wizard file system cloning system. It is designed as a disaster-recovery tool, producing complete file system images. In the event a server experiences a hard drive failure, the LiveCD can be booted, and the previous backup image can be used to recover the server to the point in time of the last backup. Kleo backups are file system aware backups. Unused blocks and empty space are not included which reduces the size of the backup image. The backup images are also passed through gzip to further compress their size. But even with compression, backup images can be quite large. For this reason, they are not suitable replacements for daily backups and are typically only done whenever significant changes are made to a server's application set or prior to applying service patches and hotfixes. Supported storage media Hard drive USB/Firewire drive CIFS Windows network share SSH server NFS server Supported file systems Known limitations Kleo does NOT support Ext4 which is the default on new Ubuntu installations The destination partition must be the same or larger than the source partition (limitation of the underlying partimage tool) File format The file format written is a gzipped partimage file. The file extension is .KB2 The backup image is broken into 2 GB chunks to make it possible to store the backups on DVD. See also Disk cloning List of disk cloning software
https://en.wikipedia.org/wiki/Numeracy
Numeracy is the ability to understand, reason with, and to apply simple numerical concepts. The charity National Numeracy states: "Numeracy means understanding how mathematics is used in the real world and being able to apply it to make the best possible decisions...It’s as much about thinking and reasoning as about 'doing sums'". Basic numeracy skills consist of comprehending fundamental arithmetical operations like addition, subtraction, multiplication, and division. For example, if one can understand simple mathematical equations such as 2 + 2 = 4, then one would be considered to possess at least basic numeric knowledge. Substantial aspects of numeracy also include number sense, operation sense, computation, measurement, geometry, probability and statistics. A numerically literate person can manage and respond to the mathematical demands of life. By contrast, innumeracy (the lack of numeracy) can have a negative impact. Numeracy has an influence on healthy behaviors, financial literacy, and career decisions. Therefore, innumeracy may negatively affect economic choices, financial outcomes, health outcomes, and life satisfaction. It also may distort risk perception in health decisions. Greater numeracy has been associated with reduced susceptibility to framing effects, less influence of nonnumerical information such as mood states, and greater sensitivity to different levels of numerical risk. Ellen Peters and her colleagues argue that achieving the benefits of numeric literacy, however, may depend on one's numeric self-efficacy or confidence in one's skills. Representation of numbers Humans have evolved to mentally represent numbers in two major ways from observation (not formal math). These representations are often thought to be innate (see Numerical cognition), to be shared across human cultures, to be common to multiple species, and not to be the result of individual learning or cultural transmission. They are: Approximate representation of numerical magni
https://en.wikipedia.org/wiki/Dewberry
The dewberries are a group of species in the genus Rubus, section Rubus, closely related to the blackberries. They are small trailing (rather than upright or high-arching) brambles with aggregate fruits, reminiscent of the raspberry, but are usually purple to black instead of red. Alternatively, they are sometimes referred to as ground berries. Description The plants do not have upright canes like some other Rubus species, but have stems that trail along the ground, putting forth new roots along the length of the stem. The stems are covered with fine spines or stickers. Around March and April, the plants start to grow white flowers that develop into small green berries. The tiny green berries grow red and then a deep purple-blue as they ripen. When the berries are ripe, they are tender and difficult to pick in any quantity without squashing them. The berries are sweet and often less seedy than blackberries. In the winter the leaves often remain on the stems, but may turn dark red. The European dewberry, Rubus caesius, grows more upright like other brambles. Its fruits are a deep, almost black, purple and are coated with a thin layer or 'dew' of waxy droplets. Thus, they appear sky-blue (caesius being Latin for pale blue). Its fruits are small and retain a markedly tart taste even when fully ripe. Species Rubus Section Caesii, European dewberry European dewberry, Rubus caesius L. Rubus Section Flagellares, American dewberries Rubus aboriginum Rydb., synonyms: Rubus almus (L.H. Bailey) L.H.Bailey Rubus austrinus L.H.Bailey Rubus bollianus L.H.Bailey Rubus clair-brownii''' L.H.Bailey Rubus decor L.H. Bailey Rubus flagellaris Willd. var. almus L.H.Bailey Rubus foliaceus L.H. Bailey Rubus ignarus L.H. Bailey Rubus ricei L.H. Bailey Aberdeen dewberry, Rubus depavitus L.H.Bailey Northern dewberry, Rubus flagellaris Willd. Swamp dewberry, Rubus hispidus L. Upland dewberry, Rubus invisus (L.H.Bailey) Britton Pacific dewberry, Rubus ursinus Cham. & Sc
https://en.wikipedia.org/wiki/Geoxyle
A geoxyle is a plant in which an enlarged, woody structure occurs beneath the surface of the ground. Such plants have developed independently in various plant lineages, mostly evolving in the Pliocene and subsequently diverging within the last two million years. In contrast to their close relatives, these plants have developed in areas with both high rainfall and a high frequency of fires. They are sometimes known as underground trees, and the areas where they grow as underground forests. The geoxylic growth forms of woody subshrubs is characterised by massive lignotubers or underground woody axes from which emerge aerial shoots which may be ephemeral. These growth forms are found in savannahs in southern Africa. It is thought they developed in tandem with the spread of savannahs which resulted in an increase in tall grasses which are easily flammable during the long dry season associated with the savannah climate. Some well-known examples of geoxyles are the sand apple (Parinari capensis), the plough-breaker (Erythrina zeyheri), the red wings (Combretum platypetalum) and the wild grape (Lannea edulis). Others are Ancylobothrys petersiana, Diospyros galpinii, Elephantorrhiza elephantina, Eugenia albanensis, Eugenia capensis, Maytenus nemorosa, Pachystigma venosum and Salacia kraussii.
https://en.wikipedia.org/wiki/Guess%20value
In mathematical modeling, a guess value is more commonly called a starting value or initial value. These are necessary for most optimization problems which use search algorithms, because those algorithms are mainly deterministic and iterative, and they need to start somewhere. One common type of application is nonlinear regression. Use The quality of the initial values can have a considerable impact on the success or lack of such of the search algorithm. This is because the fitness function or objective function (in many cases a sum of squared errors (SSE)) can have difficult shapes. In some parts of the search region, the function may increase exponentially, in others quadratically, and there may be regions where the function asymptotes to a plateau. Starting values that fall in an exponential region can lead to algorithm failure because of arithmetic overflow. Starting values that fall in the asymptotic plateau region can lead to algorithm failure because of "dithering". Deterministic search algorithms may use a slope function to go to a minimum. If the slope is very small, then underflow errors can cause the algorithm to wander, seemingly aimlessly; this is dithering. Finding value Guess values can be determined a number of ways. Guessing is one of them. If one is familiar with the type of problem, then this is an educated guess or guesstimate. Other techniques include linearization, solving simultaneous equations, reducing dimensions, treating the problem as a time series, converting the problem to a (hopefully) linear differential equation, and using mean values. Further methods for determining starting values and optimal values in their own right come from stochastic methods, the most commonly known of these being evolutionary algorithms and particularly genetic algorithms. Mathematical optimization Regression analysis Computational statistics
https://en.wikipedia.org/wiki/Dendroclimatology
Dendroclimatology is the science of determining past climates from trees (primarily properties of the annual tree rings). Tree rings are wider when conditions favor growth, narrower when times are difficult. Other properties of the annual rings, such as maximum latewood density (MXD) have been shown to be better proxies than simple ring width. Using tree rings, scientists have estimated many local climates for hundreds to thousands of years previous. By combining multiple tree-ring studies (sometimes with other climate proxy records), scientists have estimated past regional and global climates. Advantages Tree rings are especially useful as climate proxies in that they can be well-dated via dendrochronology, i.e. matching of the rings from sample to sample. This allows extension backwards in time using deceased tree samples, even using samples from buildings or from archeological digs. Another advantage of tree rings is that they are clearly demarked in annual increments, as opposed to other proxy methods such as boreholes. Furthermore, tree rings respond to multiple climatic effects (temperature, moisture, cloudiness), so that various aspects of climate (not just temperature) can be studied. However, this can be a double-edged sword. Limitations Along with the advantages of dendroclimatology are some limitations: confounding factors, geographic coverage, annular resolution, and collection difficulties. The field has developed various methods to partially adjust for these challenges. Confounding factors There are multiple climate and non-climate factors as well as nonlinear effects that impact tree ring width. Methods to isolate single factors (of interest) include botanical studies to calibrate growth influences and sampling of "limiting stands" (those expected to respond mostly to the variable of interest). Climate factors Climate factors that affect trees include temperature, precipitation, sunlight, and wind. To differentiate among these factors, sc
https://en.wikipedia.org/wiki/Irony%20punctuation
Irony punctuation is any form of notation proposed or used to denote irony or sarcasm in text. Written text, in English and other languages, lacks a standard way to mark irony, and several forms of punctuation have been proposed to fill the gap. The oldest is the percontation point in the form of a reversed question mark (), proposed by English printer Henry Denham in the 1580s for marking rhetorical questions, which can be a form of irony. Specific irony marks have also been proposed, such as in the form of an open upward arrow (△|), used by Marcellin Jobard in the 19th century, and in a form resembling a reversed question mark (), proposed by French poet Alcanter de Brahm during the 19th century. Irony punctuation is primarily used to indicate that a sentence should be understood at a second level. A bracketed exclamation point or question mark as well as scare quotes are also occasionally used to express irony or sarcasm. Percontation point The percontation point , a reversed question mark later referred to as a rhetorical question mark, was proposed by Henry Denham in the 1580s and was used at the end of a question that does not require an answer—a rhetorical question. Its use died out in the 17th century. This character can be represented using the reversed question mark (⸮) found in Unicode as U+2E2E; another character approximating it is the Arabic question mark (؟), U+061F. The modern question mark (? U+003F) is descended from the "punctus interrogativus" (described as "a lightning flash, striking from right to left"), but unlike the modern question mark, the punctus interrogativus may be contrasted with the punctus percontativus—the former marking questions that require an answer while the latter marks rhetorical questions. Irony mark In 1668, John Wilkins, in An Essay towards a Real Character and a Philosophical Language, proposed using an inverted exclamation mark to punctuate rhetorical questions. In an article dated 11 October 1841, Marcellin Job
https://en.wikipedia.org/wiki/Device%20fingerprint
A device fingerprint or machine fingerprint is information collected about the software and hardware of a remote computing device for the purpose of identification. The information is usually assimilated into a brief identifier using a fingerprinting algorithm. A browser fingerprint is information collected specifically by interaction with the web browser of the device. Device fingerprints can be used to fully or partially identify individual devices even when persistent cookies (and zombie cookies) cannot be read or stored in the browser, the client IP address is hidden, or one switches to another browser on the same device. This may allow a service provider to detect and prevent identity theft and credit card fraud, but also to compile long-term records of individuals' browsing histories (and deliver targeted advertising or targeted exploits) even when they are attempting to avoid tracking – raising a major concern for internet privacy advocates. History Basic web browser configuration information has long been collected by web analytics services in an effort to measure real human web traffic and discount various forms of click fraud. Since its introduction in the late 1990s, client-side scripting has gradually enabled the collection of an increasing amount of diverse information, with some computer security experts starting to complain about the ease of bulk parameter extraction offered by web browsers as early as 2003. In 2005, researchers at the University of California, San Diego showed how TCP timestamps could be used to estimate the clock skew of a device, and consequently to remotely obtain a hardware fingerprint of the device. In 2010, Electronic Frontier Foundation launched a website where visitors can test their browser fingerprint. After collecting a sample of 470161 fingerprints, they measured at least 18.1 bits of entropy possible from browser fingerprinting, but that was before the advancements of canvas fingerprinting, which claims to add anoth
https://en.wikipedia.org/wiki/Automatic%20Computing%20Engine
The Automatic Computing Engine (ACE) was a British early electronic serial stored-program computer design by Alan Turing. Turing completed the ambitious design in late 1945, having had experience in the years prior with the secret Colossus computer at Bletchley Park. The ACE was not built, but a smaller version, the Pilot ACE, was constructed at the National Physical Laboratory and became operational in 1950. A larger implementation of the ACE design was the MOSAIC computer which became operational in 1955. ACE also led to the Bendix G-15 and other computers. Background The project was managed by John R. Womersley, superintendent of the Mathematics Division of the National Physical Laboratory (NPL). The use of the word Engine was in homage to Charles Babbage and his Difference Engine and Analytical Engine. Turing's technical design Proposed Electronic Calculator was the product of his theoretical work in 1936 "On Computable Numbers" and his wartime experience at Bletchley Park where the Colossus computers had been successful in breaking German military codes. In his 1936 paper, Turing described his idea as a "universal computing machine", but it is now known as the Universal Turing machine. Turing was sought by Womersley to work in the NPL on the ACE project; he accepted and began work on 1 October 1945 and by the end of the year he completed his outline of his 'Proposed electronic calculator', which was the first reasonably complete design of a stored-program computer and, apart from being on a much larger scale than the final working machine, anticipated the final realisation in most important respects. However, because of the strict and long-lasting secrecy around the Bletchley Park work, he was prohibited (because of the Official Secrets Act) from explaining that he knew that his ideas could be implemented in an electronic device. The better-known EDVAC design presented in the First Draft of a Report on the EDVAC (dated 30 June 1945), by John von Neumann,
https://en.wikipedia.org/wiki/Heliconia%20%C3%97%20flabellata
Heliconia × flabellata is a species of plant in the family Heliconiaceae. It is endemic to Ecuador. Its natural habitat is subtropical or tropical moist lowland forest. It is apparently a hybrid, H. episcopalis × H. rostrata.
https://en.wikipedia.org/wiki/Voltage-gated%20calcium%20channel
Voltage-gated calcium channels (VGCCs), also known as voltage-dependent calcium channels (VDCCs), are a group of voltage-gated ion channels found in the membrane of excitable cells (e.g., muscle, glial cells, neurons, etc.) with a permeability to the calcium ion Ca2+. These channels are slightly permeable to sodium ions, so they are also called Ca2+–Na+ channels, but their permeability to calcium is about 1000-fold greater than to sodium under normal physiological conditions. At physiologic or resting membrane potential, VGCCs are normally closed. They are activated (i.e.: opened) at depolarized membrane potentials and this is the source of the "voltage-gated" epithet. The concentration of calcium (Ca2+ ions) is normally several thousand times higher outside the cell than inside. Activation of particular VGCCs allows a Ca2+ influx into the cell, which, depending on the cell type, results in activation of calcium-sensitive potassium channels, muscular contraction, excitation of neurons, up-regulation of gene expression, or release of hormones or neurotransmitters. VGCCs have been immunolocalized in the zona glomerulosa of normal and hyperplastic human adrenal, as well as in aldosterone-producing adenomas (APA), and in the latter T-type VGCCs correlated with plasma aldosterone levels of patients. Excessive activation of VGCCs is a major component of excitotoxicity, as severely elevated levels of intracellular calcium activates enzymes which, at high enough levels, can degrade essential cellular structures. Structure Voltage-gated calcium channels are formed as a complex of several different subunits: α1, α2δ, β1-4, and γ. The α1 subunit forms the ion-conducting pore while the associated subunits have several functions including modulation of gating. Channel subunits There are several different kinds of high-voltage-gated calcium channels (HVGCCs). They are structurally homologous among varying types; they are all similar, but not structurally identical. In the la
https://en.wikipedia.org/wiki/C2orf73
Uncharacterized protein C2orf73 is a protein that in humans is encoded by the C2orf73 gene. The protein is predicted to be localized to the nucleus. Gene The full gene spans a total of 53,712 base pairs and contains nine exons. The gene's location in the Human genome is on chromosome 2 at position 2p16.2 and is flanked by the genes ACYP2 and SPTBN1. There are no aliases for this gene. mRNA The primary mRNA produced by the C2or73 gene is 1921 nucleotides long. There are six other mRNA isoforms produced by alternative splicing and variation in exon length. Protein The protein has a molecular mass of 32,142 Daltons. There are four protein isoforms. The primary isoform (X1) is 287 amino acids long. C2orf73 contains a short sequence motif, GDWWSH (This motif does not yet have any known function). The protein is lysine rich and leucine poor compared to the content of the average Human gene and has a predicted isoelectric point of 9.305. Structure A 3D structure for C2orf73 has not yet been determined experimentally. A computational prediction made by I-TASSER is presented to the right. The PELE tool on Biology Workbench predicts three likely α-helices and one β-strand in the protein. Post translational modifications The GPS, NetPhos, MyHits and SUMOsp tools on ExPASy predict potential post-translational modifications for the protein. Six potential phosphorylation sites and one sumoylation site are predicted. Subcellular localization PSORT II predicts C2orf73 to be localized to the nucleus. This is supported by the predicted presence of a sumoylation site, which is involved in nuclear cytoplasmic transport. Expression GEO profiles from NCBI show that C2orf73 is weakly expressed in the following tissues in Humans: bone marrow, liver, heart, lung, brain, spinal cord, skeletal muscle, thymus, and epithelium. Regulation of expression The Genomatix El Dorado tool predicts many transcription factors to have a high binding affinity in the 1100 base pairs upstream of C2o
https://en.wikipedia.org/wiki/Neuromuscular%20medicine
Neuromuscular medicine is a subspecialty of neurology and physiatry that focuses the diagnosis and management of neuromuscular diseases. The field encompasses issues related to both diagnosis and management of these conditions, including rehabilitation interventions to optimize the quality of life of individuals with these conditions. This field encompasses disorders that impact both adults and children and which can be inherited or acquired, typically from an autoimmune disease. A neurologist or physiatrist can diagnose these diseases through a clinical history, examination, and electromyography including nerve conduction studies.  Many recent drug therapies have been developed to address the acquired neuromuscular diseases including but not limited to immune suppression and drugs that increase the neurotransmitters at the neuromuscular junction.  Gene modifying therapies are also a recent treatment branch of neuromuscular medicine with advancements made in disorders such as spinal muscular atrophy and Duchenne muscular dystrophy. See also List of neuromuscular disorders Muscle Motor neuron diseases