source
stringlengths
31
227
text
stringlengths
9
2k
https://en.wikipedia.org/wiki/Adenium%20obesum
Adenium obesum, more commonly known as a desert rose, is a poisonous species of flowering plant belonging to the tribe Nerieae of the subfamily Apocynoideae of the dogbane family, Apocynaceae. It is native to the Sahel regions south of the Sahara (from Mauritania and Senegal to Sudan), tropical and subtropical eastern and southern Africa and also the Arabian Peninsula. Other names for the flower include Sabi star, kudu, mock azalea, and impala lily. Adenium obesum is a popular houseplant and bonsai in temperate regions. Description It is an evergreen or drought-deciduous succulent shrub (which can also lose its leaves during cold spells, or according to the subspecies or cultivar). It can grow to in height, with pachycaul (disproportionately large) stems and a stout, swollen basal caudex (a rootstock that protrudes from the soil). The leaves are spirally arranged, clustered toward the tips of the shoots, simple entire, leathery in texture, long and broad. The flowers are tubular, long, with the outer portion diameter with five petals, resembling those of other related genera such as Plumeria and Nerium. The flowers tend to be red and pink, often with a whitish blush outward of the throat. Taxonomy Some taxonomies consider some other species in the genus to be subspecies of Adenium obesum. Subspecies Adenium obesum subsp. oleifolium (South Africa, Botswana) Adenium obesum subsp. socotranum (Socotra) Adenium obesum subsp. somalense (Eastern Africa) Adenium obesum subsp. swazicum (Eswatini, South Africa) Adenium swazicum is a critically endangered African species native to Eswatini and Mozambique, growing up to 0.7 m (2.29 ft) tall. Adenium somalense is also native to Africa, inhabiting Tanzania, Kenya, and Somalia, and reaching heights of 5 m (16.40 ft), which makes it the largest of these four subspecies. Adenium socotranum is native exclusively to the island of Socotra, and can grow to be 4.6 m (15 ft), but despite its small range, it is of least co
https://en.wikipedia.org/wiki/Archon%20X%20Prize
The Archon Genomics X PRIZE presented by Express Scripts for Genomics, the second X Prize offered by the X Prize Foundation, based in Playa Vista, California, was announced on October 4, 2006 stating that the prize of "$10 million will be awarded to the first team to rapidly, accurately and economically sequence 100 whole human genomes to an unprecedented level of accuracy." The 30 day evaluation phase of the competition to begin on September 5, 2013, was canceled August 22, 2013 and this cancellation was debated on March 27, 2014. Goals and history In November 2011 the prize goals were stated as: "The $10 million grand prize will be awarded to the team(s) able to sequence 100 human genomes within 30 days to an accuracy of 1 error per 1,000,000 bases, with 98% completeness, identification of insertions, deletions and rearrangements, and a complete haplotype, at an audited total cost of $1,000 per genome." The $10 million was donated by Canadian geologist and philanthropist Stewart Blusson, who co-discovered the Ekati Diamond Mine. The name "Archon" is the name of Blusson's company, which refers to the type of lithosphere beneath northern Canada. Upon cancellation, the money was returned to the Blussons because no Master Team Agreements were in place. The Archon X Prize in genomics began as a joint effort of the X Prize Foundation and the J. Craig Venter Science Foundation. The J. Craig Venter Science Foundation offered the $500,000 (US) Innovation in Genomics Science and Technology Prize in September 2003 aimed at stimulating development of less expensive and faster sequencing technology. To attract even more resources to this goal, Dr. Venter joined forces with the X Prize Foundation, wrapping his competition and prize purse into a later incarnation, The Archon Genomics X Prize presented by Express Scripts. Test genomes The 100 human genomes to be sequenced in this competition were donated by 100 centenarians (ages 100 or older) from all over the world,
https://en.wikipedia.org/wiki/Ciliary%20neurotrophic%20factor
Ciliary neurotrophic factor is a protein that in humans is encoded by the CNTF gene. The protein encoded by this gene is a polypeptide hormone and neurotrophic factor whose actions have mainly been studied in the nervous system where it promotes neurotransmitter synthesis and neurite outgrowth in certain neural populations including astrocytes. It is a hypothalamic neuropeptide that is a potent survival factor for neurons and oligodendrocytes and may be relevant in reducing tissue destruction during inflammatory attacks. A mutation in this gene, which results in aberrant splicing, leads to ciliary neurotrophic factor deficiency, but this phenotype is not causally related to neurologic disease. In addition to the predominant monocistronic transcript originating from this locus, the gene is also cotranscribed with the upstream ZFP91 gene. Cotranscription from the two loci results in a transcript that contains a complete coding region for the zinc finger protein but lacks a complete coding region for ciliary neurotrophic factor. CNTF has also been shown to be expressed by cells on the bone surface, and to reduce the activity of bone-forming cells (osteoblasts). Therapeutic applications Satiety effects In 2001, it was reported that in a human study examining the usefulness of CNTF for treatment of motor neuron disease, CNTF produced an unexpected and substantial weight loss in the study subjects. Further investigation revealed that CNTF could reduce food intake without causing hunger or stress, making it a candidate for weight control in leptin-resistant subjects, as CNTF is believed to operate like leptin, but by a non-leptin pathway. Recombinant human CNTF (Axokine) A recombinant version of human CNTF (rhCNTF), trade name Axokine, is a modified version with a 15 amino acid truncation of the C-terminus and two amino acid substitutions. It is three to five times more potent than CNTF in in vitro and in vivo assays and has improved stability properties. Like CN
https://en.wikipedia.org/wiki/Argonaute
The Argonaute protein family, first discovered for its evolutionarily conserved stem cell function, plays a central role in RNA silencing processes as essential components of the RNA-induced silencing complex (RISC). RISC is responsible for the gene silencing phenomenon known as RNA interference (RNAi). Argonaute proteins bind different classes of small non-coding RNAs, including microRNAs (miRNAs), small interfering RNAs (siRNAs) and Piwi-interacting RNAs (piRNAs). Small RNAs guide Argonaute proteins to their specific targets through sequence complementarity (base pairing), which then leads to mRNA cleavage, translation inhibition, and/or the initiation of mRNA decay. The name of this protein family is derived from a mutant phenotype resulting from mutation of AGO1 in Arabidopsis thaliana, which was likened by Bohmert et al. to the appearance of the pelagic octopus Argonauta argo. {{Infobox protein family | Symbol = Piwi | Name = Argonaute Piwi domain | image = 1u04-argonaute.png | width = | caption = An argonaute protein from Pyrococcus furiosus. PDB . PIWI domain is on the right, PAZ domain to the left. | Pfam = PF02171 | Pfam_clan = | InterPro = IPR003165 | SMART = | PROSITE = PS50822 | MEROPS = | SCOP = | TCDB = | OPM family = | OPM protein = | CDD = cd02826 | PDB = }} RNA interference RNA interference (RNAi) is a biological process in which the RNA molecules inhibit gene expression. The method of inhibition is via the destruction of specific mRNA molecules or by simply suppressing the protein translation. The RNA interference has a significant role in defending cells against parasitic nucleotide sequences. In many eukaryotes, including animals, the RNA interference pathway is found, and it is initiated by the enzyme Dicer. Dicer cleaves long double-stranded RNA (dsRNA, often found in viruses and small interfering RNA) molecules into short double stranded fragments of around 20 nucleotide siRNAs. The dsRNA is then separated into two single-stranded RNAs (ss
https://en.wikipedia.org/wiki/Hepatocyte%20growth%20factor
Hepatocyte growth factor (HGF) or scatter factor (SF) is a paracrine cellular growth, motility and morphogenic factor. It is secreted by mesenchymal cells and targets and acts primarily upon epithelial cells and endothelial cells, but also acts on haemopoietic progenitor cells and T cells. It has been shown to have a major role in embryonic organ development, specifically in myogenesis, in adult organ regeneration, and in wound healing. Function Hepatocyte growth factor regulates cell growth, cell motility, and morphogenesis by activating a tyrosine kinase signaling cascade after binding to the proto-oncogenic c-Met receptor. Hepatocyte growth factor is secreted by platelets, and mesenchymal cells and acts as a multi-functional cytokine on cells of mainly epithelial origin. Its ability to stimulate mitogenesis, cell motility, and matrix invasion gives it a central role in angiogenesis, tumorogenesis, and tissue regeneration. Structure It is secreted as a single inactive polypeptide and is cleaved by serine proteases into a 69-kDa alpha-chain and 34-kDa beta-chain. A disulfide bond between the alpha and beta chains produces the active, heterodimeric molecule. The protein belongs to the plasminogen subfamily of S1 peptidases but has no detectable protease activity. Clinical significance Human HGF plasmid DNA therapy of cardiomyocytes is being examined as a potential treatment for coronary artery disease as well as treatment for the damage that occurs to the heart after myocardial infarction. As well as the well-characterised effects of HGF on epithelial cells, endothelial cells and haemopoietic progenitor cells, HGF also regulates the chemotaxis of T cells into heart tissue. Binding of HGF by c-Met, expressed on T cells, causes the upregulation of c-Met, CXCR3, and CCR4 which in turn imbues them with the ability to migrate into heart tissue. HGF also promotes angiogenesis in ischemia injury. HGF may further play a role as an indicator for prognosis of chronicit
https://en.wikipedia.org/wiki/Apoptosis-inducing%20factor
Apoptosis inducing factor is involved in initiating a caspase-independent pathway of apoptosis (positive intrinsic regulator of apoptosis) by causing DNA fragmentation and chromatin condensation. Apoptosis inducing factor is a flavoprotein. It also acts as an NADH oxidase. Another AIF function is to regulate the permeability of the mitochondrial membrane upon apoptosis. Normally it is found behind the outer membrane of the mitochondrion and is therefore secluded from the nucleus. However, when the mitochondrion is damaged, it moves to the cytosol and to the nucleus. Inactivation of AIF leads to resistance of embryonic stem cells to death following the withdrawal of growth factors indicating that it is involved in apoptosis. Function Apoptosis Inducing Factor (AIF) is a protein that triggers chromatin condensation and DNA fragmentation in a cell in order to induce programmed cell death. The mitochondrial AIF protein was found to be a caspase-independent death effector that can allow independent nuclei to undergo apoptotic changes. The process triggering apoptosis starts when the mitochondrion releases AIF, which exits through the mitochondrial membrane, enters the cytosol, and moves to the nucleus of the cell, where it signals the cell to condense its chromosomes and fragment its DNA molecules in order to prepare for cell death. Recently, researchers have discovered that the activity of AIF depends on the type of cell, the apoptotic insult, and its DNA-binding ability. AIF also plays a significant role in the mitochondrial respiratory chain and metabolic redox reactions. Synthesis The AIF protein is located across 16 exons on the X chromosome in humans. AIF1 (the most abundant type of AIF) is translated in the cytosol and recruited to the mitochondrial membrane and intermembrane space by its N-terminal mitochondrial localization signal (MLS). Inside the mitochondrion, AIF folds into its functional configuration with the help of the cofactor flavin adenine dinuc
https://en.wikipedia.org/wiki/Chimerin%201
Chimerin 1 (CHN1), also known as alpha-1-chimerin, n-chimerin, is a protein which in humans is encoded by the CHN1 gene. Chimerin 1 is a GTPase activating protein specific for RAC GTP-binding proteins. It is expressed primarily in the brain and may be involved in signal transduction. This gene encodes GTPase-activating protein for p21-rac and a phorbol ester receptor. It plays an important role in ocular motor axon pathfinding. Function CHN1 is a three-domain protein with the N-terminal SH2 domain, the C-terminal RhoGAP domain and the central C1 domain similar to protein kinase C. When lipid diacylglycerol (DAG) binds to the C1 domain, CHN1 is transferred to the plasma membrane and negatively regulates Rho-family small GTPases RAC1 and CDC42, thus causing the morphological change of axons by pruning the ends of axon dendrites. Mutational analysis suggests that un-overlapping residues of the RhoGAP domain are involved in RAC1-binding and the RAC1-GAP activity. Regulation of the RhoGAP activity of CHN1 by phorbol esters, natural compounds mimic of the lipid second messenger DAG, presents a possible way of designing agents for therapeutics. Clinical significance Heterozygous missense mutations in this gene cause Duane's retraction syndrome 2 (DURS2).
https://en.wikipedia.org/wiki/Son%20of%20Sevenless
In cell signalling, Son of Sevenless (SOS) refers to a set of genes encoding guanine nucleotide exchange factors that act on the Ras subfamily of small GTPases. History and name The gene was so named because the Sos protein that it encoded was found to operate downstream of the sevenless gene in Drosophila melanogaster in a Ras/MAP kinase pathway. When sevenless is mutated or otherwise dysfunctional during development of the fly's ultraviolet light-sensitive compound eye, the seventh, central photoreceptor (R7) of each ommatidium fails to form. Similarly, the mammalian orthologues of Sos, SOS1 and SOS2, function downstream of many growth factor and adhesion receptors. Function Ras-GTPases act as molecular switches that bind to downstream effectors, such as the protein kinase c-Raf, and localise them to the membrane, resulting in their activation. Ras-GTPases are considered inactive when bound to guanosine diphosphate (GDP), and active when bound to guanosine triphosphate (GTP). As the name implies, Ras-GTPases possess intrinsic enzymatic activity that hydrolyses GTP to GDP and phosphate. Thus, upon binding to GTP, the duration of Ras-GTPase activity depends on the rate of hydrolysis. SOS (and other guanine nucleotide exchange factors) act by binding Ras-GTPases and forcing them to release their bound nucleotide (usually GDP). Once released from SOS, the Ras-GTPase quickly binds fresh guanine nucleotide from the cytosol. Since GTP is roughly ten times more abundant than GDP in the cytosol, this usually results in Ras activation. The normal rate of Ras catalytic GTPase (GTP hydrolysis) activity can be increased by proteins of the RasGAP family, which bind to Ras and increase its catalytic rate by a factor of one thousand - in effect, increasing the rate at which Ras is inactivated. Genetic diseases associated with SOS1 Dominant mutant alleles of SOS1 have recently been found to cause Noonan syndrome and hereditary gingival fibromatosis type 1. Noonan syndrom
https://en.wikipedia.org/wiki/Neurocalcin
Neurocalcin is a neuronal calcium-binding protein that belongs to the neuronal calcium sensor (NCS) family of proteins. It expressed in mammalian brains. It possesses a Ca2+/myristoyl switch The subclass of neurocalcins are brain-specific proteins that fit into the EF-hand superfamily of calcium binding proteins. The NCS family were defined by the photoreceptor cell-specific protein, recoverin. Neurocalcin was purified from the bovine brain by using calcium-dependent drug affinity chromatography. The amino acid sequence showed that neurocalcin has three functional calcium binding sites. It is expressed in the central nervous system, retina, and adrenal gland. With this unique pattern of expression it is thought that neurcalcin offers a different physiological role than similar proteins visinin and recoverin. Neurocalcin delta an isoform of Neurocalcin is known to regulate adult neurogenesis ( Upadhyay et al., 2019) External links NCS proteins Protein Database Page
https://en.wikipedia.org/wiki/Numerical%20linear%20algebra
Numerical linear algebra, sometimes called applied linear algebra, is the study of how matrix operations can be used to create computer algorithms which efficiently and accurately provide approximate answers to questions in continuous mathematics. It is a subfield of numerical analysis, and a type of linear algebra. Computers use floating-point arithmetic and cannot exactly represent irrational data, so when a computer algorithm is applied to a matrix of data, it can sometimes increase the difference between a number stored in the computer and the true number that it is an approximation of. Numerical linear algebra uses properties of vectors and matrices to develop computer algorithms that minimize the error introduced by the computer, and is also concerned with ensuring that the algorithm is as efficient as possible. Numerical linear algebra aims to solve problems of continuous mathematics using finite precision computers, so its applications to the natural and social sciences are as vast as the applications of continuous mathematics. It is often a fundamental part of engineering and computational science problems, such as image and signal processing, telecommunication, computational finance, materials science simulations, structural biology, data mining, bioinformatics, and fluid dynamics. Matrix methods are particularly used in finite difference methods, finite element methods, and the modeling of differential equations. Noting the broad applications of numerical linear algebra, Lloyd N. Trefethen and David Bau, III argue that it is "as fundamental to the mathematical sciences as calculus and differential equations", even though it is a comparatively small field. Because many properties of matrices and vectors also apply to functions and operators, numerical linear algebra can also be viewed as a type of functional analysis which has a particular emphasis on practical algorithms. Common problems in numerical linear algebra include obtaining matrix decomposition
https://en.wikipedia.org/wiki/Myogenin
Myogenin, is a transcriptional activator encoded by the MYOG gene. Myogenin is a muscle-specific basic-helix-loop-helix (bHLH) transcription factor involved in the coordination of skeletal muscle development or myogenesis and repair. Myogenin is a member of the MyoD family of transcription factors, which also includes MyoD, Myf5, and MRF4. In mice, myogenin is essential for the development of functional skeletal muscle. Myogenin is required for the proper differentiation of most myogenic precursor cells during the process of myogenesis. When the DNA coding for myogenin was knocked out of the mouse genome, severe skeletal muscle defects were observed. Mice lacking both copies of myogenin (homozygous-null) suffer from perinatal lethality due to the lack of mature secondary skeletal muscle fibers throughout the body. In cell culture, myogenin can induce myogenesis in a variety of non-muscle cell types. Interactions Myogenin has been shown to interact with: MDFI, POLR2C, Serum response factor Sp1 transcription factor, and TCF3.
https://en.wikipedia.org/wiki/Pattern%20recognition%20%28psychology%29
In psychology and cognitive neuroscience, pattern recognition describes a cognitive process that matches information from a stimulus with information retrieved from memory. Pattern recognition occurs when information from the environment is received and entered into short-term memory, causing automatic activation of a specific content of long-term memory. An early example of this is learning the alphabet in order. When a carer repeats ‘A, B, C’ multiple times to a child, utilizing the pattern recognition, the child says ‘C’ after they hear ‘A, B’ in order. Recognizing patterns allows us to predict and expect what is coming. The process of pattern recognition involves matching the information received with the information already stored in the brain. Making the connection between memories and information perceived is a step of pattern recognition called identification. Pattern recognition requires repetition of experience. Semantic memory, which is used implicitly and subconsciously, is the main type of memory involved with recognition. Pattern recognition is not only crucial to humans, but to other animals as well. Even koalas, who possess less-developed thinking abilities, use pattern recognition to find and consume eucalyptus leaves. The human brain has developed more, but holds similarities to the brains of birds and lower mammals. The development of neural networks in the outer layer of the brain in humans has allowed for better processing of visual and auditory patterns. Spatial positioning in the environment, remembering findings, and detecting hazards and resources to increase chances of survival are examples of the application of pattern recognition for humans and animals. There are six main theories of pattern recognition: template matching, prototype-matching, feature analysis, recognition-by-components theory, bottom-up and top-down processing, and Fourier analysis. The application of these theories in everyday life is not mutually exclusive. Pattern
https://en.wikipedia.org/wiki/Microphthalmia-associated%20transcription%20factor
Microphthalmia-associated transcription factor also known as class E basic helix-loop-helix protein 32 or bHLHe32 is a protein that in humans is encoded by the MITF gene. MITF is a basic helix-loop-helix leucine zipper transcription factor involved in lineage-specific pathway regulation of many types of cells including melanocytes, osteoclasts, and mast cells. The term "lineage-specific", since it relates to MITF, means genes or traits that are only found in a certain cell type. Therefore, MITF may be involved in the rewiring of signaling cascades that are specifically required for the survival and physiological function of their normal cell precursors. MITF, together with transcription factor EB (TFEB), TFE3 and TFEC, belong to a subfamily of related bHLHZip proteins, termed the MiT-TFE family of transcription factors. The factors are able to form stable DNA-binding homo- and heterodimers. The gene that encodes for MITF resides at the mi locus in mice, and its protumorogenic targets include factors involved in cell death, DNA replication, repair, mitosis, microRNA production, membrane trafficking, mitochondrial metabolism, and much more. Mutation of this gene results in deafness, bone loss, small eyes, and poorly pigmented eyes and skin. In human subjects, because it is known that MITF controls the expression of various genes that are essential for normal melanin synthesis in melanocytes, mutations of MITF can lead to diseases such as melanoma, Waardenburg syndrome, and Tietz syndrome. Its function is conserved across vertebrates, including in fishes such as zebrafish and Xiphophorus. An understanding of MITF is necessary to understand how certain lineage-specific cancers and other diseases progress. In addition, current and future research can lead to potential avenues to target this transcription factor mechanism for cancer prevention. Clinical significance Mutations As mentioned above, changes in MITF can result in serious health conditions. For example,
https://en.wikipedia.org/wiki/Michael%20J.%20Flynn
Michael J. Flynn (born May 20, 1934) is an American professor emeritus at Stanford University. Early life and education Flynn was born in New York City. Career Flynn proposed Flynn's taxonomy, a method of classifying parallel digital computers, in 1966. In the early 1970s, he was the founding chairman of IEEE Computer Society's Technical Committee on Computer Architecture (TCCA) and Association for Computing Machinery's Special Interest Group on Computer Architecture, ACM SIGARCH (initially SICARCH). Flynn encouraged, from the beginning, joint cooperation between the two groups which now sponsors many leading joint symposiums and conferences like ACM/IEEE International Symposium on Computer Architecture (ISCA). In 1995 he received a Harry H. Goode Memorial Award. In 2009, Flynn received an honorary doctorate from the University of Belgrade. Flynn co-founded Palyn Associates with Maxwell Paley and in 2014 is Chairman of Maxeler Technologies. Bibliography
https://en.wikipedia.org/wiki/Joel%20Lebowitz
Joel Louis Lebowitz (born May 10, 1930) is a mathematical physicist widely acknowledged for his outstanding contributions to statistical physics, statistical mechanics and many other fields of Mathematics and Physics. Lebowitz has published more than five hundred papers concerning statistical physics and science in general, and he is one of the founders and editors of the Journal of Statistical Physics, one of the most important peer-reviewed journals concerning scientific research in this area. He has been president of the New York Academy of Sciences. Lebowitz is the George William Hill Professor of Mathematics and Physics at Rutgers University. He is also an active member of the human rights community and a long-term co-chair of the Committee of Concerned Scientists. Biography Lebowitz was born in Taceva, then in Czechoslovakia, now Ukraine, in 1930 into a Jewish family. During World War II he was deported with his family to Auschwitz, where his father, his mother, and his younger sister were killed in 1944. After being liberated from the camp, he moved to United States by boat, and he studied in an Orthodox Jewish school and Brooklyn College. He earned his PhD at Syracuse University in 1956 under the supervision of Peter G. Bergmann. Then he continued his research with Lars Onsager, at Yale University, where he got a faculty position. He moved to the Stevens Institute of Technology in 1957 and to the Belfer Graduate School of Science of Yeshiva University in 1959. Finally he got a faculty position at Rutgers University in 1977, where he holds the prestigious George William Hill Professor position. During his years at the Yeshiva University and Rutgers University he has been in contact with several scientists, and artists, like Fumio Yoshimura and Kate Millett. In 1975 he founded the Journal of Statistical Physics. In 1979 he was president of the New York Academy of Sciences. He has been one of the most active supporters of dissident scientists in the forme
https://en.wikipedia.org/wiki/Superior%20cardiac%20nerve
The superior cardiac nerve arises by two or more branches from the superior cervical ganglion, and occasionally receives a filament from the trunk between the first and second cervical ganglia. It runs down the neck behind the common carotid artery, and in front of the Longus colli muscle; and crosses in front of the inferior thyroid artery, and recurrent nerve. The course of the nerves on the two sides then differs. Right nerve The right nerve, at the root of the neck, passes either in front of or behind the subclavian artery, and along the brachiocephalic trunk to the back of the arch of the aorta, where it joins the deep part of the cardiac plexus. It is connected with other branches of the sympathetic; about the middle of the neck it receives filaments from the external laryngeal nerve; lower down, one or two twigs from the vagus; and as it enters the thorax it is joined by a filament from the recurrent laryngeal nerve. Filaments from the nerve communicate with the thyroid branches from the middle cervical ganglion. Left nerve The left nerve, in the thorax, runs in front of the left common carotid artery and across the left side of the aortic arch, to the superficial part of the cardiac plexus.
https://en.wikipedia.org/wiki/Industrial%20control%20system
An industrial control system (ICS) is an electronic control system and associated instrumentation used for industrial process control. Control systems can range in size from a few modular panel-mounted controllers to large interconnected and interactive distributed control systems (DCSs) with many thousands of field connections. Control systems receive data from remote sensors measuring process variables (PVs), compare the collected data with desired setpoints (SPs), and derive command functions that are used to control a process through the final control elements (FCEs), such as control valves. Larger systems are usually implemented by supervisory control and data acquisition (SCADA) systems, or DCSs, and programmable logic controllers (PLCs), though SCADA and PLC systems are scalable down to small systems with few control loops. Such systems are extensively used in industries such as chemical processing, pulp and paper manufacture, power generation, oil and gas processing, and telecommunications. Discrete controllers The simplest control systems are based around small discrete controllers with a single control loop each. These are usually panel mounted which allows direct viewing of the front panel and provides means of manual intervention by the operator, either to manually control the process or to change control setpoints. Originally these would be pneumatic controllers, a few of which are still in use, but nearly all are now electronic. Quite complex systems can be created with networks of these controllers communicating using industry-standard protocols. Networking allow the use of local or remote SCADA operator interfaces, and enables the cascading and interlocking of controllers. However, as the number of control loops increase for a system design there is a point where the use of a programmable logic controller (PLC) or distributed control system (DCS) is more manageable or cost-effective. Distributed control systems A distributed control system (DCS
https://en.wikipedia.org/wiki/Fair%20Trade%20Certified%20Mark
The Fair Trade Certified Mark is a fair trade certification mark used primarily in the United States and Canada. It appears on products as an independent guarantee that disadvantaged producers in the developing world are getting a better deal. The Fair Trade Certified Mark is the North American equivalent of the International Fairtrade Certification Mark used in Europe, Africa, Asia, Australia and New Zealand. For a product to carry either Certification Marks, it must come from Fair Trade USA inspected and certified producer organizations. The crops must be grown and harvested in accordance with the fair trade standards set by Fair Trade USA. Some of the supply chains are also monitored by FLO-CERT to ensure the integrity of labelled products. Only Fair Trade USA (formerly "TransFair USA") licensees can use the Fair Trade Certified Mark on their products. The Fair Trade Certified Mark in the United States was introduced by TransFair USA on the American market in 1998. In 2012 a variation of the US Fair Trade certification mark was adopted with the benefit of being registered globally as a trade mark. The mark is designed to pop better on the shelf through a far simpler design and the use of color. The one basket with outstretched hands indicates sharing and a "give and take" between producers and purchases. The green signals the environmental strength of Fair Trade.
https://en.wikipedia.org/wiki/Armorial%20of%20Estonia
The coats of arms of the 15 counties of Estonia are presented below. Military Estonian Land Forces Estonian Navy Estonian Air Force Counties Harju County Hiiu County Ida-Viru County Jõgeva County Järva County Lääne County Lääne-Viru County Põlva County Pärnu County Rapla County Saare County Tartu County Valga County Viljandi County Võru County Colleges See also Flags of Estonian counties Estonia Estonia geography-related lists
https://en.wikipedia.org/wiki/List%20of%20compilers
This page is intended to list all current compilers, compiler generators, interpreters, translators, tool foundations, assemblers, automatable command line interfaces (shells), etc. Ada compilers ALGOL 60 compilers ALGOL 68 compilers cf. ALGOL 68s specification and implementation timeline Assemblers (Intel *86) Assemblers (Motorola 68*) Assemblers (Zilog Z80) Assemblers (other) BASIC compilers BASIC interpreters C compilers Notes: C++ compilers Notes: C# compilers COBOL compilers Common Lisp compilers D compilers DIBOL/DBL compilers ECMAScript interpreters Eiffel compilers Forth compilers and interpreters Fortran compilers Go compilers Haskell compilers ISLISP compilers and interpreters Java compilers Lisaac compiler Pascal compilers Perl interpreters PHP compilers PL/I compilers Python compilers and interpreters Ruby compilers and interpreters Rust compilers Smalltalk compilers Tcl interpreters DCL interpreters Rexx interpreters CLI compilers Source-to-source compilers This list is incomplete. A more extensive list of source-to-source compilers can be found here. Open source compilers Production quality, open source compilers. Amsterdam Compiler Kit (ACK) [C, Pascal, Modula-2, Occam, and BASIC] [Unix-like] Clang C/C++/Objective-C Compiler AMD Optimizing C/C++ Compiler FreeBASIC [Basic] [DOS/Linux/Windows] Free Pascal [Pascal] [DOS/Linux/Windows(32/64/CE)/MacOS/NDS/GBA/..(and many more)] GCC: C, C++ (G++), Java (GCJ), Ada (GNAT), Objective-C, Objective-C++, Fortran (GFortran), and Go (GCCGo); also available, but not in standard are: Modula-2, Modula-3, Pascal, PL/I, D, Mercury, VHDL; Linux, the BSDs, macOS, NeXTSTEP, Windows and BeOS, among others Local C compiler [C] [Linux, SPARC, MIPS] The LLVM Compiler Infrastructure which is also frequently used for research Portable C Compiler [C] [Unix-like] Open Watcom [C, C++, and Fortran] [Windows and OS/2, Linux/FreeBSD WIP] TenDRA [C/C++] [Unix-like] Tiny C Compiler
https://en.wikipedia.org/wiki/Amplitude-comparison%20monopulse
Amplitude-comparison monopulse refers to a common direction finding technique. This method is used in monopulse radar, electronic warfare and radio astronomy. Amplitude monopulse antennas are usually reflector antennas. Approach Two overlapping antenna beams are formed, which are steered in slightly different directions, usually such that they overlap at the half-power point (-3 dB-point) of the beams. By comparing the relative amplitude of the pulse in the two beams, its position in the beams can be determined with an accuracy dependent on the signal-to-noise ratio (SNR). An accuracy of a tenth of beamwidth can be achieved with an SNR of 10 dB. In most implementations, two signals are formed, one being the sum of the two beams, and the other being the difference of the two beams. The ratio of these two beams normalises the difference signal and allows the direction of arrival of the signal to be calculated. The shape of the antenna beams must be known exactly and hence the accuracy of the techniques can be affected by unwanted multipath reflections. See also Monopulse radar Phase-Comparison Monopulse
https://en.wikipedia.org/wiki/Michael%20Butler%20%28computer%20scientist%29
Michael J. Butler is an Irish computer scientist. As of 2022, he is professor of computer science and Dean of the Faculty of Engineering and Physical Sciences at the University of Southampton, England. Biography Butler was born in Ireland. He received his bachelor's degree in computer science from Trinity College, Dublin in 1988. He then took an MSc (1989) and DPhil (1992) at the Programming Research Group of the University of Oxford, working in the area of communicating sequential processes. He then worked for Broadcom in Dublin and at Åbo Akademi University in Turku, Finland with Ralph-Johan Back on refinement calculus. He joined the University of Southampton in 1995 as a lecturer, rising to reader in 2000 and then professor in the same year. He led the Dependable Systems & Software Engineering group at the School of Electronics and Computer Science, University of Southampton (inactive as of 2022). His main research is in the area of the B-Method (originated by J.-R. Abrial), especially tool support such as ProB (advanced model checking for B which allows for the simulation of Event-B machines in the Rodin/Eclipse platform), U2B (UML and B), csp2B (CSP and B), and the RODIN toolset for Event-B.
https://en.wikipedia.org/wiki/K%C3%A4hler%E2%80%93Einstein%20metric
In differential geometry, a Kähler–Einstein metric on a complex manifold is a Riemannian metric that is both a Kähler metric and an Einstein metric. A manifold is said to be Kähler–Einstein if it admits a Kähler–Einstein metric. The most important special case of these are the Calabi–Yau manifolds, which are Kähler and Ricci-flat. The most important problem for this area is the existence of Kähler–Einstein metrics for compact Kähler manifolds. This problem can be split up into three cases dependent on the sign of the first Chern class of the Kähler manifold: When the first Chern class is negative, there is always a Kähler–Einstein metric, as Thierry Aubin and Shing-Tung Yau proved independently. When the first Chern class is zero, there is always a Kähler–Einstein metric, as Yau proved in the Calabi conjecture. That leads to the name Calabi–Yau manifolds. He was awarded with the Fields Medal partly because of this work. The third case, the positive or Fano case, remained a well-known open problem for many years. In this case, there is a non-trivial obstruction to existence. In 2012, Xiuxiong Chen, Simon Donaldson, and Song Sun proved that in this case existence is equivalent to an algebro-geometric criterion called K-stability. Their proof appeared in a series of articles in the Journal of the American Mathematical Society. A proof was produced independently by Gang Tian at the same time. When first Chern class is not definite, or we have intermediate Kodaira dimension, then finding canonical metric remained as an open problem, which is called the algebrization conjecture via analytical minimal model program. Definition Einstein manifolds Suppose is a Riemannian manifold. In physics the Einstein field equations are a set of partial differential equations on the metric tensor which describe how the manifold should curve due to the existence of mass or energy, a quantity encapsulated by the stress–energy tensor . In a vacuum where there is no mass or en
https://en.wikipedia.org/wiki/Spandrel%20%28biology%29
In evolutionary biology, a spandrel is a phenotypic trait that is a byproduct of the evolution of some other characteristic, rather than a direct product of adaptive selection. Stephen Jay Gould and Richard Lewontin brought the term into biology in their 1979 paper "The Spandrels of San Marco and the Panglossian Paradigm: A Critique of the Adaptationist Programme". Adaptationism is a point of view that sees most organismal traits as adaptive products of natural selection. Gould and Lewontin sought to temper what they saw as adaptationist bias by promoting a more structuralist view of evolution. The term "spandrel" originates from architecture, where it refers to the roughly triangular spaces between the top of an arch and the ceiling. Etymology The term was coined by paleontologist Stephen Jay Gould and population geneticist Richard Lewontin in their paper "The Spandrels of San Marco and the Panglossian Paradigm: A Critique of the Adaptationist Programme" (1979). Evolutionary biologist Günter P. Wagner called the paper "the most influential structuralist manifesto". In their paper, Gould and Lewontin employed the analogy of spandrels in Renaissance architecture, such as the curved areas of masonry between arches supporting a dome that arise as a consequence of decisions about the shape of the arches and the base of the dome, rather than being designed for the artistic purposes for which they were often employed. The authors singled out properties like the necessary number of four spandrels and their specific three-dimensional shape. At the time, it was widely thought in the scientific community that everything an animal has developed that has a positive effect on that animal's fitness was due to natural selection or some adaptation. Gould and Lewontin proposed an alternative hypothesis: that due to adaptation and natural selection, byproducts are also formed. Because these byproducts of adaptations that had no real relative advantage to survival, they were termed
https://en.wikipedia.org/wiki/Paxillin
Paxillin is a protein that in humans is encoded by the PXN gene. Paxillin is expressed at focal adhesions of non-striated cells and at costameres of striated muscle cells, and it functions to adhere cells to the extracellular matrix. Mutations in PXN as well as abnormal expression of paxillin protein has been implicated in the progression of various cancers. Structure Human paxillin is 64.5 kDa in molecular weight and 591 amino acids in length. The C-terminal region of paxillin is composed of four tandem double zinc finger LIM domains that are cysteine/histidine-rich with conserved repeats; these serve as binding sites for the protein tyrosine phosphatase-PEST, tubulin and serves as the targeting motif for focal adhesions. The N-terminal region of paxillin has five highly conserved leucine-rich sequences termed LD motifs, which mediate several interactions, including that with pp125FAK and vinculin. The LD motifs are predicted to form amphipathic alpha helices, with each leucine residue positioned on one face of the alpha helix to form a hydrophobic protein-binding interface. The N-terminal region also has a proline-rich domain that has potential for Src-SH3 binding. Three N-terminal YXXP motifs may serve as binding sites for talin or v-Crk SH2. Function Paxillin is a signal transduction adaptor protein discovered in 1990 in the laboratory of Keith Burridge The C-terminal region of paxillin contains four LIM domains that target paxillin to focal adhesions. It is presumed through a direct association with the cytoplasmic tail of beta-integrin. The N-terminal region of paxillin is rich in protein–protein interaction sites. The proteins that bind to paxillin are diverse and include protein tyrosine kinases, such as Src and focal adhesion kinase (FAK), structural proteins, such as vinculin and actopaxin, and regulators of actin organization, such as COOL/PIX and PKL/GIT. Paxillin is tyrosine-phosphorylated by FAK and Src upon integrin engagement or growth factor
https://en.wikipedia.org/wiki/Intracellular%20calcium-sensing%20proteins
Intracellular calcium-sensing proteins are proteins that act in the second messenger system. Examples include: calmodulin calnexin calreticulin gelsolin
https://en.wikipedia.org/wiki/Mitogen-activated%20protein%20kinase%209
Mitogen-activated protein kinase 9 is an enzyme that in humans is encoded by the MAPK9 gene. Function The protein encoded by this gene is a member of the MAP kinase family. MAP kinases act as an integration point for multiple biochemical signals, and are involved in a wide variety of cellular processes such as proliferation, differentiation, transcription regulation and development. This kinase targets specific transcription factors, and thus mediates immediate-early gene expression in response to various cell stimuli. It is most closely related to MAPK8, both of which are involved in UV radiation-induced apoptosis, thought to be related to the cytochrome c-mediated cell death pathway. This gene and MAPK8 are also known as c-Jun N-terminal kinases. This kinase blocks the ubiquitination of tumor suppressor p53, and thus it increases the stability of p53 in nonstressed cells. Studies of this gene's mouse counterpart suggest a key role in T-cell differentiation. Four alternatively spliced transcript variants encoding distinct isoforms have been reported. Interactions Mitogen-activated protein kinase 9 has been shown to interact with: Grb2, MAPK8IP1, MAPK8IP2, MAPK8IP3 P53, and TOB1.
https://en.wikipedia.org/wiki/List%20of%20Institute%20of%20Food%20Technologists%20presidents
This a list of the people who have served as president of the Institute of Food Technologists since its establishment in 1939. 1939–41: Samuel C. Prescott 1941–42: Laurence V. Burton 1942–43: Roy C. Newton 1943–44: William V. Cruess 1944–45: Fred C. Blanck 1945–46: Fred W. Tanner 1946–47: Ellery R. Harvey 1947–48: George J. Hucker 1948–49: Helmut C. Diehl 1949–50: Carl R. Fellers 1950–51: Paul F. Sharp 1951–52: Charles N. Frey 1952–53: Bernard E. Proctor 1953–54: Berton S. Clark 1954–55: Philip K. Bates 1955–56: LaVerne E. Clifcorn 1956–57: George F. Garantz 1957–58: Emil M. Mrak 1958–59: Askel G. Olsen 1959–60: Ray B. Wakefield 1960–61: Irmi J. Hutchings 1961–62: Harold W. Schultz 1962–63: John M. Jackson 1963–64: C. Olin Ball 1964–65: Charles T. Townsend 1965–66: Maynard A. Joslyn 1966–67: John H. Nair 1967–68: George F. Stewart 1968–69: Bernard L. Oser 1969–70: Herbert E. Robinson 1970–71: Hans Lineweaver 1971–72: Richard L. Hall 1972–73: Ben F. Buchanan 1973–74: Reid T. Milner 1974–75: Charles F. Niven, Jr. 1975–76: Ernest J. Briskey 1976–77: John C. Ayres 1977–78: Howard E. Bauman 1978–79: Bernard S. Schweigert 1979–80: Walter L. Clark 1980–81: Frederick Jack Francis 1981–82: Arthur T. Schramm 1982–83: Owen R. Fennema 1983–84: Gilbert A. Leveille 1984–85: Bernard J. Liska 1985–86: Charles J. Bates 1986–87: John J. Powers 1987–88: Roy E. Morse 1988–89: Theodore P. Labuza 1989–90: Paul F. Hopper 1990–91: Daryl B. Lund 1991–92: John H. Litchfield 1992–93: David R. Lineback 1993–94: Adolph (Al) S. Clausi 1994–95: Roy G. Arnold 1995–96: Francis F. Busta 1996–97: Robert E. Smith 1997–98: Mary K. Wagner 1998–99: Bruce R. Stillings 1999–2000: Charles E. Manley 2000–01: Mary K. Schmidl 2001–02: Philip E. Nelson 2002–03: Mark R. McLellan 2003–04: C. Ann Hollingsworth 2004–05: Herbert F. Stone 2005–06: Margaret A. Lawson 2006–07: Dennis R. Heldman 2007–08: John D. Floros 2008–09: Sheri Schellhaass 2009–10: Marianne Gillette 2010-11: Robert B. Gravani 2011–12: Roger A. Cl
https://en.wikipedia.org/wiki/Olfactory%20marker%20protein
In molecular biology, olfactory marker protein is a protein involved in signal transduction. It is a highly expressed, cytoplasmic protein found in mature olfactory sensory receptor neurons of all vertebrates. OMP is a modulator of the olfactory signal transduction cascade. The crystal structure of OMP reveals a beta sandwich consisting of eight strands in two sheets with a jelly-roll topology. Three highly conserved regions have been identified as possible protein–protein interaction sites in OMP, indicating a possible role for OMP in modulating such interactions, thereby acting as a molecular switch. External links
https://en.wikipedia.org/wiki/Delta-sleep-inducing%20peptide
Delta-sleep-inducing peptide (DSIP) is a neuropeptide that when infused into the mesodiencephalic ventricle of recipient rabbits induces spindle and delta EEG activity and reduced motor activities. Its amino acid sequence is Trp-Ala-Gly-Gly-Asp-Ala-Ser-Gly-Glu (WAGGDASGE). The gene has yet to be found in rabbits, along with any receptors or precursor peptides. However, searches through BLAST have found that it aligns with a hypothetical Amycolatopsis coloradensis protein. This could indicate that DSIP has a bacterial origin. Discovery Delta-sleep-inducing peptide was first discovered in 1974 by the Swiss Schoenenberger-Monnier group who isolated it from the cerebral venous blood of rabbits in an induced state of sleep. It was primarily believed to be involved in sleep regulation due to its apparent ability to induce slow-wave sleep in rabbits, but studies on the subject have been contradictory. DSIP-like material has been found in human breast milk. Structure and interactions DSIP is an amphiphilic peptide of molecular weight 850 daltons with the amino acid motif:N-Trp-Ala-Gly-Gly-Asp-Ala-Ser-Gly-Glu-C It has been found in both free and bound forms in the hypothalamus, limbic system and pituitary as well as various peripheral organs, tissues and body fluids. In the pituitary it co-localises with many peptide and non-peptide mediators such as corticotropin-like intermediate peptide (CLIP), adrenocorticotrophic hormone (ACTH), melanocyte-stimulating hormone (MSH), thyroid-stimulating hormone (TSH) and melanin concentrating hormone (MCH). It is abundant in the gut secretory cells and in the pancreas where it co-localises with glucagon. In the brain its action may be mediated by NMDA receptors. In another study delta-sleep-inducing peptide stimulated acetyltransferase activity through α1 receptors in rats. It is unknown where DSIP is synthesized. In vitro it has been found to have a low molecular stability with a half life of only 15 minutes due to the action
https://en.wikipedia.org/wiki/Molecular%20orbital%20diagram
A molecular orbital diagram, or MO diagram, is a qualitative descriptive tool explaining chemical bonding in molecules in terms of molecular orbital theory in general and the linear combination of atomic orbitals (LCAO) method in particular. A fundamental principle of these theories is that as atoms bond to form molecules, a certain number of atomic orbitals combine to form the same number of molecular orbitals, although the electrons involved may be redistributed among the orbitals. This tool is very well suited for simple diatomic molecules such as dihydrogen, dioxygen, and carbon monoxide but becomes more complex when discussing even comparatively simple polyatomic molecules, such as methane. MO diagrams can explain why some molecules exist and others do not. They can also predict bond strength, as well as the electronic transitions that can take place. History Qualitative MO theory was introduced in 1928 by Robert S. Mulliken and Friedrich Hund. A mathematical description was provided by contributions from Douglas Hartree in 1928 and Vladimir Fock in 1930. Basics Molecular orbital diagrams are diagrams of molecular orbital (MO) energy levels, shown as short horizontal lines in the center, flanked by constituent atomic orbital (AO) energy levels for comparison, with the energy levels increasing from the bottom to the top. Lines, often dashed diagonal lines, connect MO levels with their constituent AO levels. Degenerate energy levels are commonly shown side by side. Appropriate AO and MO levels are filled with electrons by the Pauli Exclusion Principle, symbolized by small vertical arrows whose directions indicate the electron spins. The AO or MO shapes themselves are often not shown on these diagrams. For a diatomic molecule, an MO diagram effectively shows the energetics of the bond between the two atoms, whose AO unbonded energies are shown on the sides. For simple polyatomic molecules with a "central atom" such as methane () or carbon dioxide (),
https://en.wikipedia.org/wiki/Diffusing%20capacity%20for%20carbon%20monoxide
DLCO or TLCO (diffusing capacity or transfer factor of the lung for carbon monoxide (CO),) is the extent to which oxygen passes from the air sacs of the lungs into the blood. Commonly, it refers to the test used to determine this parameter. It was introduced in 1909. Mechanism This test involves measuring the partial pressure difference between inspired and expired carbon monoxide. It relies on the strong affinity and large absorption capacity of red blood cells for carbon monoxide and thus demonstrates gas uptake by the capillaries that are less dependent on cardiac output. The measurement of DLCO is affected by atmospheric pressure and/or altitude and correction factors can be calculated using the method recommended by the American Thoracic Society. Expected DLCO is also affected by the amount of hemoglobin, carboxyhemoglobin, age and sex. The correction for hemoglobin is based on the method of Cotes as recommended by the American Thoracic Society. DLCO vs TLCO Generally DLCO is measured in "ml/min/kPa" and TLCO is measured in "mmol/min/kPa". Factors affecting DLCO Decrease DLCO is decreased in any condition which affects the effective alveolar surface area: Hindrance in the alveolar wall. e.g. fibrosis, alveolitis, vasculitis Decrease of total lung area, e.g. Restrictive lung disease or lung resection (partial or total). Chronic obstructive pulmonary disease (Emphysema) due to decreased surface area in the alveoli, as well as damage to the capillary bed Pulmonary embolism Cardiac insufficiency Pulmonary hypertension Bleomycin (upon administration of more than 200 units) Anemia-due to decrease in blood volume Amiodarone high cumulative dose; more than 400 milligrams per day After chemotherapy and radiotherapy However, many modern devices compensate for the hemoglobin value of the patient (taken by blood test), and excludes it as a factor in the DLCO interpretation. Increase Factors that can increase the DLCO include polycythaemia, asthma (can als
https://en.wikipedia.org/wiki/Process%20variable
In control theory, a process variable (PV; also process value or process parameter) is the current measured value of a particular part of a process which is being monitored or controlled. An example of this would be the temperature of a furnace. The current temperature is the process variable, while the desired temperature is known as the set-point (SP). Control system use Measurement of process variables is essential in control systems to controlling a process. The value of the process variable is continuously monitored so that control may be exerted. Four commonly measured variables that affect chemical and physical processes are: pressure, temperature, level and flow. but there are in fact a large number of measurement quantities which for international purposes use the International System of Units (SI) The SP-PV error is used to exert control on a process so that the value of PV equals the value of the SP. A classic use of this is in the PID controller.
https://en.wikipedia.org/wiki/Standardized%20coefficient
In statistics, standardized (regression) coefficients, also called beta coefficients or beta weights, are the estimates resulting from a regression analysis where the underlying data have been standardized so that the variances of dependent and independent variables are equal to 1. Therefore, standardized coefficients are unitless and refer to how many standard deviations a dependent variable will change, per standard deviation increase in the predictor variable. Usage Standardization of the coefficient is usually done to answer the question of which of the independent variables have a greater effect on the dependent variable in a multiple regression analysis where the variables are measured in different units of measurement (for example, income measured in dollars and family size measured in number of individuals). It may also be considered a general measure of effect size, quantifying the "magnitude" of the effect of one variable on another. For simple linear regression with orthogonal predictors, the standardized regression coefficient equals the correlation between the independent and dependent variables. Implementation A regression carried out on original (unstandardized) variables produces unstandardized coefficients. A regression carried out on standardized variables produces standardized coefficients. Values for standardized and unstandardized coefficients can also be re-scaled to one another subsequent to either type of analysis. Suppose that is the regression coefficient resulting from a linear regression (predicting by ). The standardized coefficient simply results as , where and are the (estimated) standard deviations of and , respectively. Sometimes, standardization is done only without respect to the standard deviation of the regressor (the independent variable ). Advantages and disadvantages Standardized coefficients' advocates note that the coefficients are independent of the involved variables' units of measurement (i.e., standardized coeff
https://en.wikipedia.org/wiki/Tortilla%20Wall
The Tortilla Wall is a term given to a 14-mile (22.5 kilometer) section of United States border fence between the Otay Mesa Border Crossing in San Diego, California and the Pacific Ocean. This "San Diego wall" was completed in the early 1990s. While there are other walls at various points along the border, the Tortilla Wall is the longest to date. No other wall sections have evolved distinct names, so The Tortilla Wall is often used to describe the entire set of walled defensive structures. The Tortilla Wall is marked with graffiti, crosses, photos, pictures and remembrances of migrants who died trying to illegally enter the United States. Effectiveness The effectiveness of the wall has been significant according to U.S. Congressional testimony by Representative Ed Royce: ...apprehensions along the region with a security fence dropped from 202,000 in 1992 to 9,000 in 1994. The building of the Tortilla Wall is generally considered by Mexicans to be an unfriendly gesture. It is a symbol of the controversial immigration issue. It is argued that the wall simply forces illegal border crossings to be moved to the more dangerous area of the Arizona desert. Expansion of the wall In 2006, the U.S. Congress passed the Secure Fence Act of 2006 which authorized spending $1.2 billion to build 700 miles (1,100 km) of additional fencing on the southern border facing Mexico. Anecdotal wall stories Tunnels under the wall are still a common way to illegally cross the border. Some tunnels are quite sophisticated. One such tunnel created by smugglers ran from Tijuana to San Diego, was a half mile long, and included a concrete floor as well as electricity. Other tunnels have included steel rails, while some tunnels are simply dirt passageways or connect to sewer or drain systems. As a stunt, a circus cannon was placed on the south side of the wall and an acrobat was blasted over the wall into the Border Field State Park in the U.S. He had his passport with him. See also
https://en.wikipedia.org/wiki/Thermally%20conductive%20pad
In computing and electronics, thermal pads (also called thermally conductive pad or thermal interface pad) are pre-formed rectangles of solid material (often paraffin wax or silicone based) commonly found on the underside of heatsinks to aid the conduction of heat away from the component being cooled (such as a CPU or another chip) and into the heatsink (usually made from aluminium or copper). Thermal pads and thermal compound are used to fill air gaps caused by imperfectly flat or smooth surfaces which should be in thermal contact; they would not be needed between perfectly flat and smooth surfaces. Thermal pads are relatively firm at room temperature, but become soft and are able to fill gaps at higher temperatures. It is an alternative to thermal paste to be used as thermal interface material. AMD and Intel have included thermal pads on the bottom of heatsinks shipped with some of their processors, as they are cleaner and generally easier to install. However, thermal pads conduct heat less effectively than a minimal amount of thermal paste. See also Computer cooling Hot-melt adhesive Phase-change material Thermal adhesive Thermal paste List of thermal conductivities
https://en.wikipedia.org/wiki/MPEG%20program%20stream
Program stream (PS or MPEG-PS) is a container format for multiplexing digital audio, video and more. The PS format is specified in MPEG-1 Part 1 (ISO/IEC 11172-1) and MPEG-2 Part 1, Systems (ISO/IEC standard 13818-1/ITU-T H.222.0). The MPEG-2 Program Stream is analogous and similar to ISO/IEC 11172 Systems layer and it is forward compatible. Program streams are used on DVD-Video discs and HD DVD video discs, but with some restrictions and extensions. The filename extensions are VOB and EVO respectively. Coding structure Program streams are created by combining one or more Packetized Elementary Streams (PES), which have a common time base, into a single stream. It is designed for reasonably reliable media such as disks, in contrast to MPEG transport stream which is for data transmission in which loss of data is likely. Program streams have variable size records and minimal use of start codes which would make over the air reception difficult, but has less overhead. Program stream coding layer allows only one program of one or more elementary streams to be packaged into a single stream, in contrast to transport stream, which allows multiple programs. MPEG-2 Program stream can contain MPEG-1 Part 2 video, MPEG-2 Part 2 video, MPEG-1 Part 3 audio (MP3, MP2, MP1) or MPEG-2 Part 3 audio. It can also contain MPEG-4 Part 2 video, MPEG-2 Part 7 audio (AAC) or MPEG-4 Part 3 (AAC) audio, but they are rarely used. The MPEG-2 Program stream has provisions for non-standard data (e.g. AC-3 audio or subtitles) in the form of so-called private streams. International Organization for Standardization authorized SMPTE Registration Authority, LLC as the registration authority for MPEG-2 format identifiers. It publishes a list of compression formats which can be encapsulated in MPEG-2 transport stream and program stream. Coding details See also Elementary stream MPEG transport stream
https://en.wikipedia.org/wiki/Muirfield%20Seamount
The Muirfield Seamount is a submarine mountain located in the Indian Ocean approximately 130 kilometres (70 nautical miles) southwest of the Cocos (Keeling) Islands. The Cocos Islands are an Australian territory, and therefore the Muirfield Seamount is within Australia's Exclusive Economic Zone (EEZ). The Muirfield Seamount is a submerged archipelago, approximately in diameter and below the surface of the sea. A 1999 biological survey of the seamount performed by the Australian Commonwealth Scientific and Industrial Research Organisation (CSIRO) revealed that the area is depauperate. The Muirfield Seamount was discovered accidentally in 1973 when the cargo ship MV Muirfield (a merchant vessel named after Muirfield, Scotland) was underway in waters charted at a depth of greater than , when she suddenly struck an unknown object, resulting in extensive damage to her keel. In 1983, , a Royal Australian Navy survey ship, surveyed the area where Muirfield was damaged, and charted in detail this previously unsuspected hazard to navigation. The dramatic accidental discovery of the Muirfield Seamount is often cited as an example of limitations in the vertical datum accuracy of some offshore areas as represented on nautical chart especially on small-scale charts. More recently, in 2005 the submarine ran into an uncharted seamount about 560 kilometers (350 statute miles) south of Guam at a speed of , sustaining serious damage and killing one seaman. See also Graveyard Seamounts Jasper Seamount Muirfield Reef Mud volcano Sedlo Seamount South Chamorro Seamount
https://en.wikipedia.org/wiki/St.%20Augustine%20Monster
The St. Augustine Monster is the name given to a large carcass, originally postulated to be the remains of a gigantic octopus, that washed ashore on the United States coast near St. Augustine, Florida in 1896. It is sometimes referred to as the Florida Monster or the St. Augustine Giant Octopus and is one of the earliest recorded examples of a globster. The species that the carcass supposedly represented has been assigned the binomial names Octopus giganteus (Latin for "giant octopus") and Otoctopus giganteus (Greek prefix: oton = "ear"; "giant-eared octopus"), although these are not valid under the rules of the ICZN. A 1995 analysis concluded that the St. Augustine Monster was a large mass of the collagenous matrix of whale blubber, likely from a sperm whale. Discovery Initial stranding The carcass was first spotted on the evening of November 30, 1896, by two young boys, Herbert Coles and Dunham Coretter, while bicycling along Anastasia Island. The enormous mass was half buried in the sand, having sunk under its immense weight. The two boys thought the carcass was the remains of a beached whale, as a similar stranding had occurred two years earlier near the mouth of the Matanzas River, located several miles to the south of St. Augustine (see map). The two boys returned to St. Augustine the same day and reported their discovery to a local physician, Dr. DeWitt Webb. Webb, who was the founder of the St. Augustine Historical Society and Institute of Science, came to the beach the following day, December 1, to examine the remains. He would be the only known person of an academic background to see the specimen in situ. His first impression was that it was the remains of an animal, very mutilated, and in an advanced state of decomposition. The carcass was very pale pink, almost white, in colour, with a silver reflection in the sunlight. It was composed of a rubbery substance of a very hard consistency, such that it could only be cut with great difficulty. The part
https://en.wikipedia.org/wiki/Criticism%20of%20Windows%20Vista
Windows Vista, an operating system released by Microsoft for consumers on January 30, 2007, has been widely criticized by reviewers and users. Due to issues with new security features, performance, driver support and product activation, Windows Vista has been the subject of a number of negative assessments by various groups. Security Driver signing requirement For security reasons, 64-bit versions of Windows Vista allow only signed drivers to be installed in kernel mode. Because code executing in kernel mode enjoys wide privileges on the system, the signing requirement aims to ensure that only code with a known origin executes at this level. In order for a driver to be signed, a developer/software vendor has to obtain an Authenticode certificate with which to sign the driver. Authenticode certificates can be obtained from certificate authorities trusted by Microsoft. Microsoft trusts the certificate authority to verify the applicant's identity before issuing a certificate. If a driver is not signed using a valid certificate, or if the driver was signed using a certificate which has been revoked by Microsoft or the certificate authority, Windows will refuse to load the driver. The following criticisms/claims have been made regarding this requirement: It disallows experimentation from the hobbyist community. The required Authenticode certificates for signing Vista drivers are expensive and out of reach for small developers, usually about $400–$500/year (from Verisign). Microsoft allows developers to temporarily or locally disable the signing requirement on systems they control (by hitting F8 during boot) or by signing the drivers with self-issued certificates or by running a kernel debugger. At one time, a third-party tool called Atsiv existed that would allow any driver, unsigned or signed to be loaded. Atsiv worked by installing a signed "surrogate" driver which could be directed to load any other driver, thus circumventing the driver signing requirement. Since
https://en.wikipedia.org/wiki/Copper%20pour
In electronics, the term copper pour refers to an area on a printed circuit board filled with copper (the metal used to make connections in printed circuit boards). Copper pour is commonly used to create a ground plane. Another reason for using copper pour is to reduce the amount of etching fluid used during manufacturing. A distinctive feature of copper pour is the backoff (or stand-off) - a certain distance between the copper pour and any tracks or pads not belonging to the same electrical net. A copper pour therefore looks like it flows around other components, with the exception of pads which are connected to the copper pour using thermal connections. PCB designers today almost invariably use completely solid areas of copper pour that completely cover the remaining area outside those tracks, pads, and stand-off regions. Many early PCBs have a "hatched copper pour", sometimes called a "cherry pie lattice". While solid copper pour provides better resistive characteristics, hatched copper pour is used to balance the heat and dilatation on both sides of the board in order to avoid warping of certain substrate. Heating might cause gas bubbles between solid copper pour and certain substrates. Furthermore, it might be possible to adjust the impedance of high frequency traces by using hatched copper pour in order to reach better signal quality. See also
https://en.wikipedia.org/wiki/Society%20for%20Behavioral%20Neuroendocrinology
The Society for Behavioral Neuroendocrinology is an interdisciplinary scientific organization dedicated to the study of hormonal processes and neuroendocrine systems that regulate behavior. Publications SBN publishes the scientific journal Hormones and Behavior. External links Neuroscience organizations
https://en.wikipedia.org/wiki/Octomino
An octomino (or 8-omino) is a polyomino of order 8, that is, a polygon in the plane made of 8 equal-sized squares connected edge-to-edge. When rotations and reflections are not considered to be distinct shapes, there are 369 different free octominoes. When reflections are considered distinct, there are 704 one-sided octominoes. When rotations are also considered distinct, there are 2,725 fixed octominoes. Symmetry The figure shows all possible free octominoes, coloured according to their symmetry groups: 316 octominoes (coloured grey) have no symmetry. Their symmetry group consists only of the identity mapping. 23 octominoes (coloured red) have an axis of reflection symmetry aligned with the gridlines. Their symmetry group has two elements, the identity and the reflection in a line parallel to the sides of the squares. 5 octominoes (coloured green) have an axis of reflection symmetry at 45° to the gridlines. Their symmetry group has two elements, the identity and a diagonal reflection. 18 octominoes (coloured blue) have point symmetry, also known as rotational symmetry of order 2. Their symmetry group has two elements, the identity and the 180° rotation. 1 octomino (coloured yellow) has rotational symmetry of order 4. Its symmetry group has four elements, the identity and the 90°, 180° and 270° rotations. 4 octominoes (coloured purple) have two axes of reflection symmetry, both aligned with the gridlines. Their symmetry group has four elements, the identity, two reflections and the 180° rotation. It is the dihedral group of order 2, also known as the Klein four-group. 1 octomino (coloured orange) has two axes of reflection symmetry, both aligned with the diagonals. Its symmetry group is also the dihedral group of order 2 with four elements. 1 octomino (coloured cyan) has four axes of reflection symmetry, aligned with the gridlines and the diagonals, and rotational symmetry of order 4. Its symmetry group, the dihedral group of order 4, has eight
https://en.wikipedia.org/wiki/Selective%20calling
In a conventional, analog two-way radio system, a standard radio has noise squelch or carrier squelch, which allows a radio to receive all transmissions. Selective calling is used to address a subset of all two-way radios on a single radio frequency channel. Where more than one user is on the same channel (co-channel users), selective calling can address a subset of all receivers or can direct a call to a single radio. Selective calling features fit into two major categories—individual calling and group calling. Individual calls generally have longer time-constants: it takes more air-time to call an individual radio unit than to call a large group of radios. Selective calling is akin to the use of a lock on a door. A radio with carrier squelch is unlocked and will let any signal in. Selective calling locks out all signals except ones with the correct "key", in this case a specific digital code. Selective calling systems can overlap; a radio may have (group call) and DTMF individual calling. Selective calling prevents the user from hearing others on a shared channel. It does not eliminate interference from co-channel users (other users on the same radio channel). If two users try to talk at the same time, the signal will be affected by the other party using the channel. Some selective calling systems experience falsing. In other words, the decoder activates when a valid signal is not present. Falsing may come from a maintenance problem or poor engineering. Group calling In conventional FM two-way radio systems, the most common form of selective calling is CTCSS, which is based on a sub-audible tone. One implementation of this system is by Motorola and is called Private Line, or PL. Radios made by nearly any manufacturer will work acceptably with existing systems using CTCSS. The system allows groups of radios to remain muted while other users are talking on the channel. In business and industrial systems, as many as 50 sets of users could share the same channel w
https://en.wikipedia.org/wiki/Jeff%20Cheeger
Jeff Cheeger (born December 1, 1943, Brooklyn, New York City) is a mathematician. Cheeger is professor at the Courant Institute of Mathematical Sciences at New York University in New York City. His main interests are differential geometry and its connections with topology and analysis. Biography Cheeger graduated from Harvard University with a B.A. in 1964. He graduated from Princeton University with an M.S. in 1966 and with a PhD in 1967. He is a Silver Professor at the Courant Institute at New York University where he has worked since 1993. He worked as a teaching assistant and research assistant at Princeton University from 1966–1967, a National Science Foundation postdoctoral fellow and instructor from 1967–1968, an assistant professor from 1968 to 1969 at the University of Michigan, and an associate professor from 1969–1971 at SUNY at Stony Brook. Cheeger was a professor at SUNY, Stony Brook from 1971 to 1985, a leading professor from 1985 to 1990, and a distinguished professor from 1990 until 1992. Cheeger has also had a number of visiting positions in Brazil (1971), at the Institute for Advanced Study (1972, 1977, 1978, 1995), Harvard University (1972), the Institut des Hautes Études Scientifiques (1984–1985) and the Mathematical Sciences Research Institute (1985). He has supervised at least 13 doctoral theses and three postdoctoral fellows. He has served as a member of several American Mathematical Society committees and National Science Foundation panels. Cheeger delivered invited addresses at the International Congress of Mathematicians in 1974 and in 1986. He received the Guggenheim Fellowship in 1984. In 1998 Cheeger was elected a foreign member of the Finnish Academy of Science and Letters. Cheeger was elected a member of the United States National Academy of Sciences in 1997. His election citation read: Cheeger has discovered many of the deepest results in Riemannian geometry, such as estimates for the spectrum of the Laplace-Beltrami operato
https://en.wikipedia.org/wiki/Zero%20sound
Zero sound is the name given by Lev Landau in 1957 to the unique quantum vibrations in quantum Fermi liquids. The zero sound can no longer be thought of as a simple wave of compression and rarefaction, but rather a fluctuation in space and time of the quasiparticles' momentum distribution function. As the shape of Fermi distribution function changes slightly (or largely), zero sound propagates in the direction for the head of Fermi surface with no change of the density of the liquid. Predictions and subsequent experimental observations of zero sound was one of the key confirmation on the correctness of Landau's Fermi liquid theory. Derivation from Boltzmann transport equation The Boltzmann transport equation for general systems in the semiclassical limit gives, for a Fermi liquid, , where is the density of quasiparticles (here we ignore spin) with momentum and position at time , and is the energy of a quasiparticle of momentum ( and denote equilibrium distribution and energy in the equilibrium distribution). The semiclassical limit assumes that fluctuates with angular frequency and wavelength , which are much lower than and much longer than respectively, where and are the Fermi energy and momentum respectively, around which is nontrivial. To first order in fluctuation from equilibrium, the equation becomes . When the quasiparticle's mean free path (equivalently, relaxation time ), ordinary sound waves ("first sound") propagate with little absorption. But at low temperatures (where and scale as ), the mean free path exceeds , and as a result the collision functional . Zero sound occurs in this collisionless limit. In the Fermi liquid theory, the energy of a quasiparticle of momentum is , where is the appropriately normalized Landau parameter, and . The approximated transport equation then has plane wave solutions , with given by . This functional operator equation gives the dispersion relation for the zero sound waves with frequency
https://en.wikipedia.org/wiki/Fixed-point%20iteration
In numerical analysis, fixed-point iteration is a method of computing fixed points of a function. More specifically, given a function defined on the real numbers with real values and given a point in the domain of , the fixed-point iteration is which gives rise to the sequence of iterated function applications which is hoped to converge to a point . If is continuous, then one can prove that the obtained is a fixed point of , i.e., More generally, the function can be defined on any metric space with values in that same space. Examples A first simple and useful example is the Babylonian method for computing the square root of , which consists in taking , i.e. the mean value of and , to approach the limit (from whatever starting point ). This is a special case of Newton's method quoted below. The fixed-point iteration converges to the unique fixed point of the function for any starting point This example does satisfy (at the latest after the first iteration step) the assumptions of the Banach fixed-point theorem. Hence, the error after n steps satisfies (where we can take , if we start from .) When the error is less than a multiple of for some constant , we say that we have linear convergence. The Banach fixed-point theorem allows one to obtain fixed-point iterations with linear convergence. The requirement that is continuous is important, as the following example shows. The iteration converges to 0 for all values of . However, 0 is not a fixed point of the function as this function is not continuous at , and in fact has no fixed points. Attracting fixed points An attracting fixed point of a function is a fixed point of with a neighborhood of "close enough" points around such that for any value of in , the fixed-point iteration sequence is contained in and converges to . The basin of attraction of is the largest such neighborhood . The natural cosine function ("natural" means in radians, not degrees or other units) has exactly one f
https://en.wikipedia.org/wiki/Great%20deluge%20algorithm
The Great deluge algorithm (GD) is a generic algorithm applied to optimization problems. It is similar in many ways to the hill-climbing and simulated annealing algorithms. The name comes from the analogy that in a great deluge a person climbing a hill will try to move in any direction that does not get his/her feet wet in the hope of finding a way up as the water level rises. In a typical implementation of the GD, the algorithm starts with a poor approximation, S, of the optimum solution. A numerical value called the badness is computed based on S and it measures how undesirable the initial approximation is. The higher the value of badness the more undesirable is the approximate solution. Another numerical value called the tolerance is calculated based on a number of factors, often including the initial badness. A new approximate solution S' , called a neighbour of S, is calculated based on S. The badness of S' , b' , is computed and compared with the tolerance. If b' is better than tolerance, then the algorithm is recursively restarted with S : = S' , and tolerance := decay(tolerance) where decay is a function that lowers the tolerance (representing a rise in water levels). If b' is worse than tolerance, a different neighbour S* of S is chosen and the process repeated. If all the neighbours of S produce approximate solutions beyond tolerance, then the algorithm is terminated and S is put forward as the best approximate solution obtained. See also de:Gunter Dueck
https://en.wikipedia.org/wiki/Desert%20%28particle%20physics%29
In the Grand Unified Theory of particle physics (GUT), the desert refers to a theorized gap in energy scales, between approximately the electroweak energy scale–conventionally defined as roughly the vacuum expectation value or VeV of the Higgs field (about 246 GeV)–and the GUT scale, in which no unknown interactions appear. It can also be described as a gap in the lengths involved, with no new physics below 10−18 m (the currently probed length scale) and above 10−31 m (the GUT length scale). The idea of the desert was motivated by the observation of approximate, order of magnitude, gauge coupling unification at the GUT scale. When the values of the gauge coupling constants of the weak nuclear, strong nuclear, and electromagnetic forces are plotted as a function of energy, the 3 values appear to nearly converge to a common single value at very high energies. This was one theoretical motivation for Grand Unified Theories themselves, and adding new interactions at any intermediate energy scale generally disrupts this gauge coupling unification. The disruption arises from the new quantum fields- the new forces and particles- which introduce new coupling constants and new interactions that modify the existing Standard Model coupling constants at higher energies. The fact that the convergence in the Standard Model is actually inexact, however, is one of the key theoretical arguments against the Desert, since making the unification exact requires new physics below the GUT scale. Standard model particles All the Standard Model particles were discovered well below the energy scale of approximately 1012 eV or 1 TeV. The heaviest Standard Model particle is the top quark, with a mass of approximately 173 GeV. The desert Above these energies, desert theory predicts no particles will be discovered until reaching the scale of approximately 1025 eV. According to the theory, measurements of TeV-scale physics at the Large Hadron Collider (LHC) and the near-future International Li
https://en.wikipedia.org/wiki/Vladimir%20Vari%C4%87ak
Vladimir Varićak (sometimes also spelled Vladimir Varičak; March 1, 1865 – January 17, 1942) was a Croatian mathematician and theoretical physicist of Serbian origin. Biography Varićak, an ethnic Serb, was born on March 1, 1865, in the village of Švica near Otočac, Austrian Empire (present-day Croatia). He studied physics and mathematics at the University of Zagreb from 1883 to 1887. He made his PhD in 1889 and got his habilitation in 1895. In 1899 he became professor of mathematics in Zagreb, where he gave lectures until his death in 1942. From 1903 to 1908 he wrote on hyperbolic geometry (or Bolyai–Lobachevskian geometry). In 1910, following a 1909 publication of Sommerfeld, he applied hyperbolic geometry to the special theory of relativity. Sommerfeld, using the imaginary form of Minkowski space, had shown in his 1909 paper that the Einstein formula for combination of velocities is most clearly understandable as a formula for triangular addition on the surface of a sphere of imaginary radius. Varićak reinterpreted this result as showing that rapidity combines by the triangle rule in hyperbolic space. This is a fundamental result for the hyperbolic theory which was demonstrated later by other approaches by Robb (1911) and Borel (1913). The 1910 papers also dealt with several applications of the hyperbolic theory to optics. In 1911 Varićak was invited to speak to the Deutsche Mathematiker-Vereinigung in Karlsruhe on his work. He continued to develop the hyperbolic reinterpretation of Einstein's theory collecting his results in 1924 in a textbook, Darstellung der Relativitätstheorie im drei-dimensionalen Lobatschefskijschen Raume (Relativity in Three-Dimensional Lobachevski Space), now available in English. In the period 1909 to 1913 Varićak had correspondence with Albert Einstein concerning rotation and length contraction where Varićak's interpretations differed from those of Einstein. Concerning length contraction Varićak said that in Einstein's interpretation
https://en.wikipedia.org/wiki/Kendall%27s%20notation
In queueing theory, a discipline within the mathematical theory of probability, Kendall's notation (or sometimes Kendall notation) is the standard system used to describe and classify a queueing node. D. G. Kendall proposed describing queueing models using three factors written A/S/c in 1953 where A denotes the time between arrivals to the queue, S the service time distribution and c the number of service channels open at the node. It has since been extended to A/S/c/K/N/D where K is the capacity of the queue, N is the size of the population of jobs to be served, and D is the queueing discipline. When the final three parameters are not specified (e.g. M/M/1 queue), it is assumed K = ∞, N = ∞ and D = FIFO. First example: M/M/1 queue A M/M/1 queue means that the time between arrivals is Markovian (M), i.e. the inter-arrival time follows an exponential distribution of parameter λ. The second M means that the service time is Markovian: it follows an exponential distribution of parameter μ. The last parameter is the number of service channel which one (1). Description of the parameters In this section, we describe the parameters A/S/c/K/N/D from left to right. A: The arrival process A code describing the arrival process. The codes used are: S: The service time distribution This gives the distribution of time of the service of a customer. Some common notations are: c: The number of servers The number of service channels (or servers). The M/M/1 queue has a single server and the M/M/c queue c servers. K: The number of places in the queue The capacity of queue, or the maximum number of customers allowed in the queue. When the number is at this maximum, further arrivals are turned away. If this number is omitted, the capacity is assumed to be unlimited, or infinite. Note: This is sometimes denoted c + K where K is the buffer size, the number of places in the queue above the number of servers c. N: The calling population The size of calling source. The size of
https://en.wikipedia.org/wiki/Spreading%20activation
Spreading activation is a method for searching associative networks, biological and artificial neural networks, or semantic networks. The search process is initiated by labeling a set of source nodes (e.g. concepts in a semantic network) with weights or "activation" and then iteratively propagating or "spreading" that activation out to other nodes linked to the source nodes. Most often these "weights" are real values that decay as activation propagates through the network. When the weights are discrete this process is often referred to as marker passing. Activation may originate from alternate paths, identified by distinct markers, and terminate when two alternate paths reach the same node. However brain studies show that several different brain areas play an important role in semantic processing. Spreading activation in semantic networks as a model were invented in cognitive psychology to model the fan out effect. Spreading activation can also be applied in information retrieval, by means of a network of nodes representing documents and terms contained in those documents. Cognitive psychology As it relates to cognitive psychology, spreading activation is the theory of how the brain iterates through a network of associated ideas to retrieve specific information. The spreading activation theory presents the array of concepts within our memory as cognitive units, each consisting of a node and its associated elements or characteristics, all connected together by edges. A spreading activation network can be represented schematically, in a sort of web diagram with shorter lines between two nodes meaning the ideas are more closely related and will typically be associated more quickly to the original concept. For memory psychology, Spreading activation model means people organize their knowledge of the world based on their personal experience, which is saying those personal experiences form the network of ideas that is the person's knowledge of the world. When a word
https://en.wikipedia.org/wiki/Allegany%20Ballistics%20Laboratory
Allegany Ballistics Laboratory (ABL) located in Rocket Center, West Virginia, is a diverse industrial complex employing some 1,000 people across . The facility is a member of the Federal Laboratory Consortium and is operated by Northrop Grumman (former Alliant Techsystems) under contract with the Naval Sea Systems Command (NAVSEA). Current operation The ABL facility is a manufacturer of advanced composite structures for the F-22 Raptor and other aerospace projects. ATK also operates 6 of 11 known advanced fiber placement machines. In addition the site produces about 80 military products, including: 30mm shells for Apache helicopters, training grenades, fuze-proximity sensors, mortars and warheads, and tank ammunition. Also on the site is the Robert C. Byrd Hilltop Office Complex and the Robert C. Byrd Institute for Advanced Flexible Manufacturing. At the Robert C. Byrd Complex on the hill companies have rented space to do secure research, among them IBM (which recently acquired National Interest Security Company) who is digitizing data on hurricane cleanup, avian influenza, and weather records. It also plays a significant role in continuity of government operations. History ABL was established in 1944 on the site of a former ammunitions plant on land owned by the Army. After World War II, the plant was transferred to the Office of Scientific Research and Development and was involved in building propulsion devices and engines for the solid-rocket industry. Later in the decade, ownership of ABL was transferred to the Navy office of Naval Sea Systems Command. Since 1946 it was operated by the Hercules Powder Company. In 1956 when it was producing Altair rocket stages for Vanguard rockets, ABL was, "A subsidiary of the Navy operated by the Hercules Powder Company." The Navy now contracts out operation of the facility to ATK (Alliant Techsystems), a $3.4 billion corporation based in Edina, Minnesota. In 1998, ATK's Conventional Munitions Group was selected by
https://en.wikipedia.org/wiki/Preload%20%28software%29
preload is a free Linux program which runs as a daemon to record statistics about usage of files by more frequently-used programs. This information is then used to keep these files preloaded into memory. This results in faster application startup times as less data needs to be fetched from disk. preload is often paired with prelink. preload was written by Behdad Esfahbod and uses Markov chains in its implementation; See also Prefetching Readahead
https://en.wikipedia.org/wiki/Electroneutral%20cation-Cl
In molecular biology, the electroneutral cation-Cl (electroneutral potassium chloride cotransporter) family of proteins are a family of solute carrier proteins. This family includes the products of the Human genes: SLC12A1, SLC12A1, SLC12A2, SLC12A3, SLC12A4, SLC12A5, SLC12A6, SLC12A7, SLC12A8 and SLC12A9. The K-Cl co-transporter (KCC) mediates the coupled movement of K+ and Cl− ions across the plasma membrane of many animal cells. This transport is involved in the regulatory volume decrease in response to cell swelling in red blood cells, and has been proposed to play a role in the vectorial movement of Cl− across kidney epithelia. The transport process involves one for one electroneutral movement of K+ together with Cl−, and, in all known mammalian cells, the net movement is outward. The neuronal KCC subtype KCC2 is cell-volume insensitive and plays a unique role in maintaining low intracellular Cl−concentration, which is required in neurones for the functioning of Cl− dependent fast synaptic inhibition, mediated by certain neurotransmitters, such as gamma-aminobutyric acid (GABA) and glycine. Three isoforms of the K-Cl co-transporter have been described, termed KCC1 (SLC12A4), KCC2 (SLC12A5), and KCC3 (SLC12A6), containing 1085, 1116 and 1150 amino acids, respectively. They are predicted to have 12 transmembrane (TM) regions in a central hydrophobic domain, together with hydrophilic N- and C-termini that are likely cytoplasmic. Comparison of their sequences with those of other ion-transporting membrane proteins reveals that they are part of a new superfamily of cation-chloride co-transporters, which includes the Na-Cl and Na-K-2Cl co-transporters. KCC1 and KCC3 are widely expressed in human tissues, while KCC2 is expressed only in brain neurons, making it likely that this is the isoform responsible for maintaining low Cl− concentration in neurons. A study in the model organism C. elegans found that the KCC3 ortholog functions in glial cells to regulate animal
https://en.wikipedia.org/wiki/Diospyros%20blancoi
Diospyros blancoi, (synonym Diospyros discolor), commonly known as velvet apple, velvet persimmon, kamagong, or mabolo tree, is a tree of the genus Diospyros of ebony trees and persimmons. It produces edible fruit with a fine, velvety, reddish-brown fur-like covering. The fruit has a soft, creamy, pink flesh, with a taste and aroma comparable to peaches. It is widely distributed and native to the Philippines, but it is also native to eastern and southern Taiwan. It has also been introduced to other parts of Southeast Asia, the Pacific Islands, South Asia, the Caribbean, Florida, and other tropical regions. Cultivation It is a dioecious tropical tree that grows well in a diversity of soil, from sea level to above sea level. Seed trees are normally planted from each other; this one can be planted from from each other. It needs a good distribution of rainfall through the year. Trees that were planted by seeds could take 6 or 7 years to give out fruit, but trees that were propagated by cuttings produce fruit in 3 or 4 years. It is a very productive tree. The fact that fruits vary greatly – in shape, color, hairiness and taste – suggests that there is a great deal of genetic variation in the plant. Seedless cultivars exist, and are highly favored since in the normal varieties the large seeds occupy a considerable volume of the fruit. Timber Like that of other trees in Dyospiros, which includes ebony, Kamagong timber is extremely dense and hard and is famous for its dark color. The wood is generally used for house construction which include flooring, post, doors, and windows, among others. Finished products from kamagong wood, such as fine furniture and decoratives can be exported provided that they are properly documented and approved by the Customs authorities. Kamagong is also popular for martial arts training implements such as bokkens and eskrima sticks. Secondary metabolites The leaves of velvet apple trees have been shown to contain isoarborinol methyl
https://en.wikipedia.org/wiki/Proximodistal%20trend
The proximodistal trend is the tendency for more general functions of limbs to develop before more specific or fine motor skills. It comes from the Latin words proxim- which means "close" and "-dis-" meaning "away from", because the trend essentially describes a path from the center outward.
https://en.wikipedia.org/wiki/Mathematical%20discussion%20of%20rangekeeping
In naval gunnery, when long-range guns became available, an enemy ship would move some distance after the shells were fired. It became necessary to figure out where the enemy ship, the target, was going to be when the shells arrived. The process of keeping track of where the ship was likely to be was called rangekeeping, because the distance to the target—the range—was a very important factor in aiming the guns accurately. As time passed, train (also called bearing), the direction to the target, also became part of rangekeeping, but tradition kept the term alive. Rangekeeping is an excellent example of the application of analog computing to a real-world mathematical modeling problem. Because nations had so much money invested in their capital ships, they were willing to invest enormous amounts of money in the development of rangekeeping hardware to ensure that the guns of these ships could put their projectiles on target. This article presents an overview of the rangekeeping as a mathematical modeling problem. To make this discussion more concrete, the Ford Mk 1 Rangekeeper is used as the focus of this discussion. The Ford Mk 1 Rangekeeper was first deployed on the in 1916 during World War I. This is a relatively well documented rangekeeper that had a long service life. While an early form of mechanical rangekeeper, it does illustrate all the basic principles. The rangekeepers of other nations used similar algorithms for computing gun angles, but often differed dramatically in their operational use. In addition to long range gunnery, the launching of torpedoes also requires a rangekeeping-like function. The US Navy during World War II had the TDC, which was the only World War II-era submarine torpedo fire control system to incorporate a mechanical rangekeeper (other navies depended on manual methods). There were also rangekeeping devices for use with surface ship-launched torpedoes. For a view of rangekeeping outside that of the US Navy, there is a detailed refer
https://en.wikipedia.org/wiki/Passenger%20car%20equivalent
Passenger car equivalent (PCE) or passenger car unit (PCU) is a metric used in transportation engineering, to assess traffic-flow rate on a highway. A passenger car equivalent is essentially the impact that a mode of transport has on traffic variables (such as headway, speed, density) compared to a single car. Traffic studies and/ or analysis must be done to obtain the number of trips, which shall then be converted to PCUs based on the above standards. However, it is worth noting that every region has its own manual with regards to pcu equivalence factors. Highway capacity is measured in PCE/hour daily A common method used in the US is the density method. However, the PCU values derived from the density method are based on underlying homogeneous traffic concepts such as strict lane discipline, car following and a vehicle fleet that does not vary greatly in width. On the other hand, highways in India, carry heterogeneous traffic, where road space is shared among many traffic modes with different physical dimensions. Loose lane discipline prevails; car following is not the norm. This complicates computing of PCE. Using multiple heuristic techniques, transportation engineers convert a mixed traffic stream into a hypothetical passenger-car stream. Methods Many methods exist for determining passenger car units (PCUs) examples: homogenization coefficient, semi-empirical method, Walker's method, headway method, multiple linear regression method simulation method. Transport for London recommend the following PCU values in an urban context: Pedal cycle 0.2 Motorcycle 0.4 Car or light goods vehicle 1.0 Medium goods vehicle 1.5 Bus or coach 2.0 Heavy goods vehicle (HGV) 2.3 It may be appropriate to use different values for the same vehicle type according to circumstances. For example, in the UK in the 1960s and 1970s, bicycles were evaluated thus: on rural roads 0.5 on urban roads 0.33 on roundabouts 0.5 at traffic lights 0.2.
https://en.wikipedia.org/wiki/M%C3%B6bius%20ladder
In graph theory, the Möbius ladder , for even numbers , is formed from an by adding edges (called "rungs") connecting opposite pairs of vertices in the cycle. It is a cubic, circulant graph, so-named because (with the exception of (the utility graph ), has exactly four-cycles which link together by their shared edges to form a topological Möbius strip. Möbius ladders were named and first studied by . Properties For every even , the Möbius ladder is a nonplanar apex graph, meaning that it cannot be drawn without crossings in the plane but removing one vertex allows the remaining graph to be drawn without crossings. These graphs have crossing number one, and can be embedded without crossings on a torus or projective plane. Thus, they are examples of toroidal graphs. explores embeddings of these graphs onto higher genus surfaces. Möbius ladders are vertex-transitive – they have symmetries taking any vertex to any other vertex – but (with the exceptions of and ) they are not edge-transitive. The edges from the cycle from which the ladder is formed can be distinguished from the rungs of the ladder, because each cycle edge belongs to a single 4-cycle, while each rung belongs to two such cycles. Therefore, there is no symmetry taking a cycle edge to a rung edge or vice versa. When n ≡ 2 (mod 4), Mn is bipartite. When n ≡ 0 (mod 4), it is not bipartite. The endpoints of each rung are an even distance apart in the initial cycle, so adding each rung creates an odd cycle. In this case, because the graph is 3-regular but not bipartite, by Brooks' theorem it has chromatic number 3. show that the Möbius ladders are uniquely determined by their Tutte polynomials. The Möbius ladder M8 has 392 spanning trees; it and M6 have the most spanning trees among all cubic graphs with the same number of vertices. However, the 10-vertex cubic graph with the most spanning trees is the Petersen graph, which is not a Möbius ladder. The Tutte polynomials of the Möbius ladders may be
https://en.wikipedia.org/wiki/PAL-M
PAL-M is the analogue colour TV system used in Brazil since early 1972, making it the first South American country to broadcast in colour. It is unique among analogue TV systems in that it combines the 525-line 30 frames-per-second System M with the PAL colour encoding system (using very nearly the NTSC colour subcarrier frequency), unlike all other countries which pair PAL with 625-line systems and NTSC with 525-line systems. Colour broadcasts began on February 19, 1972, when the Globo and Bandeirantes networks transmitted the Caxias do Sul Grape Festival. Transition from black and white to colour on most programmes was not complete until 1978, and only became commonplace nationwide by 1980. Origins NTSC being the "natural" choice for countries with monochrome standard M, the choice of a different colour system poses problems of incompatibility with available hardware and the need to develop new television sets and production hardware. Walter Bruch, inventor of PAL, explains Brazil's choice of PAL over NTSC against these odds by an advertising campaign Telefunken and Philips carried out across South America in 1972, which included colour test broadcasts of popular shows (done with TV Globo) and technical demonstrations with executives of television stations. Technical specifications PAL-M signals are in general identical to North American NTSC signals, except for the encoding of the colour carrier. Both systems are based on the monochrome CCIR System M standard, therefore, PAL-M will display in monochrome with sound on an NTSC set and vice versa. Nevertheless, due to the different gamma correction values (2.2 for NTSC, 2.8 for PAL-M), gray tones will be incorrect. PAL-M is incompatible with 625-line based versions of PAL, because its frame rate, scan line, colour subcarrier and sound carrier specifications are different. It will therefore usually give a rolling and/or squashed monochrome picture with no sound on a native European PAL television, as do NTSC
https://en.wikipedia.org/wiki/Software%20engine
A software engine is a core component of a complex software system. Alternate phrases include "software core" and "software core engine", or just "core engine". Definitions The word "engine" is a metaphor of a car's engine. Thus, it is a software engine that is a complex subsystem. There is no formal guideline for what should be called an engine, but the term has become engraved in the software industry. Notable examples are; Database engines, Graphics engines, Physics engines, Search engines, Plotting engines, Game engines. Moreover, a web browser has two components referred to as engines: Browser engine JavaScript engine. Classically, an engine is something packaged as a library, such as ".sa", ".so", or ".dll", that provides functionality to the software that loads or embeds it. Engines may produce graphics, such as the Python matplotlib or the Objective-C Core Plot. Engines do not in and of themselves generally have standalone user interfaces - they are not applications. A distinguishing characteristic of an engine is its presentation as an API. Examples Engines may be used to produce higher-level services that are applications, and the application developers or the management may choose to call the service an "engine". In the context of the packaging of software components, "engine" means one thing. In the context of advertising an online service, "engine" may mean something entirely different. In the arena of "core software development", an engine is a software module that might be included in other software using a package manager such as NuGet for C#, Pipenv for Python, and Swift Package Manager for the Swift language. One seeming outlier is a search engine, such as Google Search, because it is a standalone service provided to end users. However, for the search provider, the engine is part of a distributed computing system that can encompass many data centres throughout the world. The word "engine" is evolving along with the evolutio
https://en.wikipedia.org/wiki/Calponin
Calponin is a calcium binding protein. Calponin tonically inhibits the ATPase activity of myosin in smooth muscle. Phosphorylation of calponin by a protein kinase, which is dependent upon calcium binding to calmodulin, releases the calponin's inhibition of the smooth muscle ATPase. Structure and function Calponin is mainly made up of α-helices with hydrogen bond turns. It is a binding protein and is made up of three domains. These domains in order of appearance are Calponin Homology (CH), regulatory domain (RD), and Click-23, domain that contains the calponin repeats. At the CH domain calponin binds to α-actin and filamin and binds to actin within the RD domain. Calmodulin, when activated by calcium may bind weakly to the CH domain and inhibit calponin binding with α-actin. Calponin is responsible for binding many actin binding proteins, phospholipids, and regulates the actin/myosin interaction. Calponin is also thought to negatively affect the bone making process due to being expressed in high amounts in osteoblasts.
https://en.wikipedia.org/wiki/Caldesmon
Caldesmon is a protein that in humans is encoded by the CALD1 gene. Caldesmon is a calmodulin binding protein. Like calponin, caldesmon tonically inhibits the ATPase activity of myosin in smooth muscle. This gene encodes a calmodulin- and actin-binding protein that plays an essential role in the regulation of smooth muscle and nonmuscle contraction. The conserved domain of this protein possesses the binding activities to -calmodulin, actin, tropomyosin, myosin, and phospholipids. This protein is a potent inhibitor of the actin-tropomyosin activated myosin MgATPase, and serves as a mediating factor for -dependent inhibition of smooth muscle contraction. Alternative splicing of this gene results in multiple transcript variants encoding distinct isoforms. Immunochemistry In diagnostic immunochemistry, caldesmon is a marker for smooth muscle differentiation.
https://en.wikipedia.org/wiki/Implantation%20%28embryology%29
Implantation, also known as nidation is the stage in the embryonic development of mammals in which the blastocyst hatches, attaches, adheres, and invades into the wall of the female's uterus. Implantation is the first stage of gestation, and, when successful, the female is considered to be pregnant. An implanted embryo is detected by the presence of increased levels of human chorionic gonadotropin (hCG) in a pregnancy test. The implanted embryo will receive oxygen and nutrients in order to grow. For implantation to take place the uterus must become receptive. Uterine receptivity involves changes to the endometrium, and much cross-talk between the uterus and the embryo. This stage gives a synchrony that opens a window of implantation that enables successful implantation of the viable embryo. The endocannabinoid system plays a vital role in this synchrony in the uterus, influencing uterine receptivity, and embryo implantation. The embryo expresses cannabinoid receptors early in its development that are responsive to anandamide (AEA) secreted in the uterus. AEA is produced at higher levels before implantation and is then down-regulated at the time of implantation. This signaling is of importance in the embryo-uterus crosstalk in regulating the timing of embryonic implantation and uterine receptivity. Adequate concentrations of AEA that are neither too high or too low, are needed for successful implantation. There is an extensive variation in the type of trophoblast cells, and structures of the placenta across the different species of mammals. Of the five recognised stages of implantation including two pre-implantation stages that precede placentation, the first four are similar across the species. The five stages are migration and hatching, pre-contact, attachment, adhesion, and invasion. The two pre-implantation stages are associated with the pre-implantation embryo. In humans, following the stage of hatching that takes place around four to five days after ferti
https://en.wikipedia.org/wiki/819%20line
819-line was an analog monochrome TV system developed and used in France as television broadcast resumed after World War II. Transmissions started in 1949 and were active up to 1985, although limited to France, Belgium and Luxembourg. It is associated with CCIR System E and F. History When Europe resumed TV transmissions after World War II (i.e. in the late 1940s and early 1950s) most countries standardized on 625-line television systems. The two exceptions were the British 405-line system, which had already been introduced in 1936, and the French 819-line system. During the 1940s René Barthélemy had already reached 1,015 lines and even 1,042 lines. On November 20, 1948, François Mitterrand, the then Secretary of State for Information, decreed a broadcast standard of 819 lines developed by Henri de France; broadcasting began at the end of 1949 in this higher definition format. It was used in France by TF1, and in Monaco by Tele Monte Carlo. Some 819-line TV sets were available, like the Grammont 504-A-31 from 1951 and the Philips 14TX100 multi-standard 625/819-line TV from 1952 The system was also adopted (with limited bandwidth, affecting image resolution) in 1953 in Belgium by RTB and in 1955 in Luxembourg by Télé-Luxembourg. Broadcasts were discontinued in Belgium in February 1968, and in Luxembourg in September 1971. Despite some attempts to create a color SECAM version of the 819-line system, France gradually abandoned the system in favor of the Europe-wide standard of 625-lines with the final 819-line transmissions taking place in Paris from the Eiffel Tower on 19 July 1983. Tele Monte Carlo in Monaco were the last broadcasters to transmit 819-line television, closing down their transmitter in 1985. Technical details This was arguably the world's first high-definition television system, and, by today's standards, it could be called 736i (as it had 737 lines active, but one of the lines was composed of 2 halves) with a maximum theoretical resolution of 4
https://en.wikipedia.org/wiki/BPO%20security
Information security has emerged as a significant concern for banks, mobile phone companies and other businesses that use call centers or business process outsourcing, or BPO. There have been instances of theft of personal data reported from call centers. Britain's Financial Services Authority examined standards in India in April 2005 and the Banking Code Standards Board audited eight India-based call centers in 2006, handling more than a million calls per month from the UK. the examinations did not extend to Africa-based call centers staffed by workers of Indian origin. The BCSB report stated that "Customer data is subject to the same level of security as in the UK. High risk and more complex processes are subject to higher levels of scrutiny than similar activities onshore." India's NASSCOM has said that they take breach in security extremely seriously and will assist the police in their probe. Common countermeasures There are three identifiable types of illicit activities concerning fraud emanating from call centers: Crooks who pretend to be legitimate call centres. Hackers who gain access to call centre information through illegal means Call centre agents who illegally misuse the information they have access to in call centres. 3rd and 4th party software implementation, allowing for "back-doors" to be entered remotely, sometimes under the "credentials" of security. While items 1 and 2 are mostly subject to police action, call centres can use internal procedures to minimise risk. Such mitigation measures include but are not limited to: Creating a paperless environment, preventing employees from writing down and removing information by ensuring that all work processes are done on the computer, without having to record anything on forms or notes. Prohibiting the use of cellphones and cameras on the floor. Prohibiting paper, pens and digital recording devices from being brought onto the floor. Preventing internet access for employees on the floor. Lim
https://en.wikipedia.org/wiki/Laser%20accelerometer
A laser accelerometer is an accelerometer that uses a laser to measure changes in velocity/direction. Mechanism It employs a frame with three orthogonal input axes and multiple proof masses. Each proof mass has a predetermined blanking surface. A flexible beam supports each proof mass. The flexible beam permits movement of the proof mass on its axis. A laser light source provides a light ray. The laser source has a transverse field characteristic with a central null intensity region. A mirror transmits a beam of light to a detector. The detector is positioned to be centered on the light ray and responds to the light's intensity to provide an intensity signal. The signal's magnitude is related to the intensity of the light ray. The proof mass blanking surface is centrally positioned within and normal to the light ray null intensity region to provide increased blanking of the light ray in response to transverse movement of the mass on the input axis. In response to acceleration in the direction of the input axis, the proof mass deflects the beam and moves the blanking surface in a direction transverse to the light ray to partially blank the light beam. A control responds to the intensity signal to apply a restoring force to restore the proof mass to a central position and provides an output signal proportional to the restoring force. Applications Accelerometers are added to many devices, including (smart) watches, phones and vehicles of all kinds. Accelerometers oriented vertically function as gravimeters, useful for mining. Other applications include medical diagnostics and satellite measurements for climate change studies. Lasers Basic lasers operate with a frequency range (line width) of some 500 mHz. The range is widened by small temperature changes and vibrations, and by imperfections in the laser cavity. The line width of a specialised scientific laser approaches 1mHz. History 2021 An accelerometer was announced that used infrared light to measure t
https://en.wikipedia.org/wiki/Adaptive%20response
The adaptive response is a form of direct DNA repair in E. coli that protects DNA from damage by external agents or by errors during replication. It is initiated against alkylation, particularly methylation, of guanine or thymine nucleotides or phosphate groups on the sugar-phosphate backbone of DNA. Under sustained exposure to low-level treatment with alkylating mutagens, E. coli can adapt to the presence of the mutagen, rendering subsequent treatment with high doses of the same agent less effective. This mechanism has four related genes, also known as “SOS genes”: ada, alkA, alkB, and aidB, each one working in specific residues, all regulated by ada protein. The adaptive response is mediated by the ada protein (a part of the ada regulon), which covalently transfers alkyl groups from the damaged DNA to one of its two active methyl acceptor cysteine residues: Cys69 and Cys321. Ada protein can repair the damage by transferring methyl groups from O6-methylguanine or O4-methylthymine to Cys321 and also from methylphosphotriesters to Cys69 residue through an irreversible process. It can also convert the protein from a weak to a strong activator of transcription, increasing alkylation repair activity. Function Environmental influence plays a crucial role in the developmental plasticity of genotypes due to the introduction of DNA damaging agents. This phenomenon and the defense mechanism that has evolved to protect an organism’s genotype against damage and prevent multiple phenotypes is known as the adaptive response. Since the adaptive response is able to prevent the possibility of different phenotypes it, therefore, allows organisms to minimize the stress effects it experiences from different stressors and eventually develop a resistance to the stressors. The effects of various chemical, biological, and physical genotoxic damaging agents jeopardize the genotypic integrity of all organisms; however, many evolutionary defense mechanisms have developed so that the s
https://en.wikipedia.org/wiki/Data%20synchronization
Data synchronization is the process of establishing consistency between source and target data stores, and the continuous harmonization of the data over time. It is fundamental to a wide variety of applications, including file synchronization and mobile device synchronization. Data synchronization can also be useful in encryption for synchronizing public key servers. Data synchronization is needed to update and keep multiple copies of a set of data coherent with one another or to maintain data integrity, Figure 3. For example, database replication is used to keep multiple copies of data synchronized with database servers that store data in different locations. Examples Examples include: File synchronization, such as syncing a hand-held MP3 player to a desktop computer; Cluster file systems, which are file systems that maintain data or indexes in a coherent fashion across a whole computing cluster; Cache coherency, maintaining multiple copies of data in sync across multiple caches; RAID, where data is written in a redundant fashion across multiple disks, so that the loss of any one disk does not lead to a loss of data; Database replication, where copies of data on a database are kept in sync, despite possible large geographical separation; Journaling, a technique used by many modern file systems to make sure that file metadata are updated on a disk in a coherent, consistent manner. Challenges Some of the challenges which user may face in data synchronization: data formats complexity; real-timeliness; data security; data quality; performance. Data formats complexity Data formats tend to grow more complex with time as the organization grows and evolves. This results not only in building simple interfaces between the two applications (source and target), but also in a need to transform the data while passing them to the target application. ETL (extraction transformation loading) tools can be helpful at this stage for managing data format complexities. Re
https://en.wikipedia.org/wiki/Microneurography
Microneurography is a neurophysiological method employed to visualize and record the traffic of nerve impulses that are conducted in peripheral nerves of waking human subjects. It can also be used in animal recordings. The method has been successfully employed to reveal functional properties of a number of neural systems, e.g. sensory systems related to touch, pain, and muscle sense as well as sympathetic activity controlling the constriction state of blood vessels. To study nerve impulses of an identified nerve, a fine tungsten needle microelectrode is inserted into the nerve and connected to a high input impedance differential amplifier. The exact position of the electrode tip within the nerve is then adjusted in minute steps until the electrode discriminates nerve impulses of interest. A unique feature and a significant strength of the microneurography method is that subjects are fully awake and able to cooperate in tests requiring mental attention, while impulses in a representative nerve fibre or set of nerve fibres are recorded, e.g. when cutaneous sense organs are stimulated or subjects perform voluntary precision movements. History Before the microneurography technique was developed in the late 1960s, impulses in peripheral nerves had been recorded in animal experiments alone using a technique that involved dissection and splitting the nerve. This approach is not tolerable for general use in humans although it has been pursued in one single study. Actually, the concern of nerve damage was a major obstacle for the development of microneurography because the approach of inserting a needle electrode in a human nerve was generally regarded as potentially dangerous and involving substantial risk of permanent nerve damage. The two Swedish scientists who developed the microneuropgraphy technique (Hagbarth and Vallbo) handled the medical-ethical concern by performing a large series of experiments on their own nerves during a period of about 2 years while carefully
https://en.wikipedia.org/wiki/Retrogression%20heat%20treatment
Retrogression heat treatment (RHT) is a heat treatment process that rapidly heat treats age-hardenable aluminum alloys. Mainly induction heating is used for RHT. In the past, it was mainly used for 6061 and 6063 aluminum alloys. Therefore, forming of complex shapes is possible, without creating damages like cracks. Even hard tempers (for example -T6) can be formed easily after subjecting these alloys to RHT.
https://en.wikipedia.org/wiki/Dysthanasia
In medicine, dysthanasia means "bad death" and is considered a common fault of modern medicine. Dysthanasia is a term generally used when a person is kept alive artificially, in a condition where otherwise they cannot survive, sometimes for some sort of ulterior motive. Dysthanasia occurs when a person who is dying has their biological life extended through technological means without regard to the person's quality of life. Technologies such as an implantable cardioverter defibrillator, artificial ventilation, ventricular assist devices, and extracorporeal membrane oxygenation can extend the dying process. In some cases, cardiopulmonary resuscitation can be considered a form of dysthanasia. The term was used frequently in the investigation into the death of Formula One driver Ayrton Senna in 1994. The etymology of the term is from the Greek language: δυσ, dus; "bad, difficult" + θάνατος, thanatos; "death". See also Brain death Death with Dignity National Center Euthanasia Life support Palliative care
https://en.wikipedia.org/wiki/MARS%20%28ticket%20reservation%20system%29
, which stands for Multi Access (originally Magnetic-electronic Automatic) seat Reservation System is a train ticket reservation system used by the railway companies of former Japanese National Railways, currently Japan Railways Group (JR Group) and travel agencies in Japan, developed jointly by Hitachi and the Railway Information Systems Co., Ltd (JR Systems), a JR Group company jointly owned by the seven members of the group. Outline The host computer of the system was previously located in Kokubunji, Tokyo until 2013, when it was moved to an undisclosed location in the northern part of the Kantō region. The system is managed by since 1 April 1987. Ticket offices at JR stations equipped with MARS terminals are called , selling tickets of all JR Group trains and partly highway buses and route buses and ferries. It is possible for passengers to reserve tickets of buses and trains from one month prior to the given trip. Currently the Midori no Madoguchi is named by JR Group. History MARS-1 The MARS-1 system was created by Mamoru Hosaka, Yutaka Ohno, and others at the Japanese National Railways' R&D Institute (now the Railway Technical Research Institute), and was built in 1958. It was the world's first seat reservation system for trains, and entered service in February 1960, initially only providing bookings for the Kodama and Tsubame limited express services. The MARS-1 was capable of reserving seat positions, and was controlled by a Hitachi mainframe transistor computer with a central processing unit consisting of a thousand transistors and a magnetic drum memory unit for data storage, which was where the MARS acronym originated from. In 2008, the MARS-1 system received a "One Step on Electro Technology -Look Back to the Future-" commemorative plaque from the Institute of Electrical Engineers of Japan. MARS 100/200 MARS 300 MARS 500 MARS 501 Introduced in stages between 2002 and 2004, the MARS 501 introduced the concept of an Ethernet-based client–ser
https://en.wikipedia.org/wiki/N-Myc
N-myc proto-oncogene protein also known as N-Myc or basic helix-loop-helix protein 37 (bHLHe37), is a protein that in humans is encoded by the MYCN gene. Function The MYCN gene is a member of the MYC family of transcription factors and encodes a protein with a basic helix-loop-helix (bHLH) domain. This protein is located in the cell nucleus and must dimerize with another bHLH protein in order to bind DNA. N-Myc is highly expressed in the fetal brain and is critical for normal brain development. The MYCN gene has an antisense RNA, N-cym or MYCNOS, transcribed from the opposite strand which can be translated to form a protein product. N-Myc and MYCNOS are co-regulated both in normal development and in tumor cells, so it is possible that the two transcripts are functionally related. It has been shown that the antisense RNA encodes for a protein, named NCYM, that has originated de novo and is specific to human and chimpanzee. This NCYM protein inhibits GSK3b and thus prevents MYCN degradation. Transgenic mice that harbor human MYCN/NCYM pair often show neuroblastomas with distant metastasis, which are atypical for normal mice. Thus NCYM represents a rare example of a de novo gene that has acquired molecular function and plays a major role in oncogenesis. Clinical significance Amplification and overexpression of N-Myc can lead to tumorigenesis. Excess N-Myc is associated with a variety of tumors, most notably neuroblastomas where patients with amplification of the N-Myc gene tend to have poor outcomes. MYCN can also be activated in neuroblastoma and other cancers through somatic mutation. Intriguingly, recent genome-wide H3K27ac profiling in patient-derived NB samples revealed four distinct SE-driven epigenetic subtypes, characterized by their own and specific master regulatory networks. Three of them are named after the known clinical groups: MYCN-amplified, MYCN non-amplified high-risk, and MYCN non-amplified low-risk NBs, while the fourth displays cellular featu
https://en.wikipedia.org/wiki/Orosomucoid
Orosomucoid (ORM) or alpha-1-acid glycoprotein (α1AGp, AGP or AAG) is an acute phase protein found in plasma. It is an alpha-globulin glycoprotein and is modulated by two polymorphic genes. It is synthesized primarily in hepatocytes and has a normal plasma concentration between 0.6–1.2 mg/mL (1–3% plasma protein). Plasma levels are affected by pregnancy, burns, certain drugs, and certain diseases, particularly HIV. The only established function of ORM is to act as a carrier of basic and neutrally charged lipophilic compounds. In medicine, it is known as the primary carrier of basic (negatively charged) drugs (whereas albumin carries acidic (positively charged) and neutral drugs), steroids, and protease inhibitors. Aging causes a small decrease in plasma albumin levels; if anything, there is a small increase in alpha-1-acid glycoprotein. The effect of these changes on drug protein binding and drug delivery, however, appear to be minimal. AGP shows a complex interaction with thyroid homeostasis: AGP in low concentrations was observed to stimulate the thyrotropin (TSH) receptor and intracellular accumulation of cyclic AMP. High AGP concentrations, however, inhibited TSH signalling. Alpha-1-acid glycoprotein has been identified as one of four potentially useful circulating biomarkers for estimating the five-year risk of all-cause mortality (the other three are albumin, very low-density lipoprotein particle size, and citrate). Orosomucoid increases in amount in obstructive jaundices while diminishes in hepatocellular jaundice and in intestinal infections.
https://en.wikipedia.org/wiki/Public%20analyst
Public Analysts are scientists in the British Isles whose principal task is to ensure the safety and correct description of food by testing for compliance with legislation. Most Public Analysts are also Agricultural Analysts who carry out similar work on animal feedingstuffs and fertilisers. Nowadays this includes checking that the food labelling is accurate. They also test drinking water, and may carry out chemical and biological tests on other consumer products. While much of the work is done by other scientists and technicians in the laboratory, the Public Analyst has legal responsibility for the accuracy of the work and the validity of any opinion expressed on the results reported. There is an Association of Public Analysts, which includes members with similar roles if different titles in other countries. History The office of Public Analyst was established by the Adulteration of Food and Drink Act 1860, the first three appointments being in London, Birmingham and Dublin. The first Scottish analyst was Henry Littlejohn in Edinburgh in 1862, who, with a strong medicinal background and brilliant mind, established many of the critical foundations of public analysis. The Sale of Food and Drugs Act 1875 made food analysis compulsory and the Sale of Food and Drugs Act 1899 extended its scope. Sampling officers generally operated through local public health or sanitary committees. By 1894 there were 99 public analysts overseeing 237 English and Welsh districts. The City of London Corporation had three food inspectors and a wharf and warehouse inspector in 1908. Bradford employed an inspector who made 756 visits to fish and chip shops in 1915. In the 1930s the staff in Birmingham comprised three qualified assistants, a clerk and a laboratory attendant. The Nuisances Removal (England) Act 1855 and the Public Health Act 1875 gave authority for taking food samples "at all reasonable times". Inspectors, police constables and samplers were responsible for taking food
https://en.wikipedia.org/wiki/Electronic%20authentication
Electronic authentication is the process of establishing confidence in user identities electronically presented to an information system. Digital authentication, or e-authentication, may be used synonymously when referring to the authentication process that confirms or certifies a person's identity and works. When used in conjunction with an electronic signature, it can provide evidence of whether data received has been tampered with after being signed by its original sender. Electronic authentication can reduce the risk of fraud and identity theft by verifying that a person is who they say they are when performing transactions online. Various e-authentication methods can be used to authenticate a user's identify ranging from a password to higher levels of security that utilize multifactor authentication (MFA). Depending on the level of security used, the user might need to prove his or her identity through the use of security tokens, challenge questions, or being in possession of a certificate from a third-party certificate authority that attests to their identity. Overview The American National Institute of Standards and Technology (NIST) has developed a generic electronic authentication model that provides a basic framework on how the authentication process is accomplished regardless of jurisdiction or geographic region. According to this model, the enrollment process begins with an individual applying to a Credential Service Provider (CSP). The CSP will need to prove the applicant's identity before proceeding with the transaction. Once the applicant's identity has been confirmed by the CSP, he or she receives the status of "subscriber", is given an authenticator, such as a token and a credential, which may be in the form of a username. The CSP is responsible for managing the credential along with the subscriber's enrollment data for the life of the credential. The subscriber will be tasked with maintaining the authenticators. An example of this is when a us
https://en.wikipedia.org/wiki/Innexin
Innexins are transmembrane proteins that form gap junctions in invertebrates. Gap junctions are composed of membrane proteins that form a channel permeable to ions and small molecules connecting the cytoplasm of adjacent cells. Although gap junctions provide similar functions in all multicellular organisms, it was not known what proteins invertebrates used for this purpose until the late 1990s. While the connexin family of gap junction proteins was well-characterized in vertebrates, no homologues were found in non-chordates. Innexins or related proteins are widespread among Eumetazoa, with the exception of echinoderms. Discovery Gap junction proteins with no sequence homology to connexins were initially identified in fruit flies. It was suggested that these proteins are specific invertebrate gap junctions, and they were thus named "innexins" (invertebrate analog of connexins). They were later identified in diverse invertebrates. Invertebrate genomes may contain more than a dozen innexin genes. Once the human genome was sequenced, innexin homologues were identified in humans and then in other vertebrates, indicating their ubiquitous distribution in the animal kingdom. These homologues were called "pannexins" (from the Greek pan - all, throughout, and Latin nexus - connection, bond). However, increasing evidence suggests that pannexins do not form gap junctions unless overexpressed in tissue and thus, differ functionally from innexins. Structure Innexins have four transmembrane segments (TMSs) and, like the vertebrate connexin gap junction protein, innexin subunits together form a channel (an "innexon") in the plasma membrane of the cell. Two innexons in apposed plasma membranes can form a gap junction. Innexons are made from eight subunits, instead of the six subunits of connexons. Structurally, innexins and connexins are very similar, consisting of 4 transmembrane domains, 2 extracellular and 1 intracellular loop, along with intracellular N- and C-terminal tail
https://en.wikipedia.org/wiki/Digital%20edition
A digital edition is an online magazine or online newspaper delivered in electronic form which is formatted identically to the print version. Digital editions are often called digital facsimiles to underline the likeness to the print version. Digital editions have the benefit of reduced cost to the publisher and reader by avoiding the time and the expense to print and deliver paper edition. This format is considered more environmentally friendly due to the reduction of paper and energy use. These editions also often feature interactive elements such as hyperlinks both within the publication itself and to other internet resources, search option and bookmarking, and can also incorporate multimedia such as video or animation to enhance articles themselves or for advertisement purposes. Some delivery methods also include animation and sound effects that replicate turning of the page to further enhance the experience of their print counterparts. Magazine publishers have traditionally relied on two revenue sources: selling ads and selling magazines. Additionally some publishers are using other electronic publication methods such as RSS to reach out to readers and inform them when new digital editions are available. Current technologies are generally either reader-based, requiring a download of an application and subsequent download of each edition, or browser-based, often using Macromedia Flash , requiring no application download (such as Adobe Acrobat). Some application-based readers allow users to access editions while not connected to internet. Dedicated hardware such as the Amazon Kindle and the iPad is also available for reading digital editions of select books, popular national magazines such as Time, The Atlantic, and Forbes and popular national newspapers such as the New York Times, Wall Street Journal, and Washington Post. Archives of print newspapers, in some cases dating hundreds of years back, are being digitized and made available online. Google is index
https://en.wikipedia.org/wiki/Ryanodine
Ryanodine is a poisonous diterpenoid found in the South American plant Ryania speciosa (Salicaceae). It was originally used as an insecticide. The compound has extremely high affinity to the open-form ryanodine receptor, a group of calcium channels found in skeletal muscle, smooth muscle, and heart muscle cells. It binds with such high affinity to the receptor that it was used as a label for the first purification of that class of ion channels and gave its name to it. At nanomolar concentrations, ryanodine locks the receptor in a half-open state, whereas it fully closes them at micromolar concentration. The effect of the nanomolar-level binding is that ryanodine causes release of calcium from calcium stores as the sarcoplasmic reticulum in the cytoplasm, leading to massive muscle contractions. The effect of micromolar-level binding is paralysis. This is true for both mammals and insects. See also Ryanoid, a class of insecticides with the same mechanism of action as ryanodine Ryanodine receptor
https://en.wikipedia.org/wiki/Powertype%20%28UML%29
In the Unified Modeling Language 1.x, powertype is a keyword for a specific UML stereotype, and applies to a class or dependency. Powertype shows a classifier whose instances (objects) are children of the given parent. In UML 2.x, a powertype is a metaclass whose instances are subclasses of a given class. The stereotype has been removed and the powertype is now indicated by typing the generalization set. In the UML Superstructure 2.4.1 Specification Document the following definition is given:
https://en.wikipedia.org/wiki/Babcock-Hart%20Award
The Babcock-Hart Award has been awarded since 1948 by the Institute of Food Technologists. It is given for significant contributions in food technology that resulted in public health through some aspects of nutrition. It was first named the Stephan M. Babcock Award after the agricultural chemist Stephen M. Babcock of the University of Wisconsin–Madison for his "single-grain experiment" of 1907–1911, but renamed the Babcock-Hart Award following the death of Babcock's colleague Edwin B. Hart in 1953. Award winners receive a plaque from the International Life Sciences Institute-North America, headquartered in Washington, DC and a USD 3000 honorarium. Winners
https://en.wikipedia.org/wiki/International%20Life%20Sciences%20Institute
The International Life Sciences Institute (ILSI) is a global nonprofit 501(c)(3) organization headquartered in Washington, DC, United States that publishes peer-reviewed studies on nutrition and food safety. It was founded in 1978 by Alex Malaspina, a former Coca-Cola executive (who was affiliated with ILSI until 2001), and it is partially financed by its 300+ members, which includes food and chemical corporations such as BASF, McDonald’s, Syngenta and Pepsi. In 2020, the organization's revenue was US$10.1 million. In 2021 ILSI North America rebranded as the Institute for the Advancement of Food and Nutrition Sciences (IAFNS). Structure Until 2018, ILSI operated as a member organization, whose members were exclusively food and beverage, agricultural, chemical, and pharmaceutical companies. In November 2018, ILSI restructured as a global federation of entities, consisting of non-profit organizations with public and private sector members. Based on its 2020 annual report, 70.8% of its revenue comes from membership dues and committee assessments; 19.2% from grants and contributions; and the remaining from publications and conference registration. According to ILSI’s bylaws, at least 51% of its Board of Trustees must come from the public sector (i.e. academic, government and non-governmental organization representatives). The remainder of the Board is elected from its member companies. According to ILSI, the organization receives in-kind support of time and expertise from volunteer academic, government, and non-governmental scientists. History ILSI was formed in 1978 by Alex Malaspina, who was concurrently a senior vice president at Coca-Cola from 1969 until leaving both Coca-Cola and the institute in 2001. In 1986, ILSI employee Michael Gough published a book about Dioxin and Agent Orange entitled Dioxin, Agent Orange: The Facts. The Los Angeles Times said that The Facts had "minimal scientific merit" as the book is about "toxicology, teratology, carcinogenesis,
https://en.wikipedia.org/wiki/Partition%20type
The partition type (or partition ID) in a partition's entry in the partition table inside a master boot record (MBR) is a byte value intended to specify the file system the partition contains or to flag special access methods used to access these partitions (e.g. special CHS mappings, LBA access, logical mapped geometries, special driver access, hidden partitions, secured or encrypted file systems, etc.). Overview Lists of assigned partition types to be used in the partition table in the MBR were originally maintained by IBM and Microsoft internally. When the market of PC operating systems and disk tools grew and liberated, other vendors had a need to assign special partition types to their products as well. As Microsoft neither documented all partition types already assigned by them nor wanted to maintain foreign assignments, third parties started to simply assign partition types on their own behalf in a mostly uncoordinated trial-and-error manner. This led to various conflicting assignments sometimes causing severe compatibility problems between certain products. Several industry experts including Hale Landis, Ralf D. Brown, Matthias R. Paul, and Andries E. Brouwer in the 1990s started to research partition types and published (and later synchronized) partition type lists in order to help document the industry de facto standard and thereby reduce the risk of further conflicts. Some of them also actively helped to maintain software dealing with partitions to work with the updated lists, indicated conflicts, devised additional detection methods and work-arounds for vendors, or engaged in coordinating new non-conflictive partition type assignments as well. It is up to an operating system's boot loader or kernel how to interpret the value. So the table specifies which operating systems or disk-related products introduced an ID and what file system or special partition type they mapped it to. Partitions with partition types unknown to the software should be treate
https://en.wikipedia.org/wiki/Solid-state%20drive
Solid-state drive (SSD) is a solid-state storage device that uses integrated circuit assemblies to store data persistently, typically using flash memory, and functions as secondary storage in the hierarchy of computer storage. It is also sometimes called a semiconductor storage device, a solid-state device, or a solid-state disk, even though SSDs lack the physical spinning disks and movable read–write heads used in hard disk drives (HDDs) and floppy disks. SSD also has rich internal parallelism for data processing. In comparison to hard disk drives and similar electromechanical media which use moving parts, SSDs are typically more resistant to physical shock, run silently, and have higher input/output rates and lower latency. SSDs store data in semiconductor cells. cells can contain between 1 and 4 bits of data. SSD storage devices vary in their properties according to the number of bits stored in each cell, with single-bit cells ("Single Level Cells" or "SLC") being generally the most reliable, durable, fast, and expensive type, compared with 2- and 3-bit cells ("Multi-Level Cells/MLC" and "Triple-Level Cells/TLC"), and finally, quad-bit cells ("QLC") being used for consumer devices that do not require such extreme properties and are the cheapest per gigabyte (GB) of the four. In addition, 3D XPoint memory (sold by Intel under the Optane brand) stores data by changing the electrical resistance of cells instead of storing electrical charges in cells, and SSDs made from RAM can be used for high speed, when data persistence after power loss is not required, or may use battery power to retain data when its usual power source is unavailable. Hybrid drives or solid-state hybrid drives (SSHDs), such as Intel's Hystor and Apple's Fusion Drive, combine features of SSDs and HDDs in the same unit using both flash memory and spinning magnetic disks in order to improve the performance of frequently-accessed data. Bcache achieves a similar effect purely in software, using comb
https://en.wikipedia.org/wiki/Vacuum%20packing
Vacuum packing is a method of packaging that removes air from the package prior to sealing. This method involves placing items in a plastic film package, removing air from inside and sealing the package. Shrink film is sometimes used to have a tight fit to the contents. The intent of vacuum packing is usually to remove oxygen from the container to extend the shelf life of foods and, with flexible package forms, to reduce the volume of the contents and package. Vacuum packing reduces atmospheric oxygen, limiting the growth of aerobic bacteria or fungi, and preventing the evaporation of volatile components. It is also commonly used to store dry foods over a long period of time, such as cereals, nuts, cured meats, cheese, smoked fish, coffee, and potato chips (crisps). On a more short-term basis, vacuum packing can also be used to store fresh foods, such as vegetables, meats, and liquids, because it inhibits bacterial growth. Vacuum packing greatly reduces the bulk of non-food items. For example, clothing and bedding can be stored in bags evacuated with a domestic vacuum cleaner or a dedicated vacuum sealer. This technique is sometimes used to compact household waste, for example where a charge is made for each full bag collected. Vacuum packaging products, using plastic bags, canisters, bottles, or mason jars, are available for home use. For delicate food items that might be crushed by the vacuum packing process (such as potato chips), an alternative is to replace the interior gas with nitrogen. This has the same effect of inhibiting deterioration due to the removal of oxygen. Type of Vacuum Sealers or Vacuum Packaging Machines Edge, Suction, and External Vacuum Sealers External vacuum sealers involve a bag being attached to the vacuum-sealing machine externally. The machine will remove the air and seal the bag, which is all done outside the machine. A heat sealer is often used to seal the pack. Typically these units use a dry piston vacuum pump which is often
https://en.wikipedia.org/wiki/FlashCopy
FlashCopy is an IBM feature supported on various IBM storage devices that made it possible to create, nearly instantaneously, point-in-time snapshot copies of entire logical volumes or data sets. The Hitachi Data Systems implementation providing similar function was branded as ShadowImage. Using either implementation, the copies are immediately available for both read and write access. Implementations Version 1 The first implementation of FlashCopy, Version 1 allowed entire volumes to be instantaneously “copied” to another volume by using the facilities of the newer Enterprise Storage Subsystems (ESS). Version 1 of FlashCopy had limitations however. Although the copy or “flash” of a volume occurred instantaneously, the FlashCopy commands were issued sequentially and the ESS required a brief moment to establish the new pointers. Because of this minute processing delay, the data residing on two volumes that were FlashCopied are not exactly time consistent. Version 2 FlashCopy Version 2 introduced the ability to flash individual data sets and then added support for “consistency groups”. FlashCopy consistency groups can be used to help create a consistent point-in-time copy across multiple volumes, and even across multiple ESSs, thus managing the consistency of dependent writes. FlashCopy consistency groups are used in a single-site scenario in order to create a time-consistent copy of data that can then be backed-up and sent off site, or in a multi-site Global Mirror for ESS implementation to force time consistency at the remote site. The implementation of consistency groups is not limited to FlashCopy. Global Mirror for IBM System z series (formerly known as XRC or eXtended Remote Copy) also creates consistency groups to asynchronously mirror disk data from one site to another over any distance .
https://en.wikipedia.org/wiki/VoxForge
VoxForge is a free speech corpus and acoustic model repository for open source speech recognition engines. VoxForge was set up to collect transcribed speech to create a free GPL speech corpus in order to be uses with open source speech recognition engines. The speech audio files will be 'compiled' into acoustic models for use with open source speech recognition engines such as Julius, ISIP, and Sphinx and HTK (note: HTK has distribution restrictions). VoxForge has used LibriVox as a source of audio data since 2007. See also Speech recognition in Linux List of speech recognition software
https://en.wikipedia.org/wiki/QPNC-PAGE
QPNC-PAGE, or Quantitative Preparative Native Continuous PolyAcrylamide Gel Electrophoresis, is a bioanalytical, one-dimensional, high-resolution and high-precision technique applied in biochemistry and bioinorganic chemistry to separate proteins quantitatively by isoelectric point and by continuous elution from a gel column. This variant of native gel electrophoresis and subset of preparative polyacrylamide gel electrophoresis is used by biophysicists to isolate macromolecules in solution, for example, active or native metalloproteins in biological samples or properly and improperly folded metal cofactor-containing proteins or protein isoforms in complex protein mixtures. The special feature of the QPNC-PAGE technique is based on a standardized polymerization time for acrylamide gel solutions to optimize gel properties, in particular gel stability, during an electrophoretic run. Physiological concentrations of metal ion cofactors are fractionated in quantitative amounts for further analysis. Introduction Proteins perform several functions in living organisms, including catalytic reactions and transport of molecules or ions within the cells, the organs or the whole body. The understanding of the processes in human organisms, which are mainly driven by biochemical reactions and protein-protein interactions, depends to a great extent on the ability to isolate active proteins in biological samples for more detailed examination of chemical structure and physiological function. This essential information can imply an important indication of a patient's state of health. As about 30-40% of all known proteins contain one or more metal ion cofactors (e.g., ceruloplasmin, ferritin, amyloid-beta precursor protein, matrix metalloproteinase, or metallochaperones), especially native and denatured metalloproteins have to be isolated, identified and quantified after liquid biopsy. Many of these cofactors (e.g., iron, copper, or zinc) play a key role in vital enzymatic catalyt
https://en.wikipedia.org/wiki/Iproute2
iproute2 is a collection of userspace utilities for controlling and monitoring various aspects of networking in the Linux kernel, including routing, network interfaces, tunnels, traffic control, and network-related device drivers. iproute2 is an open-source project released under the terms of version 2 of the GNU General Public License. Its development is closely tied to the development of networking components of the Linux kernel. , iproute2 is maintained by Stephen Hemminger and David Ahern. The original author, Alexey Kuznetsov, was responsible for the quality of service (QoS) implementation in the Linux kernel. iproute2 collection contains the following command-line utilities: arpd, bridge, ctstat, dcb, devlink, ip, lnstat, nstat, rdma, routef, routel, rtacct, rtmon, rtstat, ss, tc, tipc and vdpa. tc is used for traffic control. iproute2 utilities communicate with the Linux kernel using the netlink protocol. Some of the iproute2 utilities are often recommended over now-obsolete net-tools utilities that provide the same functionality. Below is a table of obsolete utilities and their iproute2 replacements. See also BusyBox ethtool TIPC
https://en.wikipedia.org/wiki/Peg%20%28unit%29
A peg is a unit of volume, typically used to measure amounts of liquor in the Indian subcontinent. Informally, a peg is an undefined measure of any alcoholic drink poured in a glass. The terms "large (badda) peg" and "small (chota) peg" are equal to 60ml and 30ml, respectively, with "peg" alone simply referring to a 60ml peg. The "chota peg" was often used by the 1940s mythical British Indian Army figure "Colonel Chinstrap" in the UK Radio Serial "ITMA" ("It's That Man Again") who was frequently having Chota Pegs of whisky (and sounded like it in his deliberately slurred speech). In India, liquor's alcohol content is typically 42.8% ABV. A 30 ml of liquor usually contains 12.84ml of pure alcohol, See also Gorkhali Peg Patiala peg
https://en.wikipedia.org/wiki/Structured%20entity%20relationship%20model
The SERM (structured entity relationship model) is an amplification of the ERM which is commonly used for data modeling. It was first proposed from Prof. Dr. Elmar J. Sinz in 1988. The SERM is commonly used in the SAP-world for the data modeling. Aims structuring of large schemes visualization of existence dependency avoidance of inconsistencies avoidance of unnecessary relationshiptypes SERM symbols SERM example Customer and article are independent entities Every order is referred to one customer. Orders without customers are illegal (order is an ER type). Customers without any orders are legal because they are independent Entities. To every order there is belonging at least one order item. Every order item is related to exactly one order. Every invoice is referred to one customer, as well. Invoices without customers are illegal. Customers without any invoice are legal. To every invoice there is belonging at least one invoice line item. Every invoice line item is related to exactly order item. An order item could be calculated or not. SERM is already in the third normal form
https://en.wikipedia.org/wiki/Superpattern
In the mathematical study of permutations and permutation patterns, a superpattern or universal permutation is a permutation that contains all of the patterns of a given length. More specifically, a k-superpattern contains all possible patterns of length k. Definitions and example If π is a permutation of length n, represented as a sequence of the numbers from 1 to n in some order, and s = s1, s2, ..., sk is a subsequence of π of length k, then s corresponds to a unique pattern, a permutation of length k whose elements are in the same order as s. That is, for each pair i and j of indexes, the ith element of the pattern for s should be less than the jthe element if and only if the ith element of s is less than the jth element. Equivalently, the pattern is order-isomorphic to the subsequence. For instance, if π is the permutation 25314, then it has ten subsequences of length three, forming the following patterns: A permutation π is called a k-superpattern if its patterns of length k include all of the length-k permutations. For instance, the length-3 patterns of 25314 include all six of the length-3 permutations, so 25314 is a 3-superpattern. No 3-superpattern can be shorter, because any two subsequences that form the two patterns 123 and 321 can only intersect in a single position, so five symbols are required just to cover these two patterns. Length bounds introduced the problem of determining the length of the shortest possible k-superpattern. He observed that there exists a superpattern of length k2 (given by the lexicographic ordering on the coordinate vectors of points in a square grid) and also observed that, for a superpattern of length n, it must be the case that it has at least as many subsequences as there are patterns. That is, it must be true that , from which it follows by Stirling's approximation that n ≥ k2/e2, where e ≈ 2.71828 is Euler's number. This lower bound was later improved very slightly by , who increased it to 1.000076k2/e2, disproving
https://en.wikipedia.org/wiki/Interactive%20Mathematics%20Program
The Interactive Mathematics Program (IMP) is a four-year, problem-based mathematics curriculum for high schools. It was one of several curricula funded by the National Science Foundation and designed around the 1989 National Council of Teachers of Mathematics (NCTM) standards. The IMP books were authored by Dan Fendel and Diane Resek, professors of mathematics at San Francisco State University, and by Lynne Alper and Sherry Fraser. IMP was published by Key Curriculum Press in 1997 and sold in 2012 to It's About Time. Curriculum Designed in response to national reports pointing to the need for a major overhaul in mathematics education, the IMP curriculum is markedly different in structure, content, and pedagogy from courses more typically found in the high school sequence. Each book of the curriculum is divided into five- to eight-week units, each having a central problem or theme. This larger problem is intended to serve as motivation for students to develop the underlying skills and concepts needed to solve it, through solving a variety of smaller related problems. There is an emphasis on asking students to work together in collaborative groups. It is hoped that communication skills will be developed; exercises aimed at this goal are embedded throughout the curriculum, through the use of group and whole class discussions, the use of writing to present and clarify mathematical solutions; in some IEP classes, formal oral presentations are required. The IMP curriculum expects students to make nearly daily use of a scientific graphing calculator. Controversy Nearly every one of these distinctive characteristics has generated controversy and placed the IMP curriculum right in the middle of the “math wars,” the conflict between those that favor more traditional curricula in mathematics education and the supporters of the reform curricula that were largely an outgrowth of the 1989 NCTM standards. IMP is among the reform curricula that have been heavily critici
https://en.wikipedia.org/wiki/Dunkaroos
Dunkaroos are a snack food from Betty Crocker, first launched in 1990. It consists of a snack-sized package containing cookies and frosting; as the name implies, the cookies are meant to be dunked into the frosting before eating. Individual snack packages contain about ten small cookies and one cubic inch of frosting. The cookies are made in a variety of shapes, including a circle with an uppercase "D" in the center (the only shape featured in the 2020 version), feet, the mascot in different poses, and a hot air balloon. Marketing The Dunkaroos mascot is a cartoon kangaroo, explaining the product's name which is a portmanteau of dunk and kangaroos. The original mascot was Sydney, a caricature of modern Australian culture, who wore a hat, vest, and tie and spoke with an Australian accent, and was voiced by John Cameron Mitchell. At the height of their popularity in 1996, a contest known as "Dunk-a-roos Kangaroo Kanga-Who Search" was held, resulting in the new mascot: Duncan, named the dunkin' daredevil. History The product was discontinued in the United States in 2012 but continued to be sold in Canada. In 2016, General Mills announced a campaign called "Smugglaroos", which encouraged Canadians travelling to the United States to bring the snack to Americans who wanted it. Dunkaroos continued being sold in Canada until January 2018, with no comment by General Mills. In December 2019 Dunkaroos were brought back unofficially by Nestlé with a chocolate-hazelnut flavour. The biscuits are shaped like a kangaroo biscuit. This is only available in Australia as Nestlé does not have the right to sell Dunkaroos worldwide. On February 3, 2020, a BuzzFeed article was published claiming that General Mills sent them exclusive info regarding a return of Dunkaroos. The official Twitter account for Dunkaroos claimed that they were scheduled to be re-released during the summer of 2020. It also used to link to the BuzzFeed article in the bio, but this was later changed to their off
https://en.wikipedia.org/wiki/Remote%20manipulator
A remote manipulator, also known as a telefactor, telemanipulator, or waldo (after the 1942 short story "Waldo" by Robert A. Heinlein which features a man who invents and uses such devices), is a device which, through electronic, hydraulic, or mechanical linkages, allows a hand-like mechanism to be controlled by a human operator. The purpose of such a device is usually to move or manipulate hazardous materials for reasons of safety, similar to the operation and play of a claw crane game. History In 1945, the company Central Research Laboratories was given the contract to develop a remote manipulator for the Argonne National Laboratory. The intent was to replace devices which manipulated highly radioactive materials from above a sealed chamber or hot cell, with a mechanism which operated through the side wall of the chamber, allowing a researcher to stand normally while working. The result was the Master-Slave Manipulator Mk. 8, or MSM-8, which became the iconic remote manipulator seen in newsreels and movies, such as the The Andromeda Strain or THX 1138. Robert A. Heinlein claimed a much earlier origin for remote manipulators. He wrote that he got the idea for "waldos" after reading a 1918 article in Popular Mechanics about "a poor fellow afflicted with myasthenia gravis ... [who] devised complicated lever arrangements to enable him to use what little strength he had." An article in Science Robotics on robots, science fiction, and nuclear accidents discusses how the science fiction waldos are now a major type of real-world robots used in the nuclear industry. See also Glovebox Dextre Doctor Octopus Teleoperation Telerobotics Master/slave (technology) Avatar (computing) Pantograph
https://en.wikipedia.org/wiki/Transgressive%20segregation
In genetics, transgressive segregation is the formation of extreme phenotypes, or transgressive phenotypes, observed in segregated hybrid populations compared to phenotypes observed in the parental lines. The appearance of these transgressive (extreme) phenotypes can be either positive or negative in terms of fitness. If both parents' favorable alleles come together, it will result in a hybrid having a higher fitness than the two parents. The hybrid species will show more genetic variation and variation in gene expression than their parents. As a result, the hybrid species will have some traits that are transgressive (extreme) in nature. Transgressive segregation can allow a hybrid species to populate different environments/niches in which the parent species do not reside, or compete in the existing environment with the parental species. Causes Genetic There are many causes for transgressive segregation in hybrids. One cause can be due to recombination of additive alleles. Recombination results in new pairs of alleles at two or more loci. These different pairs of alleles can give rise to new phenotypes if gene expression has been changed at these loci. Another cause can be elevated mutation rate. When mutation rates are high, it is more probable that a mutation will occur and cause an extreme phenotypic change. Reduced developmental stability is another cause for transgressive segregation. Developmental stability refers to the capability of a genotype to go through a constant development of a phenotype in a certain environmental setting. If there is a disturbance due to genetic or environmental factors, the genotype will be more sensitive to phenotypic changes. Another cause arises from the interaction between two alleles of two different genes, also known as the epistatic effect. Epistasis is the event when one allele at a locus prevents an allele at another locus to express its product as if it is masking its effect. Therefore, epistasis can be related to gene
https://en.wikipedia.org/wiki/Oblique%20correction
In particle physics, an oblique correction refers to a particular type of radiative correction to the electroweak sector of the Standard Model. Oblique corrections are defined in four-fermion scattering processes, ( +  →  +  ) at the CERN Large Electron–Positron Collider. There are three classes of radiative corrections to these processes: vacuum polarization corrections, vertex corrections, and box corrections. The vacuum polarization corrections are referred to as oblique corrections, since they only affect the mixing and propagation of the gauge bosons and they do not depend on which type of fermions appear in the initial or final states. (The vertex and box corrections, which depend on the identity of the initial and final state fermions, are called nonoblique corrections.) Any new particles charged under the electroweak gauge groups can contribute to oblique corrections. Therefore, the oblique corrections can be used to constrain possible new physics beyond the Standard Model. To affect the nonoblique corrections, on the other hand, the new particles must couple directly to the external fermions. The oblique corrections are usually parameterized in terms of the Peskin–Takeuchi parameters S, T, and U. See also Initial and final state radiation
https://en.wikipedia.org/wiki/Nonoblique%20correction
In four-fermion scattering processes of particle physics, a nonoblique correction, also called a direct correction, refers to a radiative correction of type  +  →  +  in the electroweak sector of the Standard Model. These corrections are being studied at the CERN LEP collider. Together with the oblique corrections, nonoblique corrections can be used to constrain models of physics beyond the Standard Model. Classes There are three classes of radiative corrections to these processes: vacuum polarization corrections, vertex corrections, and box corrections. The vertex and box corrections, which depend on the identity of the initial and final state fermions, are referred to as the non-oblique corrections. The vacuum polarization corrections are referred to as oblique corrections, since they only affect the mixing and propagation of the gauge bosons and they do not depend on which type of fermions appear in the initial or final states. Examples An example of a vertex correction is the nonuniversality (flavor dependence) of the couplings of the quarks and leptons to the charged and neutral weak currents. Another example is the anomalous magnetic dipole moment. In order to affect the nonoblique corrections, particles must couple directly to the external fermions. Such couplings are expected to be suppressed in most cases, with one exception being the vertex. See also Initial and final state radiation