source
stringlengths
31
227
text
stringlengths
9
2k
https://en.wikipedia.org/wiki/Pterocarpus%20cambodianus
Pterocarpus cambodianus is a taxonomic synonym of Pterocarpus macrocarpus that may refer to: Pterocarpus cambodianus Pterocarpus cambodianus
https://en.wikipedia.org/wiki/New%20Journal%20of%20Physics
New Journal of Physics is an online-only, open-access, peer-reviewed scientific journal covering research in all aspects of physics, as well as interdisciplinary topics where physics forms the central theme. The journal was established in 1998 and is a joint publication of the Institute of Physics and the Deutsche Physikalische Gesellschaft. It is published by IOP Publishing. The editor-in-chief is Andreas Buchleitner (Albert Ludwigs University). New Journal of Physics is part of the SCOAP3 initiative. In April 2023, on occasion of the World Quantum Day, IOP Publishing has launched a special collection of its most important articles published in the field of quantum research. The articles will be extracted from Materials for Quantum Technology, Quantum Science and Technology, New Journal of Physics and Reports on Progress in Physics. Abstracting and indexing The journal is indexed and abstracted in:
https://en.wikipedia.org/wiki/Arc%20routing
Arc routing problems (ARP) are a category of general routing problems (GRP), which also includes node routing problems (NRP). The objective in ARPs and NRPs is to traverse the edges and nodes of a graph, respectively. The objective of arc routing problems involves minimizing the total distance and time, which often involves minimizing deadheading time, the time it takes to reach a destination. Arc routing problems can be applied to garbage collection, school bus route planning, package and newspaper delivery, deicing and snow removal with winter service vehicles that sprinkle salt on the road, mail delivery, network maintenance, street sweeping, police and security guard patrolling, and snow ploughing. Arc routings problems are NP hard, as opposed to route inspection problems that can be solved in polynomial-time. For a real-world example of arc routing problem solving, Cristina R. Delgado Serna & Joaquín Pacheco Bonrostro applied approximation algorithms to find the best school bus routes in the Spanish province of Burgos secondary school system. The researchers minimized the number of routes that took longer than 60 minutes to traverse first. They also minimized the duration of the longest route with a fixed maximum number of vehicles. There are generalizations of arc routing problems that introduce multiple mailmen, for example the k Chinese Postman Problem (KCPP). Background The efficient scheduling and routing of vehicles can save industry and government millions of dollars every year. Arc routing problems have applications in school bus planning, garbage and waste and refuse collection in cities, mail and package delivery by mailmen and postal services, winter gritting and laying down salt to keep roads safe in the winter, snow plowing and removal, meter reading including remote radio frequency identification meter reading technology, street maintenance and sweeping, police patrol car route planning, and more. Basis The basic routing problem is: given a
https://en.wikipedia.org/wiki/Almost%20integer
In recreational mathematics, an almost integer (or near-integer) is any number that is not an integer but is very close to one. Almost integers are considered interesting when they arise in some context in which they are unexpected. Almost integers relating to the golden ratio and Fibonacci numbers Well-known examples of almost integers are high powers of the golden ratio , for example: The fact that these powers approach integers is non-coincidental, because the golden ratio is a Pisot–Vijayaraghavan number. The ratios of Fibonacci or Lucas numbers can also make almost integers, for instance: The above examples can be generalized by the following sequences, which generate near-integers approaching Lucas numbers with increasing precision: As n increases, the number of consecutive nines or zeros beginning at the tenths place of a(n) approaches infinity. Almost integers relating to e and Other occurrences of non-coincidental near-integers involve the three largest Heegner numbers: where the non-coincidence can be better appreciated when expressed in the common simple form: where and the reason for the squares is due to certain Eisenstein series. The constant is sometimes referred to as Ramanujan's constant. Almost integers that involve the mathematical constants and e have often puzzled mathematicians. An example is: To date, no explanation has been given for why Gelfond's constant () is nearly identical to , which is therefore considered a mathematical coincidence. See also Schizophrenic number
https://en.wikipedia.org/wiki/Janus%20Experiments
The Janus Experiments investigated the effects of exposure to neutron radiation and gamma radiation on mice and dogs. They consisted of ten large scale experiments conducted at Argonne National Laboratory from 1972 to 1989. To explore various relationships, the studies varied radiation type, dose rates, total dose and fractionation. The work formed the basis of dozens of publications in the medical literature. The original studies were funded by the United States Department of Energy. Later grants from NASA and additional funding from the Department of Energy enabled researchers at Northwestern University to make the data public through mouse and dog portals that permit radiation researchers to search for and request specific tissues from the studies' archives. These resources continue to be used in studies of radio-sensitivity, for example, at the laboratory of Gayle Woloschak at Northwestern University. Studies of the survival and causes of death of the control groups of mice and dogs, which were not exposed to radiation, were the basis of the development by S. Jay Olshansky and Bruce A. Carnes of their biodemographic theory of intrinsic mortality.
https://en.wikipedia.org/wiki/Internet%20Citizen%27s%20Band
Internet Citizen's Band (better known as ICB) is an early Internet chat program and its associated protocol. It was released in 1989. ICB is typically served on port 7326. History The first version of ICB was a program called "Forumnet" or "fn", written by University of Kentucky IT staffer Sean Carrick Casey. It was widely used at the University of Kentucky, Georgia Tech, UC Davis, MIT, University of New Mexico, Stanford University, Mills College, UC Santa Cruz, and UC Berkeley. Fn, based on a MUD software program by Casey, established the protocol and clients. Fn was used as a realtime communications channel after the 1989 Loma Prieta earthquake - Internet access from hard-hit Santa Cruz returned to service before reliable phone service did. In March 1991 the University of Kentucky changed policy and shut down the fn server. Within 2 months a new server had been created from the client software by another fn user, John Atwood Devries, and was put online now renamed ICB. This new server code, unrelated to the original server except by the common client software source, was then used as the basis of many ICB servers to follow. From 1995 to 2000 the server code was heavily rewritten for stability and additional features by Jon Luini and Michel Hoche-Mong and remains available at the ICB.net web site. ICB is still in operation with a dedicated user base. A variety of clients exist for all major operating systems. Features ICB features many standard chat program functions, including channels, private messages, and nickname registration. Most of the common clients support TCL scripting of commands and functions. Some clients (principally icbm) support scripting in Perl instead. Limitations ICB has never supported multi-server shared groups, so the number of simultaneous users has always been somewhat limited in comparison to more popular chat programs. ICB does not support transferring files or multimedia via the chat program. However, the very restrict
https://en.wikipedia.org/wiki/Cladistics%20%28journal%29
Cladistics is a bimonthly peer-reviewed scientific journal which has published research in cladistics since 1985. It is published by Wiley-Blackwell on behalf of the Willi Hennig Society. Cladistics publishes papers relevant to evolution, systematics, and integrative biology. Papers of both a conceptual or philosophical nature, discussions of methodology, empirical studies on taxonomic groups from animals to bacteria, and applications of systematics in disciplines such as genomics, paleontology and biomedical epidemiology are accepted. Five types of paper appear in the journal: reviews, regular papers, forum papers, letters to the editor, and book reviews. According to the Journal Citation Reports, the journal has a 2020 impact factor of 5.254, ranking it 10th out of 50 journals in the category "Evolutionary Biology". Its editor-in-chief is Rudolf Meier, who replaced Dennis Stevenson in 2019. External links Willi Hennig Society Evolutionary biology journals Phylogenetics Bimonthly journals English-language journals Wiley-Blackwell academic journals Academic journals established in 1985 Academic journals associated with learned and professional societies
https://en.wikipedia.org/wiki/The%20God%20Particle%20%28book%29
The God Particle: If the Universe Is the Answer, What Is the Question? is a 1993 popular science book by Nobel Prize-winning physicist Leon M. Lederman and science writer Dick Teresi. The book provides a brief history of particle physics, starting with the pre-Socratic Greek philosopher Democritus, and continuing through Isaac Newton, Roger J. Boscovich, Michael Faraday, and Ernest Rutherford and quantum physics in the 20th century. Lederman explains in the book why he gave the Higgs boson the nickname "The God Particle": In 2013, subsequent to the discovery of the Higgs boson, Lederman co-authored, with theoretical physicist Christopher T. Hill, a sequel: Beyond the God Particle which delves into the future of particle physics in the post-Higgs boson era. This book is part of a trilogy, with companions, Symmetry and the Beautiful Universe and Quantum Physics for Poets (see bibliography below). Historical context Fermilab director and subsequent Nobel physics prize winner Leon Lederman was a very prominent early supporter – some sources say the architect or proposer – of the Superconducting Super Collider project, which was endorsed around 1983, and was a major proponent and advocate throughout its lifetime. Lederman wrote his 1993 popular science book – which sought to promote awareness of the significance of such a project – in the context of the project's last years and the changing political climate of the 1990s. The increasingly moribund project was finally shelved that same year after some $2 billion of expenditure. The proximate causes of the closure were the rising US budget deficit, rising projected costs of the project, and the cessation of the Cold War, which reduced the perceived political pressure within the United States to undertake and complete high-profile science megaprojects. List of chapters Chapter 1: The Invisible Soccer Ball: This chapter uses a metaphor of a soccer game with an invisible ball to depict the process by which the existe
https://en.wikipedia.org/wiki/Assembly%20rules
Community assembly rules are a set of controversial rules in ecology, first proposed by Jared Diamond. Rules The rules were developed after more than a decade of research into the avian assemblages on islands near New Guinea. The rules assert that competition is responsible for determining the patterns of assemblage composition. Rule 1: Forbidden species combinations The first rule is "forbidden species combinations". Diamond's hypothesis was that competition, not random immigration, was the main force structuring the species composition of islands. So for example, the Bismarck black myzomela (Myzomela pammelaena) excludes the black sunbird (Nectarinia sericea). The Bismarck black myzomela lives on 23 of the 41 surveyed islands in the Bismarck Archipelago, but not on any of the 14 islands inhabited by the black sunbird. The two birds are about the same size, and both use their curved bills to sip nectar; Diamond argued that competition affects their distribution. Rule 2: Reduced niche overlap Case tested the assembly rule that species occurring together on islands should have less niche overlap than random assemblages because they have undergone specialization. His study measured niche overlap of lizards on 37 islands near Baja California and compared niche overlap to the median niche overlap of computer generated random species assemblages. Case found that 30 of the 37 islands had lower niche overlap than the random assemblages and that some of the competition is due to interspecific competition. Testing and alternative theories Testing the assembly rules is a complex process that often uses computer simulations to compare experimental data with characteristics of random assemblages of species. The rules are generally regarded as hypotheses that need to be tested on an individual basis, not as accepted conclusions. This is the reason why Diamond's results sparked nearly two decades worth of controversy in the literature, from the late seventies through th
https://en.wikipedia.org/wiki/Brainbow
Brainbow is a process by which individual neurons in the brain can be distinguished from neighboring neurons using fluorescent proteins. By randomly expressing different ratios of red, green, and blue derivatives of green fluorescent protein in individual neurons, it is possible to flag each neuron with a distinctive color. This process has been a major contribution to the field of neural connectomics. The technique was originally developed in 2007 by a team led by Jeff W. Lichtman and Joshua R. Sanes, both at Harvard University. The original technique has recently been adapted for use with other model research organisms including the fruit fly (Drosophila melanogaster), zebrafish (Danio rerio), and Arabidopsis thaliana. While earlier labeling techniques allowed for the mapping of only a few neurons, this new method allows more than 100 differently mapped neurons to be simultaneously and differentially illuminated in this manner. This leads to its characteristic multicolored appearance on imaging, earning its name and winning awards in science photography competitions. History and development Brainbow was initially developed by Jeff W. Lichtman and Joshua R. Sanes at Washington University in St. Louis, though they have moved to Harvard University since then. The team constructed Brainbow using a two-step process: first, a specific genetic construct was generated that could be recombined in multiple arrangements to produce one of either three or four colors based on the particular fluorescent proteins (XFPs) being implemented. Next, multiple copies of the same transgenic construct were inserted into the genome of the target species, resulting in the random expression of different XFP ratios and subsequently causing different cells to exhibit a variety of colorful hues. Brainbow was originally created as an improvement over more traditional neuroimaging techniques, such as Golgi staining and dye injection, both of which presented severe limitations to researche
https://en.wikipedia.org/wiki/Field-sequential%20color%20system
A field-sequential color system (FSC) is a color television system in which the primary color information is transmitted in successive images and which relies on the human vision system to fuse the successive images into a color picture. One field-sequential system was developed by Peter Goldmark for CBS, which was its sole user in commercial broadcasting. It was first demonstrated to the press on September 4, 1940, and first shown to the general public on January 12, 1950. The Federal Communications Commission adopted it on October 11, 1950, as the standard for color television in the United States, but it was later withdrawn. History The use of sequential color systems for moving images predates the invention of fully electronic television. Described at the time as "additive" rather than sequential color systems, two-color Kinemacolor, in commercial use since 1906, and its predecessor three-color format, invented by Edward Raymond Turner and patented in 1899, were both sequential natural color systems for use with motion picture film. They utilized black-and-white film and rotating color filter wheels to record the amount of each color in the scene on alternating frames of the film, so that when the frames were projected by light of similar colors at a sufficiently rapid rate, those colors blended together in the color center of the viewer's occipital lobe, producing a wider range of hues. Due to litigation by William Friese-Greene, Kinemacolor ended up in the public domain in 1915, after which several derivative sequential color processes (such as Friese-Greene's Biocolour and the original Prisma Color) were developed. Some were brought to the point of being publicly shown, but during the 1920s they could not compete with rival bipack and other subtractive color processes, which were free of color flicker and did not require special projection equipment—the final multicolored images were right there on the film as transparent coloring matter. Operation The
https://en.wikipedia.org/wiki/Minicircle
Minicircles are small (~4kb) circular replicons. They occur naturally in some eukaryotic organelle genomes. In the mitochondria-derived kinetoplast of trypanosomes, minicircles encode guide RNAs for RNA editing. In Amphidinium, the chloroplast genome is made of minicircles that encode chloroplast proteins. In vitro experimentally-derived minicircles Minicircles are small (~4kb) circular plasmid derivatives that have been freed from all prokaryotic vector parts. They have been applied as transgene carriers for the genetic modification of mammalian cells, with the advantage that, since they contain no bacterial DNA sequences, they are less likely to be perceived as foreign and destroyed. (Typical transgene delivery methods involve plasmids, which contain foreign DNA.) The smaller size of minicircles also extends their cloning capacity and facilitates their delivery into cells. Their preparation usually follows a two-step procedure: production of a 'parental plasmid' (bacterial plasmid with eukaryotic inserts) in E. coli induction of a site-specific recombinase at the end of this process but still in bacteria. These steps are followed by the excision of prokaryotic vector parts via two recombinase-target sequences at both ends of the insert recovery of the resulting minicircle (vehicle for the highly efficient modification of the recipient cell) and the miniplasmid by capillary gel electrophoresis (CGE) The purified minicircle can be transferred into the recipient cell by transfection or lipofection and into a differentiated tissue by, for instance, jet injection. Conventional minicircles lack an origin of replication, so they do not replicate within the target cells and the encoded genes will disappear as the cell divides (which can be either an advantage or disadvantage depending on whether the application demands persistent or transient expression). A novel addition to the field are nonviral self-replicating minicircles, which owe this property to the pre
https://en.wikipedia.org/wiki/DNA%20analyzer
A DNA analyzer is a device used to determine characteristics of a person's DNA. For example, genetic fingerprinting can be conducted with a portable DNA analyzer.
https://en.wikipedia.org/wiki/CPT%20Corporation
CPT Corporation was founded in 1971 by Dean Scheff in Minneapolis, Minnesota, with co-founders James Wienhold and Richard Eichhorn. CPT first designed, manufactured, and marketed the CPT 4200, a dual-cassette-tape machine that controlled a modified IBM Selectric typewriter to support text editing and word processing. The CPT 4200 was followed in 1976 by the CPT VM (Visual Memory), a partial-page display-screen dual-cassette-tape unit, and shortly thereafter by the CPT 8000, a full-page display dual-diskette desktop microcomputer that drove stand-alone daisy wheel printers. Subsequent products included (1) variants on the 8000 series; (2) the CPT 6000 series, which had a lower capacity, smaller screen, and was less expensive; (3) the CPT 9000 series, which had a larger capacity and could run IBM personal computer software; (4) the CPT Phoenix series, which had a graphical capabilities; (5) CPT PT, a software-only reduced version that ran on IBM personal computers and clones; and (6) other related products. The CPT logooriginally three letters chosen to sound well togetherbegan to be taken as an acronym for "cassette powered typewriting," and subsequently for "computer processed text," and numerous other variants. Major competition was IBM, Wang, Lanier, Xerox, and other word processing vendors. CPT Corporation was fifth in size among Minnesota-based top high-tech companies, after 3M, Honeywell, Control Data, and Medtronic. Corporate revenues grew to approximately a quarter-billion dollars per year in the mid-1980s, then declined with the proliferation of personal computers. CPT ultimately ceased major manufacturing late in the 20th century. Selected products Cassette based The CPT 4200 was a dual-cassette-tape unit with a small built-in keyboard that controlled a modified IBM Selectric typewriter. Keystrokes entered on the typewriter appeared on the paper as they were recorded on the output cassette, which formed a magnetic replica of the characters printed on
https://en.wikipedia.org/wiki/Non-random%20two-liquid%20model
The non-random two-liquid model (abbreviated NRTL model) is an activity coefficient model that correlates the activity coefficients of a compound with its mole fractions in the liquid phase concerned. It is frequently applied in the field of chemical engineering to calculate phase equilibria. The concept of NRTL is based on the hypothesis of Wilson that the local concentration around a molecule is different from the bulk concentration. This difference is due to a difference between the interaction energy of the central molecule with the molecules of its own kind and that with the molecules of the other kind . The energy difference also introduces a non-randomness at the local molecular level. The NRTL model belongs to the so-called local-composition models. Other models of this type are the Wilson model, the UNIQUAC model, and the group contribution model UNIFAC. These local-composition models are not thermodynamically consistent for a one-fluid model for a real mixture due to the assumption that the local composition around molecule i is independent of the local composition around molecule j. This assumption is not true, as was shown by Flemr in 1976. However, they are consistent if a hypothetical two-liquid model is used. Derivation Like Wilson (1964), Renon & Prausnitz (1968) began with local composition theory, but instead of using the Flory–Huggins volumetric expression as Wilson did, they assumed local compositions followed with a new "non-randomness" parameter α. The excess Gibbs free energy was then determined to be . Unlike Wilson's equation, this can predict partially miscible mixtures. However, the cross term, like Wohl's expansion, is more suitable for than , and experimental data is not always sufficiently plentiful to yield three meaningful values, so later attempts to extend Wilson's equation to partial miscibility (or to extend Guggenheim's quasichemical theory for nonrandom mixtures to Wilson's different-sized molecules) eventually yielded v
https://en.wikipedia.org/wiki/Nonsingular%20black%20hole%20models
A nonsingular black hole model is a mathematical theory of black holes that avoids certain theoretical problems with the standard black hole model, including information loss and the unobservable nature of the black hole event horizon. Avoiding paradoxes in the standard black hole model For a black hole to physically exist as a solution to Einstein's equation, it must form an event horizon in finite time relative to outside observers. This requires an accurate theory of black hole formation, of which several have been proposed. In 2007, Shuan Nan Zhang of Tsinghua University proposed a model in which the event horizon of a potential black hole only forms (or expands) after an object falls into the existing horizon, or after the horizon has exceeded the critical density. In other words, an infalling object causes the horizon of a black hole to expand, which only occurs after the object has fallen into the hole, allowing an observable horizon in finite time. This solution does not solve the information paradox, however. Alternative black hole models Nonsingular black hole models have been proposed since theoretical problems with black holes were first realized. Today some of the most viable candidates for the result of the collapse of a star with mass well above the Chandrasekhar limit include the gravastar and the dark energy star. While black holes were a well-established part of mainstream physics for most of the end of the 20th century, alternative models received new attention when models proposed by George Chapline and later by Lawrence Krauss, Dejan Stojkovic, and Tanmay Vachaspati of Case Western Reserve University showed in several separate models that black hole horizons could not form. Such research has attracted much media attention, as black holes have long captured the imagination of both scientists and the public for both their innate simplicity and mysteriousness. The recent theoretical results have therefore undergone much scrutiny and most of
https://en.wikipedia.org/wiki/Surface%20gradient
In vector calculus, the surface gradient is a vector differential operator that is similar to the conventional gradient. The distinction is that the surface gradient takes effect along a surface. For a surface in a scalar field , the surface gradient is defined and notated as where is a unit normal to the surface. Examining the definition shows that the surface gradient is the (conventional) gradient with the component normal to the surface removed (subtracted), hence this gradient is tangent to the surface. In other words, the surface gradient is the orthographic projection of the gradient onto the surface. The surface gradient arises whenever the gradient of a quantity over a surface is important. In the study of capillary surfaces for example, the gradient of spatially varying surface tension doesn't make much sense, however the surface gradient does and serves certain purposes. See also Aspect (geography) Geomorphometry#Surface gradient Derivatives Grade (slope) Spatial gradient
https://en.wikipedia.org/wiki/Jessica%20Sklar
Jessica Katherine Sklar (born 1973) is a mathematician interested in abstract algebra, recreational mathematics, mathematics and art, and mathematics and popular culture. She is a professor of mathematics at Pacific Lutheran University, and former head of the mathematics department at Pacific Lutheran. Education and career As a high school student, Sklar studied poetry at the Interlochen Arts Academy. She did her undergraduate studies at Swarthmore College, where her mother Elizabeth S. had earned a degree in English (later becoming an English professor at Wayne State University) and her father Lawrence Sklar had taught philosophy. Jessica completed a double major in English and mathematics in 1995. Next, Sklar moved to the University of Oregon for graduate study in mathematics, earning a master's degree in 1997 and completing her Ph.D. there in 2001. Her dissertation, Binomial Rings and Algebras, was supervised by Frank Wylie Anderson. She has been a faculty member in the mathematics department at Pacific Lutheran since 2001. Combining her interests in mathematics and art she is one of 24 mathematicians and artists who make up the Mathemalchemy Team. Selected publications “‘Bok bok’: exploring the game of Chicken in film,” with Jennifer F. Nordstrom. In: Handbook of the Mathematics of the Arts and Sciences. Ed. Bharath Sriraman. Springer International Publishing, Cham, 2020. “‘Elegance in design’: mathematics and the works of Ted Chiang.” In: Handbook of the Mathematics of the Arts and Sciences. Ed. Bharath Sriraman. Springer International Publishing, Cham, 2020. “Disciple” (poem). Journal of Humanistic Mathematics 7(2) (July 2017), 418. First-Semester Abstract Algebra: A Structural Approach. GNU Free Documentation License, 2017. “A confused electrician uses Smith normal form,” with Tom Edgar. Mathematics Magazine 89(1) (2016), 3–13. Mathematics in Popular Culture: Essays on Appearances in Film, Literature, Games, Television and Other Media. Jefferson, NC:
https://en.wikipedia.org/wiki/PPARGC1A
Peroxisome proliferator-activated receptor gamma coactivator 1-alpha (PGC-1α) is a protein that in humans is encoded by the PPARGC1A gene. PPARGC1A is also known as human accelerated region 20 (HAR20). It may, therefore, have played a key role in differentiating humans from apes. PGC-1α is the master regulator of mitochondrial biogenesis. PGC-1α is also the primary regulator of liver gluconeogenesis, inducing increased gene expression for gluconeogenesis. Function PGC-1α is a gene that contains two promoters, and has 4 alternative splicings. PGC-1α is a transcriptional coactivator that regulates the genes involved in energy metabolism. It is the master regulator of mitochondrial biogenesis. This protein interacts with the nuclear receptor PPAR-γ, which permits the interaction of this protein with multiple transcription factors. This protein can interact with, and regulate the activity of, cAMP response element-binding protein (CREB) and nuclear respiratory factors (NRFs) . PGC-1α provides a direct link between external physiological stimuli and the regulation of mitochondrial biogenesis, and is a major factor causing slow-twitch rather than fast-twitch muscle fiber types. Endurance exercise has been shown to activate the PGC-1α gene in human skeletal muscle. Exercise-induced PGC-1α in skeletal muscle increases autophagy and unfolded protein response. PGC-1α protein may also be involved in controlling blood pressure, regulating cellular cholesterol homeostasis, and the development of obesity. Regulation PGC-1α is thought to be a master integrator of external signals. It is known to be activated by a host of factors, including: Reactive oxygen species and reactive nitrogen species, both formed endogenously in the cell as by-products of metabolism but upregulated during times of cellular stress. Fasting can also increase gluconeogenic gene expression, including hepatic PGC-1α. It is strongly induced by cold exposure, linking this environmental stimulus to a
https://en.wikipedia.org/wiki/Prolonged%20exposure%20therapy
Prolonged exposure therapy (PE) is a form of behavior therapy and cognitive behavioral therapy designed to treat post-traumatic stress disorder. It is characterized by two main treatment procedures – imaginal and in vivo exposures. Imaginal exposure is repeated 'on-purpose' retelling of the trauma memory. In vivo exposure is gradually confronting situations, places, and things that are reminders of the trauma or feel dangerous (despite being objectively safe). Additional procedures include processing of the trauma memory and breathing retraining. Overview Prolonged exposure therapy was developed by Edna B Foa, Director of the Center for the Treatment and Study of Anxiety at the University of Pennsylvania. Prolonged exposure therapy (PE) is a theoretically based and highly effective treatment for chronic post-traumatic stress disorder (PTSD) and related depression, anxiety, and anger. PE falls under the category of "exposure-based therapy" and is supported by scientific studies which reflect its positive impact on patient symptoms. Exposure-based therapies focus on confronting the harmless cues/triggers of trauma/stress in order to unpair them from the feelings of anxiety and stress. Prolonged exposure is a flexible therapy that can be modified to fit the needs of individual clients. It is specifically designed to help clients psychologically process traumatic events and reduce trauma-induced psychological disturbances. Prolonged exposure produces clinically significant improvement in 40–75% of patients with chronic PTSD, with the only reliable predictor of treatment outcomes being pre-treatment chronic PTSD symptom severity. Prolonged exposure is rooted in the theory of emotional processing, in which new, accurate information challenges previously learned fear structures and modifies them in such a way that the new, accurate information is more readily retrieved. With PTSD, it is thought that traumatic events cause inaccurate associations to develop, betwe
https://en.wikipedia.org/wiki/Nuclear%20or%20Not%3F
Nuclear or Not? Does Nuclear Power Have a Place in a Sustainable Energy Future? is a 2007 book edited by Professor David Elliott. The book offers various views and perspectives on nuclear power. Authors include: Paul Allen from the Centre for Alternative Technology Dr Ian Fairlie, who served on the Committee Examining Radiation Risks of Internal Emitters (CERRIE) Stephen Kidd of the World Nuclear Association Professor Elliott calls for continued debate on the nuclear power issue. He has worked with the United Kingdom Atomic Energy Authority before moving to the Open University where he is Professor of Technology Policy and has developed courses on technological innovation, focusing in particular on renewable energy technology. See also List of books about nuclear issues Nuclear Power and the Environment Reaction Time Contesting the Future of Nuclear Power Non-Nuclear Futures
https://en.wikipedia.org/wiki/N-acetyltransferase%202
N-acetyltransferase 2 (arylamine N-acetyltransferase), also known as NAT2, is an enzyme which in humans is encoded by the NAT2 gene. Function This gene encodes a type of N-acetyltransferase. The NAT2 isozyme functions to both activate and deactivate arylamine and hydrazine drugs and carcinogens. Polymorphisms in this gene are responsible for the N-acetylation polymorphism in which human populations segregate into rapid, intermediate, and slow acetylator phenotypes. Polymorphisms in NAT2 are also associated with higher incidences of cancer and drug toxicity. A second arylamine N-acetyltransferase gene (NAT1) is located near NAT2. Phenotype prediction The NAT2 acetylator phenotype can be inferred from NAT2 genotype (a combination of SNPs observed in a given individual).
https://en.wikipedia.org/wiki/Ambric
Ambric, Inc. was a designer of computer processors that developed the Ambric architecture. Its Am2045 Massively Parallel Processor Array (MPPA) chips were primarily used in high-performance embedded systems such as medical imaging, video, and signal-processing. Ambric was founded in 2003 in Beaverton, Oregon by Jay Eisenlohr and Anthony Mark Jones. Eisenlohr previously founded and sold Rendition, Inc. to Micron Technology for $93M, while Jones is a leading expert in analog, digital, and system IC design and is the named inventor on over 120 U.S. patents. Jones was also the founder of a number of companies prior to Ambric, and has since co-founded Vitek IP with technology and patent expert Dan Buri in 2019. Ambric developed and introduced the Am2045 and its software tools in 2007, but fell victim to the financial crisis of 2007–2008. Ambric's Am2045 and tools remained available through Nethra Imaging, Inc., which closed in 2012. Architecture and programming model Ambric architecture is a massively parallel distributed memory multiprocessor, based on the Structural Object Programming Model. Each processor is programmed in conventional Java (a strict subset) and/or assembly code. The hundreds of processors on the chip send data and control messages to one another through an interconnect of reconfigurable, self-synchronizing channels, which provide both communication and synchronization. The model of computation is very similar to a Kahn process network with bounded buffers. Devices and tools The Am2045 device has 336 32-bit RISC-DSP fixed-point processors and 336 2-kibibyte memories, which run at up to 300 MHz. It has an Eclipse-based integrated development environment including editor, compiler, assemblers, simulator, configuration generator, source-code debugger and video/image-processing, signal-processing, and video-codec libraries. Power and performance The Am2045 delivers 1 TeraOPS (Operations Per Second) and 50 Giga-MACs (Multply-Accumulates per second) of f
https://en.wikipedia.org/wiki/Differential%20nonlinearity
Differential nonlinearity (acronym DNL) is a commonly used measure of performance in digital-to-analog (DAC) and analog-to-digital (ADC) converters. It is a term describing the deviation between two analog values corresponding to adjacent input digital values. It is an important specification for measuring error in a digital-to-analog converter (DAC); the accuracy of a DAC is mainly determined by this specification. Ideally, any two adjacent digital codes correspond to output analog voltages that are exactly one Least Significant Bit (LSB) apart. Differential non-linearity is a measure of the worst-case deviation from the ideal 1 LSB step. For example, a DAC with a 1.5 LSB output change for a 1 LSB digital code change exhibits 1⁄2 LSB differential non-linearity. Differential non-linearity may be expressed in fractional bits or as a percentage of full scale. A differential non-linearity greater than 1 LSB may lead to a non-monotonic transfer function in a DAC. It is also known as a missing code. Differential linearity refers to a constant relation between the change in the output and input. For transducers if a change in the input produces a uniform step change in the output the transducer possess differential linearity. Differential linearity is desirable and is inherent to a system such as a single-slope analog-to-digital converter used in nuclear instrumentation. Formula Effects of DNL If the DNL of an ADC is smaller than -1, missing codes appear in the transfer function, i.e. there are codes for which there is no input voltage to get the code at the ADC output. If the DNL of a DAC is bigger than 1, the transfer function of the DAC becomes non-monotonic. A non-monotonic DAC is especially not desired in closed-loop control application as it may cause stability problems, i.e. it may cause oscillations. See also Integral nonlinearity
https://en.wikipedia.org/wiki/Non-stick%20surface
A non-stick surface is engineered to reduce the ability of other materials to stick to it. Non-stick cookware is a common application, where the non-stick coating allows food to brown without sticking to the pan. Non-stick is often used to refer to surfaces coated with polytetrafluoroethylene (PTFE), a well-known brand of which is Teflon. In the twenty-first century, other coatings have been marketed as non-stick, such as anodized aluminium, silica, enameled cast iron, and seasoned cookware. Types Seasoning Cast iron, carbon steel, stainless steel and cast aluminium cookware may be seasoned before cooking by applying a fat to the surface and heating it to polymerize it. This produces a dry, hard, smooth, hydrophobic coating, which is non-stick when food is cooked with a small amount of cooking oil or fat. Fluoropolymer The modern non-stick pans were made using a coating of Teflon (polytetrafluoroethylene or PTFE). PTFE was invented serendipitously by Roy Plunkett in 1938, while working for a joint venture of the DuPont company. The substance was found to have several unique properties, including very good corrosion-resistance and the lowest coefficient of friction of any substance yet manufactured. PTFE was first used to make seals resistant to the uranium hexafluoride gas used in development of the atomic bomb during World War II, and was regarded as a military secret. Dupont registered the Teflon trademark in 1944 and soon began planning for post-war commercial use of the new product. By 1951 Dupont had developed applications for Teflon in commercial bread and cookie-making; however, the company avoided the market for consumer cookware due to potential problems associated with release of toxic gases if stove-top pans were overheated in inadequately ventilated spaces. While working at DuPont, NYU Tandon School of Engineering alumnus John Gilbert was asked to evaluate a newly developed material called Teflon. His experiments using the fluorinated polymer as a s
https://en.wikipedia.org/wiki/Alpha%20acid
Alpha acids (α acids) are a class of chemical compounds primarily of importance to the production of beer. They are found in the resin glands of the flowers of the hop plant and are the source of hop bitterness. Alpha acids may be isomerized to form iso-alpha acids by the application of heat in solution. Iso-alpha acids (iso-α-acids) are typically produced in beer from the addition of hops to the boiling wort. The degree of isomerization and the amount of bitter flavor produced by the addition of hops is highly dependent on the length of time the hops are boiled. Longer boil times will result in isomerization of more alpha acids and thus increased bitterness. Common alpha acids include humulone, adhumulone, cohumulone, posthumulone, and prehumulone. The most common iso-α-acids are cis- and trans-isohumulone. Bittering The alpha acid "rating" on hops indicates the amount of alpha acid as a percentage of total weight of the hop. Hops with a higher alpha acid content will contribute more bitterness than a lower alpha acid hop when using the same amount of hops. High alpha acid varieties of hops are more efficient for producing highly bitter beers. Alpha acid percentages vary within specific varieties depending on growing conditions, drying methods, age of the hop, and other factors. For example, this list shows the typical range of alpha acids found in some common varieties (percentages are based on total dried weight). The choice of a hop variety used in beer brewing depends on the beer style. For instance, lager styles use hop varieties with a low alpha acid content (such as Saaz and Hallertauer) while IPA styles use hop varieties with a high alpha acid content (such as Cascade, Centennial and Chinook). Anti-bacterial properties Iso-α-acids have a bacteriostatic effect on many common Gram-positive bacteria found in beer. While the iso-α-acids are very effective at preventing serious contamination from Gram-positive bacteria such as the lactic aci
https://en.wikipedia.org/wiki/Inversion%20temperature
The inversion temperature in thermodynamics and cryogenics is the critical temperature below which a non-ideal gas (all gases in reality) that is expanding at constant enthalpy will experience a temperature decrease, and above which will experience a temperature increase. This temperature change is known as the Joule–Thomson effect, and is exploited in the liquefaction of gases. Inversion temperature depends on the nature of the gas. For a van der Waals gas we can calculate the enthalpy using statistical mechanics as where is the number of molecules, is volume, is temperature (in the Kelvin scale), is Boltzmann's constant, and and are constants depending on intermolecular forces and molecular volume, respectively. From this equation, we note that if we keep enthalpy constant and increase volume, temperature must change depending on the sign of . Therefore, our inversion temperature is given where the sign flips at zero, or , where is the critical temperature of the substance. So for , an expansion at constant enthalpy increases temperature as the work done by the repulsive interactions of the gas is dominant, and so the change in kinetic energy is positive. But for , expansion causes temperature to decrease because the work of attractive intermolecular forces dominates, giving a negative change in average molecular speed, and therefore kinetic energy. See also Critical point (thermodynamics) Phase transition Joule-Thomson effect
https://en.wikipedia.org/wiki/Call%20signs%20in%20Australia
Call signs in Australia are allocated by the Australian Communications and Media Authority and are unique for each broadcast station. The use of callsigns on-air in both radio and television in Australia is optional, so many stations used other on-air identifications. Australian broadcast stations officially have the prefix VL- and originally all callsigns used that format, but since Australia has no nearby neighbours, this prefix is no longer used except in an international context. Call sign blocks for telecommunication The International Telecommunication Union has assigned Australia the following call sign blocks for all radio communication, broadcasting or transmission: While not directly related to call signs, the International Telecommunication Union (ITU) further has divided all countries assigned amateur radio prefixes into three regions; Australia is located in ITU Region 3. It is assigned ITU Zones 55, 58 and 59, with the Pacific Islands in Australian jurisdiction in Zone 60. Call signs for radio All radio call signs begin with a single-digit number indicating the state or territory, followed by two or three letters. In most cases, two letters are used for AM stations and three for FM, but there are some exceptions, such as 5UV in Adelaide, which broadcasts on an FM frequency, and 3RPH in Melbourne, which broadcasts on an AM frequency. While some AM stations retained their old call signs when moving to FM, most add an extra letter to the call sign. For instance, when 7HO Hobart became an FM station, it adopted the callsign 7HHO. Certain ABC radio stations, particularly outside of metropolitan areas, may use five-letter call signs for FM stations: xABCFM for ABC Classic FM, xABCRN for Radio National, and xABCRR for ABC Local Radio – the x being the state number. Also, SBS FM radio stations use a five-letter call sign, xSBSFM. (Sydney and Melbourne's AM stations use 2EA and 3EA, meaning Ethnic Australia.) There are a number of exceptions: For some t
https://en.wikipedia.org/wiki/Harvest%20%28wine%29
The harvesting of wine grapes (Vintage) is one of the most crucial steps in the process of wine-making. The time of harvest is determined primarily by the ripeness of the grape as measured by sugar, acid and tannin levels with winemakers basing their decision to pick based on the style of wine they wish to produce. The weather can also shape the timetable of harvesting with the threat of heat, rain, hail, and frost which can damage the grapes and bring about various vine diseases. In addition to determining the time of the harvest, winemakers and vineyard owners must also determine whether to use hand pickers or mechanical harvesters. The harvest season typically falls between August & October in the Northern Hemisphere and February & April in the Southern Hemisphere. With various climate conditions, grape varieties, and wine styles the harvesting of grapes could happen in every month of the calendar year somewhere in the world. In the New World it is often referred to as the crush. Harvest season The majority of the world's wine producing regions lie between the temperate latitudes of 30° and 50° in both hemispheres with regions lying closer to the equator typically harvesting earlier due to their warmer climates. In the Northern Hemisphere, vineyards in Cyprus begin harvesting as early as July. In California some sparkling wine grapes are harvested in late July to early August at a slightly unripe point to help maintain acidity in the wine. The majority of Northern Hemisphere harvesting occurs in late August to early October with some late harvest wine grapes being harvested throughout the autumn. In Germany, Austria, the United States and Canada, ice wine grapes can be harvested as late as January. In the Southern Hemisphere harvest can begin as early as January 1 in some of the warmer climate sites in New South Wales, Australia. The majority of Southern Hemisphere harvesting occurs between the months of February and April with some cool climate sites like Centr
https://en.wikipedia.org/wiki/Marcela%20Agoncillo
Doña Marcela Mariño de Agoncillo (née Mariño y Coronel; June 24, 1859 – May 30, 1946) was a Filipina who was the principal seamstress of the first and official flag of the Philippines, gaining her the title of "The Mother of the Philippine Flag." Marcela Coronel Mariño was the daughter of Don Francisco Diokno Mariño and Doña Eugenia Coronel Mariño, a rich family in her hometown of Taal, Batangas. She finished her studies at Santa Catalina College, Marcela acquired her learning in music and feminine crafts. At the age of 30, Marcela Coronel Mariño married Felipe Encarnacion Agoncillo, a Filipino lawyer, and a jurist, and gave birth to six children. Her marriage led an important role in Philippine history. When her husband was exiled in Hong Kong during the outbreak of the Philippine Revolution, Marcela Mariño Agoncillo and the rest of the family joined him and temporarily resided there to avoid the anti-Filipino hostilities of the occupying Spain. While in Hong Kong, General Emilio Aguinaldo requested her to sew the flag that would represent the Republic of the Philippines. Doña Marcela Mariño de Agoncillo, with her eldest daughter Lorenza and a friend Delfina Herbosa de Natividad, niece of Dr. Jose Rizal, manually sewed the flag in accordance with General Emilio Aguinaldo's design which later became the official flag of the Republic of the Philippines. While the flag itself is the perpetual legacy of Doña Marcela Mariño de Agoncillo, she is also commemorated through museums and monuments: like the marker in Hong Kong (where her family temporarily sojourned), at her ancestral home in Taal, Batangas which has been turned into a museum, in paintings by notable painters as well as through other visual arts. Early life Marcela Coronel Mariño was born on June 24, 1859, in Taal, Batangas, Philippines to Don Francisco Diokno Mariño and Doña Eugenia Coronel Mariño. She grew up in her ancestral Mariño house in Taal, Batangas built in the 1770s by her grandparents,
https://en.wikipedia.org/wiki/Karl%20Johan%20%C3%85str%C3%B6m
Karl Johan Åström (born August 5, 1934) is a Swedish control theorist, who has made contributions to the fields of control theory and control engineering, computer control and adaptive control. In 1965, he described a general framework of Markov decision processes with incomplete information, what ultimately led to the notion of a Partially observable Markov decision process. In 1995, Åström was elected as a member into the National Academy of Engineering for contributions to identification, stochastic, and adaptive control and their incorporation in control engineering practice. Biography Åström was born in Östersund, Sweden, and received his M.Sc. in Engineering Physics (1957) and PhD in Automatic Control and Mathematics (1960) from the Royal Institute of Technology (KTH) in Stockholm, where he also taught from 1955 to 1960 while working on inertial guidance for the Swedish National Defence Research Institute. In 1961 Åström joined the IBM Nordic Laboratory to work on computerized process control, with tours at IBM Research in Yorktown Heights, New York (1962) and San Jose, California (1963). After his return he led efforts in the computer control of paper manufacturing machinery. In 1965, Åström was named chair of the newly founded Department of Automatic Control at Lund University, Sweden. From 1965 to 1999 he was chair of the Department of Automatic Control at Lund University, where he is now professor emeritus. Since 2002 he has been distinguished visiting professor at the University of California, Santa Barbara. Åström is a Fellow of the IEEE, member of the Royal Swedish Academy of Sciences, vice president of the Royal Swedish Academy of Engineering Sciences (IVA), and a foreign associate of the US National Academy of Engineering. He was awarded the ASME Rufus Oldenburger medal (1985) and the International Federation of Automatic Control Quazza Medal (1987). In 1987 he was also awarded the degree Docteur Honoris Causa from l'Institut National Polytechni
https://en.wikipedia.org/wiki/Philips%20Velo
The Philips Velo was a Handheld PC. Velo 1 The initial Velo 1 was a PDA device released by Philips in 1997. The device was typical of the HPCs at the time, being powered by two AA batteries or a rechargeable NiMH battery pack. It had a back-lit, greyscale, 5.1-inch resistive touchscreen with a resolution of 480×240 pixels. Employing two bits per pixel allowed for the display of four shades of grey (16 shades with Windows CE 2.0 upgrade). Expandability was via two internal Miniature Card expansion slots, as well as a PCMCIA slot provided by an optional V-Module. The Velo Dock, a docking station, also provided a PCMCIA slot. The standard model included 4 MB of RAM and 8 MB ROM. It originally shipped with Windows CE 1.0, but in September 1997, an 8 MB model with NiMH battery pack was announced at a price of $839.99, and Windows CE 2.0 offered as a free future ROM upgrade for all newly purchased Velo 1 units. Existing customers were able to order this upgrade for $99.99. Users with 4 MB devices were strongly recommended to upgrade to 8 MB, and an upgrade programme was announced to offer 4 MB Miniature Cards at reduced prices for US customers. Unlike most other HPCs, the Velo 1 included a built-in low-power software modem, where most other devices required a PCMCIA card modem. PCMCIA cards could quickly drain batteries, so the low-power software modem helped eke out battery life. The modem operated at 19.2 Kbit/s. Other communication features included an infrared (IrDa) port and serial port, capable of a maximum transmission rate of 115,200 bit/s and 230,000 bits/s respectively. The MIPS processor in the unit, the PR31500, incorporated a digital signal processing (DSP) unit to allow the processor to take over audio processing and signalling responsibilities, eliminating the need for dedicated DSP circuitry or modem. An accompanying processor, the UCB1100, was provided for telephone line interfacing. The PR31500 also offered improved code density compared to previou
https://en.wikipedia.org/wiki/Classification%20of%20manifolds
In mathematics, specifically geometry and topology, the classification of manifolds is a basic question, about which much is known, and many open questions remain. Main themes Overview Low-dimensional manifolds are classified by geometric structure; high-dimensional manifolds are classified algebraically, by surgery theory. "Low dimensions" means dimensions up to 4; "high dimensions" means 5 or more dimensions. The case of dimension 4 is somehow a boundary case, as it manifests "low dimensional" behaviour smoothly (but not topologically); see discussion of "low" versus "high" dimension. Different categories of manifolds yield different classifications; these are related by the notion of "structure", and more general categories have neater theories. Positive curvature is constrained, negative curvature is generic. The abstract classification of high-dimensional manifolds is ineffective: given two manifolds (presented as CW complexes, for instance), there is no algorithm to determine if they are isomorphic. Different categories and additional structure Formally, classifying manifolds is classifying objects up to isomorphism. There are many different notions of "manifold", and corresponding notions of "map between manifolds", each of which yields a different category and a different classification question. These categories are related by forgetful functors: for instance, a differentiable manifold is also a topological manifold, and a differentiable map is also continuous, so there is a functor . These functors are in general neither one-to-one nor onto; these failures are generally referred to in terms of "structure", as follows. A topological manifold that is in the image of is said to "admit a differentiable structure", and the fiber over a given topological manifold is "the different differentiable structures on the given topological manifold". Thus given two categories, the two natural questions are: Which manifolds of a given type admit an additiona
https://en.wikipedia.org/wiki/Eric%20Schechter
Eric Schechter (born August 1, 1950) is an American mathematician, retired from Vanderbilt University with the title of professor emeritus. His interests started primarily in analysis but moved into mathematical logic. Schechter is best known for his 1996 book Handbook of Analysis and its Foundations, which provides a novel approach to mathematical analysis and related topics at the graduate level. Important works Schechter has authored a number of articles in analysis, differential equations, mathematical logic, and set theory. He is best known for writing two textbooks covering advanced material but written at an introductory level: (2005) Classical and Nonclassical Logics () (more information) (1996) Handbook of Analysis and its Foundations () (more information) Handbook of Analysis and its Foundations was reviewed at length by the Society for Industrial and Applied Mathematics review, which wrote: Every once in a while a book comes along that so effectively redefines an educational enterprise -- in this case, graduate mathematical training -- and so effectively reexamines the hegemony of ideas prevailing in a discipline -- in this case, mathematical analysis -- that it deserves our careful attention. This is such a book. Schechter also maintains two webpages that are frequently cited: Common Errors in Undergraduate Mathematics A Home Page for the Axiom of Choice Politics Schechter is involved in political activism of the democratic socialist variety. His mathematical homepage includes a few anti-war statements, and his political home page includes a long essay about progressive ideology. He has worked as an organizer for the Nashville Peace Coalition, protesting the wars in Iraq and Afghanistan. At a meeting for the living wage movement on Vanderbilt's campus, he remarked that it is hard to bring up politics in a non-political environment, and expressed that people did not talk much about politics in the mathematics department at Vanderbilt. Hi
https://en.wikipedia.org/wiki/Center%20for%20Molecular%20and%20Biomolecular%20Informatics
The Center for Molecular and Biomolecular Informatics (CMBI) is an academic computational biology (bioinformatics) research department based in Nijmegen (the Netherlands), and is part of the Radboudumc University Medical Center. Most of the research topics at the CMBI focus strongly on (in silico) biomedical data analysis of biomolecules, and with - because of the department's embedding in a leading Dutch hospital - a strong interest in medical and clinical scientific research lines, both fundamental and translational in nature. Specific topics of interest involve protein (structure), (meta)genomics as well as the analysis of various other types of omics data (transcriptomics, proteomics, metabolomics, amongst others), and small (bio)molecules and their drug, protein or other molecular interaction partners. Furthermore, especially machine learning (including the application of artificial neural networks), personalized medicine and FAIR data research have a lot of interest and attention by the department. As of 2023, the CMBI has become part of the Radboudumc Medical BioSciences department. History The institute was founded in 1999 at the Radboud University Nijmegen, by the previous department head prof. dr. Gerrit (Gert) Vriend. In 2008 the institute became part of the Radboudumc (i.e. the Radboud University Medical Center). In 2023, the CMBI became part of the Radboudumc Medical BioSciences department. External links Radboudumc research website See also WHAT IF software Bioinformatics organizations Organizations established in 1999 Organisations based in Gelderland 1999 establishments in the Netherlands Nijmegen
https://en.wikipedia.org/wiki/Robert%20H.%20Gray
Robert Hansen Gray (March 7, 1948 - December 6, 2021) was an American data analyst, author, and astronomer, and author of The Elusive Wow: Searching for Extraterrestrial Intelligence. Education Gray attended Shimer College, a Great Books school then located in Mount Carroll, Illinois, where he received a bachelor's degree in 1970. He went on to obtain a master's in urban planning and policy analysis from the University of Illinois at Chicago in 1980. Career Data analysis In 1984, Gray founded the company Gray Data in Chicago, which provided data analysis research services and published reference cards for microcomputer software. He continued to work as a data analyst through his company Gray Consulting. Search for Extraterrestrial Intelligence (SETI) Gray is best known for his work as an independent SETI researcher. The Atlantic called Gray "the 'Wow!' signal's most devoted seeker and chronicler, having traveled to the very ends of the earth in search of it." The Wow! signal was detected by the Ohio State University Radio Observatory (also known as Big Ear) on August 15, 1977. The signal was so pronounced in the data, and so similar to a radio signal rather than a natural source, that SETI scientist Jerry R. Ehman circled it on the computer printout in red ink and wrote "Wow!" next to it. After hearing about the Wow! signal a few years after its detection, Gray contacted the Ohio team, visited Big Ear, and spoke with Ehman, Robert S. Dixon (director of the SETI project) and John D. Kraus (the telescope's designer). In 1980, Gray began scanning the skies from his backyard in Chicago, using a 12-foot commercial telecommunications dish. He operated his small SETI radio observatory regularly beginning in 1983 and for the next 15 years, but did not find a trace of the Wow! signal. In 1987 and 1989 he led searches for the signal using the Harvard/Smithsonian META radio telescope at the Oak Ridge Observatory in Harvard, Massachusetts. In September 1995 and again in
https://en.wikipedia.org/wiki/Interleukin-4%20receptor
The interleukin 4 receptor is a type I cytokine receptor. It is a heterodimer, that is, composed of two subunits. IL4R is the human gene coding for IL-4Rα, the subunit which combines with either common gamma chain (γc, forming the type I IL4 receptor) or with IL-13Rα1 (forming the type II IL4 receptor). Function This gene encodes the alpha chain of the interleukin-4 receptor, a type I transmembrane protein that can bind interleukin 4 and interleukin 13 to regulate IgE antibody production in B cells. Among T cells, the encoded protein also can bind interleukin 4 to promote differentiation of Th2 cells. A soluble form of the encoded protein can be produced by an alternate splice variant or by proteolysis of the membrane-bound protein, and this soluble form can inhibit IL4-mediated cell proliferation and IL5 upregulation by T-cells. Allelic variations in this gene have been associated with atopy, a condition that can manifest itself as allergic rhinitis, sinusitis, asthma, or eczema. Two transcript variants encoding different isoforms, a membrane-bound and a soluble form, have been found for this gene. Interactions of IL-4 with TNFα promote structural changes to vascular endothelial cells, thus playing an important role in tissue inflammation. The binding of IL-4 or IL-13 to the IL-4 receptor on the surface of macrophages results in the alternative activation of those macrophages. Alternatively activated macrophages (AAMΦ) downregulate inflammatory mediators such as IFNγ during immune responses, particularly with regards to helminth infections. Interactions Interleukin-4 receptor has been shown to interact with SHC1. Structure The N-terminal (extracellular) portion of interleukin-4 receptor is related in overall topology to fibronectin type III modules and folds into a sandwich comprising seven antiparallel beta sheets arranged in a three-strand and a four-strand beta-pleated sheet. They are required for binding of interleukin-4 to the receptor alpha chain,
https://en.wikipedia.org/wiki/EMBRACE
EMBRACE (A European Model for Bioinformatics Research and Community Education) was a project from the years 2005 to 2010, with the objective of drawing together a wide group of European experts involved in using information technology in the biomolecular sciences. The EMBRACE Network endeavored to integrate major bioinformatics databases and software tools, using existing methods and emerging grid service technologies. Integration efforts were driven by a set of test problems representing key issues for bioinformatics service providers and end-user biologists. EMBRACE made many bioinformatics web services available to the international research community. As a result, groups throughout Europe were able to use the EMBRACE service interfaces for their own local or proprietary data and tools. The project was run from the EBI in Hinxton, England. Fred Marcus was its EU project coordinator. The EMBRACE project was funded by the European Commission within its FP6 Programme, under the thematic area "Life sciences, genomics, and biotechnology for health", contract number LHSG-CT-2004-512092.
https://en.wikipedia.org/wiki/Decipol
The decipol is a unit used to measure the perceived air quality. It was introduced by Danish professor P. Ole Fanger. One decipol (dp) is the perceived air quality (PAQ) in a space with a sensory load of one olf (one standard person) ventilated by 10 L/s. It was developed to quantify how the strength of indoor pollution sources indoors influence air quality as it is perceived by humans.
https://en.wikipedia.org/wiki/Institute%20of%20Mathematics%20of%20National%20Academy%20of%20Sciences%20of%20Armenia
The Institute of Mathematics of National Academy of Sciences of Armenia (Armenian: ) is owned and operated by the Armenian Academy of Sciences, located in Yerevan. History The Institute of Mathematics of National Academy of Sciences of Armenia originated as the Section for Mathematics and Mechanics, created within the newly formed Armenian Academy of Sciences in 1944. The section later developed into an Institute of Mathematics and Mechanics of the Armenian Academy of Sciences, whose first Director was academician Artashes Shahinian, known for his results in complex analysis. The Institute of Mathematics of Armenian Academy of Sciences separated from the latter Institute in 1971. The bearer of the office of the Director of Institute has been academician Mkhitar Djrbashian (1971-1989, 1989-1994 Honorary Director). The academicians Sergey Mergelyan, Norair Arakelian, Alexandr Talalyan, Raphayel Alexandrian, Rouben V. Ambartzumian and Anry Nersesyan also have greatly influenced the formation of the scientific profile of the Institute and largely contributed to mathematics in general. In particular Rouben V. Ambartzumian is famous for his work in Stochastic Geometry and Integral Geometry, where he created a new branch called Combinatorial Integral Geometry. He has provided solutions to a number of classical problems in particular the solution to the Buffon Sylvester problem as well as the Hilbert's fourth problem in dimensions 2 and 3. In the early years, the investigations carried out in the Institute concentrated on Function Theory. Gradually the sphere of investigations expanded and now includes Differential and Integral Equations, Functional Analysis, Probability Theory and Mathematical Statistics. Journals The institute has a journal, Izvestia NAS RA Matematika. The founder and the first Editor in Chief (1971–1990) of the journal was Mkhitar Djrbashian; under Rouben V. Ambartzumian Editor in Chief (1990 - 2010) the journal obtained international recognition and
https://en.wikipedia.org/wiki/Enthalpy%20of%20mixing
In thermodynamics, the enthalpy of mixing (also heat of mixing and excess enthalpy) is the enthalpy liberated or absorbed from a substance upon mixing. When a substance or compound is combined with any other substance or compound, the enthalpy of mixing is the consequence of the new interactions between the two substances or compounds. This enthalpy, if released exothermically, can in an extreme case cause an explosion. Enthalpy of mixing can often be ignored in calculations for mixtures where other heat terms exist, or in cases where the mixture is ideal. The sign convention is the same as for enthalpy of reaction: when the enthalpy of mixing is positive, mixing is endothermic, while negative enthalpy of mixing signifies exothermic mixing. In ideal mixtures, the enthalpy of mixing is null. In non-ideal mixtures, the thermodynamic activity of each component is different from its concentration by multiplying with the activity coefficient. One approximation for calculating the heat of mixing is Flory–Huggins solution theory for polymer solutions. Formal definition For a liquid, enthalpy of mixing can be defined as follows Where: H(mixture) is the total enthalpy of the system after mixing ΔHmix is the enthalpy of mixing xi is the mole fraction of component i in the system Hi is the enthalpy of pure i Enthalpy of mixing can also be defined using Gibbs free energy of mixing However, Gibbs free energy of mixing and entropy of mixing tend to be more difficult to determine experimentally. As such, enthalpy of mixing tends to be determined experimentally in order to calculate entropy of mixing, rather than the reverse. Enthalpy of mixing is defined exclusively for the continuum regime, which excludes molecular-scale effects (However, first-principles calculations have been made for some metal-alloy systems such as Al-Co-Cr or β-Ti). When two substances are mixed the resulting enthalpy is not an addition of the pure component enthalpies, unless the substances form
https://en.wikipedia.org/wiki/Convention%20for%20the%20Protection%20of%20Human%20Rights%20and%20Dignity%20of%20the%20Human%20Being%20with%20regard%20to%20the%20Application%20of%20Biology%20and%20Medicine
The Convention for the Protection of Human Rights and Dignity of the Human Being with regard to the Application of Biology and Medicine, otherwise known as the European Convention on Bioethics or the European Bioethics Convention, is an international instrument aiming to prohibit the misuse of innovations in biomedicine and to protect human dignity. The Convention was opened for signature on 4 April 1997 in Oviedo, Spain and is thus otherwise known as the Oviedo Convention. The International treaty is a manifestation of the effort on the part of the Council of Europe to keep pace with developments in the field of biomedicine; it is notably the first multilateral binding instrument entirely devoted to biolaw. The Convention entered into force on 1 December 1999. Characteristics The Convention provides a framework structure to preserve human dignity comprehensively across the field of bioethics. The instrument is shaped around the premise that there is a fundamental connection between human rights and biomedicine. A minimum common standard is created by the Convention and allows states to legislate for a greater degree of protection upon ratification (Article 27). In addition, judicial protection is conferred on the national courts. Therefore, there is no basis on which an individual can bring an action in relation to the Oviedo Convention alone. The Convention may only be referenced in conjunction with proceedings brought in respect of a violation of the European Convention on Human Rights. Absence of any provisions for a judicial procedure from the convention is considered to be a major weakness of the Oviedo Convention. History The rate of advancement in biomedicine caused concern to the Council of Europe that as much as development in this field instilled hope for mankind, it also posed a threat. It became the objective of the Council of Europe to set out common general standards for the protection of the dignity of the human person in relation to biomedical sci
https://en.wikipedia.org/wiki/Ferranti%20Argus
Ferranti's Argus computers were a line of industrial control computers offered from the 1960s into the 1980s. Originally designed for a military role, a re-packaged Argus was the first digital computer to be used to directly control an entire factory. They were widely used in a variety of roles in Europe, particularly in the UK, where a small number continue to serve as monitoring and control systems for nuclear reactors. Original series Blue Envoy, hearing aid computer The original concept for the computer was developed as part of the Blue Envoy missile project. This was a very long-range surface-to-air missile system with a range on the order of . To reach these ranges, the missile was "lofted" in a nearly vertical trajectory at launch, flying as quickly as possible to high altitude where it suffered less drag during the subsequent long cruise toward the target. During the vertical climb, the missile's radar would not be able to see the target, so during this period it was command guided from the ground. Argus began as a system to read the radar data, compute the required trajectory, and send that to the missile in-flight. The system not only had to develop the trajectory, but also directly controlled the control surfaces of the missile and thus had a complete control feedback system. Development was carried out by Maurice Gribble at Ferranti's Automation Division in Wythenshawe starting in 1956. The system used OC71 transistors from Mullard, originally designed for use in hearing aids. They could only be run at the low speed of 25 kHz, but this was enough for the task. Blue Envoy was cancelled in 1957 as part of the sweeping 1957 Defence White Paper. Ferranti decided to continue the development of the computer for other uses. During a visit by Prince Philip, Duke of Edinburgh in November 1957, they set up a system with an automotive headlamp connected to a handle that could be moved by hand to shine at any point on a wall, while the computer attempted to move
https://en.wikipedia.org/wiki/British%20Nutrition%20Foundation
The British Nutrition Foundation is a British registered charity and company limited by guarantee that works to share scientific knowledge and advice on diet, physical activity and health. Activities The BNF aims to give the general public, educators and organisations access to reliable information on nutrition. Its website provides details on healthy, sustainable diets, on nutrition at different life stages, on diet in relation to health issues and on putting advice on diet and nutrition into practice. Training is provided through online courses and webinars, with past webinars available on the website, together with videos of previous annual conferences. The organisation runs an annual “Healthy Eating Week” each June. The BNF manages the educational programme Food – a fact of life (FFL). This is designed to support teachers by providing accessible information on diet and health for children and young people progressively through the ages 3 years to 16+ years. It was originally launched in 1991 in a partnership with MAFF that continued until 1997. The BNF continued to develop the educational resources and from 2018 the FFL programme has been a partnership between the BNF and the Agriculture and Horticulture Development Board. The official journal of the BNF, Nutrition Bulletin, is an international, peer-reviewed journal published quarterly by Wiley. Its coverage has included review articles and news items on nutrition, but since acquisition of an impact factor and MEDLINE coverage an increasing number of papers reporting original research have been included; many articles are open-access. Governance Under its Memorandum and Articles of Association, the Board may comprise no more than 12 Trustees. Trustees are appointed by the Board and serve for a term of three years, with each Trustee able to serve for a maximum of nine years; membership of the Board is weighted towards the scientific academic community. As a charity largely funded by corporate don
https://en.wikipedia.org/wiki/Relative%20dimension
In mathematics, specifically linear algebra and geometry, relative dimension is the dual notion to codimension. In linear algebra, given a quotient map , the difference dim V − dim Q is the relative dimension; this equals the dimension of the kernel. In fiber bundles, the relative dimension of the map is the dimension of the fiber. More abstractly, the codimension of a map is the dimension of the cokernel, while the relative dimension of a map is the dimension of the kernel. These are dual in that the inclusion of a subspace of codimension k dualizes to yield a quotient map of relative dimension k, and conversely. The additivity of codimension under intersection corresponds to the additivity of relative dimension in a fiber product. Just as codimension is mostly used for injective maps, relative dimension is mostly used for surjective maps.
https://en.wikipedia.org/wiki/Delaunay%20refinement
In mesh generation, Delaunay refinements are algorithms for mesh generation based on the principle of adding Steiner points to the geometry of an input to be meshed, in a way that causes the Delaunay triangulation or constrained Delaunay triangulation of the augmented input to meet the quality requirements of the meshing application. Delaunay refinement methods include methods by Chew and by Ruppert. Chew's second algorithm Chew's second algorithm takes a piecewise linear system (PLS) and returns a constrained Delaunay triangulation of only quality triangles where quality is defined by the minimum angle in a triangle. Developed by L. Paul Chew for meshing surfaces embedded in three-dimensional space, Chew's second algorithm has been adopted as a two-dimensional mesh generator due to practical advantages over Ruppert's algorithm in certain cases and is the default quality mesh generator implemented in the freely available Triangle package. Chew's second algorithm is guaranteed to terminate and produce a local feature size-graded meshes with minimum angle up to about 28.6 degrees. The algorithm begins with a constrained Delaunay triangulation of the input vertices. At each step, the circumcenter of a poor-quality triangle is inserted into the triangulation with one exception: If the circumcenter lies on the opposite side of an input segment as the poor quality triangle, the midpoint of the segment is inserted. Moreover, any previously inserted circumcenters inside the diametral ball of the original segment (before it is split) are removed from the triangulation. Circumcenter insertion is repeated until no poor-quality triangles exist. Ruppert's algorithm Ruppert's algorithm takes a planar straight-line graph (or in dimension higher than two a piecewise linear system) and returns a conforming Delaunay triangulation of only quality triangles. A triangle is considered poor-quality if it has a circumradius to shortest edge ratio larger than some prescribed thresh
https://en.wikipedia.org/wiki/Grafting
Grafting or graftage is a horticultural technique whereby tissues of plants are joined so as to continue their growth together. The upper part of the combined plant is called the scion () while the lower part is called the rootstock. The success of this joining requires that the vascular tissues grow together. The natural equivalent of this process is inosculation. The technique is most commonly used in asexual propagation of commercially grown plants for the horticultural and agricultural trades. The scion is typically joined to the rootstock at the soil line; however, top work grafting may occur far above this line, leaving an understock consisting of the lower part of the trunk and the root system. In most cases, the stock or rootstock is selected for its roots and the scion is selected for its stems, leaves, flowers, or fruits. The scion contains the desired genes to be duplicated in future production by the grafted plant. In stem grafting, a common grafting method, a shoot of a selected, desired plant cultivar is grafted onto the stock of another type. In another common form called bud grafting, a dormant side bud is grafted onto the stem of another stock plant, and when it has inosculated successfully, it is encouraged to grow by pruning off the stem of the stock plant just above the newly grafted bud. For successful grafting to take place, the vascular cambium tissues of the stock and scion plants must be placed in contact with each other. Both tissues must be kept alive until the graft has "taken", usually a period of a few weeks. Successful grafting only requires that a vascular connection take place between the grafted tissues. Research conducted in Arabidopsis thaliana hypocotyls has shown that the connection of phloem takes place after three days of initial grafting, whereas the connection of xylem can take up to seven days. Joints formed by grafting are not as strong as naturally formed joints, so a physical weak point often still occurs at the graft
https://en.wikipedia.org/wiki/Silicone%20quaternary%20amine
Silicone quaternary amine is a chemical antimicrobial agent used in some odor-repellent socks, including Burlington Bioguard Socks. External links Antimicrobials Polymers
https://en.wikipedia.org/wiki/Triangular%20coordinates
The term triangular coordinates may refer to any of at least three related systems of coordinates in the Euclidean plane: a special case of barycentric coordinates for a triangle, in which case it is known as a ternary plot or areal coordinates, among other names Trilinear coordinates, in which the coordinates of a point in a triangle are its relative distances from the three sides Synergetics coordinates Triangle geometry Coordinate systems
https://en.wikipedia.org/wiki/MENTOR%20routing%20algorithm
The MENTOR routing algorithm is an algorithm for use in routing of mesh networks, specifically pertaining to their initial topology. It was developed in 1991 by Aaron Kershenbaum, Parviz Kermani, and George A. Grove and was published by the IEEE. Complexity Empirical observation has shown the complexity class of this algorithm to be O(N²), or quadratic. This represents "a significant improvement over currently used algorithms, [while still yielding] solutions of a quality competitive with other, much slower procedures." Methodology The algorithm assumes three things are conducive to low-"cost" (that is, minimal in distance travelled and time between destinations) topology: that paths will tend to be direct, not circuitous; that links will have a "high utilization"—that is, they will be used to nearly their maximum operating capacity; and that "long, high-capacity links [will be used] whenever possible." The overall plan is to send traffic over a direct route between the source and destination whenever the magnitude of the requirement is sufficiently large and to send it via a path within a tree in all other cases. In the former case, we are satisfying all three of our goals--we are using a direct path of high utilization and high capacity. In the latter case we are satisfying at least the last two objectives as we are aggregating traffic as much as possible. The minimum spanning tree on which traffic flows in the latter case is heuristically defined by Dijkstra's algorithm and Prim's algorithm.
https://en.wikipedia.org/wiki/Win%E2%80%93stay%2C%20lose%E2%80%93switch
In psychology, game theory, statistics, and machine learning, win–stay, lose–switch (also win–stay, lose–shift) is a heuristic learning strategy used to model learning in decision situations. It was first invented as an improvement over randomization in bandit problems. It was later applied to the prisoner's dilemma in order to model the evolution of altruism. The learning rule bases its decision only on the outcome of the previous play. Outcomes are divided into successes (wins) and failures (losses). If the play on the previous round resulted in a success, then the agent plays the same strategy on the next round. Alternatively, if the play resulted in a failure the agent switches to another action. A large-scale empirical study of players of the game rock, paper, scissors shows that a variation of this strategy is adopted by real-world players of the game, instead of the Nash equilibrium strategy of choosing entirely at random between the three options.
https://en.wikipedia.org/wiki/Smart%20toy
A smart toy is an interactive toy which effectively has its own intelligence by virtue of on-board electronics. These enable it to learn, behave according to preset patterns, and alter its actions depending upon environmental stimuli and user input. Typically, it can adjust to the abilities of the player. A modern smart toy has electronics consisting of one or more microprocessors or microcontrollers, volatile and/or non-volatile memory, storage devices, and various forms of input–output devices. It may be networked together with other smart toys or a personal computer in order to enhance its play value or educational features. Generally, the smart toy may be controlled by software which is embedded in firmware or else loaded from an input device such as a USB flash drive, Memory Stick or CD-ROM. Smart toys frequently have extensive multimedia capabilities, and these can be utilized to produce a realistic, animated, simulated personality for the toy. Some commercial examples of smart toys are Amazing Amanda, Furby and iDog. The first smart-toy was the Mego Corporation's 2-XL robot (2XL), invented in the 1970s Common confusions Smart toys are frequently confused with toys for which it is claimed that children who play with them become smarter. Examples are educational toys that may or may not provide on-board intelligence features. A toy which merely contains a media player for telling the child a story should not be classified as a smart toy even if the player contains its own microprocessor. What best distinguishes a smart toy is the way the on-board intelligence is holistically integrated into the play experience in order to create simulated human-like intelligence or its facsimile. History Smart toys have their early roots in clockworks such as those of the eighteenth and nineteenth century cuckoo clocks, music boxes of the nineteenth, and Disney audio-animatronics of the twentieth. Perhaps the biggest early contribution is from novelty and toy maker
https://en.wikipedia.org/wiki/Chromosome%20conformation%20capture
Chromosome conformation capture techniques (often abbreviated to 3C technologies or 3C-based methods) are a set of molecular biology methods used to analyze the spatial organization of chromatin in a cell. These methods quantify the number of interactions between genomic loci that are nearby in 3-D space, but may be separated by many nucleotides in the linear genome. Such interactions may result from biological functions, such as promoter-enhancer interactions, or from random polymer looping, where undirected physical motion of chromatin causes loci to collide. Interaction frequencies may be analyzed directly, or they may be converted to distances and used to reconstruct 3-D structures. The chief difference between 3C-based methods is their scope. For example, when using PCR to detect interaction in a 3C experiment, the interactions between two specific fragments are quantified. In contrast, Hi-C quantifies interactions between all possible pairs of fragments simultaneously. Deep sequencing of material produced by 3C also produces genome-wide interactions maps. History Historically, microscopy was the primary method of investigating nuclear organization, which can be dated back to 1590. In 1879, Walther Flemming coined the term chromatin. In 1883, August Weismann connected chromatin with heredity. In 1884, Albrecht Kossel discovered histones. In 1888, Sutton and Boveri proposed the theory of continuity of chromatin during the cell cycle In 1889, Wilhelm von Waldemeyer created the term "chromosome". In 1928, Emil Heitz coined the terms heterochromatin and euchromatin. In 1942, Conrad Waddington postulated the epigenetic landscapes. In 1948, Rollin Hotchkiss discovered DNA methylation. In 1953, Watson and Crick reported the double helix structure of DNA based on Rosalind Franklin's X-ray diffraction images. In 1961, Mary Lyon postulated the principle of X-inactivation. In 1973/1974, chromatin fiber was discovered. In 1975, Pierre Chambon coined the te
https://en.wikipedia.org/wiki/Systematic%20and%20Evolutionary%20Biogeography%20Association
The Systematic and Evolutionary Biogeographical Association (SEBA) promotes an open and diverse international biogeographical community by assisting in sharing biogeographical information. Enhancing communication between biogeographers of all countries, SEBA contributes to the development of the theory and practice of systematic and evolutionary biogeography. SEBA was established in 2006. It is a non-profit organization and a scientific member of the International Union of Biological Sciences. SEBA promotes the International Code of Area Nomenclature (ICAN), a standardized system of biogeographical reference.
https://en.wikipedia.org/wiki/International%20Code%20of%20Area%20Nomenclature
The International Code of Area Nomenclature (ICAN) was proposed by a group of few biogeographers to provide a universal naming system or nomenclature for areas of endemism used in contemporary biogeography. There are other proposals to palaeobiogeographic areas. The ICAN also serves as the international standard rules for proposing and using "area of endemism" names. The ICAN was ratified by the Systematic and Evolutionary Biogeographical Association (SEBA) in Paris in the only meeting of the group, during July 2007. This community was created and maintained by the same group of biogeographers that proposed the ICAN. There is little agreement on the use of different methods and systems of nomenclature in Biogeography, and this proposal represents one of the various kinds of study in this science.
https://en.wikipedia.org/wiki/Isoguanine
Isoguanine or 2-hydroxyadenine is a purine base that is an isomer of guanine. It is a product of oxidative damage to DNA and has been shown to cause mutation. It is also used in combination with isocytosine in studies of unnatural nucleic acid analogues of the normal base pairs in DNA. It is used as a nucleobase of hachimoji nucleic acids. In hachimoji DNA, it pairs with 1-methylcytosine, while in hachimoji RNA, it pairs with isocytosine.
https://en.wikipedia.org/wiki/Isocytosine
Isocytosine or 2-aminouracil is a pyrimidine base that is an isomer of cytosine. It is used in combination with isoguanine in studies of unnatural nucleic acid analogues of the normal base pairs in DNA. In particular, it is used as a nucleobase of hachimoji RNA. It can be synthesized from guanidine and malic acid. It is also used in physical chemical studies involving metal complex binding, hydrogen bonding, and tautomerism and proton transfer effects in nucleobases.
https://en.wikipedia.org/wiki/1%3A72%20scale
1:72 scale is a scale used for scale models, most commonly model aircraft, where one inch on the model equals six feet (which is seventy-two inches) in real life. The scale is popular for aircraft because sizes ranging from small fighters to large bombers are all reasonably manageable and displayable. History This scale originated with the Skybirds and Frog Penguin aircraft model ranges produced in the United Kingdom during the 1930s and was subsequently used for aircraft recognition models by the Allies of World War II. More subjects and genres are covered in this scale than any other. Widespread in Western and Eastern Europe, Japan and Latin America, the scale is extremely popular in the UK but less so in Australia, Canada, and the US, where 1:48 is more prevalent. Manufacturers of aircraft kits, past and present, in 1:72 scale include; Airfix, Frog, Novo, Revell, Italeri, Tamiya, Hasegawa, Heller, Monogram and Matchbox. In recent years this scale has also gained popularity for model military vehicles, and is also used for Japanese anime science fiction models, die-cast model cars, figures, and radio-control model ships, as well as soldiers. Die-cast models 1:72 scale is becoming increasingly popular for die-cast toys. The beginning of die-cast production in this scale can be connected to the noted Hong Kong model car producer Hongwell Company, which opened its factory dedicated to these model cars in 1997. The success of this scale in the die-cast market has multiple reasons. One of them is that these models may be precisely designed using Zamak as a basic alloy, which is a real advantage in contrast to 1:87. The other reason is that these models are sold in perspex boxes that can easily be organized into columns, thus helping collectors to keep their collection in good condition. Unlike 1:64, 1:72 die-cast models are strictly produced in true scale even if they do not fit the original box size. The number of complementary plastic parts (lamps, mirrors)
https://en.wikipedia.org/wiki/Interdental%20papilla
The interdental papilla, also known as the interdental gingiva, is the part of the gums (gingiva) that exists coronal to the free gingival margin on the mesial and distal surfaces of the teeth. The interdental papillae fill in the area between the teeth apical to their contact areas to prevent food impaction; they assume a conical shape for the anterior teeth and a blunted shape buccolingually for the posterior teeth. A missing papilla is often visible as a small triangular gap between adjacent teeth. The relationship of interdental bone to the interproximal contact point between adjacent teeth is a determining factor in whether the interdental papilla will be present. If greater than 8mm exist between the interdental bone and the interproximal contact, usually no papilla will be present. If the distance is 5mm or less, then a papilla will almost always be present. With active periodontal disease, both the marginal gingiva and attached gingiva can become enlarged, especially the interdental papillae. This enlargement results from edema occurring in the lamina propria of the tissue caused by the inflammatory response.
https://en.wikipedia.org/wiki/Bcl-2-like%20protein%201
Bcl-2-like protein 1 is a protein encoded in humans by the BCL2L1 gene. Through alternative splicing, the gene encodes both of the human proteins Bcl-xL and Bcl-xS. Function The protein encoded by this gene belongs to the Bcl-2 protein family. Bcl-2 family members form hetero- or homodimers and act as anti- or pro-apoptotic regulators that are involved in a wide variety of cellular activities. The proteins encoded by this gene are located at the outer mitochondrial membrane, and have been shown to regulate outer mitochondrial membrane channel (voltage-dependent anion channels (VDACs) opening. VDACs regulate mitochondrial membrane potential, and thus controls the production of reactive oxygen species and release of cytochrome C by mitochondria, both of which are the potent inducers of cell apoptosis. Two alternatively spliced transcript variants, which encode distinct isoforms, have been reported. The longer isoform (Bcl-xL) acts as an apoptotic inhibitor and the shorter form (Bcl-xS) acts as an apoptotic activator. Interactions BCL2-like 1 (gene) has been shown to interact with: APAF1, BAK1, BCAP31, BCL2L11, BNIP3, BNIPL, BAD, BAX, BIK, Bcl-2, HRK, IKZF3, Noxa, PPP1CA, PSEN2 RAD9A, RTN1, RTN4, and VDAC1.
https://en.wikipedia.org/wiki/Eirpac
EIRPAC is Ireland's packet switched X.25 data network. It replaced Euronet in 1984. Eirpac uses the DNIC 2724. HEAnet was first in operation via X.25 4.8Kb Eirpac connections back in 1985. By 1991 most Universities in Ireland used 64k Eirpac VPN connections. Today Eirpac is owned and operated by Eircom but does not accept new applications for Eirpac: no reference is made on the products-offering on their website They began the process of migrating existing customers using more capable forms of telecommunications back in late April 2004. In 2001 Eirpac had approximately 5,000 customers dialing in daily via switched virtual circuits although those numbers have been declining rapidly. Eirpac is still an important element for data transfer in Ireland with numerous banks (automatic teller machines), telecoms switches, pager systems and other networks that utilise permanent virtual circuits. Connecting to Eirpac can be done using a simple AT compatible modem. The dial in number is 1511 + baud rate. So for example to connect at 28,800 bit/s would be ATDT 15112880. The user would then have to authenticate with their Eirpac NUI. The NUI (Network User Identification) consists of a name and password provided by Eir. Sources External links Official website Computer networking Internet in Ireland
https://en.wikipedia.org/wiki/Centroidal%20Voronoi%20tessellation
In geometry, a centroidal Voronoi tessellation (CVT) is a special type of Voronoi tessellation in which the generating point of each Voronoi cell is also its centroid (center of mass). It can be viewed as an optimal partition corresponding to an optimal distribution of generators. A number of algorithms can be used to generate centroidal Voronoi tessellations, including Lloyd's algorithm for K-means clustering or Quasi-Newton methods like BFGS. Proofs Gersho's conjecture, proven for one and two dimensions, says that "asymptotically speaking, all cells of the optimal CVT, while forming a tessellation, are congruent to a basic cell which depends on the dimension." In two dimensions, the basic cell for the optimal CVT is a regular hexagon as it is proven to be the most dense packing of circles in 2D Euclidean space. Its three dimensional equivalent is the rhombic dodecahedral honeycomb, derived from the most dense packing of spheres in 3D Euclidean space. Applications Centroidal Voronoi tessellations are useful in data compression, optimal quadrature, optimal quantization, clustering, and optimal mesh generation. A weighted centroidal Voronoi diagrams is a CVT in which each centroid is weighted according to a certain function. For example, a grayscale image can be used as a density function to weight the points of a CVT, as a way to create digital stippling. Occurrence in nature Many patterns seen in nature are closely approximated by a centroidal Voronoi tessellation. Examples of this include the Giant's Causeway, the cells of the cornea, and the breeding pits of the male tilapia.
https://en.wikipedia.org/wiki/Mucin%20short%20variant%20S1
Mucin short variant S1, also called polymorphic epithelial mucin (PEM) or epithelial membrane antigen (EMA), is a mucin encoded by the MUC1 gene in humans. Mucin short variant S1 is a glycoprotein with extensive O-linked glycosylation of its extracellular domain. Mucins line the apical surface of epithelial cells in the lungs, stomach, intestines, eyes and several other organs. Mucins protect the body from infection by pathogen binding to oligosaccharides in the extracellular domain, preventing the pathogen from reaching the cell surface. Overexpression of MUC1 is often associated with colon, breast, ovarian, lung and pancreatic cancers. Joyce Taylor-Papadimitriou identified and characterised the antigen during her work with breast and ovarian tumors. Structure MUC1 is a member of the mucin family and encodes a membrane bound, glycosylated phosphoprotein. MUC1 has a core protein mass of 120-225 kDa which increases to 250-500 kDa with glycosylation. It extends 200-500 nm beyond the surface of the cell. The protein is anchored to the apical surface of many epithelia by a transmembrane domain. Beyond the transmembrane domain is a SEA domain that contains a cleavage site for release of the large extracellular domain. The release of mucins is performed by sheddases. The extracellular domain includes a 20 amino acid variable number tandem repeat (VNTR) domain, with the number of repeats varying from 20 to 120 in different individuals. These repeats are rich in serine, threonine and proline residues which permits heavy o-glycosylation. Multiple alternatively spliced transcript variants that encode different isoforms of this gene have been reported, but the full-length nature of only some has been determined. MUC1 is cleaved in the endoplasmic reticulum into two pieces, the cytoplasmic tail including the transmembrane domain and the extracellular domain. These domains tightly associate in a non-covalent fashion. This tight, non-covalent association is not broken by
https://en.wikipedia.org/wiki/Journal%20of%20Physics%20G
Journal of Physics G: Nuclear and Particle Physics is a peer-reviewed journal that publishes theoretical and experimental research into nuclear physics, particle physics and particle astrophysics, including all interface areas between these fields. The editor-in-chief is Jacek Dobaczewski, University of York, UK. Scope The journal publishes research articles on: theoretical and experimental topics in the physics of elementary particles and fields; intermediate-energy physics and nuclear physics; experimental and theoretical research in particle, neutrino, and nuclear astrophysics; research arising from all interface areas among these fields. Research is published in the following formats: Research Papers: Reports of original and high-quality research work; Research Notes: Contributions from individuals (or small groups) within large collaborations, containing early results of analyses, detector development, simulations, etc. which might not otherwise be published in the wider literature; Topical Reviews: Specially commissioned review articles on areas of current interest; LabTalk: Article summaries written by the researchers themselves which introduce the findings, techniques, and possible applications of their research. Abstracting and indexing information The journal is indexed in INSPEC Information Services, ISI (Science Citation Index, SciSearch, ISI Alerting Services, Current Contents/Physical, Chemical and Earth Sciences), Article@INIST, and Chemical Abstracts.
https://en.wikipedia.org/wiki/Soan%20papdi
Sohan papdi (also known as san papri, shompapri, sohan papdi, shonpapdi or patisa) is a popular dessert in the Indian subcontinent. Soan Papdi has no confirmed origin, but it is believed to have originated in the western state Maharashtra, India. According to a culinary anthropologists, the word "soan" has Persian origin and the name comes from the term sohan pashmaki. Others speculation also includes that ranges from Rajasthan, Uttar Pradesh, Punjab and West Bengal. It bears some resemblance to the Dragon's beard candy in China, Kkul-tarae in Korea, Persian pashmaki, which gave rise to the Turkish pişmaniye. It is usually cube-shaped or served as flakes, and has a crisp and flaky texture. Traditionally sold loose in rolled paper cones, modern industrial production has led it to being sold in the form of tightly formed cubes. Ingredients Its main ingredients are sugar, gram flour, flour, ghee, almond, milk, and cardamom. See also Sohan (confectionery) Sohan halwa Pişmaniye, a similar Turkish dessert Pashmak, a similar Iranian dessert
https://en.wikipedia.org/wiki/FlexCAN
FlexCAN is an embedded network architecture that extends Controller Area Network (CAN). It was designed by Dr. Juan Pimentel at Kettering University. It was inspired by FlexRay and the need to provide more deterministic behavior over the CAN network. Its focus is on redundancy at the hardware level, and time-based prioritized communication at the protocol level. Benefits of FlexCAN FlexCAN is similar to CAN except for the following improvements: Deterministic Behavior Increased reliability External links FlexCAN: A Flexible Architecture for Highly Dependable Embedded Applications Serial buses CAN bus
https://en.wikipedia.org/wiki/Gla%20domain
Vitamin K-dependent carboxylation/gamma-carboxyglutamic (GLA) domain is a protein domain that contains post-translational modifications of many glutamate residues by vitamin K-dependent carboxylation to form γ-carboxyglutamate (Gla). Proteins with this domain are known informally as Gla proteins. The Gla residues are responsible for the high-affinity binding of calcium ions. The GLA domain binds calcium ions by chelating them between two carboxylic acid residues. These residues are part of a region that starts at the N-terminal extremity of the mature form of Gla proteins, and that ends with a conserved aromatic residue. This results in a conserved Gla-x(3)-Gla-x-Cys motif that is found in the middle of the domain, and which seems to be important for substrate recognition by the carboxylase. The 3D structures of several Gla domains have been solved. Calcium ions induce conformational changes in the Gla domain and are necessary for the Gla domain to fold properly. A common structural feature of functional Gla domains is the clustering of N-terminal hydrophobic residues into a hydrophobic patch that mediates interaction with the cell surface membrane. At present, the following human Gla-containing proteins (Gla proteins) have been characterized to the level of primary structure: the blood coagulation factors II (prothrombin), VII, IX, and X, the anticoagulant proteins C and S, and the factor X-targeting protein Z. The bone Gla protein osteocalcin, the calcification-inhibiting matrix Gla protein (MGP), the cell growth regulating "growth arrest specific gene 6" protein GAS6, periostin (a factor necessary for migration and adhesion of epithelial cells), plus two proline-rich Gla-proteins (PRGPs) and two transmembrane Gla proteins (TMGPs), the functions of which are unknown. In all cases in which their function was known, the presence of the Gla residues in these proteins turned out to be essential for functional activity. Subfamilies Coagulation factor, Gla region
https://en.wikipedia.org/wiki/Youden%27s%20J%20statistic
Youden's J statistic (also called Youden's index) is a single statistic that captures the performance of a dichotomous diagnostic test. (Bookmaker) Informedness is its generalization to the multiclass case and estimates the probability of an informed decision. Definition Youden's J statistic is with the two right-hand quantities being sensitivity and specificity. Thus the expanded formula is: The index was suggested by W. J. Youden in 1950 as a way of summarising the performance of a diagnostic test; however, the formula was earlier published in Science by C. S. Pierce in 1884. Its value ranges from -1 through 1 (inclusive), and has a zero value when a diagnostic test gives the same proportion of positive results for groups with and without the disease, i.e the test is useless. A value of 1 indicates that there are no false positives or false negatives, i.e. the test is perfect. The index gives equal weight to false positive and false negative values, so all tests with the same value of the index give the same proportion of total misclassified results. While it is possible to obtain a value of less than zero from this equation, e.g. Classification yields only False Positives and False Negatives, a value of less than zero just indicates that the positive and negative labels have been switched. After correcting the labels the result will then be in the 0 through 1 range. Youden's index is often used in conjunction with receiver operating characteristic (ROC) analysis. The index is defined for all points of an ROC curve, and the maximum value of the index may be used as a criterion for selecting the optimum cut-off point when a diagnostic test gives a numeric rather than a dichotomous result. The index is represented graphically as the height above the chance line, and it is also equivalent to the area under the curve subtended by a single operating point. Youden's index is also known as deltaP' and generalizes from the dichotomous to the multiclass case as inf
https://en.wikipedia.org/wiki/DEVS
DEVS abbreviating Discrete Event System Specification is a modular and hierarchical formalism for modeling and analyzing general systems that can be discrete event systems which might be described by state transition tables, and continuous state systems which might be described by differential equations, and hybrid continuous state and discrete event systems. DEVS is a timed event system. History DEVS is a formalism for modeling and analysis of discrete event systems (DESs). The DEVS formalism was invented by Bernard P. Zeigler, who is emeritus professor at the University of Arizona. DEVS was introduced to the public in Zeigler's first book, Theory of Modeling and Simulation , in 1976, while Zeigler was an associate professor at University of Michigan. DEVS can be seen as an extension of the Moore machine formalism, which is a finite state automaton where the outputs are determined by the current state alone (and do not depend directly on the input). The extension was done by associating a lifespan with each state [Zeigler76], providing a hierarchical concept with an operation, called coupling [Zeigler84]. Since the lifespan of each state is a real number (more precisely, non-negative real) or infinity, it is distinguished from discrete time systems, sequential machines, and Moore machines, in which time is determined by a tick time multiplied by non-negative integers. Moreover, the lifespan can be a random variable; for example the lifespan of a given state can be distributed exponentially or uniformly. The state transition and output functions of DEVS can also be stochastic. Zeigler proposed a hierarchical algorithm for DEVS model simulation in 1984 [Zeigler84] which was published in Simulation journal in 1987. Since then, many extended formalism from DEVS have been introduced with their own purposes: DESS/DEVS for combined continuous and discrete event systems, P-DEVS for parallel DESs, G-DEVS for piecewise continuous state trajectory modeling of DESs, RT
https://en.wikipedia.org/wiki/Kleinschmidt%20Inc
Kleinschmidt Inc. was established in 1931 by Edward Kleinschmidt. It is a privately owned firm that provides electronic commerce, electronic data interchange, and value-added network services. Its headquarters are in Deerfield, Illinois. Edward Kleinschmidt was one of the inventors of the teleprinter, one of the first electronic commerce devices. History 1893 - Edward Ernst Kleinschmidt started working with telegraphy; 1898 - Edward E Kleinschmidt opened his own experimental shop; 1906 - George Seely joined Kleinschmidt’s shop with a partially developed block system for electric trolley car railways; 1910 - Exhibited at the Association of American Railroads Communications Convention; 1910 - Kleinschmidt started to receive multiple patents; 1914 - Kleinschmidt Electric Company was founded; 1924 - Kleinschmidt Electric merged with the Morkrum Company to form Morkrum-Kleinschmidt Corporation; 1928 - The company name was changed to Teletype Corporation; 1930 - The Teletype Corporation was sold to AT&T for $30,000,000 in stock; 1931 - Kleinschmidt Laboratories Inc. was founded; 1944 - Edward E. Kleinschmidt demonstrated his lightweight teleprinter at the Chief Signal Officer; 1949 - The Kleinschmidt 100-words-per-minute typebar page printer was made the standard for the Military; 1956 - Kleinschmidt Laboratories Inc. merged with Smith Corona which merged with Marchant Calculating Machine Company shortly thereafter, forming SCM; 1979 - Started to provide Electronic Data Interchange (EDI) and Car Location Message (CLM) services; 1986 - Hanson Trust acquired SCM Corporation. Harry S. Gaples, then Kleinschmidt division president, purchased the division from Hanson Trust. See also Edward Kleinschmidt Electronic commerce Electronic Data Interchange Enterprise Application Integration Supply Chain Management Teletypewriter Value-added network
https://en.wikipedia.org/wiki/Sodium%20bicarbonate%20rocket
A sodium bicarbonate rocket (sometimes called an Alka-Seltzer rocket) is a model rocket fashioned from a 35mm film canister and propelled by the pressure of a gas, often carbon dioxide, generated from the reaction of an acid with sodium bicarbonate. Sodium bicarbonate rockets are often used in science classes to demonstrate principles of chemistry and physics. In the experiment, a film canister is filled with water, an effervescent tablet (commonly Alka-Seltzer) is added and the canister tightly sealed. After a short time, the pressure of the carbon dioxide is great enough to cause the body of the canister to be launched into the air with a popping sound. The canister may be embellished with paper fins to resemble more closely a real rocket. Various experiments and lessons can center around the use of a bicarbonate rocket. For example, students are sometimes asked to experiment with the amounts of water and Alka-Seltzer to find the combination which propels the rocket the greatest distance. Alternatively they may derive equations to calculate the speed and velocity of the rocket from the distance it travels. In rocketry, a chemical reaction rapidly creates gas that is expelled in one direction from its container (the rocket engine); momentum forces the rocket in the opposite direction. The alka-seltzer rocket experiment demonstrates Newton's third law. The film canister rocket has a buildup of gas that wants to come out of the weakest spot making all the gas come out at once through the hole at the bottom. The gas comes out from the underside and pushes the rocket up. After it gets pushed up, air resistance slows it down and gravity pulls it down to earth. The film canister accelerates quickly because it has very little mass. The film canister rocket uses a solid fuel mixed with a liquid fuel to create a gas that escapes out of the bottom. The gas is carbon dioxide (CO2), the liquid is water (H2O), and the solid is an effervescent tablet. When the H2O is mixed wi
https://en.wikipedia.org/wiki/Genie%20TV
Genie TV (; formerly Olleh TV) is an IPTV service provided by South Korea's main telecom operator KT. It broadcasts VOD including TV programs, movies, children's programs, sports, documentaries and animations. On 4 October 2022, KT rebranded its IPTV service as Genie TV, 11 years after the Olleh TV brand was introduced. The rebranding is part of the telecom's aggressive effort to push forward its content business as growth in the traditional telecom sector slows. Olleh TV for PlayStation 3 Since November 20, 2007, following a partnership between KT and Sony Computer Entertainment, Olleh TV is made available for PlayStation 3 owners in South Korea. Once the PlayStation 3's firmware version 2.10 installed in the system, a new icon, called "TV", appears in the XMB. The user is asked to agree a terms of service form then to download the 13MB Olleh TV player. Olleh TV is also a provider of Spanish television programming. Olleh TV signed a deal with HDTV Inc. to bring three new markets into the Spanish broadcasting circuit. According to Pr Newswire, “This new market includes Tampa, Florida; Charleston, South Carolina; and Palm Springs, California.” The purpose of signing this agreement it to attract a wide range of Hispanic audiences. “Olleh TV already serves eight markets reaching over 3.5 million Hispanic viewers.” “There are many Hispanic shows slated to air on these affiliates. The popular shows include ‘Papparazzi TV Sensacional,’ ‘Maria Elvira Live!’ ‘Esta Noche Tu Night’ with Alexis Valdes and ‘Los Implicados,’ among others." Olleh TV is sure to be a hit among the Hispanic community because they are commitment to providing programming catered to the Hispanic community and culture.
https://en.wikipedia.org/wiki/External%20memory%20interface
An external memory interface is a bus protocol for communication from an integrated circuit, such as a microprocessor, to an external memory device located on a circuit board. The memory is referred to as external because it is not contained within the internal circuitry of the integrated circuit and thus is externally located on the circuit board. The external memory interface enables the processor to interface with third level caches, peripherals, and external memory. Some common external memory interfaces include: DDR DDR2 GDDR
https://en.wikipedia.org/wiki/Security%20log
A security log is used to track security-related information on a computer system. Examples include: Windows Security Log Internet Connection Firewall security log According to Stefan Axelsson, "Most UNIX installations do not run any form of security logging software, mainly because the security logging facilities are expensive in terms of disk storage, processing time, and the cost associated with analyzing the audit trail, either manually or by special software." See also Audit trail Server log Log management and intelligence Web log analysis software Web counter Data logging Common Log Format Syslog
https://en.wikipedia.org/wiki/Flashsort
Flashsort is a distribution sorting algorithm showing linear computational complexity for uniformly distributed data sets and relatively little additional memory requirement. The original work was published in 1998 by Karl-Dietrich Neubert. Concept Flashsort is an efficient in-place implementation of histogram sort, itself a type of bucket sort. It assigns each of the input elements to one of buckets, efficiently rearranges the input to place the buckets in the correct order, then sorts each bucket. The original algorithm sorts an input array as follows: Using a first pass over the input or a priori knowledge, find the minimum and maximum sort keys. Linearly divide the range into buckets. Make one pass over the input, counting the number of elements which fall into each bucket. (Neubert calls the buckets "classes" and the assignment of elements to their buckets "classification".) Convert the counts of elements in each bucket to a prefix sum, where is the number of elements in bucket or less. ( and .) Rearrange the input so all elements of each bucket are stored in positions where . Sort each bucket using insertion sort. Steps 1–3 and 6 are common to any bucket sort, and can be improved using techniques generic to bucket sorts. In particular, the goal is for the buckets to be of approximately equal size ( elements each), with the ideal being division into quantiles. While the basic algorithm is a linear interpolation sort, if the input distribution is known to be non-uniform, a non-linear division will more closely approximate this ideal. Likewise, the final sort can use any of a number of techniques, including a recursive flash sort. What distinguishes flash sort is step 5: an efficient in-place algorithm for collecting the elements of each bucket together in the correct relative order using only words of additional memory. Memory efficient implementation The Flashsort rearrangement phase operates in cycles. Elements start out "un
https://en.wikipedia.org/wiki/TI-RTOS
TI-RTOS is an embedded tools ecosystem created and offered by Texas Instruments (TI) for use across a range of their embedded system processors. It includes a real-time operating system (RTOS) component named TI-RTOS Kernel (formerly named SYS/BIOS, which evolved from DSP/BIOS), networking connectivity stacks, power management, file systems, instrumentation, and inter-processor communications like DSP/BIOS Link. It is free and open-source software, released under a BSD license. TI-RTOS can be used within TI's Code Composer Studio integrated development environment (IDE), IAR Systems' IAR Embedded Workbench, and the GNU Compiler Collection (GCC). Separate versions of TI-RTOS are provided to support TI's MSP43x (including MSP432), SimpleLink Wireless MCU, Sitara, Tiva C, C2000, and C6000 lines of embedded devices. TI-RTOS provides system services to an embedded application such as preemptive multitasking, memory management and real-time analysis. TI-RTOS can be used in different microprocessors, with different processing and memory constraints. It is supported by Secure Sockets Layer (SSL) and Transport Layer Security (TLS) libraries such as wolfSSL. History The roots of TI-RTOS were originally developed by Spectron Microsystems (a subsidiary of Dialogic Corporation) as the first RTOS developed specifically for digital signal processors and was named SPOX. Spectron eventually also developed a second product named BIOSuite that included a real-time kernel and various associated tools. Spectron Microsystems was eventually acquired by Texas Instruments and the SPOX and BIOSuite products were merged into one microkernel product named DSP/BIOS. The DSP/BIOS RTOS product underwent significant changes to its application programming interface (API) in version 6.0. With the release of version 6.3 in August 2010, DSP/BIOS was renamed SYS/BIOS to reflect its support for microcontrollers beyond DSPs. With the release of version 6.40 in April 2014, SYS/BIOS was renamed TI-RTO
https://en.wikipedia.org/wiki/Queue%20automaton
A queue machine, queue automaton, or pullup automaton (PUA) is a finite state machine with the ability to store and retrieve data from an infinite-memory queue. It is a model of computation equivalent to a Turing machine, and therefore it can process the same class of formal languages. Theory A queue machine can be defined as a six-tuple where is a finite set of states; is the finite set of the input alphabet; is the finite queue alphabet; is the initial queue symbol; is the start state; is the transition function. A machine configuration is an ordered pair of its state and queue contents , where denotes the Kleene closure of . The starting configuration on an input string is defined as , and the transition from one configuration to the next is defined as: where is a symbol from the queue alphabet, is a sequence of queue symbols (), and . Note the "first-in-first-out" property of the queue in the relation. The machine accepts a string if after a finite number of transitions the starting configuration evolves to exhaust the string (reaching the null string ), or otherwise stated, if Turing completeness We can prove that a queue machine is equivalent to a Turing machine by showing that a queue machine can simulate a Turing machine and vice versa. A Turing machine can be simulated by a queue machine that keeps a copy of the Turing machine's contents in its queue at all times, with two special markers: one for the Turing machine's head position, and one for the end of the tape; its transitions simulate those of the Turing machine by running through the whole queue, popping off each of its symbols and re-enqueing either the popped symbol, or, near the head position, the equivalent of the Turing machine transition's effect. A queue machine can be simulated by a Turing machine, but more easily by a multi-tape Turing machine, which is known to be equivalent to a normal single-tape machine. The simulating queue machine reads input on one tape an
https://en.wikipedia.org/wiki/R%C3%BCdiger%20Gamm
Rüdiger Gamm (born 10 July 1971) is a German "mental calculator". He attained the ability to mentally evaluate large arithmetic expressions at the age of 21. He can also speak backwards, and calculate calendars. Featured on the Discovery Channel program The Real Superhumans, he was examined by Allan Snyder, an expert on savants, who concluded that Gamm's ability was not a result of savant syndrome but connected to genetics. In terms of mental calculations, Rüdiger's most notable talent is the ability to memorize large powers. In the 2008 Mental Calculation World Cup in Leipzig, he recited 81100, which took approximately 2 minutes and 30 seconds. In the tournament itself, he performed strongly, finishing in 5th position overall. He also held a seminar in 2012 at the BOLDTalks event at DUCTAC (Dubai). Early life Rüdiger Gamm was born on 10 July 1971 in Welzheim, Germany. Gamm stated that he learnt how to speak backwards before learning how to speak forwards which prompted classmates to tease him or avoid him. Gamm was a self-proclaimed underachiever at school and stated "I was the worst in my class at maths. I failed my exam six times and hated school a lot. The only thing I was interested in was bodybuilding. I wanted to be like Arnold Schwarzenegger, rather than a mathematician." Gamm recalls, shortly after leaving college, listening to the radio and calculating alongside a champion mathematician. After Gamm answered the calculations faster than the champion, he started training his brain into the field of mental math and, a year later, appeared on the German TV game show Wetten, dass..? (in the United States it is known as Wanna Bet? and in the United Kingdom known as You Bet!.) Gamm won the show with the highest score achieved and received a prize of 8,400 Deutsche Marks.
https://en.wikipedia.org/wiki/Anglo-French%20Conference%20on%20Time-keeping%20at%20Sea
The Anglo-French Conference on Time-keeping at Sea was a conference held in London in June 1917. History The Conference established the nautical date line and adopted an ideal form of the terrestrial time zone system for use at sea. It recommended that time changes required by changes of longitude be made in one-hour steps. This recommendation was adopted between 1920 and 1925 by all major fleets, including British, French and American. The rules applied to almost all naval ships and to many non-naval ships. Nevertheless, up to the Second World War, the old practice of keeping local apparent time prevailed on many independent merchant ships. Current usage The nautical date line is implied but not explicitly drawn on time zone maps. It follows the 180° meridian except where it is interrupted by territorial waters adjacent to land, forming gaps: it is a pole-to-pole dashed line. Ships are required to adopt the standard time of a country when they are within its territorial waters, but must revert to international time zones (15° wide pole-to-pole gores) as soon as they leave territorial waters. The 15° gore that is offset from GMT or UT1 (not UTC) by twelve hours is bisected by the nautical date line into two 7.5° gores that differ from GMT by ±12 hours. In reality nautical time zones are used only for radio communication etc. Internally on the ship, e.g. for work and meal hours, the ship may use a suitable time of its own choosing. This includes fixed installations such as oil rigs. For example, the Norwegian Ekofisk oil rigs are located in international water at longitude 3°E but use Norwegian time. See also Marine chronometer Meridians (geography) France–United Kingdom relations 1917 in the United Kingdom 1917 in France 20th-century diplomatic conferences Time zones Nautical terminology
https://en.wikipedia.org/wiki/Nautical%20time
Nautical time is a maritime time standard established in the 1920s to allow ships on high seas to coordinate their local time with other ships, consistent with a long nautical tradition of accurate celestial navigation. Nautical time divides the globe into 24 nautical time zones with hourly clock offsets, spaced at 15 degrees by longitudinal coordinate, with no political deviation. Nautical time keeping dates back to the early 20th century as a standard way to keep time at sea, although it largely only applied to military fleets pre–World War 2. This time-keeping method is only used for radio communications and to account for slight inaccuracies that using Greenwich Mean Time (GMT) may lead to during navigation of the high seas. It is typically only used for trans-oceanic travel, as captains will often not change the timekeeping for short distances such as channels or inland seas. History of nautical time Establishment Prior to 1920, ships kept solar time on the high seas by setting their clocks at night or at the morning sight so that, given the ship's speed and direction, it would be 12 o'clock when the sun crossed the ship's meridian. The establishment of nautical standard times, nautical standard time zones and the nautical date line were recommended by the Anglo-French Conference on Time-keeping at Sea in 1917. The conference recommended that the standard apply to all ships, both military and civilian. These zones were adopted by all major fleets between 1920 and 1925 but not by many independent merchant ships until World War II. Letter suffixes Around 1950, a letter suffix was added to the zone description, assigning Z to the zero zone, and A–M (except J) to the east and N–Y to the west (J may be assigned to local time in non-nautical applications – zones M and Y have the same clock time but differ by 24 hours: a full day). These can be vocalized using the NATO phonetic alphabet which pronounces the letter Z as Zulu, leading to the use of the term "Zul
https://en.wikipedia.org/wiki/Urethral%20sphincters
The urethral sphincters are two muscles used to control the exit of urine in the urinary bladder through the urethra. The two muscles are either the male or female external urethral sphincter and the internal urethral sphincter. When either of these muscles contracts, the urethra is sealed shut. The external urethral sphincter originates at the ischiopubic ramus and inserts into the intermeshing muscle fibers from the other side. It is controlled by the deep perineal branch of the pudendal nerve. Activity in the nerve fibers constricts the urethra. The internal sphincter muscle of urethra: located at the bladder's inferior end and the urethra's proximal end at the junction of the urethra with the urinary bladder. The internal sphincter is a continuation of the detrusor muscle and is made of smooth muscle, therefore it is under involuntary or autonomic control. This is the primary muscle for prohibiting the release of urine. The female or male external sphincter muscle of urethra (sphincter urethrae): located in the deep perineal pouch, at the bladder's distal inferior end in females, and inferior to the prostate (at the level of the membranous urethra) in males. It is a secondary sphincter to control the flow of urine through the urethra. Unlike the internal sphincter muscle, the external sphincter is made of skeletal muscle, therefore it is under voluntary control of the somatic nervous system. Function and sex differences In males and females, both internal and external urethral sphincters function to prevent the release of urine. The internal urethral sphincter controls involuntary urine flow from the bladder to the urethra, whereas the external urethral sphincter controls voluntary urine flow from the bladder to the urethra. Any damage to these muscles can lead to urinary incontinence. In males, the internal urethral sphincter has the additional function of preventing the flow of semen into the male bladder during ejaculation. Females do have a more el
https://en.wikipedia.org/wiki/Internal%20urethral%20sphincter
The internal urethral sphincter is a urethral sphincter muscle which constricts the internal urethral orifice. It is located at the junction of the urethra with the urinary bladder and is continuous with the detrusor muscle, but anatomically and functionally fully independent from it. It is composed of smooth muscle, so it is under the control of the autonomic nervous system, specifically the sympathetic nervous system. Function This is the primary muscle for maintaining continence of urine, a function shared with the external urethral sphincter which is under voluntary control. It prevents urine leakage as the muscle is tonically contracted via sympathetic fibers traveling through the inferior hypogastric plexus and vesical nervous plexus. Specifically, it is controlled by the hypogastric nerve, predominantly via the alpha-1 adrenergic receptor. During urination, the preganglionic neurons of this sympathetic pathway are inhibited via signals arising in the pontine micturition center and traveling through the descending reticulospinal tracts, allowing the muscle to relax. During ejaculation, the muscle contracts to prevent reflux of semen into the urinary bladder, a phenomenon called retrograde ejaculation. Spasms of the internal urethral sphincter are associated with penile erection. Because the internal urethral sphincter is under involuntary control, it is believed to play a role in paruresis, in which a person who perceives oneself to be under observation is unable to urinate. See also Detrusor muscle External sphincter muscle of female urethra External sphincter muscle of male urethra
https://en.wikipedia.org/wiki/Odlyzko%E2%80%93Sch%C3%B6nhage%20algorithm
In mathematics, the Odlyzko–Schönhage algorithm is a fast algorithm for evaluating the Riemann zeta function at many points, introduced by . The main point is the use of the fast Fourier transform to speed up the evaluation of a finite Dirichlet series of length N at O(N) equally spaced values from O(N2) to O(N1+ε) steps (at the cost of storing O(N1+ε) intermediate values). The Riemann–Siegel formula used for calculating the Riemann zeta function with imaginary part T uses a finite Dirichlet series with about N = T1/2 terms, so when finding about N values of the Riemann zeta function it is sped up by a factor of about T1/2. This reduces the time to find the zeros of the zeta function with imaginary part at most T from about T3/2+ε steps to about T1+ε steps. The algorithm can be used not just for the Riemann zeta function, but also for many other functions given by Dirichlet series. The algorithm was used by to verify the Riemann hypothesis for the first 1013 zeros of the zeta function.
https://en.wikipedia.org/wiki/Laser%20Science%20and%20Technology%20Centre
Laser Science and Technology Centre (LASTEC) is a laboratory of the Defence Research & Development Organization (DRDO). Located in Delhi, it is the main DRDO lab involved in the development of lasers and related technologies. LASTEC functions under the DRDO Directorate of Electronics & Computer Science. History LASTEC is the oldest laboratory in DRDO. It was established in 1950 as Defence Science Laboratory (DSL) which was a nucleus laboratory with the objective to conduct research in frontier areas of physics, chemistry and mathematics with a special focus on lasers and opto-electronics. On 9 April 1960, DSL was shifted to Metcalfe House and inaugurated by then Defence Minister V. K. Krishna Menon, in the presence of Pt. Jawaharlal Nehru. Its first Director was S. P. Chakravarti, the father of Electronics and Telecommunication engineering in India, who later founded LRDE, DRDL and DLRL. With time, many DSC activities were given to newly formed, specialized DRDO laboratories. DSL served as a precursor for as many as 15 present DRDO labs, including DRDL, SSPL, INMAS, FRL, ISSA, DESIDOC, DIFR, SAG, ITM etc. In 1982, the laboratory moved to a new technical building in Metcalfe House complex and was renamed as Defence Science Centre. The centre consolidated its R&D activities towards more specific and application oriented areas, such as liquid fuel technology, spectroscopy, crystallography, system engineering, biotechnology etc. DSL now started concentrating in the area of pure sciences such as physics, chemistry and mathematics. The chemistry division developed G-fuel and unsymmetrical dimethylhydrazine (UDMH) for rockets and missiles. The Mathematics division conducted simulation and calculations related to the missile program, while the Physics division took the lead by launching many activities. DSC was also given a new charter of duties with its major thrust on lasers. Intensive work commenced on solid-state lasers, carbon dioxide lasers, ALARM, laser rangefi
https://en.wikipedia.org/wiki/Ethnic%20flag
An ethnic flag is a flag that symbolizes a certain ethnic group. Ethnic flags are often introduced to the ethnic community through the respective cultural or political ethnic movements. They are popular among diasporas, ethnic minorities, and some ethnic majorities, especially in multiethnic countries. History Like the concept of a state's national flag itself, that of an "ethnic flag" is modern, first arising in the late 19th century; strictly speaking, the national flags of nation states are themselves "ethnic flags", and often so used by ethnic minorities in neighbouring states, especially in the context of irredentism (e.g. the flag of the Republic of Albania used as an "ethnic Albanian flag" by Kosovar Albanians). Ethnic flags are often used in irredentism, representing the "national flag" of a proposed or unrecognized state. The first such flags were designed at the end of the 19th century, such as the Basque flag (1894) or the "Flag of Zion" used to symbolize Zionism from 1898, which became the national flag of Israel 50 years later. Most early ethnic flags imply a connection with an unrecognized state claimed by the respective ethnicities, such as the flag of Kurdistan which originates as the flag of the Republic of Ararat (1927). A flag of the Hispanic People was designed in 1932. Alternatively, an "ethnic flag" may represent a Pan-nationalism, such as the Pan-Arab flag which originates as the flag of the Arab Revolt during World War I, the proposed flag of Pan-Slavism (1848), Pan-Iranism or Pan-Turkism. The concept of using ethnic flags to symbolize ethnic groups within a multiethnic state, not necessarily connected with irredentism, became popular in the later 20th century, such as the Australian Aboriginal flag (1971), the Assyrian flag (1971), the flag of the Romani people (1971), the Berber flag (1970s), the Sami flag (1986) or the national Māori flag (1990). Designing ethnic or tribal flags has become very popular since the 1990s, especially
https://en.wikipedia.org/wiki/DSP/BIOS%20Link
DSP/BIOS Link or DSPLINK is an interprocessor or inter-process communication (IPC) scheme to pass messages and data in multiprocessing systems. In the case of the DaVinci digital signal processor (DSP) family from Texas Instruments, this scheme allows passing messages and data between an ARM client and a DSP server. DSPLINK can be used to implement a layer of software abstraction called a remote procedure call (RPC) that allows a remote function on the DSP to appear as local function calls in the ARM application code. The Codec Engine IPC communication layer is implemented using a RPC call scheme built on DSP/BIOS LINK. Overview DSP/BIOS LINK is implemented using shared memory and internal interrupts from the ARM to the DSP and vice versa. The shared memory protocol for IPC is implemented as follows: The ARM and DSP are programmed to a predetermined memory address where a message will be sent from the ARM to the DSP; and another for messages sent from the DSP to the ARM. One processor sends messages to the other by writing the message into the pre-determined address and then sending an interrupt to signal the other processor that a new message is available. When transferring data buffers, only a pointer to a given buffer needs to be passed since the buffer resides in shared memory that is accessible to both the processors. ARM buffer addresses must be translated into physical addresses when being presented to the DSP, as the DSP does not have an MMU or a concept of virtual addressing. Once the processor receiving the message has read it, it marks a flag in shared memory to indicate that the message memory is now available to be rewritten with another message. The DSP included in many DaVinci-based devices generally runs TI's DSP/BIOS RTOS. When multiple, heterogeneous cores are included in the device (e.g. DM644x), DSP/BIOS Link drivers run on both the ARM processor and the DSP to provide communication between the two. Operating systems A number of ARM9 ope
https://en.wikipedia.org/wiki/Nif%20gene
The nif genes are genes encoding enzymes involved in the fixation of atmospheric nitrogen into a form of nitrogen available to living organisms. The primary enzyme encoded by the nif genes is the nitrogenase complex which is in charge of converting atmospheric nitrogen (N2) to other nitrogen forms such as ammonia which the organism can use for various purposes. Besides the nitrogenase enzyme, the nif genes also encode a number of regulatory proteins involved in nitrogen fixation. The nif genes are found in both free-living nitrogen-fixing bacteria and in symbiotic bacteria associated with various plants. The expression of the nif genes is induced as a response to low concentrations of fixed nitrogen and oxygen concentrations (the low oxygen concentrations are actively maintained in the root environment of host plants). The first Rhizobium genes for nitrogen fixation (nif) and for nodulation (nod) were cloned in the early 1980s by Gary Ruvkun and Sharon R. Long in Frederick M. Ausubel's laboratory. Regulation In most bacteria, regulation of nif genes transcription is done by the nitrogen sensitive NifA protein. When there isn't enough fixed nitrogen available for the organism's use, NtrC triggers NifA expression, and NifA activates the rest of the nif genes. If there is a sufficient amount of reduced nitrogen or oxygen is present, another protein is activated: NifL. NifL inhibits NifA activity resulting in the inhibition of nitrogenase formation. NifL is regulated by the products of glnD and glnK. The nif genes can be found on bacterial chromosomes, but in symbiotic bacteria they are often found on plasmids or symbiosis islands with other genes related to nitrogen fixation (such as the nod genes). Examples in nature The expression and regulation of nif genes, while sharing common features in all or most of the nitrogen-fixing organisms in nature, have distinct characters and qualities that differ from one diazotroph to another. Examples of nif gene structure and
https://en.wikipedia.org/wiki/XDAIS%20algorithms
XDAIS or eXpressDsp Algorithm Interoperability Standard is a standard for algorithm development by Texas Instruments for the TMS320 DSP family. The standard was first introduced in 1999 and was created to facilitate integration of DSP algorithms into systems without re-engineering cost. The XDAIS standard address the issues of algorithm resource allocation and consumption on a DSP. Algorithms that comply with the standard are tested and awarded an "eXpressDSP-compliant" mark upon successful completion of the test. The standard consists of a set of general rules and guidelines that should be applied to all algorithms. For instance, all XDAIS compliant algorithms must implement an Algorithm Interface, called IALG. For those algorithms utilizing DMA, the IDMA interface must be implemented. Further, specific rules are provided for each family of TI DSP. Problems are often caused in algorithm by hard-coding access to system resources that are used by other algorithms. XDAIS prohibits the use of this type of hard-coding. Instead, XDAIS requires a standard API for the application to call a particular algorithm class. This API is defined in the xDM standard, also referred to as the VISA APIs (video, imaging, speech and audio). A XDAIS developer's kit provides the standard itself, example code, and a demonstration. Benefits of XDAIS over non-standardised approaches include: Significant reduction in integration time as algorithms do not trash each other's resources Easy comparison of algorithms from multiple different sources in the same application Access to broad range of compliant algorithms available from multiple TI DSP Third Parties eliminates need to custom develop complex algorithms Algorithms work out-of-the-box with eXpressDSP Multimedia Framework Products, such as Codec Engine (TI) See also eXpressDsp External links XpressDSP Algorithm Standard – xDAIS Developer’s Kit and xDM TMS320 DSP Algorithm Standard Developer’s Guide Texas Instruments Dig
https://en.wikipedia.org/wiki/Spontaneous%20potential
Spontaneous potentials are often measured down boreholes for formation evaluation in the oil and gas industry, and they can also be measured along the Earth's surface for mineral exploration or groundwater investigation. The phenomenon and its application to geology was first recognized by Conrad Schlumberger, Marcel Schlumberger, and E.G. Leonardon in 1931, and the first published examples were from Romanian oil fields. Physics Spontaneous potentials (SP) are usually caused by charge separation in clay or other minerals, due to presence of semi-permeable interface impeding the diffusion of ions through the pore space of rocks, or by natural flow of a conducting fluid through the rocks. The origin of SP across formation can be attributed to two processes involving the movement of ions: Streaming potential (Ek) Electrochemical potential (Ec) Streaming potential originates from the flow of an electrolyte (water) over naturally charged solids (i.e., surfaces that acquired electrokinetic or zeta potential). The streaming potential appears when mud filtrate is forced into the formation under the differential pressure between mud column and formation. The streaming potential is produced when the flow takes place across mud-cake in front of permeable formations, across permeable formations being invaded, and across shale beds. It is generally accepted that the streaming potential across the mud-cake is compensated by that across the shale. As such, in most cases, the spontaneous potential measured is only related to the electrochemical potential. Electrochemical potential (EC) is the sum of liquid junction or diffusion potential (EJ), and membrane potential (EM) Liquid junction potential is established at the direct contact of the mud filtrate and formation water at the edge of the invaded formation. Ions Na+ and Cl− diffuse from either solution to the other, but at different rate due to different mobilities. Na+ tends to be less mobile due to its affinity for water
https://en.wikipedia.org/wiki/Endophenotype
In genetic epidemiology, endophenotype (or intermediate phenotype) is a term used to separate behavioral symptoms into more stable phenotypes with a clear genetic connection. By seeing the EP notion as a special case of a larger collection of multivariate genetic models, which may be fitted using currently accessible methodology, it is possible to maximize its valuable potential lessons for etiological study in psychiatric disorders. The concept was coined by Bernard John and Kenneth R. Lewis in a 1966 paper attempting to explain the geographic distribution of grasshoppers. They claimed that the particular geographic distribution could not be explained by the obvious and external "exophenotype" of the grasshoppers, but instead must be explained by their microscopic and internal "endophenotype". The endophenotype idea represents the influence of two important conceptual currents in biology and psychology research. An adequate technology would be required to perceive the endophenotype, which represents an unobservable latent entity that cannot be directly observed with the unaided naked eye. In the investigation of anxiety and affective disorders, the endophenotype idea has gained popularity. The next major use of the term was in psychiatric genetics, to bridge the gap between high-level symptom presentation and low-level genetic variability, such as single nucleotide polymorphisms. It is therefore more applicable to more heritable disorders, such as bipolar disorder and schizophrenia. Through their impact on the growth and operation of the vital components of the nervous system, such as neurons, transmitter systems, and neural networks, genes have an impact on complex behavior. Therefore, heritable differences in mental abilities may be caused by changes in the code describing the shape and operation of the underlying neural network. One significant expression of this idea is believed to be the many cognitive deficiencies seen in ADHD, making them ideal candidates f
https://en.wikipedia.org/wiki/Kleene%27s%20T%20predicate
In computability theory, the T predicate, first studied by mathematician Stephen Cole Kleene, is a particular set of triples of natural numbers that is used to represent computable functions within formal theories of arithmetic. Informally, the T predicate tells whether a particular computer program will halt when run with a particular input, and the corresponding U function is used to obtain the results of the computation if the program does halt. As with the smn theorem, the original notation used by Kleene has become standard terminology for the concept. Definition The definition depends on a suitable Gödel numbering that assigns natural numbers to computable functions (given as Turing machines). This numbering must be sufficiently effective that, given an index of a computable function and an input to the function, it is possible to effectively simulate the computation of the function on that input. The predicate is obtained by formalizing this simulation. The ternary relation takes three natural numbers as arguments. is true if encodes a computation history of the computable function with index when run with input , and the program halts as the last step of this computation history. That is, first asks whether is the Gödel number of a finite sequence of complete configurations of the Turing machine with index , running a computation on input . If so, then asks if this sequence begins with the starting state of the computation and each successive element of the sequence corresponds to a single step of the Turing machine. If it does, finally asks whether the sequence ends with the machine in a halting state. If all three of these questions have a positive answer, then is true, otherwise, it is false. The predicate is primitive recursive in the sense that there is a primitive recursive function that, given inputs for the predicate, correctly determines the truth value of the predicate on those inputs. There is a corresponding primitive
https://en.wikipedia.org/wiki/Read-only%20Turing%20machine
A read-only Turing machine or two-way deterministic finite-state automaton (2DFA) is class of models of computability that behave like a standard Turing machine and can move in both directions across input, except cannot write to its input tape. The machine in its bare form is equivalent to a deterministic finite automaton in computational power, and therefore can only parse a regular language. Theory We define a standard Turing machine by the 9-tuple where is a finite set of states; is the finite set of the input alphabet; is the finite tape alphabet; is the left endmarker; is the blank symbol; is the transition function; is the start state; is the accept state; is the reject state. So given initial state reading symbol , we have a transition defined by which replaces with , transitions to state , and moves the "read head" in direction (left or right) to read the next input. In our 2DFA read-only machine, however, always. This model is now equivalent to a DFA. The proof involves building a table which lists the result of backtracking with the control in any given state; at the start of the computation, this is simply the result of trying to move past the left endmarker in that state. On each rightward move, the table can be updated using the old table values and the character that was in the previous cell. Since the original head-control had some fixed number of states, and there is a fixed number of states in the tape alphabet, the table has fixed size, and can therefore be computed by another finite state machine. This machine, however, will never need to backtrack, and hence is a DFA. Variants Several variants of this model are also equivalent to DFAs. In particular, the nondeterministic case (in which the transition from one state can be to multiple states given the same input) is reducible to a DFA. Other variants of this model allow more computational complexity. With a single infinite stack the model can parse (at least) any
https://en.wikipedia.org/wiki/Beurling%E2%80%93Lax%20theorem
In mathematics, the Beurling–Lax theorem is a theorem due to and which characterizes the shift-invariant subspaces of the Hardy space . It states that each such space is of the form for some inner function . See also H2
https://en.wikipedia.org/wiki/Posterior%20scrotal%20arteries
The posterior scrotal arteries are branches of the internal pudendal artery. Function The anterior scrotal arteries supply part of the scrotum in men. See also Anterior scrotal arteries
https://en.wikipedia.org/wiki/Tetrakis%28dimethylamido%29titanium
Tetrakis(dimethylamino)titanium (TDMAT), also known as Titanium(IV) dimethylamide, is a chemical compound. The compound is generally classified as a metalorganic species, meaning that its properties are strongly influenced by the organic ligands but the compound lacks metal-carbon bonds. It is used in chemical vapor deposition to prepare titanium nitride (TiN) surfaces and in atomic layer deposition as a titanium dioxide precursor. The prefix "tetrakis" refers the presence of four of the same ligand, in this case dimethylamides. Preparation and properties Tetrakis(dimethylamino)titanium is a conventional Ti(IV) compound in the sense that it is tetrahedral and diamagnetic. Unlike the many alkoxides, the diorganoamides of titanium are monomeric and thus at least somewhat volatile. It is prepared from titanium tetrachloride (which is also tetrahedral, diamagnetic, and volatile) by treatment with lithium dimethylamide: TiCl4 + 4 LiNMe2 → Ti(NMe2)4 + 4 LiCl Like many amido complexes, TDMAT is quite sensitive toward water, and its handling requires air-free techniques. The ultimate products of its hydrolysis is titanium dioxide and dimethylamine: Ti(NMe2)4 + 2 H2O → TiO2 + 4 HNMe2 In a related reaction, the compound undergoes exchange with other amines, evolving dimethylamine. See also Metalorganic chemical vapor deposition (MOCVD), general process which includes using TDMAT but also uses many other gases to layer other substances.
https://en.wikipedia.org/wiki/Rapidly%20exploring%20random%20tree
A rapidly exploring random tree (RRT) is an algorithm designed to efficiently search nonconvex, high-dimensional spaces by randomly building a space-filling tree. The tree is constructed incrementally from samples drawn randomly from the search space and is inherently biased to grow towards large unsearched areas of the problem. RRTs were developed by Steven M. LaValle and James J. Kuffner Jr. They easily handle problems with obstacles and differential constraints (nonholonomic and kinodynamic) and have been widely used in autonomous robotic motion planning. RRTs can be viewed as a technique to generate open-loop trajectories for nonlinear systems with state constraints. An RRT can also be considered as a Monte-Carlo method to bias search into the largest Voronoi regions of a graph in a configuration space. Some variations can even be considered stochastic fractals. RRTs can be used to compute approximate control policies to control high dimensional nonlinear systems with state and action constraints. Description An RRT grows a tree rooted at the starting configuration by using random samples from the search space. As each sample is drawn, a connection is attempted between it and the nearest state in the tree. If the connection is feasible (passes entirely through free space and obeys any constraints), this results in the addition of the new state to the tree. With uniform sampling of the search space, the probability of expanding an existing state is proportional to the size of its Voronoi region. As the largest Voronoi regions belong to the states on the frontier of the search, this means that the tree preferentially expands towards large unsearched areas. The length of the connection between the tree and a new state is frequently limited by a growth factor. If the random sample is further from its nearest state in the tree than this limit allows, a new state at the maximum distance from the tree along the line to the random sample is used instead of the rand
https://en.wikipedia.org/wiki/Enhanced%20other%20networks
Enhanced other networks (EON) is a radio system used to deliver traffic information to enabled devices. It is a component of the European Radio Data System (RDS). The system delivers traffic information (TP and TA) to enabled devices, by interrupting the current stream of media (radio, cd, etc.) and sends the traffic message.
https://en.wikipedia.org/wiki/Badge%20of%20shame
A badge of shame, also a symbol of shame, a mark of shame or a stigma, is typically a distinctive symbol required to be worn by a specific group or an individual for the purpose of public humiliation, ostracism or persecution. The term is also used metaphorically, especially in a pejorative sense, to characterize something associated with a person or group as shameful. In England, under the Poor Act 1697, paupers in receipt of parish relief were required to wear a badge of blue or red cloth on the shoulder of the right sleeve in an open and visible manner, in order to discourage people from collecting relief unless they were desperate, as while many would be willing to collect relief, few would be willing to do so if required to wear the "shameful" mark of the poor in public. The yellow badge that Jews were required to wear in parts of Europe during the Middle Ages, and later in Nazi Germany and German-occupied Europe, was effectively a badge of shame, as well as identification. Other identifying marks may include making shamed people go barefoot. The biblical "Mark of Cain" can be interpreted as synonymous with a badge of shame. History Depilation Punitive depilation of men, especially burning off pubic hair, was intended as a mark of shame in ancient Mediterranean cultures where male body hair was valued. Women who committed adultery have also been forced to wear specific icons or marks, or had their hair shorn, as a badge of shame. Many women who fraternized with the occupiers in German-occupied Europe had their heads shaved by angry mobs of their peers after liberation by the Allies of World War II. During World War II, the Nazis also used head shaving as a mark of shame to punish Germans like the youthful non-conformists known as the Edelweiss Pirates. Clothing In Ancient Rome, both men and women originally wore the toga, but over time matrons adopted the stola as the preferred form of dress, while prostitutes retained the toga. Later, under the Lex J
https://en.wikipedia.org/wiki/Webtag
Webtag is an on-line bioinformatics tool providing oligonucleotide sequences (usually called tags or anchors) that are absent from a specified genome. These tags can be appended to gene specific primers for reverse transcriptase polymerase chain reaction (RT-PCR) experiments, circumventing genomic DNA contamination. Background RT-PCR is a technique used for the detection of even very low copy mRNA transcripts. The sensitivity of the technique also makes it susceptible to DNA contamination. Since PCR is unable to distinguish between cDNA targets and genomic DNA contamination, false positives and/or erroneous quantitative results are possible. In order to overcome genomic DNA contamination in transcriptional studies, reverse template-specific polymerase chain reaction, a modification of RT-PCR is used. The possibility of using tags whose sequences are not found in the genome further improves reverse specific polymerase chain reaction experiments. The use of anchors, or tags, in the 5' region of a gene specific primer or poly-T tail allows for RNA-specific amplification, and constitutes a viable strategy. Techniques such as RS-PCR and (EXACT) RT-PCR are based on the integration of such tags (unique sequences not present in genomic DNA) in the 5' end of the first strand cDNA, permitting RNA-specific amplification without loss of sensitivity. Webtag This web based service builds on the Tagenerator tool, but is very fast because all tags are pre-generated and stored in a database. It is also a significant improvement since Webtag takes into account the interactions of the tag with the primers to be used in the experiment. Having it as a web based service also means that the molecular biologist doesn't have to download and install software with all the dependencies on their own computer. Webtag generates tags that combine genome absence with good priming properties for RT-PCR based experiments. The use of such tags will deliberately not result in PCR amplification of g
https://en.wikipedia.org/wiki/HABU%20equivalent
The HABU equivalent is a unit of measurement used by United States Department of Defense's High Performance Computing Modernization Program to evaluate the performance of large computers systems. "The [HPCMP method for measuring system performance] is as follows: the ratio of time [for a given benchmark application] at a target processor count provides a relative measure of the system's performance on that application test case compared with the DoD standard system, stated in Habu-equivalents. Habu, the first DoD standard system, is an IBM POWER3 formerly located at the US Naval Oceanographic Office (NAVO) Major Shared Resource Center. One Habu-equivalent is the performance of 1,024 system-under-study processors compared with 1,024 Habu processors.