source stringlengths 31 227 | text stringlengths 9 2k |
|---|---|
https://en.wikipedia.org/wiki/Solar%20Cookers%20International | Solar Cookers International (SCI) is a 501(c)(3) nonprofit, non-governmental organization that works to improve human and environmental health by supporting the expansion of effective carbon-free solar cooking in world regions of greatest need. SCI leads through advocacy, research, and strengthening the capacity of the global solar cooking movement. SCI has consultative status with the United Nations Economic and Social Council (ECOSOC) and was founded in 1987.
Recognition
Solar Cookers International won an Ashden Award in 2002 for their work with solar cookers in Kenya. In August 2006, SCI was the winner of the World Renewable Energy Award. SCI was named as a winner in the Keeling Curve Prize for sustainable planet solutions in August 2021. SCI won for its work “improving human and environmental health by supporting the expansion of effective carbon-free solar cooking in world regions of greatest need” in the Social and Cultural Pathways category. SCI has also been honored by entities such as the California State Legislature and the Center for African Peace and Conflict Resolution.
History
Solar Cookers International was founded in 1987 as Solar Box Cookers International. Barbara Kerr and Sherry Cole partnered with other supporters to form this organization.
SCI produced and distributed manuals describing the construction and use of solar box-style cookers. They became advocates of how solar cooking could be incorporated into development and relief agency programs. SCI's role evolved into networking with other solar cooking organizations worldwide. They hosted forums for dialog including co-sponsoring three international solar cooking conferences with the University of the Pacific, US, in 1992, the National University of Costa Rica in 1994 and the deemed university, Coimbatore, India in 1997.
SCI also administered a series of solar cooking field projects. Since 1995, SCI has managed or co-managed solar cooking projects in the Nyakach district, Kenya; in the K |
https://en.wikipedia.org/wiki/Subject%20and%20object%20%28philosophy%29 | The distinction between subject and object is a basic idea of philosophy.
A subject is a being that exercises agency, undergoes conscious experiences, and is situated in relation to other things that exist outside itself; thus, a subject is any individual, person, or observer
An object is any of the things observed or experienced by a subject, which may even include other beings (thus, from their own points of view: other subjects)
A simple common differentiation for subject and object is: an observer versus a thing that is observed. In certain cases involving personhood, subjects and objects can be considered interchangeable where each label is applied only from one or the other point of view. Subjects and objects are related to the philosophical distinction between subjectivity and objectivity: the existence of knowledge, ideas, or information either dependent upon a subject (subjectivity) or independent from any subject (objectivity).
Etymology
In English the word object is derived from the Latin objectus (p.p. of obicere) with the meaning "to throw, or put before or against", from ob-, "against", and the root jacere, "to throw". Some other related English words include objectify (to reify), objective (a future reference), and objection (an expression of protest). Subject uses the same root, but with the prefix sub-, meaning "under".
Terms and usage
Broadly construed, the word object names a maximally general category, whose members are eligible for being referred to, quantified over and thought of. Terms similar to the broad notion of object include thing, being, entity, item, existent, term, unit, and individual.
In ordinary language, one is inclined to call only a material object "object". In certain contexts, it may be socially inappropriate to apply the word object to animate beings, especially to human beings, while the words entity and being are more acceptable.
Some authors use object in contrast to property; that is to say, an object is an entity tha |
https://en.wikipedia.org/wiki/AStA%20Advances%20in%20Statistical%20Analysis | AStA Advances in Statistical Analysis is a peer-reviewed mathematics journal published quarterly by Springer Science+Business Media and the German Statistical Society. It was established in 2007, and covers statistical theory, methods, methodological developments, as well as probability and mathematics applications. Coverage is organized into three broad areas: statistical applications, statistical methodology, and review articles. The editor were Göran Kauermann (2009–2019) and Stefan Lang (2009–2014). In 2022 the editor are Thomas Kneib and Yarema Okhrin.
Abstracting and indexing
The journal is abstracted and indexed in:
According to the Journal Citation Reports, the journal has a 2020 impact factor of 1.160. |
https://en.wikipedia.org/wiki/G%C3%A9za%20Fodor%20%28mathematician%29 | Géza Fodor (6 May 1927 in Szeged – 28 September 1977 in Szeged) was a Hungarian mathematician, working in set theory. He proved Fodor's lemma on stationary sets, one of the most important, and most used results in set theory. He was a professor at the Bolyai Institute of Mathematics at the Szeged University. He was vice-president, then president of the Szeged University. He was elected a corresponding member of the Hungarian Academy of Sciences.
1927 births
1977 deaths
Members of the Hungarian Academy of Sciences
Set theorists
20th-century Hungarian mathematicians
People from Szeged |
https://en.wikipedia.org/wiki/Zoidogamy | Zooidogamy is a type of plant reproduction in which male gametes (antherozoids) swim in a path of water to the female gametes (archegonium). Zoidogamy is found in algae, bryophytes, pteridophytes, and some gymnosperms (others use siphonogamy). Zoidogamy relates to evolution, as it provides a pathway from wind-borne abiotic pollination and similar mechanisms to fluid-based mechanisms used in most animals. |
https://en.wikipedia.org/wiki/Symmetric%20spectrum | In algebraic topology, a symmetric spectrum X is a spectrum of pointed simplicial sets that comes with an action of the symmetric group on such that the composition of structure maps
is equivariant with respect to . A morphism between symmetric spectra is a morphism of spectra that is equivariant with respect to the actions of symmetric groups.
The technical advantage of the category of symmetric spectra is that it has a closed symmetric monoidal structure (with respect to smash product). It is also a simplicial model category. A symmetric ring spectrum is a monoid in ; if the monoid is commutative, it's a commutative ring spectrum. The possibility of this definition of "ring spectrum" was one of motivations behind the category.
A similar technical goal is also achieved by May's theory of S-modules, a competing theory. |
https://en.wikipedia.org/wiki/Sony%20NEX-F3 | The Sony α NEX-F3 is an entry level rangefinder-styled digital mirrorless interchangeable-lens camera announced by Sony on 17 May 2012. |
https://en.wikipedia.org/wiki/List%20of%20members%20of%20the%20U.S.%20Congress%20who%20support%20or%20oppose%20SOPA/PIPA | The Stop Online Piracy Act (SOPA) and the PROTECT IP Act (PIPA) are two proposed draft laws that were being considered by the United States Congress. Their stated goals are to increase the ability of U.S. law enforcement to fight online trafficking in copyrighted intellectual property and counterfeit goods, and give the U.S. government and copyright holders additional tools to curb access to "rogue websites dedicated to infringing or counterfeit goods", especially those registered outside the United States.
Supporters of the laws argued that they are needed to protect the intellectual property of owners of content. Opponents of the laws argued that they endanger free speech and free expression by harmfully regulating the internet.
Supporters of SOPA/PIPA
SOPA is the bill under consideration by the U.S. House of Representatives. PIPA is the related bill under consideration by the U.S. Senate.
SOPA Sponsors
The Stop Online Piracy Act was introduced by Representative Lamar Smith (R-TX).
Sponsors at introduction
SOPA was initially co-sponsored by 12 Representatives:
Howard Berman (D-CA)
Mary Bono Mack (R-CA)
Steve Chabot (R-OH)
John Conyers (D-MI)
Ted Deutch (D-FL)
Elton Gallegly (R-CA)
Bob Goodlatte (R-VA)
Adam Schiff (D-CA)
Marsha Blackburn (R-TN)
Timothy Griffin (R-AR)Withdrew Sponsorship 1/18/2012
Dennis A. Ross (R-FL)Withdrew Sponsorship 1/18/2012The Rachel Maddow Show, Jan 18, 12
Lee Terry (R-NE)Withdrew Sponsorship, 1/18/2012
Subsequent sponsors
After its initial introduction, other Representatives became co-sponsors of SOPA:
Mark Amodei [R-NV2]
Joe Baca [D-CA43]
John Barrow [D-GA12]
Karen Bass [D-CA33]
John Carter [R-TX31]Withdrew Sponsorship 1/24/2012
Judy Chu [D-CA32]
Jim Cooper [D-TN5]
Peter T. King [R-NY3]
John Larson [D-CT1]
Ben R. Luján [D-NM3]Withdrew Sponsorship 1/23/2012
Thomas Marino [R-PA10]
Alan Nunnelee [R-MS1]
Bill Owens [D-NY23]
Steve Scalise [R-LA1]Withdrew Sponsorship 1/23/2012
Brad Sherman [D-CA27]
Debbie Was |
https://en.wikipedia.org/wiki/Brain | The brain (or encephalon) is an organ that serves as the center of the nervous system in all vertebrate and most invertebrate animals. The brain is the largest cluster of neurons in the body and is typically located in the head, usually near organs for special senses such as vision, hearing and olfaction. It is the most specialized and energy-consuming organ in the body, responsible for complex sensory perception, motor control, endocrine regulation and the development of intelligence.
While invertebrate brains arise from paired segmental ganglia (each of which is only responsible for the respective body segment) of the ventral nerve cord, vertebrate brains develop axially from the midline dorsal nerve cord as a vesicular enlargement at the rostral end of the neural tube, with centralized control over all body segments. All vertebrate brains can be embryonically divided into three parts: the forebrain (prosencephalon, subdivided into telencephalon and diencephalon), midbrain (mesencephalon) and hindbrain (rhombencephalon, subdivided into metencephalon and myelencephalon). The spinal cord, which directly interacts with somatic functions below the head, can be considered a caudal extension of the myelencephalon enclosed inside the vertebral column. Together, the brain and spinal cord constitute the central nervous system in all vertebrates.
In humans, the cerebral cortex contains approximately 14–16 billion neurons, and the estimated number of neurons in the cerebellum is 55–70 billion. Each neuron is connected by synapses to several thousand other neurons, typically communicating with one another via root-like protrusions called dendrites and long fiber-like extensions called axons, which are usually myelinated and carry trains of rapid micro-electric signal pulses called action potentials to target specific recipient cells in other areas of the brain or distant parts of the body. The prefrontal cortex, which controls executive functions, is particularly well devel |
https://en.wikipedia.org/wiki/OneID | OneID was a digital security service based in Redwood City, California. OneID sold a digital identity system that claimed to provide security across all devices using public-key cryptography instead of passwords. The technology is utilized by non-profits, such as Salsa Labs, to increase the frequency and security of online donations. OneID now operates as a subsidiary program of Neustar following its acquisition in 2016.
History
The company was founded in 2011 by serial entrepreneur, Steve Kirsch. Kirsch recruited engineers Jim Fenton, Adam Back, and Bobby Beckmann to create the flagship product, which was launched in early 2012. Following the launch, the company raised US$7 million in venture capital financing from Menlo Park-based venture capital firm Khosla Ventures.
Following their Series A, Alex Doll took over the position as CEO. Doll was previously an executive-in-residence as Khosla Ventures, and before that was a founding executive at PGP Corp. With Doll's appointment, founder Kirsch moved into the CTO role. Shirish Sathaye, general partner at Khosla Ventures, and Jonathan Heiliger were added to OneID's board of directors.
Following a growth period in late 2013, the company appointed Kirsch CEO.
In 2014, Fast Company named OneID one of the top 10 most innovative companies in finance.
In August 2016, OneID was acquired by Neustar.
See also
Multi-factor authentication |
https://en.wikipedia.org/wiki/HECT%20domain | In molecular biology, the HECT domain is a protein domain found in ubiquitin-protein ligases. The name HECT comes from 'Homologous to the E6-AP Carboxyl Terminus'. Proteins containing this domain at the C terminus include ubiquitin-protein ligase, which regulates ubiquitination of CDC25. Ubiquitin-protein ligase accepts ubiquitin from an E2 ubiquitin-conjugating enzyme in the form of a thioester, and then directly transfers the ubiquitin to targeted substrates. A cysteine residue is required for ubiquitin-thiolester formation. Human thyroid receptor interacting protein 12 (TRIP12), which also contains this domain, is a component of an ATP-dependent multisubunit protein that interacts with the ligand binding domain of the thyroid hormone receptor. It could be an E3 ubiquitin-protein ligase. Human E6AP ubiquitin-protein ligase interacts with the E6 protein of the cancer-associated Human papillomavirus type 16 and Human papillomavirus type 18. The E6/E6-AP complex binds to and targets the p53 tumour-suppressor protein for ubiquitin-mediated proteolysis. |
https://en.wikipedia.org/wiki/The%20Math%20Book | The Math Book (Sterling Publishing, 2009. ) is a book by American author Clifford A. Pickover.
Summary
The book contains 250 one-page articles on milestones in the history of math. Each article is followed by a related full-page color image.
Reception
The book has consistently received good reviews.
The book has been praised by Martin Gardner.
The book is the winner of the Von Neumann Prize.
The book has been praised by Boing Boing. |
https://en.wikipedia.org/wiki/Hepatocyte%20nuclear%20factors | Hepatocyte nuclear factors (HNFs) are a group of phylogenetically unrelated transcription factors that regulate the transcription of a diverse group of genes into proteins. These proteins include blood clotting factors and in addition, enzymes and transporters involved with glucose, cholesterol, and fatty acid transport and metabolism.
Function
As the name suggests, hepatocyte nuclear factors are expressed predominantly in the liver. However HNFs are also expressed and play important roles in a number of other tissues so that the name hepatocyte nuclear factor is somewhat misleading. Nevertheless, the liver is the only tissue in which a significant number of different HNFs are expressed at the same time. In addition, there are a number of genes which contain multiple promoter and enhancer regions each regulated by a different HNF. Furthermore, efficient expression of these genes require synergistic activation by multiple HNFs. Hence hepatocyte nuclear factors function to ensure liver specific expression of certain genes.
As is the case with many transcription factors, HNFs regulate the expression of a wide variety of target genes and therefore functions. These functions (and especially functions involving the liver) include development and metabolic homeostasis of the organism. For example, HNFs influence expression of the insulin gene as well as genes involved in glucose transport and metabolism. In embryo development, HNF4α is thought to have an important role in the development of the liver, kidney, and intestines.
Disease implication
Variants of the genes can cause several relatively rare forms of MODY, an inherited, early onset form of diabetes. Mutations in the HNF4α, HNF1α, or HNF1β genes are linked to MODY 1, MODY 3, and MODY 5 respectively. Mutations in HNF genes are also associated with a number of others diseases including hepatic adenomas and renal cysts.
Members
The following is a list of human hepatocyte nuclear factors (see als |
https://en.wikipedia.org/wiki/CAINE%20Linux | CAINE Linux (Computer Aided INvestigative Environment) is an Italian Linux live distribution managed by Giovanni "Nanni" Bassetti. The project began in 2008 as an environment to foster digital forensics and incidence response (DFIR), with several related tools pre-installed.
Purpose
CAINE is a professional open source forensic platform that integrates software tools as modules along with powerful scripts in a graphical interface environment. Its operational environment was designed with the intent to provide the forensic professional all the tools required to perform the digital forensic investigate process (preservation, collection, examination and analysis). CAINE is a live Linux distribution so it can be booted from removable media (flash drive) or from an optical disk and run in memory. It can also be installed onto a physical or virtual system. In Live mode, CAINE can operate on data storage objects without having to boot up a supporting operating system. The latest version 11.0 can boot on UEFI/UEFI+Secure and Legacy BIOS allowing CAINE to be used on information systems that boot older operating systems (e.g. Windows NT) and newer platforms (Linux, Windows 10).
Requirements
CAINE is based on Ubuntu 18.04 64-bit, using Linux kernel 5.0.0-32. CAINE system requirements to run as a live disc are similar to Ubuntu 18.04. It can run on a physical system or in a virtual machine environment such as VMware Workstation.
Supported platforms
The CAINE Linux distribution has numerous software applications, scripts and libraries that can be used in a graphical or command line environment to perform forensic tasks. CAINE can perform data analysis of data objects created on Microsoft Windows, Linux and some Unix systems. One of the key forensic features since version 9.0 is that it sets all block devices by default to read-only mode. Write-blocking is a critical methodology to ensure that disks are not subject to writing operations by the operating system or forensic |
https://en.wikipedia.org/wiki/International%20Electrical%20Congress | The International Electrical Congress was a series of international meetings, from 1881 - 1904, in the then new field of applied electricity. The first meeting was initiated by the French government, including official national representatives, leading scientists, and others. Subsequent meetings also included official representatives, leading scientists, and others. Primary aims were to develop reliable standards, both in relation to electrical units and electrical apparatus.
Historical background
In 1881, both within and across countries, different electrical units were being used. There were at least 12 different units of electromotive force, 10 different units
of electric current and 15 different units of resistance.
A number of international Congresses were held, and sometimes referred to as International Electrical Congress, Electrical Conference, and similar variations. Secondary sources make different judgments about how to classify the Congresses. In this article, the Congresses with representatives from national governments are identified as International Electrical Congress. Other Congresses — often addressing the same issues — are identified here as Concurrent Related International Electrical Congresses. Some of these related conferences were devoted to preparing for an International Electrical Congress.
In 1906 the International Electrotechnical Commission was created. Congresses were organised under its auspices were also sometimes referred to as International Electrical Congress. In this article, Congresses organized by the Commission are listed under International Electrotechnical Congresses, while other related Congresses are listed under Related International Electrotechnical Conferences.
International Electrical Congress
1881 in Paris
Held from 15 September-5 October 1881, in connection with the International Exposition of Electricity. Adolphe Cochery, Minister of Posts and Telegraphs of the French
Government, was the Chairman. At the Congre |
https://en.wikipedia.org/wiki/Heart%20nanotechnology | Heart nanotechnology is the "Engineering of functional systems at the molecular scale" ("Nanotechnology Research").
Nanotechnology
Nanotechnology deals with structures and materials that are approximately one to one-hundred nanometers in length. At this microscopic level, quantum mechanics take place and are in effect, resulting in behaviors that would seem quite strange compared to what humans see with the naked eye (regular matter). Nanotechnology is used for a wide variety of fields of technology, ranging from energy to electronics to medicine. In the category of medicine, nanotechnology is still relatively new and has not yet been widely adopted by the field. It is possible that nanotechnology could be the new breakthrough of medicine and may eventually be the solution and cure for many of the health problems that humans encounter. Nanotechnology may lead to the cure for illnesses such as the common cold, diseases, and cancer. It is already starting to be used as a treatment for some serious health issues; more specifically it is being used to treat the heart and cancer.
Nanomedicine
Nanotechnology in the field of medicine is more commonly referred to as nanomedicine. Nanomedicine that deals with helping the heart is really starting to take off and gain in popularity compared to most of the other fields that nanomedicine currently has to offer. There are several heart problems that nanotechnology has promising evidence of being effective in the treatment of heart disease in the near future.
Examples
It should hopefully be able to treat heart valves that are defective; and detect and treat arterial plaque in the heart ("Nanotechnology Made Clear"). Nanomedicine should be able to help heal the hearts of people that have already been victims of heart disease and heart attacks. On the other hand, it will also play a key role in finding people with a high risk of having heart disease, and will be able to help prevent heart attacks from happening in the first p |
https://en.wikipedia.org/wiki/Rasmus%20Bartholin | Rasmus Bartholin (; Latinized: Erasmus Bartholinus; 13 August 1625 – 4 November 1698) was a Danish physician and grammarian.
Biography
Bartholin was born in Roskilde. He was the son of Caspar Bartholin the Elder (1585–1629) and Anna Fincke, daughter of the mathematician Thomas Fincke.
As part of his studies, he travelled in Europe for ten years. He stayed in the Netherlands, England, France and Italy. In 1647, he took a Master's degree at the University of Copenhagen. In 1654, he received a Doctoral degree at the University of Padua.
He was a professor at the University of Copenhagen, first in Geometry, later in Medicine. He was also dean of the faculty of medicine, librarian, and rector. He wrote, in Latin, the first grammar of the Danish language, the 1657 De studio lingvæ danicæ.
Rasmus Bartholin is remembered especially for his discovery (1669) of the double refraction of a light ray by Iceland spar (calcite). He published an accurate description of the phenomenon, but since the physical nature of light was poorly understood at the time, he was unable to explain it. It was only after Thomas Young proposed the wave theory of light, c. 1801 that an explanation became possible.
Personal
He was a younger brother of Thomas Bartholin (1616–1680). |
https://en.wikipedia.org/wiki/Basket%20weave%20knot | The basket weave knots are a family of bend and lanyard knots with a regular pattern of over–one, under–one. All of these knots are rectangular and lie in a plane. They are named after plait-woven baskets, which have a similar appearance.
Construction
A basket weave knot is made up of two sets of parallel lines drawn inside a rectangle such that the lines meet at the edges of the rectangle. For a true basket weave knot that can be tied with two strands, the number of intersections in each direction cannot have a common divisor. Within this constraint, there is no theoretical upper limit to the size of a basket weave knot. Thus, a knot that has two intersections in one direction can be lengthened with any odd number in the perpendicular direction. If the dimension n in the smaller direction is odd, it is always possible to construct a knot with n + 2 intersections in the other dimension. However, large basket weave knots have a tendency to twist and curl because they are completely flat.
A basket weave knot can be tied from a single strand by first forming a bight in the middle of the line. The ends near the bight become the standing ends. This method will keep the knot in one plane only for knots in which the standing ends enter the same side; these knots are called bosun's knots because they can be tied in a lanyard. For knots in which the standing ends enter from different sides of the rectangle, the bight will wrap across one side of the knot after it is set.
Any basket weave knot that can be tied from two strands can be drawn as an endless knot by connecting the standing ends together and the working ends together. An example of this can be seen in the carrick mat.
If a basket weave knot is tied with a flat line such as ribbon instead of a round line such as rope or cord, the method of turning the line at the edges affects the final appearance. Deflecting the line will form a series of bights or scallops along the edge, while folding it over will leave the |
https://en.wikipedia.org/wiki/Internal%20transcribed%20spacer | Internal transcribed spacer (ITS) is the spacer DNA situated between the small-subunit ribosomal RNA (rRNA) and large-subunit rRNA genes in the chromosome or the corresponding transcribed region in the polycistronic rRNA precursor transcript.
Across life domains
In bacteria and archaea, there is a single ITS, located between the 16S and 23S rRNA genes. Conversely, there are two ITSs in eukaryotes: ITS1 is located between 18S and 5.8S rRNA genes, while ITS2 is between 5.8S and 28S (in opisthokonts, or 25S in plants) rRNA genes. ITS1 corresponds to the ITS in bacteria and archaea, while ITS2 originated as an insertion that interrupted the ancestral 23S rRNA gene.
Organization
In bacteria and archaea, the ITS occurs in one to several copies, as do the flanking 16S and 23S genes. When there are multiple copies, these do not occur adjacent to one another. Rather, they occur in discrete locations in the circular chromosome. It is not uncommon in bacteria to carry tRNA genes in the ITS.
In eukaryotes, genes encoding ribosomal RNA and spacers occur in tandem repeats that are thousands of copies long, each separated by regions of non-transcribed DNA termed intergenic spacer (IGS) or non-transcribed spacer (NTS).
Each eukaryotic ribosomal cluster contains the 5' external transcribed spacer (5' ETS), the 18S rRNA gene, the ITS1, the 5.8S rRNA gene, the ITS2, the 26S or 28S rRNA gene, and finally the 3' ETS.
During rRNA maturation, ETS and ITS pieces are excised. As non-functional by-products of this maturation, they are rapidly degraded.
Use in phylogenetic inference
Sequence comparison of the eukaryotic ITS regions is widely used in taxonomy and molecular phylogeny because of several favorable properties:
It is routinely amplified thanks to its small size associated to the availability of highly conserved flanking sequences.
It is easy to detect even from small quantities of DNA due to the high copy number of the rRNA clusters.
It undergoes rapid concerted evolut |
https://en.wikipedia.org/wiki/Hilbert%27s%20twenty-fourth%20problem | Hilbert's twenty-fourth problem is a mathematical problem that was not published as part of the list of 23 problems known as Hilbert's problems but was included in David Hilbert's original notes. The problem asks for a criterion of simplicity in mathematical proofs and the development of a proof theory with the power to prove that a given proof is the simplest possible.
The 24th problem was rediscovered by German historian Rüdiger Thiele in 2000, noting that Hilbert did not include the 24th problem in the lecture presenting Hilbert's problems or any published texts. Hilbert's friends and fellow mathematicians Adolf Hurwitz and Hermann Minkowski were closely involved in the project but did not have any knowledge of this problem.
This is the full text from Hilbert's notes given in Rüdiger Thiele's paper. The section was translated by Rüdiger Thiele.
In 2002, Thiele and Larry Wos published an article on Hilbert's twenty-fourth problem with a discussion about its relation to various issues in automated reasoning, logic, and mathematics. |
https://en.wikipedia.org/wiki/Lions%E2%80%93Magenes%20lemma | In mathematics, the Lions–Magenes lemma (or theorem) is the result in the theory of Sobolev spaces of Banach space-valued functions, which provides a criterion for moving a time derivative of a function out of its action (as a functional) on the function itself.
Statement of the lemma
Let X0, X and X1 be three Hilbert spaces with X0 ⊆ X ⊆ X1. Suppose that X0 is continuously embedded in X and that X is continuously embedded in X1, and that X1 is the dual space of X0. Denote the norm on X by || · ||X, and denote the action of X1 on X0 by . Suppose for some that is such that its time derivative . Then is almost everywhere equal to a function continuous from into , and moreover the following equality holds in the sense of scalar distributions on :
The above equality is meaningful, since the functions
are both integrable on .
See also
Aubin–Lions lemma
Notes
It is important to note that this lemma does not extend to the case where is such that its time derivative for . For example, the energy equality for the 3-dimensional Navier–Stokes equations is not known to hold for weak solutions, since a weak solution is only known to satisfy and (where is a Sobolev space, and is its dual space, which is not enough to apply the Lions–Magnes lemma (one would need , but this is not known to be true for weak solutions). |
https://en.wikipedia.org/wiki/Probabilistic%20method | In mathematics, the probabilistic method is a nonconstructive method, primarily used in combinatorics and pioneered by Paul Erdős, for proving the existence of a prescribed kind of mathematical object. It works by showing that if one randomly chooses objects from a specified class, the probability that the result is of the prescribed kind is strictly greater than zero. Although the proof uses probability, the final conclusion is determined for certain, without any possible error.
This method has now been applied to other areas of mathematics such as number theory, linear algebra, and real analysis, as well as in computer science (e.g. randomized rounding), and information theory.
Introduction
If every object in a collection of objects fails to have a certain property, then the probability that a random object chosen from the collection has that property is zero.
Similarly, showing that the probability is (strictly) less than 1 can be used to prove the existence of an object that does not satisfy the prescribed properties.
Another way to use the probabilistic method is by calculating the expected value of some random variable. If it can be shown that the random variable can take on a value less than the expected value, this proves that the random variable can also take on some value greater than the expected value.
Alternatively, the probabilistic method can also be used to guarantee the existence of a desired element in a sample space with a value that is greater than or equal to the calculated expected value, since the non-existence of such element would imply every element in the sample space is less than the expected value, a contradiction.
Common tools used in the probabilistic method include Markov's inequality, the Chernoff bound, and the Lovász local lemma.
Two examples due to Erdős
Although others before him proved theorems via the probabilistic method (for example, Szele's 1943 result that there exist tournaments containing a large number of Hamilton |
https://en.wikipedia.org/wiki/Light-induced%20voltage%20alteration | Light-induced voltage alteration (LIVA) is a semiconductor analysis technique that uses a laser or infrared light source to induce voltage changes in a device while scanning the beam of light across its surface. The technique relies upon the generation of electron-hole pairs in the semiconductor material when exposed to photons.
Theory of operation
The device to be analyzed is biased using a constant current power supply. As the light source is scanned across the surface of the silicon, electron-hole pairs are generated. This causes subtle alterations of the operating characteristics of the device, which may result in slight changes to the power supply voltage. Any changes that are detected in the power supply voltage are noted and correlated with the position of the light source on the device. This allows the physical locations corresponding to power supply fluctuations to be mapped onto an image of the device. This provides the device analyst with specific locations at which the device may be examined for defects.
Notes |
https://en.wikipedia.org/wiki/Haasiophis | Haasiophis, consisting of the sole species Haasiophis terrasanctus, is an extinct genus of snakes with hind limbs. It is one of three genera of Cenomanian snakes known to have possessed hindlimbs.
Etymology
The genus was named in honor of the late paleontologist Georg Haas, who first began work in the fossils of Ein Yabrud and started the description of the genus before he died, plus the Greek "ophis", for snake . The specific names is from the Latin "terrasanctus" meaning "of the holy land".
Specimen material
H. terrasanctus is known from a single fossil discovered at Ein Yabrud in the Judean hills, near Ramallah, 20 km north of Jerusalem, in the central West Bank. This site also produced the type and only specimen for the sister genus Pachyrhachis. This site is a limestone carbonate which deposited in a low-energy marine platform environment. The only known specimen of H. terrasanctus, housed in the collections of the Hebrew University of Jerusalem as specimen HUJ-Pal. EJ 695., measures 88 cm, missing only the tip of the tail, and is considered a small snake.
Skull
The skull is well preserved though slightly compressed. In the type description the cranium is described as showing a mix of basal characters, like those found in pipe snakes, and advanced traits found in macrostomatans. The cranium shows a small premaxilla which bears no teeth, while the maxilla, palatine, pterygoid, and dentar bones host 73-75 teeth. The preorbital area of the skull, being small and slender, is similar to the preorbital area of pythonine snakes. The contact between the prefrontal and frontal bones of the skull are very similar to ones found in the sister genus Pachyrhachis and the unrelated Dinilysia. Due to the tail of the specimen being missing, the total vertebrae is 199, plus fragments of a 200th, with the rib cage extending from the 5th to the 154th vertebrae. Lymphapophyses (forked 'cloacal' ribs) are found beginning on the 155th vertebrae and continuing to the 160th. Th |
https://en.wikipedia.org/wiki/XOR%20swap%20algorithm | In computer programming, the exclusive or swap (sometimes shortened to XOR swap) is an algorithm that uses the exclusive or bitwise operation to swap the values of two variables without using the temporary variable which is normally required.
The algorithm is primarily a novelty and a way of demonstrating properties of the exclusive or operation. It is sometimes discussed as a program optimization, but there are almost no cases where swapping via exclusive or provides benefit over the standard, obvious technique.
The algorithm
Conventional swapping requires the use of a temporary storage variable. Using the XOR swap algorithm, however, no temporary storage is needed. The algorithm is as follows:
X := X XOR Y; // XOR the values and store the result in X
Y := Y XOR X; // XOR the values and store the result in Y
X := X XOR Y; // XOR the values and store the result in X
Since XOR is a commutative operation, either X XOR Y or Y XOR X can be used interchangeably in any of the foregoing three lines. Note that on some architectures the first operand of the XOR instruction specifies the target location at which the result of the operation is stored, preventing this interchangeability. The algorithm typically corresponds to three machine-code instructions, represented by corresponding pseudocode and assembly instructions in the three rows of the following table:
In the above System/370 assembly code sample, R1 and R2 are distinct registers, and each operation leaves its result in the register named in the first argument. Using x86 assembly, values X and Y are in registers eax and ebx (respectively), and places the result of the operation in the first register.
However, in the pseudocode or high-level language version or implementation, the algorithm fails if x and y use the same storage location, since the value stored in that location will be zeroed out by the first XOR instruction, and then remain zero; it will not be "swapped with itself". This is not the same as |
https://en.wikipedia.org/wiki/Joseph-Alfred%20Serret | Joseph-Alfred Serret (; August 30, 1819 – March 2, 1885) was a French mathematician who was born in Paris, France, and died in Versailles, France.
See also
Frenet–Serret formulas
Books by J.-A. Serret
Traité de trigonométrie (Gautier-Villars, 1880)
Cours de calcul differentiel et integral t. 1 (Gauthier-Villars, 1900)
Cours de calcul differentiel et integral t. 2 (Gauthier-Villars, 1900)
Cours d'algèbre supérieure. Tome I (Gauthier-Villars, 1877)
Cours d'algèbre supérieure. Tome II (Gauthier-Villars, 1879)
External links
1819 births
1885 deaths
19th-century French mathematicians
École Polytechnique alumni
Members of the French Academy of Sciences
Differential geometers |
https://en.wikipedia.org/wiki/Certificate%20Transparency | Certificate Transparency (CT) is an Internet security standard for monitoring and auditing the issuance of digital certificates.
The security of HTTPS depends on the trust that certificates are only given out by the certificate authority that was requested by the owner of some website or IT infrastructure. Certificate Transparency has the potential to expose certificates that were given out without them being requested by the genuine owner, such as malicious certificates by a compromised certificate authority, which happened in 2010 at DigiNotar.
is a standard defining a system of public logs that seek to eventually record all certificates issued by publicly trusted certificate authorities, allowing efficient identification of mistakenly or maliciously issued certificates.
Technical overview
The certificate transparency system consists of a system of append-only certificate logs. Logs are operated by many parties, including browser vendors and certificate authorities. Certificates that support certificate transparency must include one or more signed certificate timestamps (SCTs), which is a promise from a log operator to include the certificate in their log within a maximum merge delay (MMD). At some point within the maximum merge delay, the log operator adds the certificate to their log. Each entry in a log references the hash of a previous one, forming a Merkle tree. The signed tree head (STH) references the current root of the Merkle tree.
Mandatory certificate transparency
Some browsers require TLS certificates to have proof of being logged with certificate transparency, either through SCTs embedded into the certificate, an extension during the TLS handshake, or through OCSP:
Log sharding
Due to the large quantities of certificates issued with the Web PKI, certificate transparency logs can grow to contain many certificates. This large quantity of certificates can cause strain on logs. Temporal sharding is a method to reduce the strain on logs by sharding |
https://en.wikipedia.org/wiki/IoSafe | ioSafe is a manufacturer of disaster protected hard drives and network attached storage (NAS) appliances. The company was founded in 2004 and is based in Roseville, California. ioSafe's storage systems are optimized for heat from fire and complete submersion in fresh or saltwater with the ability to recover data located on the disk drive inside.
History
ioSafe, Inc. was incorporated in 2005 by the primary inventor and CEO, Robb Moore. In response to independent research regarding data loss and catastrophic business loss from fire, flood and other natural disasters their first product (released in 2005) was the ioSafe S1, the first external hard drive to have both fire and flood protection. The ioSafe S1 hard drive was later redesigned and re-released in a dual drive RAID configuration called the S2.
In 2006, in response to enterprise business demand for a fire and flood resistant network storage, the ioSafe R4 was introduced. Shortly after this NAS appliance entered production, ioSafe was granted and published patents in May 2007 for their FloSafe (fireproof) and HydroSafe (waterproof) technology.
The company was purchased by Vancouver, Washington-based CRU in July 2018. This acquisition was a strategic combination of the hardware safeguards provided by ioSafe and CRU which could subsequently address the combined and non-overlapping customer bases of the two companies.
Company Milestones
2004 — ioSafe introduces first fire and flood protected external hard drive
2005 — ioSafe incorporates in the State of California
April 2006 — Awarded Golden Mousetrap Award by Design News magazine
November 2006 — Introduces a Network Attached Storage system with shock-protection, fire protection and water submersion protection
May 2007 — Patents granted and published for protecting active electronic devices inside a fire safe
November 2007 — Patents granted and published for protecting active electronics from water
July 2008 — Introduces a 3.5” internal hard drive with integra |
https://en.wikipedia.org/wiki/Ideotype | In systematics, an ideotype is a specimen identified as belonging to a specific taxon by the author of that taxon, but collected from somewhere other than the type locality.
The concept of ideotype in plant breeding was introduced by Donald in 1968 to describe the idealized appearance of a plant variety. It literally means 'a form denoting an idea'. According to Donald, an ideotype is a biological model which is expected to perform or behave in a particular manner within a defined environment: "a crop ideotype is a plant model, which is expected to yield a greater quantity or quality of grain, oil or other useful product when developed as a cultivar." Donald and Hamblin (1976) proposed the concepts of isolation, competition and crop ideotypes. Market ideotype, climatic ideotype, edaphic ideotype, stress ideotype and disease/pest ideotypes are its other concepts. The term ideotype has the following synonyms: model plant type, ideal model plant type and ideal plan type.
The term is also used in cognitive science and cognitive psychology, where Ronaldo Vigo (2011, 2013, 2014) introduced it to refer to a type of concept metarepresentation that is a compound memory trace consisting of the structural information detected by humans in categorical stimuli.
Notes
Molecular biology
Botanical nomenclature |
https://en.wikipedia.org/wiki/ACOT11 | Acyl-coenzyme A thioesterase 11 also known as StAR-related lipid transfer protein 14 (STARD14) is an enzyme that in humans is encoded by the ACOT11 gene. This gene encodes a protein with acyl-CoA thioesterase activity towards medium (C12) and long-chain (C18) fatty acyl-CoA substrates which relies on its StAR-related lipid transfer domain. Expression of a similar murine protein in brown adipose tissue is induced by cold exposure and repressed by warmth. Expression of the mouse protein has been associated with obesity, with higher expression found in obesity-resistant mice compared with obesity-prone mice. Alternative splicing results in two transcript variants encoding different isoforms.
Structure
The ACOT11 gene is located on the 1st chromosome, with its specific localization being 1p32.3. It contains 18 exons.
The protein encoded by this gene contains 258 amino acids, and forms a homodimer with another chain. Its theoretical weight is 26.67 kDa. The protein contains a StAR-related transfer domain, which is a domain responsible for binding to lipids. There are 4 known ligands that bind to this homodimer: polyethylene glycol, chlorine, glycerol, and a form of TCEP.
Function
The protein encoded by the ACOT11 gene is part of a family of Acyl-CoA thioesterases, which catalyze the hydrolysis of various Coenzyme A esters of various molecules to the free acid plus CoA. These enzymes have also been referred to in the literature as acyl-CoA hydrolases, acyl-CoA thioester hydrolases, and palmitoyl-CoA hydrolases. The reaction carried out by these enzymes is as follows:
CoA ester + H2O → free acid + coenzyme A
These enzymes use the same substrates as long-chain acyl-CoA synthetases, but have a unique purpose in that they generate the free acid and CoA, as opposed to long-chain acyl-CoA synthetases, which ligate fatty acids to CoA, to produce the CoA ester. The role of the ACOT- family of enzymes is not well understood; however, it has been suggested that they play a cr |
https://en.wikipedia.org/wiki/Maimonachaetetes | Maimonachaetetes is a thalloid red alga of uncertain placement. |
https://en.wikipedia.org/wiki/Trading%20blows | Trading blows or trading licks is an endurance test in which the participants (usually two boys or young men) take turns, alternating between administering a blow to an opponent and assuming the agreed exposed position (e.g. bending over an object or grabbing the ankles) to endure the next one, using the same implement (e.g. a fraternity paddle), until only the winner can still bring himself to endure the gradually increasing pain in the progressively tormented target part of their anatomy (usually the posterior, in which case it is a form of spanking or the cheeks), which in the interest of fairness should be covered by a common uniform. This can be anything from regular jeans or pants, underwear, and finally, bared (naked) buttocks. As the blows are not given by the same person but by the parties themselves, the strongest-armed one actually has an unfair (but not always decisive) physical advantage.
Such rather macho displays of willpower, restraining the instinct to avoid pain, can serve various purposes, including:
a physical punishment, especially for quarreling, possibly the origin of the practice (this ritualized alternative to a more disabling or even lethal duel should also make its participants realize the futility of physical aggression)
a motivation test, especially as part of an initiation process, such as hazing
an obedience test, as in certain paddle games (possibly really an excuse for the rather sadistic amusement of the seniors)
a duel, either personal or as champions representing similar, especially rivaling, groups; in certain fraternities, refusing such a challenge may result in exclusion from the membership, even for an alumnus
as a game, either to 'proudly' display one's tenacity (often to impress some audience) or in the pursuit of a sadist and/or masochistic, erotic or pain-addicted, kick.
as a sexual fetish
Another game with the same name is often played among boys or young men, where two people agree on a place to hit the other (e.g. th |
https://en.wikipedia.org/wiki/Thermoguttaceae | Thermoguttaceae is a family of bacteria.
See also
List of bacterial orders
List of bacteria genera |
https://en.wikipedia.org/wiki/Public%20data | Public data may refer to:
Open data
any data that inadvertently becomes public affecting information privacy
Open data
Information privacy |
https://en.wikipedia.org/wiki/Shiri%20Artstein | Shiri Artstein-Avidan (, born 28 September 1978) is an Israeli mathematician who in 2015 won the Erdős Prize. She specializes in convex geometry and asymptotic geometric analysis, and is a professor of mathematics at Tel Aviv University.
Education and career
Artstein was born in Jerusalem, the daughter of mathematician Zvi Artstein. She graduated summa cum laude from Tel Aviv University in 2000, with a bachelor's degree in mathematics, and completed her PhD at Tel Aviv University in 2004 under the supervision of Vitali Milman, with a dissertation on Entropy Methods. She worked from 2004 to 2006 as a Veblen Research Instructor in Mathematics at Princeton University and as a researcher at the Institute for Advanced Study before returning to Tel Aviv as a faculty member in 2006.
Recognition
Artstein won the Haim Nessyahu Prize in Mathematics, an annual dissertation award of the Israel Mathematical Union, in 2006.
In 2008 she won the Krill Prize for Excellence in Scientific Research, from the
Wolf Foundation.
In 2015 she won the Anna and Lajos Erdős Prize in Mathematics. The award cited her "solution of Shannon's long standing problem on monotonicity of entropy (with K. Ball, F. Barthe and A. Naor), profound and unexpected development of the concept of duality, Legendre and Fourier transform from axiomatic viewpoint (with V. Milman) and discovery of an astonishing link between Mahler's conjecture in convexity theory and an isoperimetric-type inequality involving symplectic capacities (with R. Karasev and Y. Ostrover)".
Selected publications
Her research publications include: |
https://en.wikipedia.org/wiki/Value%20at%20risk | Value at risk (VaR) is a measure of the risk of loss of investment/Capital. It estimates how much a set of investments might lose (with a given probability), given normal market conditions, in a set time period such as a day. VaR is typically used by firms and regulators in the financial industry to gauge the amount of assets needed to cover possible losses.
For a given portfolio, time horizon, and probability p, the p VaR can be defined informally as the maximum possible loss during that time after excluding all worse outcomes whose combined probability is at most p. This assumes mark-to-market pricing, and no trading in the portfolio.
For example, if a portfolio of stocks has a one-day 95% VaR of $1 million, that means that there is a 0.05 probability that the portfolio will fall in value by more than $1 million over a one-day period if there is no trading. Informally, a loss of $1 million or more on this portfolio is expected on 1 day out of 20 days (because of 5% probability).
More formally, p VaR is defined such that the probability of a loss greater than VaR is (at most) (1-p) while the probability of a loss less than VaR is (at least) p. A loss which exceeds the VaR threshold is termed a "VaR breach".
It is important to note that, for a fixed p, the p VaR does not assess the magnitude of loss when a VaR breach occurs and therefore is considered by some to be a questionable metric for risk management. For instance, assume someone makes a bet that flipping a coin seven times will not give seven heads. The terms are that they win $100 if this does not happen (with probability 127/128) and lose $12,700 if it does (with probability 1/128). That is, the possible loss amounts are $0 or $12,700. The 1% VaR is then $0, because the probability of any loss at all is 1/128 which is less than 1%. They are, however, exposed to a possible loss of $12,700 which can be expressed as the p VaR for any p ≤ 0.78125% (1/128).
VaR has four main uses in finance: risk management |
https://en.wikipedia.org/wiki/Estrogenic%20fat | Estrogenic fat is a form of adipose tissue (or subcutaneous fat) which develops under the influence of estrogen, and particularly estradiol, in women.
Natural physiology
Estrogenic fat is a feminine secondary sex characteristic which develops at puberty and is maintained by estradiol throughout a woman's fertile years.
A special form of estrogenic fat is the iliac (hip) fat layer, which normally occurs below the iliac crest in females of childbearing age. Its cells contain a wider variety of fatty acids than most adipose tissues do. During the middle trimester of fetal development, when certain long-chain fatty acids are needed for organ development, the mother's iliac fat layer supplies these acids. Women who lack this normal layer are at increased risk of giving birth to underdeveloped newborns.
Estrogen Levels Pre- and Post-Menopause
There is a correlation between estradiol and estrone presence in adipose tissue in both pre- and post-menopausal women. Pre-menopausal women have higher levels of hormones including estrogen. After menopause, estrogenic fat diminishes, and lower levels of both estradiol and estrone are found in breast adipocytes, with a more pronounced decrease in estradiol levels. |
https://en.wikipedia.org/wiki/Herpes%20simplex%20virus%20protein%20vmw65 | Vmw65, also known as VP16 or α-TIF (Trans Inducing Factor) is a trans-acting protein that forms a complex with the host transcription factors Oct-1 and HCF to induce immediate early gene transcription in the herpes simplex viruses.
VP16 is a strong transactivator and is often used in Y2H systems as the activation domain of the system. |
https://en.wikipedia.org/wiki/Diffuse%20series | The diffuse series is a series of spectral lines in the atomic emission spectrum caused when electrons jump between the lowest p orbital and d orbitals of an atom. The total orbital angular momentum changes between 1 and 2. The spectral lines include some in the visible light, and may extend into ultraviolet or near infrared. The lines get closer and closer together as the frequency increases never exceeding the series limit. The diffuse series was important in the development of the understanding of electron shells and subshells in atoms. The diffuse series has given the letter d to the d atomic orbital or subshell.
The diffuse series has values given by
The series is caused by transitions from the lowest P state to higher energy D orbitals.
One terminology to identify the lines is: 1P-mD But note that 1P just means the lowest P state in the valence shell of an atom and that the modern designation would start at 2P, and is larger for higher atomic numbered atoms.
The terms can have different designations, mD for single line systems, mδ for doublets and md for triplets.
Since the Electron in the D subshell state is not the lowest energy level for the alkali atom (the S is) the diffuse series will not show up as absorption in a cool gas, however it shows up as emission lines.
The Rydberg correction is largest for the S term as the electron penetrates the inner core of electrons more.
The limit for the series corresponds to electron emission, where the electron has so much energy it escapes the atom.
In alkali metals the P terms are split and . This causes the spectral lines to be doublets, with a constant spacing between the two parts of the double line.
This splitting is called fine structure. The splitting is larger for atoms with higher atomic number. The splitting decreases towards the series limit. Another splitting occurs on the redder line of the doublet. This is because of splitting in the D level and . Splitting in the D level has a lesser |
https://en.wikipedia.org/wiki/ITF-6 | ITF-6 is the implementation of an Interleaved 2 of 5 (ITF) barcode to encode a addon to ITF-14 and ITF-16 barcodes. Originally was developed as a part of JIS specification for Physical Distribution Center. Instead of ITF-14, it wasn’t standardized by ISO Committee but it is widely used to encode additional data to Global Trade Item Number such as items quantity or container weight.
History
In 1983, the Logistics Symbol Committee proposed the Interleaved 2 of 5 barcode as a method to improve the JAN code. In 1985, a logistics symbol JIS drafting committee was set up at the Distribution System Development Center, and the final examination was started toward JIS. Then in 1987 it was standardized as JIS-X-0502, a standard physical distribution barcode symbol ITF-14/16/6.
The ITF barcode has an add-on version for displaying the weight, etc., and it is possible to encode a 5-digits numerical value and 6-th check character as ITF-6 after ITF-14 or ITF-16(obsolete in 2010).
Currently ITF-6 isn’t standardized by ISO Committee and it is used only as a part of JIS standards. However, it is widely used by manufacturers to encode additional data and it is supported by wide range of barcode scanners
Uses
Despite the fact that ITF-6 barcode isn’t included into ISO standards, it is widely used as add-on to encode items quantity in package or item weight. At this time, it is used only with ITF-14 (Global Trade Item Number), but up to 2010 it was used with standardized only in Japan ITF-16 (Extended Symbology for Physical Distribution).
From the left, ITF-6 contains 5 significant digits and the last one is control digit, which is calculated same way as UPC checksums. If a decimal point is required, the decimal point is between the 3rd and 4th digits:
NNNNN(C/D) - without decimal point;
NNN.NN(C/D) - with decimal point.
ITF-6 is supported by various barcode generating software and barcode scanners.
Checksum
Checksum is calculated as other UPC checksums:
Example for the first |
https://en.wikipedia.org/wiki/Lists%20of%20naval%20flags | Naval flags, both Naval jacks and naval ensigns, are a subset of Maritime flags flown by naval forces.
There are several lists of naval flags, organised by present or former country:
Current countries
Australia - List of Australian flags § Royal Australian Navy
Bangladesh - List of Bangladeshi flags § Military
Belgium - List of Belgian flags § Military
Croatia - List of Croatian flags § Maritime flags
Hungary - List of Hungarian flags § Naval flags
India - List of Indian flags § Navy
Ireland - List of flags of Ireland § Naval service
Japan - List of Japanese flags § Self-Defense Force and Imperial Army/Navy
Latvia - List of Latvian flags § Military flags
Norway - List of flags of Norway § Flags of the Navy
Poland - List of Polish naval and maritime flags, List of Polish flags § Navy
Russia - List of Russian navy flags
Thailand - List of flags of the Royal Thai Armed Forces § Royal Thai Navy
United Kingdom - List of British flags § Naval service
United States - Flags of the United States Armed Force § Maritime flags
Former countries
USSR - List of USSR navy flags
Yugoslavia - List of Yugoslav flags § Military flags
See also
Maritime flag
Naval jack
Naval ensign
Flag of the United States Navy
Blue Ensign
Red Ensign
White Ensign
List of flags
:Category:Lists and galleries of flags
Lists of lists
List of Bangladeshi flags |
https://en.wikipedia.org/wiki/Sunviridae | Sunviridae is a family of negative-strand RNA viruses in the order Mononegavirales. Snakes serve as natural hosts. The family includes the single genus Sunshinevirus which includes the single species Reptile sunshinevirus 1.The family was formed to accommodate the Sunshine Coast virus (SunCV), previously referred to as "Sunshine virus", a novel virus discovered in Australian pythons. The name derives from the geographic origin of the first isolate on the Sunshine Coast of Queensland, Australia.
Genome
Sunshineviruses have a nonsegmented, negative-sense, single-stranded RNA genome. The total length of the genome is 17 kbp. The genome encodes seven proteins. |
https://en.wikipedia.org/wiki/Wilf%20equivalence | In the study of permutations and permutation patterns, Wilf equivalence is an equivalence relation on permutation classes.
Two permutation classes are Wilf equivalent when they have the same numbers of permutations of each possible length, or equivalently if they have the same generating functions. The equivalence classes for Wilf equivalence are called Wilf classes; they are the combinatorial classes of permutation classes. The counting functions and Wilf equivalences among many specific permutation classes are known.
Wilf equivalence may also be described for individual permutations rather than permutation classes. In this context, two permutations are said to be Wilf equivalent if the principal permutation classes formed by forbidding them are Wilf equivalent. |
https://en.wikipedia.org/wiki/Windows%20Rally | Windows Rally is a set of technologies from Microsoft intended to simplify the setup and maintenance of wired and wireless network-connected devices. They aim to increase reliability and security of connectivity for users who connect the devices to the Internet or to computers running Microsoft Windows. These technologies provide control of network quality of service (QoS) and diagnostics for data sharing, communications, and entertainment. Windows Rally technologies provide provisioning for the following devices:
Wireless access points, PCs, and servers
Network printers, projectors, printer bridges, digital still cameras, and game consoles
Digital media receivers, network media players, set-top boxes, digital photo frames, and PDAs
Windows Rally technologies
Windows Rally includes the following set of technologies:
LLTD
The Link Layer Topology Discovery (LLTD) protocol enables applications to discover devices and determine network topology. In Windows Vista, it enables a graphical view of all the devices in the network on the Network Map. For Windows XP computers to appear on the Network Map, the LLTD Responder must be downloaded and installed. Devices that provide audio or video playback or that are bandwidth sensitive can implement the QoS Extension part of the protocol so that they receive prioritized streams and that changes in available bandwidth have less impact on the playback experience.
qWAVE
Windows Vista includes qWAVE, a pre-configured quality of service API for time-dependent multimedia data, such as audio or video streams. qWAVE uses different packet priority schemes for real-time flows (such as multimedia packets) and best-effort flows (such as file downloads or e-mails) to ensure that real-time data gets delayed as little as possible, while providing a high-quality channel for other data packets. qWAVE-enabled applications together with devices that implement the LLTD QoS Extensions aim to improve an end user's experience of streaming vide |
https://en.wikipedia.org/wiki/Frontoparietal%20scale | In reptiles, the frontoparietal scales are scales located behind the eyes, between the frontal scales to the front and the parietal scales to the back. |
https://en.wikipedia.org/wiki/Convergent%20beam%20electron%20diffraction | Convergent beam electron diffraction (CBED) is an electron diffraction technique where a convergent or divergent beam (conical electron beam) of electrons is used to study materials.
History
CBED was first introduced in 1939 by Kossel and Möllenstedt. The development of the Field Emission Gun (FEG) in the 1970s, the Scanning Transmission Electron Microscopy (STEM), energy filtering devices and so on, made possible smaller probe diameters and larger convergence angles, and all this made CBED more popular. In the seventies, CBED was being used for the determination of the point group and space group symmetries by Goodman and Lehmpfuh, and Buxton, and starting in 1985, CBED was used by Tanaka et al. for studying crystals structure.
Applications
By using CBED, the following information can be obtained:
parameters of the crystal lattice, sample thickness
strain distribution
defects such as stacking faults, dislocations, grain boundaries, three-dimensional deformations, lattice displacements
crystal symmetry information - by looking at the symmetries that appear in the CBED disks, point group and space group determination are performed.
Parameters
Positions of the CBED disks are the same as the positions of the Bragg peaks and are given approximately by the relation:
where is the distance between the crystallographic planes , is the Bragg angle, is an integer, and is the wavelength of the probing electrons.
The beam convergence semi-angle - is controlled by the C2 aperture. The probing beam convergence semi-angle, , is of the order of milliradians, ranging from 0.1˚ to 1˚. For small convergence semi-angle, the CBED disks do not overlap with each other, whereas for larger semi-convergence angles, the disks overlap.
The diameter of a CBED disk is given by the beam convergence semi-angle :
Defocus : The distance between the crossover of the probing beam and the position of the specimen is called the defocus distance . The sample can be moved along t |
https://en.wikipedia.org/wiki/M.%20Riesz%20extension%20theorem | The M. Riesz extension theorem is a theorem in mathematics, proved by Marcel Riesz during his study of the problem of moments.
Formulation
Let be a real vector space, be a vector subspace, and be a convex cone.
A linear functional is called -positive, if it takes only non-negative values on the cone :
A linear functional is called a -positive extension of , if it is identical to in the domain of , and also returns a value of at least 0 for all points in the cone :
In general, a -positive linear functional on cannot be extended to a -positive linear functional on . Already in two dimensions one obtains a counterexample. Let and be the -axis. The positive functional can not be extended to a positive functional on .
However, the extension exists under the additional assumption that namely for every there exists an such that
Proof
The proof is similar to the proof of the Hahn–Banach theorem (see also below).
By transfinite induction or Zorn's lemma it is sufficient to consider the case dim .
Choose any . Set
We will prove below that . For now, choose any satisfying , and set , , and then extend to all of by linearity. We need to show that is -positive. Suppose . Then either , or or for some and . If , then . In the first remaining case , and so
by definition. Thus
In the second case, , and so similarly
by definition and so
In all cases, , and so is -positive.
We now prove that . Notice by assumption there exists at least one for which , and so . However, it may be the case that there are no for which , in which case and the inequality is trivial (in this case notice that the third case above cannot happen). Therefore, we may assume that and there is at least one for which . To prove the inequality, it suffices to show that whenever and , and and , then . Indeed,
since is a convex cone, and so
since is -positive.
Corollary: Krein's extension theorem
Let E be a real linear space, and let K ⊂ E be a convex cone. Let x ∈ E\(− |
https://en.wikipedia.org/wiki/Genie%20TV | Genie TV (; formerly Olleh TV) is an IPTV service provided by South Korea's main telecom operator KT. It broadcasts VOD including TV programs, movies, children's programs, sports, documentaries and animations.
On 4 October 2022, KT rebranded its IPTV service as Genie TV, 11 years after the Olleh TV brand was introduced. The rebranding is part of the telecom's aggressive effort to push forward its content business as growth in the traditional telecom sector slows.
Olleh TV for PlayStation 3
Since November 20, 2007, following a partnership between KT and Sony Computer Entertainment, Olleh TV is made available for PlayStation 3 owners in South Korea. Once the PlayStation 3's firmware version 2.10 installed in the system, a new icon, called "TV", appears in the XMB. The user is asked to agree a terms of service form then to download the 13MB Olleh TV player.
Olleh TV is also a provider of Spanish television programming. Olleh TV signed a deal with HDTV Inc. to bring three new markets into the Spanish broadcasting circuit. According to Pr Newswire, “This new market includes Tampa, Florida; Charleston, South Carolina; and Palm Springs, California.” The purpose of signing this agreement it to attract a wide range of Hispanic audiences. “Olleh TV already serves eight markets reaching over 3.5 million Hispanic viewers.” “There are many Hispanic shows slated to air on these affiliates. The popular shows include ‘Papparazzi TV Sensacional,’ ‘Maria Elvira Live!’ ‘Esta Noche Tu Night’ with Alexis Valdes and ‘Los Implicados,’ among others." Olleh TV is sure to be a hit among the Hispanic community because they are commitment to providing programming catered to the Hispanic community and culture. |
https://en.wikipedia.org/wiki/Aleksei%20Chernavskii | Aleksei Viktorovich Chernavskii (or Chernavsky or Černavskii) (Алексей Викторович Чернавский, born January 17, 1938, in Moscow) is a Russian mathematician, specializing in differential geometry and topology.
Biography
Chernavskii completed undergraduate study at the Faculty of Mechanics and Mathematics of Moscow State University in 1959. He enrolled in graduate school at the Steklov Institute of Mathematics. In 1964 he defended his Candidate of Sciences (PhD) thesis, written under the under the guidance of Lyudmila Keldysh, on the topic Конечнократные отображения многообразий (Finite-fold mappings of manifolds). In 1970 he defended his Russian Doctor of Sciences (habilitation) thesis Гомеоморфизмы и топологические вложения многообразий (Homeomorphisms and topological embeddings of manifolds). In 1970 he was an Invited Speaker at the International Congress of Mathematicians in Nice.
Chernavskii worked as a senior researcher at the Steklov Institute until 1973 and from 1973 to 1980 at Yaroslavl State University. From 1980 to 1985 he was a senior researcher at the Moscow Institute of Physics and Technology.
Since 1985 he is employed the Kharkevich Institute for Information Transmission Problems of the Russian Academy of Sciences. Since 1993 he has been working part-time as a professor at the Department of Higher Geometry and Topology, Faculty of Mechanics and Mathematics, Moscow State University. He wrote a textbook on differential differential geometry for advanced students.
Chernavskii's theorem
Chernavskii's theorem (1964): If and are n-manifolds and is a discrete, open, continuous mapping of into then the branch set = { x: x is an element of and fails to be a local homeomorphism at x} satisfies dimension () ≤ n – 2.
Selected publications |
https://en.wikipedia.org/wiki/Known-plaintext%20attack | The known-plaintext attack (KPA) is an attack model for cryptanalysis where the attacker has access to both the plaintext (called a crib) and its encrypted version (ciphertext). These can be used to reveal further secret information such as secret keys and code books. The term "crib" originated at Bletchley Park, the British World War II decryption operation, where it was defined as:
History
The usage "crib" was adapted from a slang term referring to cheating (e.g., "I cribbed my answer from your test paper"). A "crib" originally was a literal or interlinear translation of a foreign-language text—usually a Latin or Greek text—that students might be assigned to translate from the original language.
The idea behind a crib is that cryptologists were looking at incomprehensible ciphertext, but if they had a clue about some word or phrase that might be expected to be in the ciphertext, they would have a "wedge," a test to break into it. If their otherwise random attacks on the cipher managed to sometimes produce those words or (preferably) phrases, they would know they might be on the right track. When those words or phrases appeared, they would feed the settings they had used to reveal them back into the whole encrypted message to good effect.
In the case of Enigma, the German High Command was very meticulous about the overall security of the Enigma system and understood the possible problem of cribs. The day-to-day operators, on the other hand, were less careful. The Bletchley Park team would guess some of the plaintext based upon when the message was sent, and by recognizing routine operational messages. For instance, a daily weather report was transmitted by the Germans at the same time every day. Due to the regimented style of military reports, it would contain the word Wetter (German for "weather") at the same location in every message. (Knowing the local weather conditions helped Bletchley Park guess other parts of the plaintext as well.) Other operators, too, |
https://en.wikipedia.org/wiki/Image%20rectification | Image rectification is a transformation process used to project images onto a common image plane. This process has several degrees of freedom and there are many strategies for transforming images to the common plane. Image rectification is used in computer stereo vision to simplify the problem of finding matching points between images (i.e. the correspondence problem), and in geographic information systems to merge images taken from multiple perspectives into a common map coordinate system.
In computer vision
Computer stereo vision takes two or more images with known relative camera positions that show an object from different viewpoints. For each pixel it then determines the corresponding scene point's depth (i.e. distance from the camera) by first finding matching pixels (i.e. pixels showing the same scene point) in the other image(s) and then applying triangulation to the found matches to determine their depth.
Finding matches in stereo vision is restricted by epipolar geometry: Each pixel's match in another image can only be found on a line called the epipolar line.
If two images are coplanar, i.e. they were taken such that the right camera is only offset horizontally compared to the left camera (not being moved towards the object or rotated), then each pixel's epipolar line is horizontal and at the same vertical position as that pixel. However, in general settings (the camera does move towards the object or rotate) the epipolar lines are slanted. Image rectification warps both images such that they appear as if they have been taken with only a horizontal displacement and as a consequence all epipolar lines are horizontal, which slightly simplifies the stereo matching process. Note however, that rectification does not fundamentally change the stereo matching process: It searches on lines, slanted ones before and horizontal ones after rectification.
Image rectification is also an equivalent (and more often used) alternative to perfect camera coplanarity. Even |
https://en.wikipedia.org/wiki/SHORT%20syndrome | SHORT syndrome is a medical condition in which affected individuals have multiple birth defects in different organ systems.
It was characterized in 1975.
Presentation
SHORT is an acronym for short stature, hyperextensibility of joints and/or inguinal hernia, ocular depression, Rieger anomaly and teething delay. Other characteristics common in SHORT syndrome are a triangular face, a prominent forehead, small chin with a dimple, a loss of fat under the skin (lipodystrophy), prominent ears (but no low implantation or posterior localisation), hearing loss and delayed speech. Facial lipodystrophy may be evident during birth and later on in the chest and higher extremities, but it usually won't affect buttocks and legs. Diabetes has been observed in ⅔ of the affected after they turn 15.
Diagnosis
Diagnosis is based on facial characteristics and molecular genetic testing that will show a mutation on gene PIK3R1 (5q13.1), which codifies the regulating alpha subunit of phosphatidylinositol 3-kinase. This mutation can alter the PI3K/AKT/mTOR signal route, which plays an important role in cell growth and proliferation.
Treatment
Treatment involves multiple disciplines.
-Screening for insulin resistance during late childhood stage.
-Glucose intolerance and diabetes mellitus can be treated with a different diet and lifestyle changes.
-Regular eye checkups are recommended in order to keep vision.
-Dental anomalies can be treated with common methods (protheses, crown, etc.) |
https://en.wikipedia.org/wiki/Pop-up%20satellite%20archival%20tag | Pop-up satellite archival tags (PSATs) are used to track movements of (usually large, migratory) marine animals. A PSAT (also commonly referred to as a PAT tag) is an archival tag (or data logger) that is equipped with a means to transmit the collected data via the Argos satellite system. Though the data are physically stored on the tag, its major advantage is that it does not have to be physically retrieved like an archival tag for the data to be available making it a viable, fishery independent tool for animal behavior and migration studies. They have been used to track movements of ocean sunfish, marlin, blue sharks, bluefin tuna, swordfish and sea turtles to name a few species. Location, depth, temperature, oxygen levels, and body movement data are used to answer questions about migratory patterns, seasonal feeding movements, daily habits, and survival after catch and release, for examples.
A satellite tag is generally constructed of several components: a data-logging section, a release section, a float, and an antenna. The release sections include an energetically popped off release section or a corrosive pin that is actively corroded on a preset date or after a specified period of time. Some limitations of using satellite tags are their depth limitations (2000m), their costs ($499–$4000+), their vulnerability to loss by environmental issues (biofouling), or premature release through ingestion by a predator.
There are two methods of underwater geolocation that PSATs employ. The first method is through light based geolocation which uses the length of the day and a noon time calculation to estimate the tags location while underwater. This method has a functional depth limitation of light penetration which can be as shallow as a few meters to upwards of hundreds of meters. Geolocation estimates based on light are usually coupled with additional satellite data like sea surface temperature or other available data input such as bathymetry, land avoidance, a |
https://en.wikipedia.org/wiki/Nuclear%20fallout | Nuclear fallout is the residual radioactive material propelled into the upper atmosphere following a nuclear blast, so called because it "falls out" of the sky after the explosion and the shock wave has passed. It commonly refers to the radioactive dust and ash created when a nuclear weapon explodes. The amount and spread of fallout is a product of the size of the weapon and the altitude at which it is detonated. Fallout may get entrained with the products of a pyrocumulus cloud and fall as black rain (rain darkened by soot and other particulates, which fell within 30–40 minutes of the atomic bombings of Hiroshima and Nagasaki). This radioactive dust, usually consisting of fission products mixed with bystanding atoms that are neutron-activated by exposure, is a form of radioactive contamination.
Types of fallout
Fallout comes in two varieties. The first is a small amount of carcinogenic material with a long half-life. The second, depending on the height of detonation, is a large quantity of radioactive dust and sand with a short half-life.
All nuclear explosions produce fission products, un-fissioned nuclear material, and weapon residues vaporized by the heat of the fireball. These materials are limited to the original mass of the device, but include radioisotopes with long lives. When the nuclear fireball does not reach the ground, this is the only fallout produced. Its amount can be estimated from the fission-fusion design and yield of the weapon.
Global fallout
After the detonation of a weapon at or above the fallout-free altitude (an air burst), fission products, un-fissioned nuclear material, and weapon residues vaporized by the heat of the fireball condense into a suspension of particles 10 nm to 20 µm in diameter. This size of particulate matter, lifted to the stratosphere, may take months or years to settle, and may do so anywhere in the world. Its radioactive characteristics increase the statistical cancer risk. Elevated atmospheric radioactivity rem |
https://en.wikipedia.org/wiki/Time%20Sharing%20Option | Time Sharing Option (TSO) is an interactive time-sharing environment for IBM mainframe operating systems, including OS/360 MVT, OS/VS2 (SVS), MVS, OS/390, and z/OS.
Use
In computing, time-sharing is a design technique that allows many people to use a computer system concurrently and independently—without interfering with each other. Each TSO user is isolated; it appears to each one that they are the only user of the system.
TSO is most commonly used by mainframe system administrators and programmers. It provides:
A text editor
Batch job support, including completion notification
Debuggers for some programming languages used on System/360 and later IBM mainframes
Support for other vendors' end-user applications, for example for querying IMS and DB2 databases
TSO interacts with users in either a line-by-line mode or in a full screen, menu-driven mode. In the line-by-line mode, the user enters commands by typing them in at the keyboard; in turn, the system interprets the commands, and then displays responses on the terminal screen. But most mainframe interaction is actually via ISPF, which allows for customized menu-driven interaction. This combination is called TSO/ISPF. TSO can also provide a Unix-style environment on OS/390 and z/OS via the UNIX System Services command shell, with or without ISPF.
TSO commands can be embedded in REXX execs or CLISTs, which can run interactively or in batch.
TSO eliminated the need to punch cards on a keypunch machine, and send card decks to the computer room to be read by a card reading machine.
History
When it was originally introduced in 1971, IBM considered time-sharing an "optional feature", as compared to standard batch processing, and hence offered TSO as an option for OS/360 MVT. With the introduction of MVS in 1974, IBM made it a standard component of their top-end mainframe operating system. TSO/E ("Time Sharing Option/Extensions") is a set of extensions to the original TSO. TSO/E is a base element of z/OS. Befo |
https://en.wikipedia.org/wiki/Key%20%28cryptography%29 | A key in cryptography is a piece of information, usually a string of numbers or letters that are stored in a file, which, when processed through a cryptographic algorithm, can encode or decode cryptographic data. Based on the used method, the key can be different sizes and varieties, but in all cases, the strength of the encryption relies on the security of the key being maintained. A key's security strength is dependent on its algorithm, the size of the key, the generation of the key, and the process of key exchange.
Scope
The key is what is used to encrypt data from plaintext to ciphertext. There are different methods for utilizing keys and encryption.
Symmetric cryptography
Symmetric cryptography refers to the practice of the same key being used for both encryption and decryption.
Asymmetric cryptography
Asymmetric cryptography has separate keys for encrypting and decrypting. These keys are known as the public and private keys, respectively.
Purpose
Since the key protects the confidentiality and integrity of the system, it is important to be kept secret from unauthorized parties. With public key cryptography, only the private key must be kept secret, but with symmetric cryptography, it is important to maintain the confidentiality of the key. Kerckhoff's principle states that the entire security of the cryptographic system relies on the secrecy of the key.
Key sizes
Key size is the number of bits in the key defined by the algorithm. This size defines the upper bound of the cryptographic algorithm's security. The larger the key size, the longer it will take before the key is compromised by a brute force attack. Since perfect secrecy is not feasible for key algorithms, researches are now more focused on computational security.
In the past, keys were required to be a minimum of 40 bits in length, however, as technology advanced, these keys were being broken quicker and quicker. As a response, restrictions on symmetric keys were enhanced to be greater in s |
https://en.wikipedia.org/wiki/IEC%2062304 | IEC 62304 – medical device software – software life cycle processes is an international standard published by the International Electrotechnical Commission (IEC). The standard specifies life cycle requirements for the development of medical software and software within medical devices. It has been adopted as national standards and therefore can be used as a benchmark to comply with regulatory requirements.
Implications of IEC 62304 for software
The IEC 62304 standard calls out certain cautions on using software, particularly SOUP (software of unknown pedigree or provenance). The standard spells out a risk-based decision model on when the use of SOUP is acceptable, and defines testing requirements for SOUP to support a rationale on why such software should be used.
Contents
General requirements
Quality management system
Risk management
Software safety classification
Software development process
Software development planning
Software requirements analysis
Software architectural design
Software detailed design
Software unit implementation and verification
Software integration and integration testing
Software system testing
Software release
Effect of safety classification on required development process documentation
Software maintenance process
Establish software maintenance plan
Problem and modification analysis
Modification implementation
Software risk management process
Analysis of software contributing to hazardous situations
Risk control measures
Verification of risk control measures
Risk management of software changes
Security and reliability through software quality
Software configuration management process
Configuration identification
Change control
Configuration status accounting
Software problem resolution process
Prepare problem reports
Investigate the problem
Advise relevant parties
Use change control process
Maintain records
Analyse problems for trends
Verify software problem resolution
Test documen |
https://en.wikipedia.org/wiki/Finite%20Legendre%20transform | The finite Legendre transform (fLT) transforms a mathematical function defined on the finite interval into its Legendre spectrum.
Conversely, the inverse fLT (ifLT) reconstructs the original function from the components of the Legendre spectrum and the Legendre polynomials, which are orthogonal on the interval [−1,1]. Specifically, assume a function x(t) to be defined on an interval [−1,1] and discretized into N equidistant points on this interval. The fLT then yields the decomposition of x(t) into its spectral Legendre components,
where the factor (2k + 1)/N serves as normalization factor and Lx(k) gives the contribution of the k-th Legendre polynomial to x(t) such that (ifLT)
The fLT should not be confused with the Legendre transform or Legendre transformation used in thermodynamics and quantum physics.
Legendre filter
The fLT of a noisy experimental outcome s(t) and the subsequent application of the inverse fLT (ifLT) on an appropriately truncated Legendre spectrum of s(t) gives a smoothed version of s(t). The fLT and incomplete ifLT thus act as a filter. In contrast to the common Fourier low-pass filter which transmits low frequency harmonics and filters out high frequency harmonics, the Legendre lowpass transmits signal components proportional to low degree Legendre polynomials, while signal components proportional to higher degree Legendre polynomials are filtered out. |
https://en.wikipedia.org/wiki/Gauss%27s%20lemma%20%28Riemannian%20geometry%29 | In Riemannian geometry, Gauss's lemma asserts that any sufficiently small sphere centered at a point in a Riemannian manifold is perpendicular to every geodesic through the point. More formally, let M be a Riemannian manifold, equipped with its Levi-Civita connection, and p a point of M. The exponential map is a mapping from the tangent space at p to M:
which is a diffeomorphism in a neighborhood of zero. Gauss' lemma asserts that the image of a sphere of sufficiently small radius in TpM under the exponential map is perpendicular to all geodesics originating at p. The lemma allows the exponential map to be understood as a radial isometry, and is of fundamental importance in the study of geodesic convexity and normal coordinates.
Introduction
We define the exponential map at by
where is the unique geodesic with and tangent and is chosen small enough so that for every the geodesic is defined. So, if is complete, then, by the Hopf–Rinow theorem, is defined on the whole tangent space.
Let be a curve differentiable in such that and . Since , it is clear that we can choose . In this case, by the definition of the differential of the exponential in applied over , we obtain:
So (with the right identification ) the differential of is the identity. By the implicit function theorem, is a diffeomorphism on a neighborhood of . The Gauss Lemma now tells that is also a radial isometry.
The exponential map is a radial isometry
Let . In what follows, we make the identification .
Gauss's Lemma states:
Let and . Then,
For , this lemma means that is a radial isometry in the following sense: let , i.e. such that is well defined.
And let . Then the exponential remains an isometry in , and, more generally, all along the geodesic (in so far as is well defined)! Then, radially, in all the directions permitted by the domain of definition of , it remains an isometry.
Proof
Recall that
We proceed in three steps:
: let us construct a curve
such tha |
https://en.wikipedia.org/wiki/The%20Tao%20of%20Programming | The Tao of Programming is a book written in 1987 by Geoffrey James. Written in a tongue-in-cheek style spoof of classic Taoist texts such as the Tao Te Ching and Zhuangzi which belies its serious message, it consists of a series of short anecdotes divided into nine "books":
The Silent Void
The Ancient Masters
Design
Coding
Maintenance
Management
Corporate Wisdom
Hardware and Software
Epilogue
Geoffrey James wrote two other books on this theme, The Zen of Programming (978-0931137099) in 1988 and Computer Parables: Enlightenment in the Information Age (978-0931137136) in 1989.
See also
Hacker koan |
https://en.wikipedia.org/wiki/Decimal32%20floating-point%20format | In computing, decimal32 is a decimal floating-point computer numbering format that occupies 4 bytes (32 bits) in computer memory.
It is intended for applications where it is necessary to emulate decimal rounding exactly, such as financial and tax computations. Like the binary16 format, it is intended for memory saving storage.
Decimal32 supports 7 decimal digits of significand and an exponent range of −95 to +96, i.e. to ±. (Equivalently, to .) Because the significand is not normalized (there is no implicit leading "1"), most values with less than 7 significant digits have multiple possible representations; , etc. Zero has 192 possible representations (384 when both signed zeros are included).
Decimal32 floating point is a relatively new decimal floating-point format, formally introduced in the 2008 version of IEEE 754 as well as with ISO/IEC/IEEE 60559:2011.
Representation of decimal32 values
IEEE 754 allows two alternative representation methods for decimal32 values.
The standard does not specify how to signify which representation is used,
for instance in a situation where decimal32 values are communicated between systems.
In one representation method, based on binary integer decimal (BID),
the significand is represented as binary coded positive integer.
The other, alternative, representation method is based on
densely packed decimal (DPD) for most of the
significand (except the most significant digit).
Both alternatives provide exactly the same range of representable numbers: 7 digits of significand and possible exponent values.
In both encodings, BID and DPD, the 2 most significant exponent bits, and the 4 most significant bits of the significand, are combined to 5 bits. The position of the 5 bits in the combination field varies, but otherwise the encoding is identical. 5 bits suffice instead of 6, because the 2 MSBs from the exponent only encode values from 0 to 2 (3 possible values), and the 4 MSBs of the significand represent a decimal digit bet |
https://en.wikipedia.org/wiki/Asperity%20%28materials%20science%29 | In materials science, asperity, defined as "unevenness of surface, roughness, ruggedness" (from the Latin asper—"rough"), has implications (for example) in physics and seismology. Smooth surfaces, even those polished to a mirror finish, are not truly smooth on a microscopic scale. They are rough, with sharp, rough or rugged projections, termed "asperities". Surface asperities exist across multiple scales, often in a self affine or fractal geometry. The fractal dimension of these structures has been correlated with the contact mechanics exhibited at an interface in terms of friction and contact stiffness.
When two macroscopically smooth surfaces come into contact, initially they only touch at a few of these asperity points. These cover only a very small portion of the surface area. Friction and wear originate at these points, and thus understanding their behavior becomes important when studying materials in contact. When the surfaces are subjected to a compressive load, the asperities deform through elastic and plastic modes, increasing the contact area between the two surfaces until the contact area is sufficient to support the load.
The relationship between frictional interactions and asperity geometry is complex and poorly understood. It has been reported that an increased roughness may under certain circumstances result in weaker frictional interactions while smoother surfaces may in fact exhibit high levels of friction owing to high levels of true contact.
The Archard equation provides a simplified model of asperity deformation when materials in contact are subject to a force. Due to the ubiquitous presence of deformable asperities in self affine hierarchical structures, the true contact area at an interface exhibits a linear relationship with the applied normal load.
See also
Surface roughness
Burnishing (metal) |
https://en.wikipedia.org/wiki/Respiratory%20tract%20antimicrobial%20defense%20system | The respiratory tract antimicrobial defense system is a layered defense mechanism which relies on components of both the innate and adaptive immune systems to protect the lungs and the rest of the respiratory tract against inhaled microorganisms.
In the first line of defense, inhaled bacteria are trapped by mucus and are swept toward the pharynx and are swallowed. Bacteria which penetrate the mucous layer are dealt with a second line of defense which includes antimicrobial peptides that are secreted by the surface epithelium of the respiratory tract which kill many strains of bacteria. Those bacteria that are resistant to antimicrobial peptides are killed by a variety of reactive oxygen species produced by phagocytes. In a third line of defense and as a last resort, persistent bacterial infections which escape the innate immune system are eliminated by the adaptive immune system.
Lactoferrin
Lactoferrin (LF) is a multifunctional protein which is an essential part of the respiratory tract antimicrobial defense system. LF proteolysis produces the small peptides lactoferricin and kaliocin-1 both with antimicrobial activity.
Reactive oxygen species, oxygenated compounds
Phagocytes possess a superoxide-producing NADPH oxidase enzyme complex. Other cells in the respiratory tract also produce superoxide and hydrogen peroxide through the activity of dual oxidase 2 proteins, also known as Duox2.
The superoxide generated by these enzymes complexes dismutates into hydrogen peroxide which in turn is used by myeloperoxidase to produce bactericidal hypochlorous acid. In addition, the submucosal glands of the respiratory tract secrete myeloperoxidase and lactoperoxidase (LPO) that catalyzes the oxidation of thiocyanate and detoxify hydrogen peroxide or ROS to the antimicrobial hypothiocyanite.
The oxygenated compounds produced by the lactoperoxidase system does not attack DNA and is not mutagenic and is known to be safe. Hypothiocyanite generated through this pathway |
https://en.wikipedia.org/wiki/28%20%28number%29 | 28 (twenty-eight) is the natural number following 27 and preceding 29.
In mathematics
It is a composite number; a square-prime, of the form (p2,q) where q is a higher prime. It is the third of this form and of the specific form (22.q), with proper divisors being 1, 2, 4, 7, and 14.
Twenty-eight is the second perfect number - it is the sum of its proper divisors: . As a perfect number, it is related to the Mersenne prime 7, since . The next perfect number is 496, the previous being 6.
Though perfect, 28 is not the aliquot sum of any other number other than itself, and so; unusually, is not part of a multi-number aliquot sequence.
The next perfect number, 496, has the single Aliquot sum, 652 which leads to multiple aliquot sequencing.
Twenty-eight is the sum of the totient function for the first nine integers.
Since the greatest prime factor of is 157, which is more than 28 twice, 28 is a Størmer number.
Twenty-eight is a harmonic divisor number, a happy number, a triangular number, a hexagonal number, a Leyland number of the second kind and a centered nonagonal number.
It appears in the Padovan sequence, preceded by the terms 12, 16, 21 (it is the sum of the first two of these).
It is also a Keith number, because it recurs in a Fibonacci-like sequence started from its decimal digits: 2, 8, 10, 18, 28...
Twenty-eight is the ninth and last number in early Indian magic square of order 3.
There are twenty-eight convex uniform honeycombs.
Twenty-eight is the only positive integer that has a unique Kayles nim-value.
Twenty-eight is the only known number that can be expressed as a sum of the first nonnegative (or positive) integers (), a sum of the first primes () and a sum of the first nonprimes (), and it is unlikely that any other number has this property.
There are twenty-eight oriented diffeomorphism classes of manifolds homeomorphic to the 7-sphere.
There are 28 elements of the cuboid: 8 vertices, 12 edges, 6 faces, 2 3-dimensional elements (interior a |
https://en.wikipedia.org/wiki/Vernon%27s%20verbal-perceptual%20model | Vernon's verbal-perceptual model is a theory about the structure of intelligence proposed by Philip E. Vernon in 1964 (Vernon, 1964, 1965). It was influenced by the theory of g factor.
Vernon puts emphasis on the g factor in all the mental abilities. He extracted the g factor from an ability test, then found that could be divided into two separate parts. He named those two orthogonal group factors as verbal-educational factor (v:ed) and perceptual-mechanical skill factor (k:m).
v:ed factor: verbal and educational abilities
k:m factor: spatial, practical, and mechanical abilities
This is a rough distinction between verbal and non-verbal intelligence measurement.
Vernon also said that v:ed and k:m can represent different education and cultural experience. The v:ed factor may come from school life and k:m factor comes from skills forming in non-school time. Vernon considered that there might be a third special factor named mathematical skill, but it can be included in perceptual-mechanical (k:m) factor.
Difference from Horn and Cattell's Gf-Gc theory
Vernon's model about intelligence looks similar to the fluid-crystallized (Gf-Gc) intelligence theory because they both agree with g factor and have two more different dimensions on intelligence structure. In fact, Gf-Gc model has more broad factors such as special visualization (Gv), retrieval (Gr) or speed factor (Gs).
"One reason why the Horn-Cattell model has more broad group factors than Vernon's is simply that they analysed a very much larger and more carefully chosen battery of tests. This probably explains why the Vernon model does not contain any group factors corresponding to Cattell's Gr (and why it had to be added by Johnson and Bouchard, 2005, when they tried to fit the Vernon model to a much larger data set)." IQ and human intelligence
See also
g-VPR model |
https://en.wikipedia.org/wiki/Parasitic%20worm | Parasitic worms, also known as helminths, are large macroparasites; adults can generally be seen with the naked eye. Many are intestinal worms that are soil-transmitted and infect the gastrointestinal tract. Other parasitic worms such as schistosomes reside in blood vessels.
Some parasitic worms, including leeches and monogeneans, are ectoparasites thus, they are not classified as helminths, which are endoparasites.
Parasitic worms live in and feed in living hosts. They receive nourishment and protection while disrupting their hosts' ability to absorb nutrients. This can cause weakness and disease in the host, and poses a global health and economic problem. Parasitic worms cannot reproduce entirely within their host's body; they have a life cycle that includes some stages that need to take place outside of the host. Helminths are able to survive in their mammalian hosts for many years due to their ability to manipulate the host's immune response by secreting immunomodulatory products. All parasitic worms produce eggs during reproduction. These eggs have a strong shell that protects them against a range of environmental conditions. The eggs can therefore survive in the environment for many months or years.
Many of the worms referred to as helminths are intestinal parasites. An infection by a helminth is known as helminthiasis, helminth infection, or intestinal worm infection. There is a naming convention which applies to all helminths: the ending "-asis" (or in veterinary science: "-osis") is added at the end of the name of the worm to denote the infection with that particular worm. For example, Ascaris is the name of a type of helminth, and ascariasis is the name of the infection caused by that helminth.
Taxonomy
Helminths are a group of organisms which share a similar form but are not necessarily related as part of evolution. The term "helminth" is an artificial term. There is no real consensus on the taxonomy (or groupings) of the helminths, particularly wi |
https://en.wikipedia.org/wiki/Variance%20of%20the%20mean%20and%20predicted%20responses | In regression, mean response (or expected response) and predicted response, also known as mean outcome (or expected outcome) and predicted outcome, are values of the dependent variable calculated from the regression parameters and a given value of the independent variable. The values of these two responses are the same, but their calculated variances are different.
The concept is a generalization of the distinction between the standard error of the mean and the sample standard deviation.
Background
In straight line fitting, the model is
where is the response variable, is the explanatory variable, εi is the random error, and and are parameters. The mean, and predicted, response value for a given explanatory value, xd, is given by
while the actual response would be
Expressions for the values and variances of and are given in linear regression.
Variance of the mean response
Since the data in this context is defined to be (x, y) pairs for every observation, the mean response at a given value of x, say xd, is an estimate of the mean of the y values in the population at the x value of xd, that is . The variance of the mean response is given by
This expression can be simplified to
where m is the number of data points.
To demonstrate this simplification, one can make use of the identity
Variance of the predicted response
The predicted response distribution is the predicted distribution of the residuals at the given point xd. So the variance is given by
The second line follows from the fact that is zero because the new prediction point is independent of the data used to fit the model. Additionally, the term was calculated earlier for the mean response.
Since (a fixed but unknown parameter that can be estimated), the variance of the predicted response is given by
Confidence intervals
The confidence intervals are computed as . Thus, the confidence interval for predicted response is wider than the interval for mean response. This is expected i |
https://en.wikipedia.org/wiki/Bangstad%20syndrome | Bangstad syndrome is a severe, inherited congenital disorder associated with abnormalities of the cell membrane.
It was characterized in 1989 by H. J. Bangstad.
Presentation
Presenting at birth, features of the disorder include moderately severe IUGR, microcephaly, craniosynostosis, moderately severe post-uterine growth retardation, deafness, deep-set eyes, cryptorchidism, truncal obesity and acanthosis nigricans, small teeth, prognathism, dislocated radial heads without generalized skeletal dysplasia, however, tall vertebrae, moderate mental retardation, hypothyroidism, insulin resistance, hypoparathyroidism.
Diagnosis
Treatment |
https://en.wikipedia.org/wiki/Electronics%20For%20You | Electronics For You magazine is India's first monthly publication for electronics engineers. It was first conceptualised at IIT Madras in 1969 by Ramesh Chopra, and was published by EFY Enterprises Pvt Ltd headed by S.P Chopra and Veena Khanna.
The publisher of this magazine currently manages multiple magazines, annual events, and around 30 book titles. The company also provides hands-on training courses, and manufactures and markets Do-It-Yourself electronics projects and hobby kits. It has partnered with Mouser Electronics for their entire IoT series in India. The magazine has partnered with ELCINA to conduct events that recognise and award innovative technology companies. Additionally, the magazine sponsors the Electronics For You Prize, an award given to a student at IIT Madras each year.
Electronics For You magazine has a history of being collected and saved by engineers and technologists across India. |
https://en.wikipedia.org/wiki/Event-related%20potential | An event-related potential (ERP) is the measured brain response that is the direct result of a specific sensory, cognitive, or motor event. More formally, it is any stereotyped electrophysiological response to a stimulus. The study of the brain in this way provides a noninvasive means of evaluating brain functioning.
ERPs are measured by means of electroencephalography (EEG). The magnetoencephalography (MEG) equivalent of ERP is the ERF, or event-related field. Evoked potentials and induced potentials are subtypes of ERPs.
History
With the discovery of the electroencephalogram (EEG) in 1924, Hans Berger revealed that one could measure the electrical activity of the human brain by placing electrodes on the scalp and amplifying the signal. Changes in voltage can then be plotted over a period of time. He observed that the voltages could be influenced by external events that stimulated the senses. The EEG proved to be a useful source in recording brain activity over the ensuing decades. However, it tended to be very difficult to assess the highly specific neural process that are the focus of cognitive neuroscience because using pure EEG data made it difficult to isolate individual neurocognitive processes. Event-related potentials (ERPs) offered a more sophisticated method of extracting more specific sensory, cognitive, and motor events by using simple averaging techniques.
In 1935–1936, Pauline and Hallowell Davis recorded the first known ERPs on awake humans and their findings were published a few years later, in 1939. Due to World War II not much research was conducted in the 1940s, but research focusing on sensory issues picked back up again in the 1950s. In 1964, research by Grey Walter and colleagues began the modern era of ERP component discoveries when they reported the first cognitive ERP component, called the contingent negative variation (CNV). Sutton, Braren, and Zubin (1965) made another advancement with the discovery of the P3 component. Over the next |
https://en.wikipedia.org/wiki/Soul%20Electronics | Soul Electronics (stylized as SOUL Electronics or simply SOUL) is an audio equipment company. Founded in 2010, it produces various lines of wireless, Bluetooth-enabled headphones, earbuds, and speakers.
History
Soul Electronics was founded in 2010. The entity was originally known as Signeo USA, an American subsidiary of the Hong Kong-based Signeo Design International. In January 2011 at that year's Consumer Electronics Show (CES), the company introduced SOUL by Ludacris, a line of headphones designed in collaboration with the rapper, Ludacris. The line featured five different pairs of headphones, which were released to the public later in 2011. By that time, the company had become known as Soul Electronics.
In 2012, Soul Electronics partnered with sprinter, Usain Bolt, on the design of a new line of headphones and earbuds designed to be used while running or exercising. In the following two years, the company entered into a number of other sponsored partnerships with athletes, including Brendan Schaub, Tim Tebow, and Alex Fowler. In 2015, the company introduced the Combat+ Sync headphones which came with a built-in walkie-talkie feature. At CES 2018, the company introduced a new artificial intelligence feature in its "Run Free Pro Bio" and "Blade" models that would offer users live, in-ear coaching while running.
In 2019, the company introduced four new products: the Ultra Wireless over-ear headphones, the ST-XX wireless earbuds, the ST-XS2 wireless earphones, and the S-Storm portable Bluetooth speaker. Four additional earbud and earphone lines were released in 2020, including SYNC Pro, the SYNC ANC (active noise-canceling), S-Fit, and S-Gear.
In 2023, the company introduced their first open-ear style headphones called the Openear Series which include Openear 2, Openear Plus, Openear S-Clip and Openear S-Free.
Products
Soul Electronics produces a number of audio products, generally targeted at runners and other consumers who exercise. Its current product |
https://en.wikipedia.org/wiki/Cleaning%20station | A cleaning station is a location where aquatic life congregate to be cleaned by smaller creatures. Such stations exist in both freshwater and marine environments, and are used by animals including fish, sea turtles and hippos, referred to as clients.
The cleaning process includes the removal of parasites from the animal's body (both externally and internally), and is performed by various smaller animals including cleaner shrimp and numerous species of cleaner fish, especially wrasses and gobies (Elacatinus spp.), collectively referred to as cleaners.
When the animal approaches a cleaning station, it will open its mouth wide or position its body in such a way as to signal that it needs to be cleaned. The cleaner fish will then remove and eat the parasites from the skin, even swimming into the mouth and gills of any fish being cleaned. This is a form of cleaning symbiosis. How predator clients recognize cleaners is still uncertain. It has been hypothesized that color, size, and pattern indicate to clients that an organism is a cleaner. For example, cleaning gobies tend to exhibit full-body lateral stripes, unlike their non-cleaning counterparts, which exhibit shorter lateral stripes. Cleaners also tend to be smaller because in fish species, usually juveniles are cleaners.
Cleaning stations may be associated with coral reefs, located either on top of a coral head or in a slot between two outcroppings. Other cleaning stations may be located under large clumps of floating seaweed or at an accepted point in a river or lagoon. Cleaning stations are an exhibition of mutualism between cleaners and clients.
Cleaner fish may also impact species diversity around coral reefs. Some clients have smaller home ranges and can only access one cleaning station. Clients with larger home ranges are able to access a variety of cleaning stations and are capable of choosing between cleaning stations. Visitor clients travel long distances to a cleaning station and are not local to the |
https://en.wikipedia.org/wiki/Sailfish%20OS | Sailfish OS is a Linux-based operating system based on free software, and open source projects such as Mer as well as including a closed source UI. The project is being developed by the Finnish company Jolla.
The OS first shipped with the original Jolla Phone in 2013; while its sale stopped in 2016, it was supplied with software updates until the end of 2020. It also shipped with Jolla Tablet in 2015 and from other vendors licensing the OS. The OS is ported by community enthusiasts to third-party mobile devices including smartphones and tablet computers. Sailfish OS can be used for many kinds of devices.
History and development
The OS is an evolved continuation of the Linux MeeGo OS previously developed by alliance of Nokia and Intel which itself relies on combined Maemo and Moblin. The MeeGo legacy is contained in the Mer core in about 80% of its code; the Mer name thus expands to MEego Reconstructed. This base is extended by Jolla with a custom user interface and default applications. Jolla and MERproject.org follow a meritocratic system to avoid the mistakes that led to the MeeGo project's then-unanticipated discontinuation.
The main elements for include:
Technically stronger OS core
Improved Android application compatibility
Support for ARM and Intel architectures, including the Intel Atom x3 processor, or any platform with kernel useable (settle-able) for MER core stack (also called middleware of Sailfish).
Design to provide visibility in the UI for digital content providers and to enable OS level integration for mobile commerce
Strong multitasking (one of the most important advantage of the OS and declared to be the best one on the market)
Strong privacy and personalization features
Enhanced user interface with new UI/UX features, including simpler swipe access to main functions, enhanced notifications and events views.
Software architecture
The and the Sailfish software development kit (SDK) are based on the Linux kernel and Mer. includes |
https://en.wikipedia.org/wiki/Schr%C3%B6dinger%20field | In quantum mechanics and quantum field theory, a Schrödinger field, named after Erwin Schrödinger, is a quantum field which obeys the Schrödinger equation. While any situation described by a Schrödinger field can also be described by a many-body Schrödinger equation for identical particles, the field theory is more suitable for situations where the particle number changes.
A Schrödinger field is also the classical limit of a quantum Schrödinger field, a classical wave which satisfies the Schrödinger equation. Unlike the quantum mechanical wavefunction, if there are interactions between the particles the equation will be nonlinear. These nonlinear equations describe the classical wave limit of a system of interacting identical particles.
The path integral of a Schrödinger field is also known as a coherent state path integral, because the field itself is an annihilation operator whose eigenstates can be thought of as coherent states of the harmonic oscillations of the field modes.
Schrödinger fields are useful for describing Bose–Einstein condensation, the Bogolyubov–de Gennes equation of superconductivity, superfluidity, and many-body theory in general. They are also a useful alternative formalism for nonrelativistic quantum mechanics.
A Schrödinger field is the nonrelativistic limit of a Klein–Gordon field.
Summary
A Schrödinger field is a quantum field whose quanta obey the Schrödinger equation. In the classical limit, it can be understood as the quantized wave equation of a Bose Einstein condensate or a superfluid.
Free field
A Schrödinger field has the free field Lagrangian
When is a complex valued field in a path integral, or equivalently an operator with canonical commutation relations, it describes a collection of identical non-relativistic bosons. When is a Grassmann valued field, or equivalently an operator with canonical anti-commutation relations, the field describes identical fermions.
External potential
If the particles interact with an extern |
https://en.wikipedia.org/wiki/PocketBook%20International | PocketBook is a multinational company which produces e-book readers based on E Ink technology (an electronic paper technology) under the PocketBook brand. The company was founded in 2007 in Kyiv, Ukraine, and its headquarters were shifted to Lugano, Switzerland in 2012. These devices enable users to browse, buy, download, and read e-books, newspapers, magazines and other digital media via wireless networking to the PocketBook Store.
Development and manufacture
The devices are assembled by factories such as Foxconn, Wisky and Yitoa and shipped out to more than 40 countries.
Device list
Models from the same family sometimes use the same model numbers, such as:
Touch HD PB631...
Touch HD 2 PB631-2...
Sales geography
The company's products are sold in 35 countries worldwide – in Central, Eastern, and Western Europe, in Baltic and Commonwealth of Independent States countries, as well as in Australia, Israel, New Zealand, and others. , PocketBook claimed it had sold over two million e-reader and tablet devices.
History
See also
Comparison of e-book readers
Comparison of tablet computers
Tablet computer |
https://en.wikipedia.org/wiki/Leibniz%20integral%20rule | In calculus, the Leibniz integral rule for differentiation under the integral sign states that for an integral of the form
where and the integrands are functions dependent on the derivative of this integral is expressible as
where the partial derivative indicates that inside the integral, only the variation of with is considered in taking the derivative. It is named after Gottfried Leibniz.
In the special case where the functions and are constants and with values that do not depend on this simplifies to:
If is constant and , which is another common situation (for example, in the proof of Cauchy's repeated integration formula), the Leibniz integral rule becomes:
This important result may, under certain conditions, be used to interchange the integral and partial differential operators, and is particularly useful in the differentiation of integral transforms. An example of such is the moment generating function in probability theory, a variation of the Laplace transform, which can be differentiated to generate the moments of a random variable. Whether Leibniz's integral rule applies is essentially a question about the interchange of limits.
General form: differentiation under the integral sign
The right hand side may also be written using Lagrange's notation as:
Stronger versions of the theorem only require that the partial derivative exist almost everywhere, and not that it be continuous. This formula is the general form of the Leibniz integral rule and can be derived using the fundamental theorem of calculus. The (first) fundamental theorem of calculus is just the particular case of the above formula where is constant, and does not depend on
If both upper and lower limits are taken as constants, then the formula takes the shape of an operator equation:
where is the partial derivative with respect to and is the integral operator with respect to over a fixed interval. That is, it is related to the symmetry of second derivatives, but invol |
https://en.wikipedia.org/wiki/Sexual%20dimorphism%20measures | Although the subject of sexual dimorphism is not in itself controversial, the measures by which it is assessed differ widely. Most of the measures are used on the assumption that a random variable is considered so that probability distributions should be taken into account. In this review, a series of sexual dimorphism measures are discussed concerning both their definition and the probability law on which they are based. Most of them are sample functions, or statistics, which account for only partial characteristics, for example the mean or expected value, of the distribution involved. Further, the most widely used measure fails to incorporate an inferential support.
Introduction
It is widely known that sexual dimorphism is an important component of the morphological variation in biological populations (see, e.g., Klein and Cruz-Uribe, 1984; Oxnard, 1987; Kelley, 1993). In higher Primates, sexual dimorphism is also related to some aspects of the social organization and behavior (Alexander et al., 1979; Clutton-Brock, 1985). Thus, it has been observed that the most dimorphic species tend to polygyny and a social organization based on male dominance, whereas in the less dimorphic species, monogamy and family groups are more common. Fleagle et al. (1980) and Kay (1982), on the other hand, have suggested that the behavior of extinct species can be inferred on the basis of sexual dimorphism and, e.g. Plavcan and van Schaick (1992) think that sex differences in size among primate species reflect processes of an ecological and social nature. Some references on sexual dimorphism regarding human populations can be seen in Lovejoy (1981), Borgognini Tarli and Repetto (1986) and Kappelman (1996).
These biological facts do not appear to be controversial. However, they are based on a series of different sexual dimorphism measures, or indices. Sexual dimorphism, in most works, is measured on the assumption that a random variable is being taken into account. This means that th |
https://en.wikipedia.org/wiki/Sterilization%20%28microbiology%29 | Sterilization (or sterilisation) refers to any process that removes, kills, or deactivates all forms of life (particularly microorganisms such as fungi, bacteria, spores, and unicellular eukaryotic organisms) and other biological agents such as prions present in or on a specific surface, object, or fluid. Sterilization can be achieved through various means, including heat, chemicals, irradiation, high pressure, and filtration. Sterilization is distinct from disinfection, sanitization, and pasteurization, in that those methods reduce rather than eliminate all forms of life and biological agents present. After sterilization, an object is referred to as being sterile or aseptic.
Applications
Foods
One of the first steps toward modernized sterilization was made by Nicolas Appert, who discovered that application of heat over a suitable period slowed the decay of foods and various liquids, preserving them for safe consumption for a longer time than was typical. Canning of foods is an extension of the same principle and has helped to reduce food borne illness ("food poisoning"). Other methods of sterilizing foods include food irradiation and high pressure (pascalization).
Medicine and surgery
In general, surgical instruments and medications that enter an already aseptic part of the body (such as the bloodstream, or penetrating the skin) must be sterile. Examples of such instruments include scalpels, hypodermic needles, and artificial pacemakers. This is also essential in the manufacture of parenteral pharmaceuticals.
Preparation of injectable medications and intravenous solutions for fluid replacement therapy requires not only sterility but also well-designed containers to prevent entry of adventitious agents after initial product sterilization.
Most medical and surgical devices used in healthcare facilities are made of materials that are able to go under steam sterilization. However, since 1950, there has been an increase in medical devices and instruments made of m |
https://en.wikipedia.org/wiki/Sloan%20letters | Sloan letters, designed by Louise Sloan in 1959, are a set of optotypes used to test visual acuity generally used in Snellen charts and logMAR charts.
This set of optotypes consists of ten specially formed "letters", C, D, H, K, N, O, R, S, V and Z. These letters, unlike the ones used in older Snellen charts, are designed to give acuity testing results that are comparable to tests made using Landolt rings.
Computer fonts for macOS and Microsoft Windows operating systems are available for research purposes. The fonts are based on Louise Sloan's design, which has been designated the US standard for acuity testing by the National Academy of Sciences, National Research Council, Committee on Vision (1980, Adv Ophthalmol, 41, 103–148).
See also
Eye chart
Landolt C
Lea test
Snellen chart |
https://en.wikipedia.org/wiki/CD%20V-700 | The CD V-700 (often written as "CDV-700") is a Geiger counter employing a probe equipped with a Geiger–Müller tube, manufactured by several companies under contract to United States federal civil defense agencies in the 1950s and 1960s. While all models adhere to a similar size, shape, coloring and form-factor, there were substantial differences between various models and manufacturers over the years the CD V-700 was in production. Many of the earlier units required the use of now-obsolete high-voltage batteries, and were declared obsolete by the end of the 1970s.
Tens of thousands of these units were distributed to US state civil defense agencies. Even though large numbers have been sold off as surplus to civilian users, many remain in use with first responders and state emergency management agencies today.
Characteristics
Case and contents
Most models of the CD V-700 are constructed using a two-piece case made of die-cast and stamped aluminum with a distinctive yellow paint (John Deere Yellow), a Civil Defense “CD” decal and check source. The upper, die-cast part of the case contains a groove around the outer edge for a rubber gasket that renders the case water-tight. These have often deteriorated over the years but can be replaced with a section of rubber bead as used for mounting household screen windows in their frames. This section of the case also houses the meter, the printed circuit board and the battery holders. Mounted to the top of the upper case is a carry handle in which the probe clips in for storage, a connector for a headphone and the control knob.
The inside of the unit contains high voltage electronics of up to 900 volts, so care is recommended when operating with the case open. Inside the lower, stamped aluminum portion of the case is a printed diagram corresponding to the make and model of the unit that the lower case shipped with. In the decades since these units were made, it is common to find case bottoms have been switched between dif |
https://en.wikipedia.org/wiki/Chaperone%20code | The chaperone code refers to post-translational modifications of molecular chaperones that control protein folding. Whilst the genetic code specifies how DNA makes proteins, and the histone code regulates histone-DNA interactions, the chaperone code controls how proteins are folded to produce a functional proteome.
The chaperone code refers to the combinatorial array of post-translational modifications (enzymes add chemical modifications to amino acids that change their properties) —i.e. phosphorylation, acetylation, ubiquitination, methylation, etc.—that are added to molecular chaperones to modulate their activity. Molecular chaperones are proteins specialized in folding and unfolding of the other cellular proteins, and the assembly and dismantling of protein complexes. This is critical in the regulation of protein-protein interactions and many cellular functions. Because post-translational modifications are marks that can be added and removed rapidly, they provide an efficient mechanism to explain the plasticity observed in proteome organization during cell growth and development.
The chaperone code concept posits that combinations of post-translational modifications at the surface of chaperones, including phosphorylation, acetylation, methylation, ubiquitination, control protein folding/unfolding and protein complex assembly/disassembly by modulating:
1) chaperone-substrate affinity and specificity
2) chaperone ATPase and therefore its refolding activity
3) chaperone localization
4) chaperone-co-chaperone interaction.
Levels of the Chaperone Code
The Chaperone code is incredibly complex with multiple layers of potential regulation. Studies of the chaperone code may include:
Level 1: Understanding the role and regulation of single PTMs on a single chaperone
Level 2: Cross-talk of different PTMs on a single amino acid or between PTMs on different amino acids (on a single chaperone)
Level 3: Understanding of why chaperone paralogs have different PTMs
|
https://en.wikipedia.org/wiki/Stochastic%20equicontinuity | In estimation theory in statistics, stochastic equicontinuity is a property of estimators (estimation procedures) that is useful in dealing with their asymptotic behaviour as the amount of data increases. It is a version of equicontinuity used in the context of functions of random variables: that is, random functions. The property relates to the rate of convergence of sequences of random variables and requires that this rate is essentially the same within a region of the parameter space being considered.
For instance, stochastic equicontinuity, along with other conditions, can be used to show uniform weak convergence, which can be used to prove the convergence of extremum estimators.
Definition
Let be a family of random functions defined from , where is any normed metric space. Here might represent a sequence of estimators applied to datasets of size n, given that the data arises from a population for which the parameter indexing the statistical model for the data is θ. The randomness of the functions arises from the data generating process under which a set of observed data is considered to be a realisation of a probabilistic or statistical model. However, in , θ relates to the model currently being postulated or fitted rather than to an underlying model which is supposed to represent the mechanism generating the data. Then is stochastically equicontinuous if, for every and , there is a such that:
Here B(θ, δ) represents a ball in the parameter space, centred at θ and whose radius depends on δ. |
https://en.wikipedia.org/wiki/Autler%E2%80%93Townes%20effect | In spectroscopy, the Autler–Townes effect (also known as AC Stark effect), is a dynamical Stark effect corresponding to the case when an oscillating electric field (e.g., that of a laser) is tuned in resonance (or close) to the transition frequency of a given spectral line, and resulting in a change of the shape of the absorption/emission spectra of that spectral line. The AC Stark effect was discovered in 1955 by American physicists Stanley Autler and Charles Townes.
It is the AC equivalent of the static Stark effect which splits the spectral lines of atoms and molecules in a constant electric field. Compared to its DC counterpart, the AC Stark effect is computationally more complex.
While generally referring to atomic spectral shifts due to AC fields at any (single) frequency, the effect is more pronounced when the field frequency is close to that of a natural atomic or molecular dipole transition. In this case, the alternating field has the effect of splitting the two bare transition states into doublets or "dressed states" that are separated by the Rabi frequency. Alternatively, this can be described as a Rabi oscillation between the bare states which are no longer eigenstates of the atom–field Hamiltonian. The resulting fluorescence spectrum is known as a Mollow triplet.
The AC Stark splitting is integral to several phenomena in quantum optics, such as electromagnetically induced transparency and Sisyphus cooling. Vacuum Rabi oscillations have also been described as a manifestation of the AC Stark effect from atomic coupling to the vacuum field.
History
The AC Stark effect was discovered in 1955 by American physicists Stanley Autler and Charles Townes while at Columbia University and Lincoln Labs at the Massachusetts Institute of Technology. Before the availability of lasers, the AC Stark effect was observed with radio frequency sources. Autler and Townes' original observation of the effect used a radio frequency source tuned to 12.78 and 38.28 MHz, corr |
https://en.wikipedia.org/wiki/CENBOL | In astronomy, CENBOL (derived from "CENtrifugal pressure supported BOundary Layer) is a model developed by the astrophysicist Sandip Chakrabarti and collaborators to explain the region of an accretion flow around a black hole.
Centrifugal force dominated boundary layer
Because centrifugal force l2/r3 increases very rapidly compared to the gravitational force (which goes as 1/r2) as the distance r decreases, matter feels increasing centrifugal force as it approaches a black hole. Thus the matter initially slows down, typically through a shock transition, and then accelerates again to become a supersonic flow.
The importance of CENBOL is that it behaves like a boundary layer of a black hole. This region is located between the shock and the innermost sonic point of an accretion flow. CENBOL becomes hot due to sudden reduction of radial kinetic energy and the flow is puffed up, since hot gas can fight against gravitational. In a certain sense it behaves like a thick accretion disk (ion pressure supported torus if the accretion rate is low; or radiation pressure supported torus, if the accretion rate is high), except that it also has radial velocity, which original models of thick disk did not have. Because it is hot, the electrons transfer their thermal energy to the photons. In other words CENBOL inverse Comptonizes low energy X-rays or soft photons (also called seed photons) and produces very high energy X-rays (also called hard photons). Just as a boundary layer, it also produces jets and outflows.
The observed spectrum of a black hole accretion disk is in reality partly coming from a Keplerian disk (viscous Shakura and Sunyaev (1973) type disk) in the form of multi-colour black body emission. But the power-law component primarily comes from the CENBOL. In presence of high accretion rates in the Keplerian component, the CENBOL can be cooled down so that the spectrum may be totally dominated by the low energy X-rays. The spectrum goes to a 'so-called' soft sta |
https://en.wikipedia.org/wiki/SUPRENUM | SUPRENUM (, ) was a German research project to develop a parallel computer from 1985 through 1990. It was a major effort which was aimed at developing a national expertise in massively parallel processing both at hardware and at software level.
Although the Suprenum-1 computer was the fastest massively parallel MIMD computer in the world during a period in 1992, the project was set and is considered a commercial failure.
History
Funded by the Federal Ministry for Research and Technology (BMFT), the SUPRENUM project began in 1985 and BMFT funding continued until 1990 when a fully configured 256-node prototype Suprenum-1 machine was available.
The project's inception in 1985 was preceded by a definition phase lasting more than one year in which ideas were gathered, concepts were formed and project partners were selected.
The project was two-tiered, of which only the first step was taken. In particular it was planned the following:
Suprenum 1 subproject: production of a high-speed MIMD computer
Suprenum 2 subproject: expanding the core applications and algorithmic service classes to include complex and dynamic grid structures; data-dependent adaptive procedures, irregular and highly dimensional grids, Monte Carlo methods based on grid structures, non-grid applications, etc., development of innovative language concepts which support automatic load distribution (particularly with dynamic grid structures) to the multiple-processor structure, investigation of alternative interconnecting structures (other topologies, variable interconnection networks) in particular with regard to dynamic grid structures and automatic load distribution strategies, new processor technologies (VLSI, GaAs and so on).
The mandate accompanying the funding was to create a project that included both a research and a commercial side. To this end, the SUPRENUM Supercomputer GmbH was founded in Bonn. The SUPRENUM Supercomputer GmbH's charge was to manage the whole enterprise, to contribute to th |
https://en.wikipedia.org/wiki/Barbara%20Burke%20Hubbard | Barbara Burke Hubbard (born 1948) is an American science journalist, mathematics popularizer, textbook author, and book publisher, known for her books on wavelet transforms and multivariable calculus.
Life
Burke Hubbard is the daughter of Los Angeles Times reporter Vincent J. Burke, and spent a year in high school living in Moscow when Burke was stationed there in 1964. She was an undergraduate at Harvard University, initially majoring in biology but switching to English, and graduating in 1969. She became a science writer for the Massachusetts Institute of Technology and a journalist for The Ithaca Journal, and was the 1981 winner of the AAAS Westinghouse Science Journalism Award in the small newspaper category, for her articles on acid rain in The Ithaca Journal.
She married mathematician John H. Hubbard, with whom she has four children, and with her family has split her time between Ithaca, New York and Marseille, France, with shorter-term stays elsewhere.
Books
Burke Hubbard is the author of a popular mathematics book on wavelet transforms, originally published in French as Ondes et ondelettes: la saga d’un outil mathématique (Pour la Science, 1995). It won the of the Société mathématique de France, and Hubbard became the first winner of this prize who was not French. The English edition of the same book, The world according to wavelets: the story of a mathematical technique in the making, was published in 1996 by A K Peters, with a second edition in 1998. It was also translated into German by M. Basler as Wavelets: Die Mathematik der kleinen Wellen (Birkhäuser, 1997). With her husband, she wrote a textbook on multivariate calculus, Vector calculus, linear algebra, and differential forms: A unified approach (Prentice Hall, 1999; 5th ed., 2015). She has also translated the book Biochronological correlations by Jean Guex from French into English.
In 2001, Burke Hubbard founded the mathematics book publisher Matrix Editions. |
https://en.wikipedia.org/wiki/Front%20controller | The front controller software design pattern is listed in several pattern catalogs and is related to the design of web applications. It is "a controller that handles all requests for a website," which is a useful structure for web application developers to achieve flexibility and reuse without code redundancy.
Instruction
Front controllers are often used in web applications to implement workflows. While not strictly required, it is much easier to control navigation across a set of related pages (for instance, multiple pages used in an online purchase) from a front controller than it is to assign individual pages responsibility for navigation.
The front controller may be implemented as a Java object, or as a script in a scripting language such as PHP, Raku, Python or Ruby that is called for every request of a web session. This script would handle all tasks that are common to the application or the framework, such as session handling, caching and input filtering. Based on the specific request, it would then instantiate further objects and call methods to handle the required tasks.
The alternative to a front controller is the usage of page controllers mapped to each site page or path. Although this may cause each individual controller to contain duplicate code, the page-controller approach delivers a high degree of specialization.
Examples
Several web-tier application frameworks implement the front controller pattern:
Apache Struts
ASP.NET MVC
Cairngorm framework in Adobe Flex
Cro or Bailador frameworks in Raku
Drupal
MVC frameworks written in PHP, such as Yii, CakePHP, Laravel, Symfony, CodeIgniter and Laminas
Spring Framework
Yesod, written in Haskell
Implementation
Front controllers may divided into three components:
XML mapping: files that map requests to the class that will handle the request processing.
Request processor: used for request processing and modifying or retrieving the appropriate model.
Flow manager: determines what will be shown on |
https://en.wikipedia.org/wiki/121st%20meridian%20west | The meridian 121° west of Greenwich is a line of longitude that extends from the North Pole across the Arctic Ocean, North America, the Pacific Ocean, the Southern Ocean, and Antarctica to the South Pole.
The 121st meridian west forms a great circle with the 59th meridian east.
From Pole to Pole
Starting at the North Pole and heading south to the South Pole, the 121st meridian west passes through:
{| class="wikitable plainrowheaders"
! scope="col" width="130" | Co-ordinates
! scope="col" | Country, territory or sea
! scope="col" | Notes
|-
| style="background:#b0e0e6;" |
! scope="row" style="background:#b0e0e6;" | Arctic Ocean
| style="background:#b0e0e6;" |
|-
|
! scope="row" |
| Northwest Territories — Prince Patrick Island
|-
| style="background:#b0e0e6;" |
! scope="row" style="background:#b0e0e6;" | M'Clure Strait
| style="background:#b0e0e6;" |
|-
|
! scope="row" |
| Northwest Territories — Banks Island
|-
| style="background:#b0e0e6;" |
! scope="row" style="background:#b0e0e6;" | Amundsen Gulf
| style="background:#b0e0e6;" |
|-valign="top"
|
! scope="row" |
| Northwest Territories — passing through the Great Bear Lake British Columbia — from , passing through Tumbler Ridge
|-valign="top"
|
! scope="row" |
| Washington — passing west of Cle Elum Oregon — from California — from , passing through Modesto (at )
|-
| style="background:#b0e0e6;" |
! scope="row" style="background:#b0e0e6;" | Pacific Ocean
| style="background:#b0e0e6;" |
|-
| style="background:#b0e0e6;" |
! scope="row" style="background:#b0e0e6;" | Southern Ocean
| style="background:#b0e0e6;" |
|-
|
! scope="row" | Antarctica
| Unclaimed territory
|-
|}
See also
120th meridian west
122nd meridian west
w121 meridian west |
https://en.wikipedia.org/wiki/HP%20Compaq%20tc1100 | The HP Compaq TC1100 is a tablet PC sold by Hewlett-Packard that was the follow-up to the Compaq TC1000. The TC1100 had either an Intel Celeron or an Intel Pentium M chip set and could be upgraded up to 2 gigabytes of memory. The switch from Transmeta Crusoe processors to the Pentium M and the ability to add memory was welcomed after numerous complaints about the poor performance of the TC1000. The TC1100 was the last version from HP in this style of tablet. It was replaced by the HP Compaq TC4200, which featured a more traditional one-piece design.
Design
The TC1100 has a 10.4 inch LCD display and pressure-sensitive pen that shares the same basic design as its predecessor, the TC1000. It has a design often referred to as a hybrid tablet, as it has the properties of both a convertible and slate tablet. All the necessary hardware components are stored within the casing of the display and digitizer. This allows it to work with or without a keyboard attached. With the keyboard attached, it can be used as a laptop, with the display rising by an adjustable hinge behind the keyboard. By rotating the display, the keyboard can fold inside the unit; or the keyboard can be removed entirely. Either method lets the user write on the screen easily, using a virtual keyboard or handwriting-recognition application for occasional character entry. This versatility gave the product a loyal following when there were no similar designs on the market.
HP also made an optional docking station, which connected to the port on the bottom of the tablet, which allowed the user to easily connect their tablet to charge, while also offering a VGA output, an Ethernet jack, audio input and output jacks, 4 USB 2.0 ports and a laptop-style CD/DVD drive. All of the ports, excluding the USB ports, are shared with the ones on the tablet, the audio ports having priority over the ones on the right side of the tablet.
Hardware
The range of processors includes Pentium-M 1.0 GHz, 1.1 GHz, and 1.2 GHz. |
https://en.wikipedia.org/wiki/Boletus%20reticulatus | Boletus reticulatus (alternately known as Boletus aestivalis (Paulet) Fr.), and commonly referred to as the summer cep is a basidiomycete fungus of the genus Boletus. It occurs in deciduous forests of Europe, where it forms a symbiotic mycorrhizal relationship with species of oak (Quercus). The fungus produces fruiting bodies in the summer months which are edible and popularly collected. The summer cep was formally described by Jacob Christian Schäffer as Boletus reticulatus in 1774, which took precedence over B. aestivalis as described by Jean-Jacques Paulet in 1793.
Taxonomy
German naturalist Jacob Christian Schäffer described the summer cep as Boletus reticulatus in 1774, in his series on fungi of Bavaria and the Palatinate, Fungorum qui in Bavaria et Palatinatu circa Ratisbonam nascuntur icones. French mycologist Jean-Jacques Paulet described it as Le grand Mousseux (Tubiporus aestivalis) in 1793, adding that it was delicious with chicken fricassee and could be found in the Bois de Boulogne in summer. the species name the species name is derived from the Latin aestas "summer". Swedish mycologist Elias Magnus Fries followed Paulet, using Boletus aestivalis in 1838.
The two names have been used in literature for many years.
Boletus reticulatus is classified in Boletus section Boletus, alongside close relatives such as B. aereus, B. edulis, and B. pinophilus. A genetic study of the four European species found that B. reticulatus was sister to B. aereus. More extensive testing of worldwide taxa revealed that B. reticulatus was most closely related to two lineages that had been classified as B. edulis from southern China and Korea/northern China respectively. The common ancestor of these three species was related to a lineage consisting of B. aereus and the genetically close B. mamorensis. Molecular analysis suggests that the B. aereus/mamorensis and B. reticulatus/Chinese B. "edulis" lineages diverged around 6 to 7 million years ago.
The British Mycological So |
https://en.wikipedia.org/wiki/Kac%E2%80%93Bernstein%20theorem | The Kac–Bernstein theorem is one of the first characterization theorems of mathematical statistics. It is easy to see that if the random variables and are independent and normally distributed with the same variance, then their sum and difference are also independent. The Kac–Bernstein theorem states that the independence of the sum and difference of two independent random variables characterizes the normal distribution (the Gauss distribution). This theorem was proved independently by Polish-American mathematician Mark Kac and Soviet mathematician Sergei Bernstein.
Formulation
Let and are independent random variables. If and are independent then and have normal distributions (the Gaussian distribution).
Generalization
A generalization of the Kac–Bernstein theorem is the Darmois–Skitovich theorem, in which instead of sum and difference linear forms from n independent random variables are considered. |
https://en.wikipedia.org/wiki/Hurwitz%20polynomial | In mathematics, a Hurwitz polynomial, named after Adolf Hurwitz, is a polynomial whose roots (zeros) are located in the left half-plane of the complex plane or on the imaginary axis, that is, the real part of every root is zero or negative. Such a polynomial must have coefficients that are positive real numbers. The term is sometimes restricted to polynomials whose roots have real parts that are strictly negative, excluding the imaginary axis (i.e., a Hurwitz stable polynomial).
A polynomial function P(s) of a complex variable s is said to be Hurwitz if the following conditions are satisfied:
1. P(s) is real when s is real.
2. The roots of P(s) have real parts which are zero or negative.
Hurwitz polynomials are important in control systems theory, because they represent the characteristic equations of stable linear systems. Whether a polynomial is Hurwitz can be determined by solving the equation to find the roots, or from the coefficients without solving the equation by the Routh–Hurwitz stability criterion.
Examples
A simple example of a Hurwitz polynomial is:
The only real solution is −1, because it factors as
In general, all quadratic polynomials with positive coefficients are Hurwitz.
This follows directly from the quadratic formula:
where, if the discriminant b2−4ac is less than zero, then the polynomial will have two complex-conjugate solutions with real part −b/2a, which is negative for positive a and b.
If the discriminant is equal to zero, there will be two coinciding real solutions at −b/2a. Finally, if the discriminant is greater than zero, there will be two real negative solutions,
because for positive a, b and c.
Properties
For a polynomial to be Hurwitz, it is necessary but not sufficient that all of its coefficients be positive (except for quadratic polynomials, which also imply sufficiency). A necessary and sufficient condition that a polynomial is Hurwitz is that it passes the Routh–Hurwitz stability criterion. A given polynomia |
https://en.wikipedia.org/wiki/Juku%20E5101 | Juku E5101 was a personal computer targeted at Estonian schools which was released in 1988. The computer had monochrome display, a mouse and basic LAN capabilities, it ran CP/M 2.2 based EKDOS and had a Soviet Intel 8080A clone KR580VM80A for CPU.
Juku E5101 was developed by and the Institute of Cybernetics of the Academy of Sciences of Estonia, test batch of 100 was produced in cooperation with factory in 1986. The computer initially used tape recorder as storage and was reported as first computer in USSR to have mouse attached. In a multibus (Soviet I41) compatible expansion slot one could also connect 32 KiB memory expansion cards or ROM cartridges.
Juku E5104 production of which started in December 1988 was upgraded to use dual 5.25 inch diskette drive and drivers for printers. Despite relabelling it to "intellectual terminal for real-time system E5104", the label presented on main unit remained E5101.
During first two years of serial production around 2000 Jukus were produced and last batch of 500 was ordered by Estonian Ministry of Education in 1992. Altogether 3000 Jukus were produced at Narva, plant (from Russian "Балтиец", Baltiyets), 2500 of them for school use.
In 1991 many, if not all, bigger (at least 100 pupils) Estonian schools had a computer classroom that was furnished with those machines and Epson LX800 printers.
Although the production was delayed four years and computers delivered were technologically outdated, Jukus did enable Estonia to "gain a head start in mass school computerization" by providing early access to computers and a standardized study environment. Despite conceived lack of end user skills and the shortage of computer professionals, there were schools having dedicated teachers and students themselves writing software for Jukus during extra hours at computer class, often convincing schools to lend computers home for summer vacation.
In general tens of thousands students got their first computing experience with Juku "much |
https://en.wikipedia.org/wiki/SM-1420 | The SM-1420 (CM-1420) is a 16 bit DEC PDP-11/45 minicomputer clone, and the successor to SM-4 in Soviet Bloc countries. Under the direction of Minpribor it was produced in the Soviet Union and Bulgaria from 1983 onwards, and is more than twice as fast as its predecessor. Its closest western counterpart is the DEC PDP-11/45, which means that the Soviet technology trailed 11 years behind compared to the Digital Equipment Corporation equivalent machine.
The standard package includes 256 KiB MOS memory, two RK-06 disks, two TU-10 decks, CM-6315 barrel or DZM-180 dot-matrix printer from Mera Blonie (Poland), VT52 compatible or VTA-2000-15 (BTA 2000-15) VT100 compatible terminals from Mera Elzab.
See also
History of computing in the Soviet Union
List of Soviet computer systems
SM EVM |
https://en.wikipedia.org/wiki/Rutilated%20quartz | Rutilated quartz is a variety of quartz which contains acicular (needle-like) inclusions of rutile. It is used for gemstones. These inclusions mostly look golden, but they also can look silver, copper red or deep black. They can be distributed randomly or in bundles, which sometimes are arranged star-like, and they can be sparse or dense enough to make the quartz body nearly opaque. While otherwise inclusions often reduce the value of a crystal, rutilated quartz is valued for the quality and beauty of these inclusions. |
https://en.wikipedia.org/wiki/In-place%20algorithm | In computer science, an in-place algorithm is an algorithm that operates directly on the input data structure without requiring extra space proportional to the input size. In other words, it modifies the input in place, without creating a separate copy of the data structure. An algorithm which is not in-place is sometimes called not-in-place or out-of-place.
In-place can have slightly different meanings. In its strictest form, the algorithm can only have a constant amount of extra space, counting everything including function calls and pointers. However, this form is very limited as simply having an index to a length array requires bits. More broadly, in-place means that the algorithm does not use extra space for manipulating the input but may require a small though nonconstant extra space for its operation. Usually, this space is , though sometimes anything in is allowed. Note that space complexity also has varied choices in whether or not to count the index lengths as part of the space used. Often, the space complexity is given in terms of the number of indices or pointers needed, ignoring their length. In this article, we refer to total space complexity (DSPACE), counting pointer lengths. Therefore, the space requirements here have an extra factor compared to an analysis that ignores the length of indices and pointers.
An algorithm may or may not count the output as part of its space usage. Since in-place algorithms usually overwrite their input with output, no additional space is needed. When writing the output to write-only memory or a stream, it may be more appropriate to only consider the working space of the algorithm. In theoretical applications such as log-space reductions, it is more typical to always ignore output space (in these cases it is more essential that the output is write-only).
Examples
Given an array of items, suppose we want an array that holds the same elements in reversed order and to dispose of the original. One seemingly simp |
https://en.wikipedia.org/wiki/Smart%20label | Smart Label, also called Smart Tag, is an extremely flat configured transponder under a conventional print-coded label, which includes chip, antenna and bonding wires as a so-called inlay. The labels, made of paper, fabric or plastics, are prepared as a paper roll with the inlays laminated between the rolled carrier and the label media for use in specially-designed printer units.
In many processes in logistics and transportation, the barcode, or the 2D-barcode, is well established as the key means for identification in short distance. Whereas the automation of such optical coding is limited in appropriate distance for reading success and usually requires manual operation for finding the code or scanner gates that scan all the surface of a coded object, the RFID-inlay allows for better tolerance in fully automated reading from a certain specified distance. However, the mechanical vulnerability of the RFID-inlay is higher than the ordinary label, which has its weaknesses in its resistance to scratch.
Thus, the smartness of the smart label is earned in compensation of typical weaknesses with the combination of the technologies of plain text, optical character recognition and radio code.
Smart Label Processing
The processing of these labels is basically as with ordinary labels in all stages of production and application, except the inlay is inserted in an automated processing step to ensure identical positioning for each label and careful processing to prevent any damage to the bonding.
The printing is processed in two steps, including
normal ink-jet printing, except the space with the bonded chip, with clearly intelligible text and
either barcode or 2D barcode for later semi-automatic reading with handheld readers or fix-mount scanners
writing coherently concatenated information to the RFID-chip
reading the written information on the RFID-chip subsequently in the printer for control purpose (read after write)
Classification
Chip Labels
Customisation of smart l |
https://en.wikipedia.org/wiki/Dualizing%20sheaf | In algebraic geometry, the dualizing sheaf on a proper scheme X of dimension n over a field k is a coherent sheaf together with a linear functional
that induces a natural isomorphism of vector spaces
for each coherent sheaf F on X (the superscript * refers to a dual vector space). The linear functional is called a trace morphism.
A pair , if it is exists, is unique up to a natural isomorphism. In fact, in the language of category theory, is an object representing the contravariant functor from the category of coherent sheaves on X to the category of k-vector spaces.
For a normal projective variety X, the dualizing sheaf exists and it is in fact the canonical sheaf: where is a canonical divisor. More generally, the dualizing sheaf exists for any projective scheme.
There is the following variant of Serre's duality theorem: for a projective scheme X of pure dimension n and a Cohen–Macaulay sheaf F on X such that is of pure dimension n, there is a natural isomorphism
.
In particular, if X itself is a Cohen–Macaulay scheme, then the above duality holds for any locally free sheaf.
Relative dualizing sheaf
Given a proper finitely presented morphism of schemes , defines the relative dualizing sheaf or as the sheaf such that for each open subset and a quasi-coherent sheaf on , there is a canonical isomorphism
,
which is functorial in and commutes with open restrictions.
Example:
If is a local complete intersection morphism between schemes of finite type over a field, then (by definition) each point of has an open neighborhood and a factorization , a regular embedding of codimension followed by a smooth morphism of relative dimension . Then
where is the sheaf of relative Kähler differentials and is the normal bundle to .
Examples
Dualizing sheaf of a nodal curve
For a smooth curve C, its dualizing sheaf can be given by the canonical sheaf .
For a nodal curve C with a node p, we may consider the normalization with two points x, y identified. L |
https://en.wikipedia.org/wiki/Light%20effects%20on%20circadian%20rhythm | Light effects on circadian rhythm are the effects that light has on circadian rhythm.
Most animals and other organisms have "built-in clocks" in their brains that regulate the timing of biological processes and daily behavior. These "clocks" are known as circadian rhythms. They allow maintenance of these processes and behaviors relative to the 24-hour day/night cycle in nature. Although these rhythms are maintained by the individual organisms, their length does vary somewhat individually. Therefore, they must, either continually or repeatedly, be reset to synchronize with nature's cycle. In order to maintain synchronization ("entrainment") to 24 hours, external factors must play some role. The human circadian rhythm occurs typically in accordance with nature's cycle. The average activity rhythm cycle is 24.18 hours in adulthood but is shortened as age increases. One of the various factors that influence this entrainment is light exposure to the eyes. When an organism is exposed to a specific wavelength of light stimulus at certain times throughout the day, the hormone melatonin is suppressed, or prevented from being secreted by the pineal gland.
Mechanism
Light first passes into a mammal's system through the retina, then takes one of two paths: the light gets collected by rod cells and cone cells and the retinal ganglion cells (RGCs), or it is directly collected by these RGCs.
The RGCs use the photopigment melanopsin to absorb the light energy. Specifically, this class of RGCs being discussed is referred to as "intrinsically photosensitive," which just means they are sensitive to light. There are five known types of intrinsically photosensitive retinal ganglion cells (ipRGCs): M1, M2, M3, M4, and M5. Each of these differently ipRGC types have different melanopsin content and photosensitivity. These connect to amacrine cells in the inner plexiform layer of the retina. Ultimately, via this retinohypothalamic tract (RHT) the suprachiasmatic nucleus (SCN) of |
https://en.wikipedia.org/wiki/Guidances%20for%20statistics%20in%20regulatory%20affairs | Guidances for statistics in regulatory affairs refers to specific documents or guidelines that provide instructions, recommendations, and standards pertaining to the application of statistical methodologies and practices within the regulatory framework of industries such as pharmaceuticals and medical devices. These guidances serve as a reference for statisticians, researchers, and professionals involved in designing, conducting, analyzing, and reporting studies and trials in compliance with regulatory requirements. These documents embody the prevailing perspectives of regulatory agencies on specific subjects. It is worth noting that in the United States, the term "Guidances" is used, while in Europe, the term "Guidelines" is employed.
Regulatory affairs, alternatively referred to as government affairs, constitutes a profession within regulated sectors like pharmaceuticals and medical devices. Professionals, including statisticians, in these fields are expected to incorporate regulatory guidance into their work practices.
Statisticians operating in regulated environments, such as the pharmaceutical and healthcare industries, are required to possess a comprehensive understanding of the regulatory requirements influencing the design, execution, analysis, and reporting of their studies.
Regulatory guidance pertinent to the pharmaceutical and medical devices industries can be found at both international and regional/national levels. Examples of regulatory bodies include the European Medicines Agency (EMA) in Europe, the Medicines and Healthcare products Regulatory Agency (MHRA) in the United Kingdom, the Institute for Quality and Efficiency in Health Care (IQWiG) in Germany, the Food and Drug Administration (FDA) in the United States, and the Pharmaceuticals and Medical Devices Agency (PMDA) in Japan.
Additionally, statistical regulatory guidance is available for general topics such as Good Clinical Practice (ICH E6(R2)), as well as specific areas explicitly related |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.