source stringlengths 31 203 | text stringlengths 28 2k |
|---|---|
https://en.wikipedia.org/wiki/Institute%20of%20Himalayan%20Bioresource%20Technology | C.S.I.R - Institute of Himalayan Bioresource Technology or CSIR-IHBT established in 1983 is a constituent laboratory of Council of Scientific and Industrial Research. This institute located in Palampur, Kangra, Himachal Pradesh, India is engaged in various advanced research aspects of Himalayan Bio-resources and modern biology. It has also been imparting Ph.D. in Biological and Chemical Sciences.
Location and facilities
Situated among the Dhauladhar ranges, CSIR-IHBT is the only laboratory of the Council of Scientific and Industrial Research in the State of Himachal Pradesh (H.P.), India. Institute has a focused research mandate on bioresources for catalysizing bioeconomy in a sustainable manner.
The institute has state-of the art laboratories; remote sensing and mapping facilities; internationally recognised herbarium; animal house facility; pilot plants in nutraceuticals, essential oil and herbals; farms and polyhouses.
The scientists propel the research and work to discover and find solutions to new challenging problems faced by society. International collaborations further strengthens scientific interactions at a global scale.
Research
Promoting industrial growth through technological interventions is a constant endeavour and several technologies developed by the institute are transferred to industries. For socio- economic upliftment, regular training programmes and advisory services are rendered to farmers, floriculturists, tea planters and small entrepreneurs involved in food processing sector. The Institute has been recognised as one of the Incubation Centres by MSME GoI and in the area of Affordable Health Care by DSIR. The Institute encourages industries to share the technological problems faced them, such that efforts could be made in developing a viable solution.
Confidentiality is strictly maintained. Work on plant adaptation studies and high altitude medicinal plants are further strengthened by the field lab ”Centre for High Altitude Biology |
https://en.wikipedia.org/wiki/Lauren%20Williams%20%28mathematician%29 | Lauren Kiyomi Williams (born 1978) is an American mathematician known for her work on cluster algebras, tropical geometry, algebraic combinatorics, amplituhedra, and the positive Grassmannian. She is Dwight Parker Robinson Professor of Mathematics at Harvard University.
Education
Williams's father is an engineer; her mother is third-generation Japanese American. She grew up in Los Angeles, where her interest in mathematics was sparked by winning a fourth-grade mathematics contest. She was the valedictorian of Palos Verdes Peninsula High School in 1996, and while there participated in summer research at the Massachusetts Institute of Technology with Satomi Okazaki, a student of her eventual advisor, Richard P. Stanley. She graduated magna cum laude from Harvard University in 2000 with a A.B. in mathematics, and received her PhD in 2005 at the Massachusetts Institute of Technology under the supervision of Stanley. Her dissertation was titled Combinatorial Aspects of Total Positivity.
Work
After postdoctoral positions at the University of California, Berkeley and Harvard, Williams rejoined the Berkeley mathematics department as an assistant professor in 2009, and was promoted to associate professor in 2013 and then full professor in 2016.
Starting in the fall of 2018, she rejoined the Harvard mathematics department as a full professor, making her the second ever tenured female math professor at Harvard. The first, Sophie Morel, left Harvard in 2012.
Along with colleagues O. Mandelshtam (her former student, now an assistant professor at University of Waterloo) and S. Corteel, in 2018 Williams developed a new characterization of both symmetric and nonsymmetric Macdonald polynomials using the combinatorial exclusion process.
Awards
In 2012, she became one of the inaugural fellows of the American Mathematical Society. She is the 2016 winner of the Association for Women in Mathematics and Microsoft Research Prize in Algebra and Number Theory. In 2022 she was awarded a |
https://en.wikipedia.org/wiki/Snapback%20%28electrical%29 | Snapback is a mechanism in a bipolar transistor in which avalanche breakdown or impact ionization provides a sufficient base current to turn on the transistor. It is used intentionally in the design of certain ESD protection devices integrated onto semiconductor chips. It can also be a parasitic failure mechanism when activated inadvertently, outwardly appearing much like latchup in that the chip seems to suddenly blow up when a high voltage is applied.
Snapback is initiated by a small current from collector to base. In the case of ESD protection devices, this current is caused by avalanche breakdown due to a sufficiently large voltage applied across the collector-base junction. In the case of parasitic failures, the initiating current may result from inadvertently turning on the bipolar transistor and a sufficiently large voltage across the collector and base causing impact ionization, with some of the generated carriers then acting as the initiating current as they flow into the base. Once this initiating current flows into the base, the transistor turns on and the collector voltage decreases to the snapback holding voltage. This voltage happens at the point where the processes of base current generation and the bipolar transistor turning on are in balance: the collector-emitter current of the bipolar transistor decreases the collector voltage, which results in a lower electric field, which results in a smaller impact ionization or avalanche current and thus smaller base current, which weakens the bipolar action.
See also
Avalanche breakdown
Bipolar transistor
Impact ionization
References
Semiconductors
Electrical breakdown |
https://en.wikipedia.org/wiki/Distinguished%20limit | In mathematics, a distinguished limit is an appropriately chosen scale factor used in the method of matched asymptotic expansions.
External links
Singular perturbation theory, Scholarpedia
Differential equations
Asymptotic analysis |
https://en.wikipedia.org/wiki/Survey%20camp | Survey Camp is an army tradition that was discontinued in the later part of twentieth century but was reinstated in 2002 across the universities of the world with a whole new structure. It is the civil engineering training course for two weeks usually after completion four semesters of bachelor of technology that consists of 8 days working in the field and 6 days of map preparation in the computer lab. Experts say that survey camp provides necessary foundation for civil engineers. Each day in the course there are at least 8 hours of working in the field. Students are divided into groups and they get out with practising surveyors and use their equipment out in the field. A camp incharge teacher appoints group leaders for each group; the leaders are responsible for all the works of his particular group and the equipment. In the computer lab, students learn applications such as AutoCAD and Carlson Survey. The students use these programs to take data collected from the field to develop topographic maps of the particular area. The basic aim of the survey camp is to know various works carried out in the industrial field by surveying, which includes determining the topography of particular area with the help of survey work, map study and reconnaissance work. The methods used for surveying are traversing, levelling and contouring
Instruments
The instruments used include:
Theodolite (transit)
Total Station and Prism
Compass (prismatic and surveyor)
Ranging rods
Measuring Tape
Levelling Staff
Tripod stand (for optical instruments)
Level (instrument)
Practicals
The survey practicals generally performed in Survey camp are listed below:
Simple levelling
Fly levelling
Trigonometric levelling
To determine and draw RL's of longitudinal and cross section of road.
Measuring horizontal angles by Ordinary, Repetition and Reiteration methods
Compass traversing
Gale traverse
Topographic map preparation
Procedure
The working rule used in survey camp is quite simple, all the gr |
https://en.wikipedia.org/wiki/CryptoNote | CryptoNote is an application layer protocol designed for use with cryptocurrencies that aims to solve specific problems identified in Bitcoin. Namely:
Traceability of transactions
The proof-of-work function (see Bitcoin network)
Irregular emission
Hardcoded constants
Bulky scripts
Financial privacy
The protocol powers several decentralized privacy-oriented cryptocurrencies, including Monero, MobileCoin and Safex Cash.
Nothing is known about the original author of CryptoNote, "Nicolas van Saberhagen." Its mathematical component and motivation are described in the article "CryptoNote Whitepaper", released in two editions: in 2012 and in 2013. Launched in the summer of 2012, Bytecoin was the first cryptocurrency to use this technology. Later, several teams launched their networks, based on the Bytecoin code.
Emission
Just like in Bitcoin, miners are rewarded for finding solutions. But the stepped release curve characteristic of Bitcoin has been replaced with a smooth one in CryptoNote: the reward decreases with each block.
One implementation of the CryptoNote protocol has resulted in a non-smooth emission curve, specifically, the S-curve of the Safex Blockchain, which was designed to match the Diffusion of Innovations technology adoption curve theory.
See also
Monero (cryptocurrency)
References
Application layer protocols
Cryptocurrencies
Digital currencies |
https://en.wikipedia.org/wiki/Given-When-Then | Given-When-Then (GWT) is a semi-structured way to write down test cases. They can either be tested manually or automated as browser tests with tools like Selenium and Cucumber.
It derives its name from the three clauses used, which start with the words given, when and then. Given describes the preconditions and initial state before the start of a test and allows for any pre-test setup that may occur. When describes actions taken by a user during a test. Then describes the outcome resulting from actions taken in the when clause.
The Given-When-Then was proposed by Dan North in 2006, as part of behavior-driven development.
See also
Acceptance test-driven development
Acceptance testing
Behavior-driven development
Cucumber syntax
Hoare triple
References
Software testing |
https://en.wikipedia.org/wiki/National%20Automotive%20Innovation%20Centre | The National Automotive Innovation Centre (NAIC), sometimes referred to as the National Automotive Innovation Campus, is a building at the University of Warwick in the United Kingdom. It is a joint venture between the University's WMG, Jaguar Land Rover and Tata Motors. The building was opened by HRH The Prince of Wales on 18 February 2020.
History
The NAIC will be a focus for automotive research, combining expertise from industry, university academics and supply chain companies. It is intended to support advances in technology to reduce dependency on fossil fuels and vehicle emissions. The 150 million project is funded by a range of partners: £50 million of the project funding is being provided by the University of Warwick's development partners, Jaguar Land Rover, with additional funding from WMG and Tata Motors European Technical Centre. The remaining £15 million is from the government's Higher Education Funding Council for England.
The four-storey building, designed by Cullinan Studio, will have a total floorspace of 33,964 square metres. It is estimated that around 1,060 people will be employed at the site. Facilities in the building are expected to include research and development facilities, an engineering hall, a virtual reality lab, design workshops, teaching facilities, offices, meeting rooms, library, outdoor car viewing garden, a publicly accessible showroom and café.
Coventry City Council approved the plans on 26 June 2014. Construction is expected to start in 2014 with the building set to be completed by 2016.
The building is named The Prof. Lord Bhattacharyya Building after Lord Bhattacharyya. It is on the main road of the Warwick campus, which was renamed Lord Bhattacharyya Way in 2018.
References
External links
Automotive industry in the United Kingdom
Engineering education in the United Kingdom
Engineering research institutes
Research institutes in the West Midlands (county)
Innovation in the United Kingdom
University of Warwick |
https://en.wikipedia.org/wiki/Two-variable%20logic | In mathematical logic and computer science, two-variable logic is the fragment of first-order logic where formulae can be written using only two different variables. This fragment is usually studied without function symbols.
Decidability
Some important problems about two-variable logic, such as satisfiability and finite satisfiability, are decidable. This result generalizes results about the decidability of fragments of two-variable logic, such as certain description logics; however, some fragments of two-variable logic enjoy a much lower computational complexity for their satisfiability problems.
By contrast, for the three-variable fragment of first-order logic without function symbols, satisfiability is undecidable.
Counting quantifiers
The two-variable fragment of first-order logic with no function symbols is known to be decidable even with the addition of counting quantifiers, and thus of uniqueness quantification. This is a more powerful result, as counting quantifiers for high numerical values are not expressible in that logic.
Counting quantifiers actually improve the expressiveness of finite-variable logics as they allow to say that there is a node with neighbors, namely . Without counting quantifiers variables are needed for the same formula.
Connection to the Weisfeiler-Leman algorithm
There is a strong connection between two-variable logic and the Weisfeiler-Leman (or color refinement) algorithm. Given two graphs, then any two nodes have the same stable color in color refinement if and only if they have the same type, that is, they satisfy the same formulas in two-variable logic with counting.
References
Model theory
Systems of formal logic |
https://en.wikipedia.org/wiki/Fragment%20%28logic%29 | In mathematical logic, a fragment of a logical language or theory is a subset of this logical language obtained by imposing syntactical restrictions on the language. Hence, the well-formed formulae of the fragment are a subset of those in the original logic. However, the semantics of the formulae in the fragment and in the logic coincide, and any formula of the fragment can be expressed in the original logic.
The computational complexity of tasks such as satisfiability or model checking for the logical fragment can be no higher than the same tasks in the original logic, as there is a reduction from the first problem to the other. An important problem in computational logic is to determine fragments of well-known logics such as first-order logic that are as expressive as possible yet are decidable or more strongly have low computational complexity. The field of descriptive complexity theory aims at establishing a link between logics and computational complexity theory, by identifying logical fragments that exactly capture certain complexity classes.
References
Mathematical logic |
https://en.wikipedia.org/wiki/Finite%20Legendre%20transform | The finite Legendre transform (fLT) transforms a mathematical function defined on the finite interval into its Legendre spectrum.
Conversely, the inverse fLT (ifLT) reconstructs the original function from the components of the Legendre spectrum and the Legendre polynomials, which are orthogonal on the interval [−1,1]. Specifically, assume a function x(t) to be defined on an interval [−1,1] and discretized into N equidistant points on this interval. The fLT then yields the decomposition of x(t) into its spectral Legendre components,
where the factor (2k + 1)/N serves as normalization factor and Lx(k) gives the contribution of the k-th Legendre polynomial to x(t) such that (ifLT)
The fLT should not be confused with the Legendre transform or Legendre transformation used in thermodynamics and quantum physics.
Legendre filter
The fLT of a noisy experimental outcome s(t) and the subsequent application of the inverse fLT (ifLT) on an appropriately truncated Legendre spectrum of s(t) gives a smoothed version of s(t). The fLT and incomplete ifLT thus act as a filter. In contrast to the common Fourier low-pass filter which transmits low frequency harmonics and filters out high frequency harmonics, the Legendre lowpass transmits signal components proportional to low degree Legendre polynomials, while signal components proportional to higher degree Legendre polynomials are filtered out.
References
Further reading
Discrete transforms
Digital signal processing
Numerical analysis |
https://en.wikipedia.org/wiki/Temporary%20adjustments%20of%20theodolites | Temporary adjustments are a set of operations which are performed on a theodolite to make it ready for taking observations. These include its initial setting up on a tripod or other stand, centering, levelling up and focusing of eyepiece.
Initial setting
The initial setting operation includes fixing the theodolite on a tripod, along with approximate levelling and centering over the station mark. For setting up the instrument, the tripod is placed over the station with its legs widely spread so that the centre of the tripod head lies above the station point and its head approximately level (by eye estimation). The instrument is then fixed with the tripod by screwing through the trivet.
The height of the instrument should be such that observer can see through telescope conveniently. After this, a plumb bob is suspended from the bottom of the instrument and it should approximately align with the station mark.
Centering
Centering means bringing the vertical axis of the theodolite exactly over the station mark. Exact centering is done by using the shifting head of the instrument. During this, first the screw-clamping ring of the sliding head is loosened and the upper plate of the shifting head is slid over the lower one until the plumb bob is exactly over the station mark. After the exact centering, the screw clamping ring is tightened. This can be done by means of a forced centering plate or tribrach. An optical or laser plummet is normally used for the most accurate setting. The centering and levelling of the instrument is interactive and iterative; a re-levelling may change the centering, so error each is eliminated successively until negligible.
Levelling
Leveling of an instrument is done to make its vertical axis adjusted with respect to the apparent force of gravity at the station.
For two spirit vials at right angles:
Bring one of the level tubes parallel to any two of the foot screws, by rotating the upper part of the instrument.
The bubble is brought to |
https://en.wikipedia.org/wiki/Decidability%20of%20first-order%20theories%20of%20the%20real%20numbers | In mathematical logic, a first-order language of the real numbers is the set of all well-formed sentences of first-order logic that involve universal and existential quantifiers and logical combinations of equalities and inequalities of expressions over real variables. The corresponding first-order theory is the set of sentences that are actually true of the real numbers. There are several different such theories, with different expressive power, depending on the primitive operations that are allowed to be used in the expression. A fundamental question in the study of these theories is whether they are decidable: that is, is there an algorithm that can take a sentence as input and produce as output an answer "yes" or "no" to the question of whether the sentence is true in the theory.
The theory of real closed fields is the theory in which the primitive operations are multiplication and addition; this implies that, in this theory, the only numbers that can be defined are the real algebraic numbers. As proven by Tarski, this theory is decidable; see Tarski–Seidenberg theorem and Quantifier elimination. Current implementations of decision procedures for the theory of real closed fields are often based on quantifier elimination by cylindrical algebraic decomposition.
Tarski's exponential function problem concerns the extension of this theory to another primitive operation, the exponential function. It is an open problem whether this theory is decidable, but if Schanuel's conjecture holds then the decidability of this theory would follow. In contrast, the extension of the theory of real closed fields with the sine function is undecidable since this allows encoding of the undecidable theory of integers (see Richardson's theorem).
Still, one can handle the undecidable case with functions such as sine by using algorithms that do not necessarily terminate always. In particular, one can design algorithms that are only required to terminate for input formulas that are robus |
https://en.wikipedia.org/wiki/Permanent%20adjustments%20of%20theodolites | The permanent adjustments of theodolites are made to establish fixed relationship between the instrument's fundamental lines. The fundamental lines or axis of a transit theodolite include the following:-
Vertical axis
Axis of plate levels
Axis of telescope
Line of collimation
Horizontal axis
Axis of altitude bubble and the vernier should read zero.
These adjustments once made last for a long time. These are important for accuracy of observations taken from the instrument. The permanent adjustments in case of transit theodolite are:-
Horizontal axis adjustment.
The horizontal axis must be perpendicular to the vertical axis.
Vertical circle index adjustment.
The vertical circle must read zero when the line of collimation is horizontal.
Adjustment of altitude level.
The axis of altitude level must be parallel to the line of collimation.
Collimation adjustment.
The line of collimation or line of sight should coincide with axis of the telescope. The line of sight should also be perpendicular to the horizontal axis at its intersection with the vertical axis . Also, the optical axis, the axis of the objective slide, and the line of sight should coincide.
Adjustment of horizontal plate levels.
The axis of plate levels must be perpendicular to the vertical axis.
See also
Temporary adjustments of theodolite
Ranging rods
Tape (surveying)
References
Surveying
Civil engineering |
https://en.wikipedia.org/wiki/Glossary%20of%20set%20theory | This is a glossary of set theory.
Greek
!$@
A
B
C
D
E
F
G
H
I
See proper, below.
J
K
L
M
N
O
P
Q
R
S
T
U
V
W
XYZ
See also
Glossary of Principia Mathematica
List of topics in set theory
Set-builder notation
References
Set theory
Set theory
Wikipedia glossaries using description lists |
https://en.wikipedia.org/wiki/Gesellschaft%20zur%20Erhaltung%20alter%20und%20gef%C3%A4hrdeter%20Haustierrassen | The or GEH is a German national association for the conservation of historic and endangered domestic animal breeds.
History
The GEH was founded on 5 December 1981 in the Rottal, in Lower Bavaria in southern Germany. It has about 2100 members. Since it was founded, no domestic livestock breed has become extinct in Germany.
Activities
The GEH co-operates with other national and international organisations for the conservation of biodiversity. It publishes an annual Rote Liste or red list of endangered breeds of livestock, which attributes one of four categories of conservation risk to domestic breeds of cattle, dogs, goats, horses, pigs, rabbits and sheep, of chickens, ducks, geese and turkeys, and of bees; listing of domestic pigeon breeds is in preparation. Some breeds from outside Germany are listed separately. The four levels of risk are:
I: , "extremely endangered"
II: , "seriously endangered"
III: , "endangered"
, "alert"
The risk level is calculated using a formula that takes into account five criteria: the number of breeding animals or breeding females; the percentage of pure-bred matings; the five-year trend in breed numbers; the number of breeders or herds; and the interval between generations of the animal.
The GEH also publishes, in conjunction with the , the German national association of poultry breeders, a separate list of the historic poultry breeds and colour varieties that were raised in Germany before 1930. The same levels of conservation risk are assigned as in the main red list.
Endangered breed of the year
Since 1984 the GEH has each year named one or more animal breeds as "endangered breed of the year". To date, these have been:
Note
References
Agriculture in Germany
Biodiversity
Conservation and environmental foundations
Sustainable agriculture
Rare breed conservation
Animal welfare organisations based in Germany
1981 establishments in Germany |
https://en.wikipedia.org/wiki/Place-permutation%20action | In mathematics, there are two natural interpretations of the place-permutation action of symmetric groups, in which the group elements act on positions or places. Each may be regarded as either a left or a right action, depending on the order in which one chooses to compose permutations. There are just two interpretations of the meaning of "acting by a permutation " but these lead to four variations, depending whether maps are written on the left or right of their arguments. The presence of so many variations often leads to confusion. When regarding the group algebra of a symmetric group as a diagram algebra it is natural to write maps on the right so as to compute compositions of diagrams from left to right.
Maps written on the left
First we assume that maps are written on the left of their arguments, so that compositions take place from right to left. Let be the symmetric group on letters, with compositions computed from right to left.
Imagine a situation in which elements of act on the “places” (i.e., positions) of something. The places could be vertices of a regular polygon of sides, the tensor positions of a simple tensor, or even the inputs of a polynomial of variables. So we have places, numbered in order from 1 to , occupied by objects that we can number . In short, we can regard our items as a word of length in which the position of each element is significant. Now what does it mean to act by “place-permutation” on ? There are two possible answers:
an element can move the item in the th place to the th place, or
it can do the opposite, moving an item from the th place to the th place.
Each of these interpretations of the meaning of an “action” by (on the places) is equally natural, and both are widely used by mathematicians. Thus, when encountering an instance of a "place-permutation" action one must take care to determine from the context which interpretation is intended, if the author does not give specific formulas.
Consider the firs |
https://en.wikipedia.org/wiki/Nvidia%20NVENC | Nvidia NVENC (short for Nvidia Encoder) is a feature in Nvidia graphics cards that performs video encoding, offloading this compute-intensive task from the CPU to a dedicated part of the GPU. It was introduced with the Kepler-based GeForce 600 series in March 2012.
The encoder is supported in many livestreaming and recording programs, such as vMix, Wirecast, Open Broadcaster Software (OBS) and Bandicam, as well as video editing apps, such as Adobe Premiere Pro or DaVinci Resolve. It also works with Share game capture, which is included in Nvidia's GeForce Experience software.
Until March 2023 consumer-targeted GeForce graphics cards officially support no more than 3 simultaneously encoding video streams, regardless of the count of the cards installed, but this restriction can be circumvented on Linux and Windows systems by applying an unofficial patch to the drivers. Doing so also unlocks NVIDIA Frame Buffer Capture (NVFBC), a fast desktop capture API that uses the capabilities of the GPU and its driver to accelerate capture. Professional cards support between 3 and unrestricted simultaneous streams per card, depending on card model and compression quality., the restrictions were loosened in 2023 allowing up to 5 simultaneously encoding video streams.
Nvidia chips also feature an onboard decoder, NVDEC (short for Nvidia Decoder), to offload video decoding from the CPU to a dedicated part of the GPU.
Versions
NVENC has undergone several hardware revisions since its introduction with the first Kepler GPU (GK104).
First generation, Kepler GK1xx
The first generation of NVENC, which is shared by all Kepler-based GPUs, supports H.264 high-profile (YUV420, I/P/B frames, CAVLC/CABAC), H.264 SVC Temporal Encode VCE, and Display Encode Mode (DEM).
NVidia's documentation states a peak encoder throughput of 8× realtime at a resolution of 1920×1080 (where the baseline "1×" equals 30Hz). Actual throughput varies on the selected preset, user-controlled parameters and setti |
https://en.wikipedia.org/wiki/%CE%97%20set | In mathematics, an η set (eta set) is a type of totally ordered set introduced by that generalizes the order type η of the rational numbers.
Definition
If is an ordinal then an set is a totally ordered set in which for any two subsets and of cardinality less than , if every element of is less than every element of then there is some element greater than all elements of and less than all elements of .
Examples
The only non-empty countable η0 set (up to isomorphism) is the ordered set of rational numbers.
Suppose that κ = ℵα is a regular cardinal and let X be the set of all functions f from κ to {−1,0,1} such that if f(α) = 0 then f(β) = 0 for all β > α, ordered lexicographically. Then X is a ηα set. The union of all these sets is the class of surreal numbers.
A dense totally ordered set without endpoints is an ηα set if and only if it is ℵα saturated.
Properties
Any ηα set X is universal for totally ordered sets of cardinality at most ℵα, meaning that any such set can be embedded into X.
For any given ordinal α, any two ηα sets of cardinality ℵα are isomorphic (as ordered sets). An ηα set of cardinality ℵα exists if ℵα is regular and Σβ<α 2ℵβ ≤ ℵα.
References
English translation in
Order theory |
https://en.wikipedia.org/wiki/Mildly%20context-sensitive%20grammar%20formalism | In computational linguistics, the term mildly context-sensitive grammar formalisms refers to several grammar formalisms that have been developed in an effort to provide adequate descriptions of the syntactic structure of natural language.
Every mildly context-sensitive grammar formalism defines a class of mildly context-sensitive grammars (the grammars that can be specified in the formalism), and therefore also a class of mildly context-sensitive languages (the formal languages generated by the grammars).
Background
By 1985, several researchers in descriptive and mathematical linguistics had provided evidence against the hypothesis that the syntactic structure of natural language can be adequately described by context-free grammars.
At the same time, the step to the next level of the Chomsky hierarchy, to context-sensitive grammars, appeared both unnecessary and undesirable.
In an attempt to pinpoint the exact formal power required for the adequate description of natural language syntax, Aravind Joshi characterized "grammars (and associated languages) that are only slightly more powerful than context-free grammars (context-free languages)".
He called these grammars mildly context-sensitive grammars and the associated languages mildly context-sensitive languages.
Joshi’s characterization of mildly context-sensitive grammars was biased toward his work on tree-adjoining grammar (TAG).
However, together with his students Vijay Shanker and David Weir, Joshi soon discovered that TAGs are equivalent, in terms of the generated string languages, to the independently introduced head grammar (HG).
This was followed by two similar equivalence results, for linear indexed grammar (LIG) and combinatory categorial grammar (CCG), which showed that the notion of mild context-sensitivity is a very general one and not tied to a specific formalism.
The TAG-equivalent formalisms were generalized by the introduction of linear context-free rewriting systems (LCFRS).
These grammars de |
https://en.wikipedia.org/wiki/Henry%20Quastler | Henry Quastler (November 11, 1908 – July 4, 1963) was an Austrian physician and radiologist who became a pioneer in the field of information theory applied to biology after emigrating to America. His work with Sidney Dancoff led to the publication of what is now commonly called Dancoff's Law.
Life
Quastler spent his early career in Vienna as a doctor. He received his medical degree in Vienna in 1932, focusing on histology and radiology. He met his wife, Gertrude Quastler, a milliner, when she came to him for treatment for tuberculosis. They married in 1933. The couple moved to Albania when King Zog asked for Quastler to train up radiologists. While there he also worked on malaria. Quastler's malaria expertise earned him a place on the International Health Board. As World War II approached in 1939, the couple left Albania and traveled to America. Within a year Quastler was working as a radiologist at New Rochelle Hospital in New York. In 1942, the Quastlers relocated to Urbana, Illinois, where Henry was employed as chief radiologist at Carle Hospital Clinic. While in Illinois, Gertrude Quastler studied art. She soon became a noted artist. Henry also painted as an amateur. According to his sister Johanna, the couple sometimes exhibited together.
In 1949 Quastler gave up his medical practice to concentrate on science. Heinz von Foerster, who knew Quastler well, said that he became even more interested in radiation after the invention of the atomic bomb, which he considered "a horrifying human catastrophe". Foerster recalled Quastler pondering: "[Quastler asked] 'Can I now, as a working person, find out what damage has been done by the radiation of atomic bombs?'—that was his research question. Thus he started to conduct experiments on radiation damage in living organisms." ... Henry Quastler learned the basic concepts and formalisms of information theory with a speed that was almost unbelievable. And why? Because he needed this instrument urgently.".
In the 194 |
https://en.wikipedia.org/wiki/Wood%20science | Wood science, commonly referred to as wood sciences, is a scientific discipline that predominantly investigates elements associated with the formation, composition and macro- and microstructure of wood. It additionally delves into the biological, chemical, physical, and mechanical properties and characteristics of wood, as a natural lignocellulosic material.
A deep understanding of wood plays a pivotal role in various endeavors, such as the processing of wood, the production of wood-based materials like particleboard, fiberboard, OSB, plywood and other materials, as well as the utilization of wood and wood-based materials in construction and a wide array of products, including pulpwood, furniture, engineered wood products such as glued laminated timber, CLT, LVL, PSL, as well as pellets, briquettes, and numerous other products.
History
Initial comprehensive investigations in the field of wood science emerged at the start of the 20th century. The advent of contemporary wood research commenced in 1910, when the Forest Products Laboratory (FPL) was established in Madison, Wisconsin, USA. The Forest Products Laboratory played a fundamental role in wood science providing scientific research on wood and wood products in partnership with academia, industry, local and other institutions in North and South America and worldwide.
In the following years, many wood research institutes came into existence across almost all industrialized nations. A general overview of these institutes and laboratories is shown below:
1913: Institute of Wood and Pulp Chemistry Eberswalde (today's Eberswalde University for Sustainable Development), Germany
1913: Forest Products Laboratory Montreal, Canada
1918: Forest Products Laboratory Vancouver, Canada
1919: Forest Products Laboratory Melbourne, Australia
1923: Forest Products Research Laboratory, Princes Risborough, Great Britain
1929: Institute for Wood Science and Technology, Leningrant, St. Petersburg, USSR
1933: Centre Technique |
https://en.wikipedia.org/wiki/Single-layer%20materials | In materials science, the term single-layer materials or 2D materials refers to crystalline solids consisting of a single layer of atoms. These materials are promising for some applications but remain the focus of research. Single-layer materials derived from single elements generally carry the -ene suffix in their names, e.g. graphene. Single-layer materials that are compounds of two or more elements have -ane or -ide suffixes. 2D materials can generally be categorized as either 2D allotropes of various elements or as compounds (consisting of two or more covalently bonding elements).
It is predicted that there are hundreds of stable single-layer materials. The atomic structure and calculated basic properties of these and many other potentially synthesisable single-layer materials, can be found in computational databases. 2D materials can be produced using mainly two approaches: top-down exfoliation and bottom-up synthesis. The exfoliation methods include sonication, mechanical, hydrothermal, electrochemical, laser-assisted, and microwave-assisted exfoliation.
Single element materials
C: graphene and graphyne
Graphene
Graphene is a crystalline allotrope of carbon in the form of a nearly transparent (to visible light) one atom thick sheet. It is hundreds of times stronger than most steels by weight. It has the highest known thermal and electrical conductivity, displaying current densities 1,000,000 times that of copper. It was first produced in 2004.
Andre Geim and Konstantin Novoselov won the 2010 Nobel Prize in Physics "for groundbreaking experiments regarding the two-dimensional material graphene". They first produced it by lifting graphene flakes from bulk graphite with adhesive tape and then transferring them onto a silicon wafer.
Graphyne
Graphyne is another 2-dimensional carbon allotrope whose structure is similar to graphene's. It can be seen as a lattice of benzene rings connected by acetylene bonds. Depending on the content of the acetylene groups, gr |
https://en.wikipedia.org/wiki/Hydrogenotroph | Hydrogenotrophs are organisms that are able to metabolize molecular hydrogen as a source of energy.
An example of hydrogenotrophy is performed by carbon dioxide-reducing organisms which use CO2 and H2 to produce methane (CH4) by the following reaction:
CO2 + 4H2 → CH4 + 2H2O
Other hydrogenotrophic metabolic pathways include acetogenesis, sulfate reduction, and other hydrogen oxidizing bacteria. Those that metabolize methane are called methanogenic. Hydrogenotrophs belong to a group of organisms known as methanogens, organisms that carry out anaerobic processes that are responsible for the production of methane through carbon dioxide reduction. Methanogens also include a group of organisms called methylotrophs, organisms that can use single-carbon molecules or molecules with no carbon-carbon bonds.
Background Information
Hydrogenotrophic bacteria were first experimented with by NASA in the 1960s in order to find a replenishable food source. Hydrogenotrophic bacteria have been found to have a high protein and carbohydrate content and have been a guiding principle in developing sustainable agricultural methods. Experimentation has revealed that hydrogenotrophic bacteria can convert carbon dioxide into food more rapidly than plants, making them an efficient and sustainable alternative to implement into plant-based high-protein diets and as a substitute in products that use plant extracts and oils.
Hydrogenotrophs are commonly found in the human gut, along with other fermentative bacteria which live in symbiosis with one another. They are also found in soils and in sediments of freshwater and marine ecosystems around the world.
See also
Single cell protein
References
Environmental microbiology
Prokaryotes |
https://en.wikipedia.org/wiki/Weighted%20automaton | In theoretical computer science and formal language theory, a weighted automaton or weighted finite-state machine is a generalization of a finite-state machine in which the edges have weights, for example real numbers or integers. Finite-state machines are only capable of answering decision problems; they take as input a string and produce a Boolean output, i.e. either "accept" or "reject". In contrast, weighted automata produce a quantitative output, for example a count of how many answers are possible on a given input string, or a probability of how likely the input string is according to a probability distribution. They are one of the simplest studied models of quantitative automata.
The definition of a weighted automaton is generally given over an arbitrary semiring , an abstract set with an addition operation and a multiplication operation . The automaton consists of a finite set of states, a finite input alphabet of characters and edges which are labeled with both a character in and a weight in . The weight of any path in the automaton is defined to be the product of weights along the path, and the weight of a string is the sum of the weights of all paths which are labeled with that string. The weighted automaton thus defines a function from to .
Weighted automata generalize deterministic finite automata (DFAs) and nondeterministic finite automata (NFAs), which correspond to weighted automata over the Boolean semiring, where addition is logical disjunction and multiplication is logical conjunction. In the DFA case, there is only one accepting path for any input string, so disjunction is not applied. When the weights are real numbers and the outgoing weights for each state add to one, weighted automata can be considered a probabilistic model and are also known as probabilistic automata. These machines define a probability distribution over all strings, and are related to other probabilistic models such as Markov decision processes and Markov chains.
Weig |
https://en.wikipedia.org/wiki/Whale%20feces | Whale feces, the excrement of whales, has a vital role in the ecology of oceans, earning whales the title of "marine ecosystem engineers." This significant ecological role stems from the nutrients and compounds found in whale feces, which have far-reaching effects on marine life.
Nitrogen and iron chelate released by cetacean species offer significant benefits to the marine food chain and contribute to long-term carbon sequestration. Additionally, whale feces contains a wealth of information about the health, natural history, and ecology of individual animals or groups. This source of information includes DNA, hormones, toxins, and various other chemicals. Studying whale feces provides valuable insights into the lives of these marine creatures, aiding scientists in understanding their behaviors, diets, and overall well-being. Furthermore, the nutrients released through whale feces play a vital role in marine ecosystems, supporting phytoplankton growth, enhancing the food chain, and contributing to the overall health of the oceans.
In addition to feces, the digestive system of sperm whales produces ambergris, a solid, waxy, flammable substance of a dull grey or blackish color which can be found floating on the sea or washed up on the coast.
Description
Whales excrete plumes of liquid feces that are flocculent in nature, consisting of loose aggregations of particles. These feces, often found floating on the sea surface after being excreted underwater before it dissociates, contain undigested hard objects such as squid beaks. Fecal samples are characterized by color, odor, texture, and buoyancy, providing valuable information about the health and ecology of whales. Flatulence has been recorded in whales.
Ecological significance
Nutrient cycling and carbon sequestration
One of the crucial roles of whale feces is in nutrient cycling, particularly nitrogen circulation in the ocean. Whales transport more nitrogen through their feces in certain regions than all the r |
https://en.wikipedia.org/wiki/Asus | ASUSTeK Computer Inc. (, , , ; ; stylized as ASUSTeK or ASUS) is a Taiwanese multinational computer, phone hardware and electronics manufacturer headquartered in Beitou District, Taipei, Taiwan. Its products include desktop computers, laptops, netbooks, mobile phones, networking equipment, monitors, wi-fi routers, projectors, motherboards, graphics cards, optical storage, multimedia products, peripherals, wearables, servers, workstations and tablet PCs. The company is also an original equipment manufacturer (OEM).
Asus is the world's 5th-largest PC vendor by unit sales as of January 2023. Asus appears in BusinessWeek "InfoTech 100" and "Asia's Top 10 IT Companies" rankings, and it ranked first in the IT Hardware category of the 2008 Taiwan Top 10 Global Brands survey with a total brand value of billion.
Asus has a primary listing on the Taiwan Stock Exchange under the ticker code 2357 and formerly had a secondary listing on the London Stock Exchange under the ticker code ASKD.
Name
The company is usually referred to as ASUS or Huáshuò in Chinese (, literally "Eminence by the Chinese"). According to the company website, the name ASUS originates from Pegasus, the winged horse of Greek mythology. Only the last four letters of the word were used to give the name a high position in alphabetical listings.
As its marketing taglines, ASUS has used Rock Solid. Heart Touching (2003–2009) and subsequently Inspiring Innovation Persistent Perfection (2009–2013). Since 2013, the company's tagline has been In Search of Incredible.
History
Asus was founded in Taipei in 1989 by T.H. Tung, Ted Hsu, Wayne Hsieh and M.T. Liao, all four having previously worked at Acer as hardware engineers. At this time, Taiwan had yet to establish a leading position in the computer hardware business. Intel Corporation would supply any new processors to more established companies like IBM first, and Taiwanese companies would have to wait for approximately six months after IBM received their engine |
https://en.wikipedia.org/wiki/ALL.Net | ALL.Net (Amusement Linkage Live Network) is an arcade video game network communication system and digital distribution system made by Sega Corporation. It is similar to the Taito NESiCAxLive game distribution systems and NESYS arcade network; the player smart card system is similar to the Konami e-AMUSEMENT system. It enables arcade games to be connected via the Internet, enabling communication battles, national rankings, and storage of play data.
Development
ALL.net was developed by Sega in 2004. It was created as a method of allowing players to save player profiles, player rankings, high scores, create online rankings and have competitive online play. The system was based on the previous VF.net created by Sega for Virtua Fighter 4 in 2001. Initially, the service was only available in Japan, but following a trial in Hong Kong in 2008, the service has been extended to other parts of Asia in 2010. The system has been rolled out to South Korea, Taiwan, Singapore, China.
ALL.Net was further developed as ALL.Net P-ras to allow digital distribution of arcade games, as well as for software updates. ALL.Net P-ras allows profit sharing with the arcade operators, with Sega renting games for free, while the operator pays the cost of the hardware, with all revenues from players being split between Sega and the arcade operator.
ALL.Net games
See List of ALL.Net games
See also
Digital distribution in video games
List of Sega arcade system boards
References
External links
Products introduced in 2004
2004 establishments in Japan
Products introduced in 2010
Internet properties established in 2004
Sega hardware
Video game distribution
Online content distribution
Online video game services |
https://en.wikipedia.org/wiki/Head-in-pillow%20defect | In the assembly of integrated circuit packages to printed circuit boards, a head-in-pillow defect (HIP or HNP) is a failure of the soldering process. For example, in the case of a ball grid array (BGA) package, the pre-deposited solder ball on the package and the solder paste applied to the circuit board may both melt, but the melted solder does not join. A cross-section through the failed joint shows a distinct boundary between the solder ball on the part and the solder paste on the circuit board, rather like a section through a head resting on a pillow.
The defect can be caused by surface oxidation or poor wetting of the solder, or by distortion of the integrated circuit package or circuit board by the heat of the soldering process. This is particularly a concern when using lead-free solder, which requires higher processing temperature.
Since the warping of the circuit board or integrated circuit may disappear when the board cools, an intermittent fault may be created. Diagnosis of head-in-pillow defects may require use of X-rays or EOTPR (Electro Optical Terahertz Pulse Reflectometry), since the solder joints are hidden between the integrated circuit package and the printed circuit board.
See also
Ball grid array
References
Further reading
Electronics manufacturing
Engineering failures
Metallurgy
Soldering defects
Soldering |
https://en.wikipedia.org/wiki/Kingman%27s%20subadditive%20ergodic%20theorem | In mathematics, Kingman's subadditive ergodic theorem is one of several ergodic theorems. It can be seen as a generalization of Birkhoff's ergodic theorem.
Intuitively, the subadditive ergodic theorem is a kind of random variable version of Fekete's lemma (hence the name ergodic). As a result, it can be rephrased in the language of probability, e.g. using a sequence of random variables and expected values. The theorem is named after John Kingman.
Statement of theorem
Let be a measure-preserving transformation on the probability space , and let be a sequence of functions such that (subadditivity relation). Then
for -a.e. x, where g(x) is T-invariant. If T is ergodic, then g(x) is a constant.
Applications
Taking recovers Birkhoff's pointwise ergodic theorem.
Kingman's subadditive ergodic theorem can be used to prove statements about Lyapunov exponents. It also has applications to percolations and probability/random variables.
References
External links
Theorem proof (Steele)
Ergodic theory |
https://en.wikipedia.org/wiki/CalPhotos | CalPhotos is an online database of natural history photographs, including many useful for identifying wildlife. It is maintained by the University of California, Berkeley. Its images are used by many universities, government agencies, websites, and others, including a partnership with the Encyclopedia of Life. Plant identifications are user-submitted, but a process for review and correction is available.
External links
CalPhotos database
References
Image-sharing websites
Online databases
University of California |
https://en.wikipedia.org/wiki/Corank | In mathematics, corank is complementary to the concept of the rank of a mathematical object, and may refer to the dimension of the left nullspace of a matrix, the dimension of the cokernel of a linear transformation of a vector space, or the number of elements of a matroid minus its rank.
Left nullspace of a matrix
The corank of an matrix is where is the rank of the matrix. It is the dimension of the left nullspace and of the cokernel of the matrix.
Cokernel of a linear transformation
Generalizing matrices to linear transformations of vector spaces, the corank of a linear transformation is the dimension of the cokernel of the transformation, which is the quotient of the codomain by the image of the transformation.
Matroid
For a matroid with elements and matroid rank , the corank or nullity of the matroid is . In the case of linear matroids this coincides with the matrix corank. In the case of graphic matroids the corank is also known as the circuit rank or cyclomatic number.
Linear algebra
Matroid theory |
https://en.wikipedia.org/wiki/Sony%20NEX-F3 | The Sony α NEX-F3 is an entry level rangefinder-styled digital mirrorless interchangeable-lens camera announced by Sony on 17 May 2012.
References
http://www.dpreview.com/products/sony/slrs/sony_nexf3/specifications
NEX-F3
NEX-F3
Live-preview digital cameras
Cameras introduced in 2012 |
https://en.wikipedia.org/wiki/Sony%20NEX-3N | The Sony α NEX-3N is an entry level rangefinder-styled digital mirrorless interchangeable lens camera announced by Sony on 20 February 2013.
See also
List of smallest mirrorless cameras
References
http://www.dpreview.com/products/sony/slrs/sony_nex3n/specifications
NEX-3N
NEX-3N
Live-preview digital cameras
Cameras introduced in 2013 |
https://en.wikipedia.org/wiki/Sony%20NEX-5R | The Sony α NEX-5R is a mid-range rangefinder-styled digital mirrorless interchangeable lens camera announced by Sony on 29 August 2012.
See also
Sony NEX-5
Sony NEX-7
List of Sony E-mount cameras
List of smallest mirrorless cameras
References
http://www.dpreview.com/products/sony/slrs/sony_nex5r/specifications
NEX-5R
NEX-5R
Live-preview digital cameras
Cameras introduced in 2012 |
https://en.wikipedia.org/wiki/Sony%20NEX-5T | The Sony α NEX-5T is a mid-range rangefinder-styled digital mirrorless interchangeable lens camera announced by Sony on 27 August 2013.
See also
Sony NEX-5
Sony NEX-7
List of Sony E-mount cameras
List of smallest mirrorless cameras
References
http://www.dpreview.com/products/sony/slrs/sony_nex5t/specifications
NEX-5T
NEX-5T
Live-preview digital cameras
Cameras introduced in 2013 |
https://en.wikipedia.org/wiki/Yik%20Yak | Yik Yak is a pseudonymous social media smartphone application that initially launched in 2013 and relaunched in 2021. The app, which is available for iOS and (formerly) Android, allows college students to create and view discussion threads within a radius (termed "Yaks" by the application). It is similar to other anonymous sharing apps such as Nearby, but differs from others such as Whisper in that it is intended for sharing primarily with those in proximity to the user.
Despite strong levels of growth in 2013 and 2014, following several bouts of heavy criticism in the media over the dissemination of racism, antisemitism, sexism and the facilitation of cyber-bullying, the service saw stagnation in the growth of its user base. In 2016 alone, user downloads fell 76% compared to 2015. Failing to maintain user engagement, Yik Yak announced on April 28, 2017, that the service would close in the coming week. For $1 million, Block, Inc. (formally Square, Inc.), purchased Yik Yak's intellectual property and hired several of its former employees. On August 15, 2021, Yik Yak announced via their official website that they were making a comeback, with the app available for download on iOS and now recently available on Android. However, as of an update in March, 2023, the app is no longer available on Android, with no indication from Yik Yak regarding when, or if, it will return to the platform.
History
The co-founders, Tyler Droll and Brooks Buffington, are both graduates from Furman University in Greenville, South Carolina. The two started collaborating when they were placed into the same class where they learned how to code iPhone apps. After graduating from Furman University, they decided to go full-time with their project. Droll dropped out of medical school just before it started and Buffington put his finance career on hold. The two released the app in November 2013, and twelve months later, Yik Yak was ranked as the ninth most downloaded social media app in the United |
https://en.wikipedia.org/wiki/Cloud-based%20design%20and%20manufacturing | Cloud-based design and manufacturing (CBDM) refers to a service-oriented networked product development model in which service consumers are able to configure products or services and reconfigure manufacturing systems through Infrastructure-as-a-Service (IaaS), Platform-as-a-Service (PaaS), Hardware-as-a-Service (HaaS), and Software-as-a-Service (SaaS).
Adapted from the original cloud computing paradigm and introduced into the realm of computer-aided product development, Cloud-Based Design and Manufacturing is gaining significant momentum and attention from both academia and industry.
Cloud-based design and manufacturing includes two aspects: cloud-based design and cloud-based manufacturing. Another related concept is cloud manufacturing that is more general and popular.
Cloud-Based Design (CBD) refers to a networked design model that leverages cloud computing, service-oriented architecture (SOA), Web 2.0 (e.g., social network sites), and semantic web technologies to support cloud-based engineering design services in distributed and collaborative environments.
Cloud-Based Manufacturing (CBM) refers to a networked manufacturing model that exploits on-demand access to a shared collection of diversified and distributed manufacturing resources to form temporary, reconfigurable production lines which enhance efficiency, reduce product lifecycle costs, and allow for optimal resource allocation in response to variable-demand customer generated tasking.
The enabling technologies for Cloud-Based Design and Manufacturing include cloud computing, Web 2.0, Internet of Things (IoT), and service-oriented architecture (SOA).
History
The term cloud-based design and manufacturing (CBDM) was initially coined by Dazhong Wu, David Rosen, and Dirk Schaefer at Georgia Tech in 2012 for the purpose of articulating a new paradigm for digital manufacturing and design innovation in distributed and collaborative settings. The main objective of CBDM is to further reduce time and cost assoc |
https://en.wikipedia.org/wiki/%C5%81ukasiewicz%E2%80%93Moisil%20algebra | Łukasiewicz–Moisil algebras (LMn algebras) were introduced in the 1940s by Grigore Moisil (initially under the name of Łukasiewicz algebras) in the hope of giving algebraic semantics for the n-valued Łukasiewicz logic. However, in 1956 Alan Rose discovered that for n ≥ 5, the Łukasiewicz–Moisil algebra does not model the Łukasiewicz logic. A faithful model for the ℵ0-valued (infinitely-many-valued) Łukasiewicz–Tarski logic was provided by C. C. Chang's MV-algebra, introduced in 1958. For the axiomatically more complicated (finite) n-valued Łukasiewicz logics, suitable algebras were published in 1977 by Revaz Grigolia and called MVn-algebras. MVn-algebras are a subclass of LMn-algebras, and the inclusion is strict for n ≥ 5. In 1982 Roberto Cignoli published some additional constraints that added to LMn-algebras produce proper models for n-valued Łukasiewicz logic; Cignoli called his discovery proper Łukasiewicz algebras.
Moisil however, published in 1964 a logic to match his algebra (in the general n ≥ 5 case), now called Moisil logic. After coming in contact with Zadeh's fuzzy logic, in 1968 Moisil also introduced an infinitely-many-valued logic variant and its corresponding LMθ algebras. Although the Łukasiewicz implication cannot be defined in a LMn algebra for n ≥ 5, the Heyting implication can be, i.e. LMn algebras are Heyting algebras; as a result, Moisil logics can also be developed (from a purely logical standpoint) in the framework of Brower’s intuitionistic logic.
Definition
A LMn algebra is a De Morgan algebra (a notion also introduced by Moisil) with n-1 additional unary, "modal" operations: , i.e. an algebra of signature where J = { 1, 2, ... n-1 }. (Some sources denote the additional operators as to emphasize that they depend on the order n of the algebra.) The additional unary operators ∇j must satisfy the following axioms for all x, y ∈ A and j, k ∈ J:
if for all j ∈ J, then x = y.
(The adjective "modal" is related to the [ultimate |
https://en.wikipedia.org/wiki/Timeline%20of%20computational%20mathematics | This is a timeline of key developments in computational mathematics.
1940s
Monte Carlo simulation (voted one of the top 10 algorithms of the 20th century) invented at Los Alamos by von Neumann, Ulam and Metropolis.
Dantzig introduces the simplex algorithm (voted one of the top 10 algorithms of the 20th century).
First hydro simulations at Los Alamos occurred.
Ulam and von Neumann introduce the notion of cellular automata.
A routine for the Manchester Baby written to factor a large number (2^18), one of the first in computational number theory. The Manchester group would make several other breakthroughs in this area.
LU decomposition technique first discovered.
1950s
Hestenes, Stiefel, and Lanczos, all from the Institute for Numerical Analysis at the National Bureau of Standards, initiate the development of Krylov subspace iteration methods. Voted one of the top 10 algorithms of the 20th century.
Equations of State Calculations by Fast Computing Machines introduces the Metropolis–Hastings algorithm. Also, important earlier independent work by Alder and S. Frankel.
Enrico Fermi, Stanislaw Ulam, John Pasta, and Mary Tsingou, discover the Fermi–Pasta–Ulam–Tsingou problem.
In network theory, Ford & Fulkerson compute a solution to the maximum flow problem.
Householder invents his eponymous matrices and transformation method (voted one of the top 10 algorithms of the 20th century).
Molecular dynamics invented by Alder and Wainwright
John G.F. Francis and Vera Kublanovskaya invent QR factorization (voted one of the top 10 algorithms of the 20th century).
1960s
First recorded use of the term "finite element method" by Ray Clough, to describe the methods of Courant, Hrenikoff and Zienkiewicz, among others. See also here.
Using computational investigations of the 3-body problem, Minovitch formulates the gravity assist method.
Molecular dynamics was invented independently by Aneesur Rahman.
Cooley and Tukey re-invent the Fast Fourier transform (voted on |
https://en.wikipedia.org/wiki/Process%20network%20synthesis | Process network synthesis (PNS) is a method to represent a process structure in a 'directed bipartite graph'. Process network synthesis uses the P-graph method to create a process structure. The scientific aim of this method is to find optimum structures.
Process network synthesis uses a bipartite graph method P-graph and employs combinatorial rules to find all feasible network solutions (maximum structure) and links raw materials to desired products related to the given problem. With a branch and bound optimisation routine and by defining the target value an optimum structure can be generated that optimises a chosen target function.
Process Network Synthesis was originally developed to solve chemical process engineering processes. Target value as well as the structure can be changed depending on the field of application. Thus many more fields of application followed.
Applications
At Pannon University software the tools PNS Editor and PNS Studio were programmed to generate maximum structure of processes. This software includes the p-graph method and MSG, SSG and ABB branch and bound algorithms to detect optimum structures within the maximum available process flows.
PNS is used in different applications where it can be used to find optimum process structures like:
Process engineering: Chemical process designs and the Synthesis of chemical processes is applied in different case studies.
Optimum energy technology networks for regional and urban energy systems: In case of regional and urban energy planning the financially most feasible solution for resource systems is selected as target value. With this setting material- and energy flows, energy demand and cost of technologies are considered and the optimum technology network can be found. Simultaneously the robustness of technologies due to price changes and limitations in resource availability can be identified.
Evacuation routes in buildings: The aim is to find optimal routes to evacuate buildings depending o |
https://en.wikipedia.org/wiki/Digital%20Signage%20Federation | Digital Signage Federation (DSF) is an American non-profit organization dedicated to the advancement of the Digital Signage industry. It was established and named by Angelo R. Varrone, then founded in 2010 by a number of industry professionals to form a trade association focused on education, networking, and standards development. The purposes of the DSF are to support and promote the common business interests of the interactive technologies industry, the digital signage industry, and the digital out-of-home network industry (DOOH).
The DSF was formed in March 2010 and officially incorporated on May 12, 2010. Since its incorporation, the DSF has supported the research of, and improved legislation, regulation, code and standards that affect the sale and use of, digital signage and other products and services within the Association Industries (digital signage, interactive technologies, and DOOH).
Its current structure includes an Executive committee formed by a Chairman, Vice-Chairman, Treasurer/Secretary, and Immediate Past Chairman; and a rotating Board of At-Large Directors composed of 14 industry professionals.
History
The DSF was formed by a number of industry professionals to form a trade association focused on education, networking, and standards development. The group of individuals who started the organization came from attendees and exhibitors of the "Digital Signage Expo" (DSE) trade show in 2010. The initial group of leaders was appointed and membership was free for the first year. The DSE and the DSF created a funding agreement and the DSF recognized the DSE as its official industry trade show. Also in 2010, the board adopted the Digital Signage Experts Group certification program as its official certification and received discounts for members. The DSF also held its first election of board members.
The original structure of the DSF was that there were independent governing councils based on industry sector. These "mini-associations" were to appoin |
https://en.wikipedia.org/wiki/OpenWireless.org | The Open Wireless Movement hosted at OpenWireless.org is an Internet activism project which seeks to increase Internet access by encouraging people and organizations to configure or install software on their own wireless router to offer a separate public guest network or to make a single public wireless access point. If many people did this, then a ubiquitous global public wireless network would be created which would achieve and surpass the goal of increasing Internet access.
History
The project was initiated in November 2012 by a coalition of ten advocacy groups including the Electronic Frontier Foundation (EFF), Fight for the Future, Free Press, Internet Archive, NYCwireless, Open Garden, OpenITP, the Open Spectrum Alliance, the Open Technology Institute, and the Personal Telco Project. EFF representative Adi Kamdar commented, "We envision a world where sharing one's Internet connection is the norm. A world of open wireless would encourage privacy, promote innovation, and largely benefit the public good. And everyone—users, businesses, developers, and Internet service providers—can get involved." As of September 2016, seventeen groups have joined the project, adding Engine, Mozilla, Noisebridge, the Open Rights Group, OpenMedia International, Sudo Room, and the Center for Media Justice.
The project uses various strategies to encourage and assist people to make their Internet connections available for public use. It explains the benefits and drawbacks of the effects on society and on the owners of routers, answers questions regarding safety and legality, guides novice users in configuring their routers, and provides firmware for novices to install on their routers.
Router
The EFF created a router firmware called OpenWireless, a fork of CeroWRT, which is a branch of the OpenWrt firmware. which anyone may volunteer to install on their router to make it work for the OpenWireless.org project. This firmware was first shared at the 2014 Hackers on Planet Earth confe |
https://en.wikipedia.org/wiki/AN/GYK-12 | The AN/GYK-12 is an obsolete 32-bit minicomputer developed by Litton Industries for the United States Army. The AN/GYK-12 is a militarized version of the L-3050 computer ruggedized for use in the TACFIRE tactical fire direction system. The design dates from the 1960s.
In 1980, the Army introduced the Nebula instruction set architecture (MIL-STD-1862), intended as an upgrade to the AN/GYK-12. Nebula is also a 32-bit architecture with 32-bit addressing mode and instructions optimized for running programs written in Ada.
Description
The basic system consists of three rack-mounted modules: The CPU, the IOU (Input/Output unit), and the MCMU. Modules are mounted vertically and are wide by high.
The AN/GYK-12 has a 32-bit instruction format and operates on data from one bit to a doubleword (64 bits) in size. Only fixed-point binary arithmetic is provided.
The system can have a maximum of 32768 pages—64 million words—of 2 μs memory (256 MB). Physically memory is divided into banks of 8 KW to minimize memory contention. Logically it is divided into pages of 2 KW each. Access to pages is controlled by 16 Page Control and Address Registers per program level, allowing an individual program to directly address 128 KB at one time.
The AN-GYK-12 CPU features 64 hardware priority program levels, numbered from 0 (highest) to 63 (lowest). One task can run at each level. Level 0 is reserved for "power off". Level 1 is power-on restart. Level 2 handles hardware and program errors. Level 63 is entered after completion of a boot load. Therefore the system can support a maximum of 60 user tasks. Tasks are scheduled preemptively—a task will run until an error occurs, until it gives up control to another task, or until successful completion of an input/output operation transfers control to the so-called normal program level.
Each task also has a privilege level, which controls the operations it is allowed to perform. Level '00'b is used for non-privileged programs. Leve |
https://en.wikipedia.org/wiki/Solaborate | Solaborate /ˈsōˈlabəˌrāte/ is a communication and collaboration platform for professionals and consumers. It allows them to do video calls, share documents, screencast and screen share, remotely monitor their home or office. Solaborate is a combination of the words "social" and "collaboration". Solaborate has both software and hardware in the form of the HELLO Messenger and HELLO Communication Device.
History
Solaborate's CEO Labinot Bytyqi, founded the company in 2012. Solaborate is headquartered in Clearwater, Florida.
In April 2013, Solaborate launched its private beta, and announced it has raised $1 million from angel investors.
In October 2013, at TechEd Las Vegas, Solaborate officially launched its public beta.
HELLO
In July 2016, Solaborate launched a Kickstarter campaign for HELLO 1. HELLO is a voice controlled device that can be used for video conferencing, wireless screensharing, live broadcasting, and security surveillance. The device runs on Android and is engineered with an array of four smart microphones, a 4K video sensor and quad-core processor. It also features a built-in accelerometer and tilting lens.
In January 2018, Solaborate announced HELLO 2 at CES 2018.
Awards
People's Choice Award at SAP Palo Alto
In 2013, Solaborate has been awarded the people's choice award at SAP's Palo Alto, during the 10th startup forum.
Best in Show Award at WebRTC Conference and Expo
In 2013, Solaborate has received the "Best in Show" award at the WebRTC Conference and Expo hosted by TMC in Atlanta, Ga
Technology
Solaborate is built in Microsoft Windows Azure Cloud Computing, using HTML5, WebSockets and WebRTC protocol for real time communication. Solaborate runs on SAP HANA platform, to provide users with real-time information on their network.
Features
Features include:
profiles for professionals, companies, products, and services
messaging, chat, and video
screen sharing
creating blogs
real-time analytics
file sharing and document organization
In Octo |
https://en.wikipedia.org/wiki/Walsh%20Report%20%28cryptography%29 | The Walsh Report was an Australian cryptography policy review undertaken for the Australian government by Gerard Walsh, initially released in 1996 and then withdrawn before its sale to the public. Electronic Frontiers Australia (EFA) obtained a redacted copy under freedom of information laws and published it on EFA's website. Subsequently, an unredacted copy was found and the redacted parts were added to the EFA copy.
Policy review
The Walsh Report was an Australian cryptography policy review undertaken at the request of the Secretary of the Attorney-General's Department by Gerard Walsh, the former deputy director of the Australian Security Intelligence Organisation (ASIO). The report included a broad analysis of cryptography issues from an Australian context.
The report, titled Review of Policy relating to Encryption Technologies, is popularly called the Walsh Report.
In his report, Walsh found that there was a lack of coordination in the government over the establishment of cryptography policy. Walsh also reported no clarity as to which department and which minister was responsible for cryptographic policy. Consequently, there was a danger that policy would be developed without being coordinated.
The main advice given by Walsh in the report was that major legislative action to safeguard law-enforcement or national security was not required at the time.
No specific options were recommended in the report for legislation on cryptography, nor did the report recommend mandatory key recovery.
Recommendations in the report for minor legislative and other actions included:
establishment of a summary law on intrusive investigative powers
to consider the setting up of an additional and more serious offence when cryptography is used to obstruct a criminal investigation
to consider establishing a power to allow police to demand encryption keys
a key recovery or escrow scheme, as had been advocated by the United States, not be established by Australia.
History
Bac |
https://en.wikipedia.org/wiki/Mesodinium%20rubrum | Mesodinium rubrum (or Myrionecta rubra) is a species of ciliates. It constitutes a plankton community and is found throughout the year, most abundantly in spring and fall, in coastal areas. Although discovered in 1908, its scientific importance came into light in the late 1960s when it attracted scientists by the recurrent red colouration it caused by forming massive blooms, that cause red tides in the oceans.
Unlike typical protozoans, M. rubrum can make its own nutrition by photosynthesis. The unusual autotrophic property was discovered in 2006 when genetic sequencing revealed that the photosynthesising organelles, plastids, were derived from the principal food of the ciliate, the photosynthetic algae called cryptomonads (or cryptophytes). It is, thus, both autotrophic and heterotrophic. This nature also indicates that it is an example of endosymbiosis, supporting the endosymbiotic theory, as well as the concept of stealing of cell organelles called kleptoplastidy. Moreover, M. rubrum represents additional endosymbiosis by transferring its plastids to its predators, the dinoflagellate planktons belonging to the genus Dinophysis.
In 2009, a new species of Gram-negative bacteria called Maritalea myrionectae was discovered from a cell culture of M. rubrum.
Description
M. rubrum is a free-living marine ciliate. It is reddish in colour and form dark-red mass during blooming. Its body is almost spherical, looking like a miniature sunflower with its radiating hair-like cilia on its body surface. It measures up to 100 μm in length and 75 μm in width. The body is superficially divided into two lobes due to formation of a constriction at the centre. The constriction gives rise to a larger anterior lobe and a smaller posterior lobe. The cilia arise from the constriction. Using the cilia it can jump about 10-20 times its body length in one movement. Its nucleus is prominently situated at the centre, and is surrounded by organelles mostly derived from algae. For example, i |
https://en.wikipedia.org/wiki/Randomness%20merger | In extractor theory, a randomness merger is a function which extracts randomness out of a set of random variables, provided that at least one of them is uniformly random. Its name stems from the fact that it can be seen as a procedure which "merges" all the variables into one, preserving at least some of the entropy contained in the uniformly random variable. Mergers are currently used in order to explicitly construct randomness extractors.
Intuition and definition
Consider a set of random variables, , each distributed over at least one of which is uniformly random; but it is not known which one. Furthermore, the variables may be arbitrarily correlated: they may be functions of one another, they may be constant, and so on. However, since at least one of them is uniform, the set as a whole contains at least bits of entropy.
The job of the merger is to output a new random variable, also distributed over , that retains as much of that entropy as possible. Ideally, if it were known which of the variables is uniform, it could be used as the output, but that information is not known. The idea behind mergers is that by using a small additional random seed, it is possible to get a good result even without knowing which one is the uniform variable.
Definition (merger):
A function is called an -merger if for every set of random variables distributed over , at least one of which is uniform, the distribution of has smooth min-entropy . The variable denotes the uniform distribution over bits, and represents a truly random seed.
In other words, by using a small uniform seed of length , the merger returns a string which is -close to having at least min-entropy; this means that its statistical distance from a string with min-entropy is no larger than .
Reminder: There are several notions of measuring the randomness of a distribution; the min-entropy of a random variable is defined as the largest such that the most probable value of occurs with probability no mor |
https://en.wikipedia.org/wiki/Lines%20of%20non-extension | In the field of biomechanics, the lines of non-extension are notional lines running across the human body along which body movement causes neither stretching or contraction. Discovered by Arthur Iberall in work beginning in the 1940s, as part of research into space suit design, they have been further developed by Dava Newman in the development of the Space Activity Suit.
They were originally mapped by Iberall by drawing a series of circles over a portion of the body and then watching their deformations as the wearer walked around or performed various tasks. The circles deform into ellipses as the skin stretches over the moving musculature, and these deformations were recorded. After a huge number of such measurements the data is then examined to find all of the possible deformations of the circles, and more importantly, the non-moving points on them where the original circle and the deformed ellipse intersect (at four points per circle). By mapping these points over the entire body, a series of lines are produced.
These lines may then be used to direct the placement of tension elements in a spacesuit to enable constant suit pressure regardless of the motion of the body.
References
Anatomy
Biomechanics
Spacesuits |
https://en.wikipedia.org/wiki/PlantUML | PlantUML is an open-source tool allowing users to create diagrams from a plain text language. Besides various UML diagrams, PlantUML has support for various other software development related formats (such as Archimate, Block diagram, BPMN, C4, Computer network diagram, ERD, Gantt chart, Mind map, and WBD), as well as visualisation of JSON and YAML files.
The language of PlantUML is an example of a domain-specific language. Besides its own DSL, PlantUML also understands AsciiMath, Creole, DOT, and LaTeX. It uses Graphviz software to lay out its diagrams and Tikz for LaTeX support. Images can be output as PNG, SVG, LaTeX and even ASCII art. PlantUML has also been used to allow blind people to design and read UML diagrams.
Applications that use PlantUML
There are various extensions or add-ons that incorporate PlantUML.
Atom has a community maintained PlantUML syntax highlighter and viewer.
Confluence wiki has a PlantUML plug-in for Confluence Server, which renders diagrams on-the-fly during a page reload. There is an additional PlantUML plug-in for Confluence Cloud.
Doxygen integrates diagrams for which sources are provided after the \startuml command.
Eclipse has a PlantUML plug-in.
Google Docs has an add-on called PlantUML Gizmo that works with the PlantUML.com server.
IntelliJ IDEA can create and display diagrams embedded into Markdown (built-in) or in standalone files (using a plugin).
LaTeX using the Tikz package has limited support for PlantUML.
LibreOffice has Libo_PlantUML extension to use PlantUML diagrams.
MediaWiki has a PlantUML plug-in which renders diagrams in pages as SVG or PNG.
Microsoft Word can use PlantUML diagrams via a Word Template Add-in. There is an additional Visual Studio Tools for Office add-in called PlantUML Gizmo that works in a similar fashion.
NetBeans has a PlantUML plug-in.
Notepad++ has a PlantUML plug-in.
Org-mode has a PlantUML org-babel support.
Rider has a PlantUML plug-in.
Visual Studio Code has various Pla |
https://en.wikipedia.org/wiki/Nemesis%202%20%28MSX%29 | is a side-scrolling shoot 'em up video game released for the MSX computer in 1987 by Konami. The game is a sequel to Nemesis, the MSX version of Gradius, but is unrelated to the arcade game Gradius II (which used the Roman numeral 'II'). This version was ported to the X68000 computer under the name , with some graphical and aural enhancements.
In a departure from other games, instead of controlling Vic Viper, the available ship is called Metalion. Unlike other titles, this game has a heavier focus on story, which is told by cut-scenes. The gameplay is mostly unchanged from the rest of the series, though there are some powerups that temporally gives the ship some enhancements. Also, when the bosses are being defeated, if the Metalion flies where they are, a mini-level can be accessed in order to obtain new permanent upgrades, if the mini levels are successfully cleared.
Plot
The Director General of Space Science Agency Dr. Venom was exiled to Planet Sard for a failed coup d'état. In the year 6665, he escapes and invades Planet Nemesis and the seven planets it controls with the help of Bacterion. The Nemesis High Council sends James Burton, ex-pilot of the Vic Viper, to pilot Metalion and attack Dr. Venom and the Bacterion invaders. The game takes place during the year 6666.
Nemesis '90 Kai
This X68000 port is essentially an enhanced remake of Nemesis 2 with graphical quality on par with Gradius III.
It includes two new stages exclusive to this version of the game, and four new bosses (two of which replace the rematches fought in the MSX version.)
Some people still prefer the original for its charm and color scheme.
Ports
Aside from being remade as Nemesis '90 Kai, Nemesis 2 was also ported to mobile phones in 2006 and Sony PSP in 2007 as part of the Salamander Portable collection.
Gradius 2 was re-released for Wii's Virtual Console in 2009, for Project EGG in 2015, and for Wii U Virtual Console in 2016 in Japan.
References
External links
Extensive Japanese |
https://en.wikipedia.org/wiki/Comparison%20of%203D%20printers | This article contains a list of 3D printers.
Metrics
Some important 3D printer metrics include:
Print technology: 3D printing processes encompasses a variety of methods which each have their own unique challenges.
Length of production run and support: 3D printers often require adjustments and parts replacement. A long production run often means that spare parts are also available.
Maximum build volume: Defines how large parts that are possible to print on a given printer. Often measured in millimeters, but sometimes in inches or centimeters instead.
Minimum layer resolution: Defines the resolution of the print (usually the vertical resolution). Often measured in micrometers (µm). The actual resolution of a printer can usually be adjusted within an interval.
Print speed: Defines how fast the printer is, and is usually measured in millimeters per second (mm/s). The actual speed of a printer can usually be adjusted within an interval.
Kit or assembled: Printers are usually sold either pre-assembled, partially assembled or as complete kits. Kits are usually less expensive, but also usually require more fitting and calibration.
Open source: Some printer designs are released into the open source domain, which can include for instance hardware specifications, CAD files, board schematics and firmware files. Open source printer designs are often popular in online do it yourself-communities.
Noise level: Measured in decibels (dB), and can vary greatly in home-printers from 15 dB to 75 dB. Some main sources of noise in filament printers are fans, motors and bearings, while in resin printers the fans usually are responsible for most of the noise. Some methods for dampening the noise from a printer may be to install vibration isolation, use larger diameter fans, perform maintenance and lubrication, or using a soundproofing enclosure.
Commercial printers
See also
List of 3D printer manufacturers
3D printing processes
References
3D |
https://en.wikipedia.org/wiki/Insider%20investment%20strategy | The insider investment strategy is an investment strategy that follows the buying and selling decisions of so-called "insiders" in a stock market. The primary insiders have an advantage because they have access to more information about issues that could affect the current and future value of stock, which is known as an "information advantage." However, in the world there are only a few investment funds that follow the insider trades, both of them were established in 2011.
In the United States, Catalyst Capital Advisors LLC manages Catalyst Insider Buying Fund. This fund is a large-cap, long-only equity fund that only invests in companies where corporate insiders are buying their own company's stock on the open market. In Europe, Dovre Forvaltning UAB manages Dovre Inside Nordic fund.
Insider trading studies
A Lorie-Niederhoffer study indicates that proper and prompt analysis of data on insider trading can be profitable.
In 2014, Dovre Forvaltning shared his analysis on Insider Influence in the Nordic Region. The company analyzed these different yearly portfolios (both for Purchases and Sales):
Had an insider Purchases in the past 1/3/6 months.
Had only insider Purchases in the past 1/3/6 months.
Last insider Transaction in the past 1/3/6 months was an insider Purchase.
Had an insider Sales in the past 1/3/6 months.
Had only insider Sales in the past 1/3/6 months.
Last insider Transaction in the past 1/3/6 months was an insider Sale.
Only transactions above 80.000 SEK were included (33% of all insider trades were excluded because they were too small). If there were no Purchases/Sales in 1,3,6 months after a company's inclusion, it was excluded from the portfolio. All stocks are equally weighted. The analysis showed that:
Highest out performance is in Small Caps insider Purchase portfolios. Smaller in Mid Caps Purchase portfolios. And smallest in Large Caps insider Purchase portfolios.
'Insider effect' fades away with longer holding horizon.
Sell |
https://en.wikipedia.org/wiki/Trinity%20%28supercomputer%29 | Trinity (or ATS-1) is a United States supercomputer built by the National Nuclear Security Administration (NNSA) for the Advanced Simulation and Computing Program (ASC). The aim of the ASC program is to simulate, test, and maintain the United States nuclear stockpile.
History
December 2013, The National Energy Research Scientific Computing Center (NERSC) and The Alliance for Computing at Extreme Scale (ACES) releases a joint RFP with technical requirements for Trinity.
July 2014, Cray announces that they were awarded the $174 Million contract by the National Nuclear Security Administration to provide a next generation supercomputer to Los Alamos National Laboratory.
June 2015, Haswell Partition installation begins.
November 2015, Trinity appears on the Supercomputing Top500 list at #6.
June 2016, Knights Landing Partition installation begins.
November 2016, Trinity falls to #10 on the Top500 list.
July 2017, The Haswell and KNL partitions are merged.
November 2018, Trinity regains #6 spot on the Top500 list.
Trinity technical specifications
Compute Tier
Trinity was built in 2 stages. The first stage incorporated the Intel Xeon Haswell processor while the second stage added a significant performance increase using the Intel Xeon Phi Knights Landing Processor.
There are 301,952 Haswell and 678,912 Knights Landing processors in the combined system, yielding a total peak performance of over 40 PF/s (petaflops)
Storage Tiers
There are 5 primary storage tiers; Memory, Burst Buffer, Parallel File System, Campaign Storage, and Archive.
Memory
2 PiB of DDR4 DRAM provide physical memory for the machine. Each processor also has DRAM built on to the tile, providing additional memory capacity. The data in this tier is highly transient and is typically in residence for only a few seconds, being overwritten continuously.
Burst Buffer
Cray supplies the three hundred XC40 Data Warp blades that each contain 2 Burst Buffer nodes and 4 SSD drives. There is a total |
https://en.wikipedia.org/wiki/RISC-V | RISC-V (pronounced "risk-five",) is an open standard instruction set architecture (ISA) based on established reduced instruction set computer (RISC) principles. Unlike most other ISA designs, RISC-V is provided under royalty-free open-source licenses. A number of companies are offering or have announced RISC-V hardware; open source operating systems with RISC-V support are available, and the instruction set is supported in several popular software toolchains.
As a RISC architecture, the RISC-V ISA is a load–store architecture. Its floating-point instructions use IEEE 754 floating-point. Notable features of the RISC-V ISA include: instruction bit field locations chosen to simplify the use of multiplexers in a CPU, a design that is architecturally neutral, and a fixed location for the sign bit of immediate values to speed up sign extension.
The instruction set is designed for a wide range of uses. The base instruction set has a fixed length of 32-bit naturally aligned instructions, and the ISA supports variable length extensions where each instruction can be any number of 16-bit parcels in length. Subsets support small embedded systems, personal computers, supercomputers with vector processors, and warehouse-scale 19 inch rack-mounted parallel computers.
The instruction set specification defines 32-bit and 64-bit address space variants. The specification includes a description of a 128-bit flat address space variant, as an extrapolation of 32 and 64 bit variants, but the 128-bit ISA remains "not frozen" intentionally, because there is yet so little practical experience with such large memory systems.
Unlike other academic designs which are typically optimized only for simplicity of exposition, the designers intended that the RISC-V instruction set be usable for practical computers. As of June 2019, version 2.2 of the user-space ISA and version 1.11 of the privileged ISA are frozen, permitting software and hardware development to proceed. The user-space ISA, now re |
https://en.wikipedia.org/wiki/Rational%20monoid | In mathematics, a rational monoid is a monoid, an algebraic structure, for which each element can be represented in a "normal form" that can be computed by a finite transducer: multiplication in such a monoid is "easy", in the sense that it can be described by a rational function.
Definition
Consider a monoid M. Consider a pair (A,L) where A is a finite subset of M that generates M as a monoid, and L is a language on A (that is, a subset of the set of all strings A∗). Let φ be the map from the free monoid A∗ to M given by evaluating a string as a product in M. We say that L is a rational cross-section if φ induces a bijection between L and M. We say that (A,L) is a rational structure for M if in addition the kernel of φ, viewed as a subset of the product monoid A∗×A∗ is a rational set.
A quasi-rational monoid is one for which L is a rational relation: a rational monoid is one for which there is also a rational function cross-section of L. Since L is a subset of a free monoid, Kleene's theorem holds and a rational function is just one that can be instantiated by a finite state transducer.
Examples
A finite monoid is rational.
A group is a rational monoid if and only if it is finite.
A finitely generated free monoid is rational.
The monoid M4 generated by the set {0,e, a,b, x,y} subject to relations in which e is the identity, 0 is an absorbing element, each of a and b commutes with each of x and y and ax = bx, ay = by = bby, xx = xy = yx = yy = 0 is rational but not automatic.
The Fibonacci monoid, the quotient of the free monoid on two generators {a,b}∗ by the congruence aab = bba.
Green's relations
The Green's relations for a rational monoid satisfy D = J.
Properties
Kleene's theorem holds for rational monoids: that is, a subset is a recognisable set if and only if it is a rational set.
A rational monoid is not necessarily automatic, and vice versa. However, a rational monoid is asynchronously automatic and hyperbolic.
A rational monoid is a regul |
https://en.wikipedia.org/wiki/Rosenbrock%20system%20matrix | In applied mathematics, the Rosenbrock system matrix or Rosenbrock's system matrix of a linear time-invariant system is a useful representation bridging state-space representation and transfer function matrix form. It was proposed in 1967 by Howard H. Rosenbrock.
Definition
Consider the dynamic system
The Rosenbrock system matrix is given by
In the original work by Rosenbrock, the constant matrix is allowed to be a polynomial in .
The transfer function between the input and output is given by
where is the column of and is the row of .
Based in this representation, Rosenbrock developed his version of the PBH test.
Short form
For computational purposes, a short form of the Rosenbrock system matrix is more appropriate and given by
The short form of the Rosenbrock system matrix has been widely used in H-infinity methods in control theory, where it is also referred to as packed form; see command pck in MATLAB. An interpretation of the Rosenbrock System Matrix as a Linear Fractional Transformation can be found in.
One of the first applications of the Rosenbrock form was the development of an efficient computational method for Kalman decomposition, which is based on the pivot element method. A variant of Rosenbrock’s method is implemented in the minreal command of Matlab and
GNU Octave.
References
1967 introductions
Control theory
Matrices |
https://en.wikipedia.org/wiki/Lode%20coordinates | Lode coordinates or Haigh–Westergaard coordinates . are a set of tensor invariants that span the space of real, symmetric, second-order, 3-dimensional tensors and are isomorphic with respect to principal stress space. This right-handed orthogonal coordinate system is named in honor of the German scientist Dr. Walter Lode because of his seminal paper written in 1926 describing the effect of the middle principal stress on metal plasticity. Other examples of sets of tensor invariants are the set of principal stresses or the set of kinematic invariants . The Lode coordinate system can be described as a cylindrical coordinate system within principal stress space with a coincident origin and the z-axis parallel to the vector .
Mechanics invariants
The Lode coordinates are most easily computed using the mechanics invariants. These invariants are a mixture of the invariants of the Cauchy stress tensor, , and the stress deviator, , and are given by
which can be written equivalently in Einstein notation
where is the Levi-Civita symbol (or permutation symbol) and the last two forms for are equivalent because is symmetric ().
The gradients of these invariants can be calculated by
where is the second-order identity tensor and is called the Hill tensor.
Axial coordinate
The -coordinate is found by calculating the magnitude of the orthogonal projection of the stress state onto the hydrostatic axis.
where
is the unit normal in the direction of the hydrostatic axis.
Radial coordinate
The -coordinate is found by calculating the magnitude of the stress deviator (the orthogonal projection of the stress state into the deviatoric plane).
where
{| class="toccolours collapsible collapsed" width="50%" style="text-align:left"
!Derivation
|-
|The relation that can be found by expanding the relation
and writing in terms of the isotropic and deviatoric parts while expanding the magnitude of
.
Because is isotropic and is deviatoric, their product is zero. Which leav |
https://en.wikipedia.org/wiki/Athanogene | This term, derived from the Greek for "against death" ('athánatos), was incorporated into name of the gene Bcl-2-associated athanogene 1 (BAG-1; alias HAP46/BAG-1M) upon discovery of its ability to confer transfected cells with resistance to apoptosis.
References
Genes
Apoptosis
Cloning |
https://en.wikipedia.org/wiki/SUPS | In computational neuroscience, SUPS (for Synaptic Updates Per Second) or formerly CUPS (Connections Updates Per Second) is a measure of a neuronal network performance, useful in fields of neuroscience, cognitive science, artificial intelligence, and computer science.
Computing
For a processor or computer designed to simulate a neural network SUPS is measured as the product of simulated neurons and average connectivity (synapses) per neuron per second:
Depending on the type of simulation it is usually equal to the total number of synapses simulated.
In an "asynchronous" dynamic simulation if a neuron spikes at Hz, the average rate of synaptic updates provoked by the activity of that neuron is . In a synchronous simulation with step the number of synaptic updates per second would be . As has to be chosen much smaller than the average interval between two successive afferent spikes, which implies , giving an average of synaptic updates equal to . Therefore, spike-driven synaptic dynamics leads to a linear scaling of computational complexity O(N) per neuron, compared with the O(N2) in the "synchronous" case.
Records
Developed in the 1980s Adaptive Solutions' CNAPS-1064 Digital Parallel Processor chip is a full neural network (NNW). It was designed as a coprocessor to a host and has 64 sub-processors arranged in a 1D array and operating in a SIMD mode. Each sub-processor can emulate one or more neurons and multiple chips can be grouped together. At 25 MHz it is capable of 1.28 GMAC.
After the presentation of the RN-100 (12 MHz) single neuron chip at Seattle 1991 Ricoh developed the multi-neuron chip RN-200. It had 16 neurons and 16 synapses per neuron. The chip has on-chip learning ability using a proprietary backdrop algorithm. It came in a 257-pin PGA encapsulation and drew 3.0 W at a maximum. It was capable of 3 GCPS (1 GCPS at 32 MHz).
In 1991-97, Siemens developed the MA-16 chip, SYNAPSE-1 and SYNAPSE-3 Neurocomputer. The MA-16 was a fast matrix-matrix mu |
https://en.wikipedia.org/wiki/Conservation-induced%20extinction | Efforts to save endangered species may, paradoxically, lead to conservation-induced extinction of other species. This mostly threatens the parasite and pathogen species that are highly host-specific to critically endangered hosts. When the last individuals of a host species are captured for the purpose of captive breeding and reintroduction programs, they typically undergo anti-parasitic treatments to increase survival and reproductive success. This practice may unintentionally result in the extinction of the species antagonistic to the target species, such as certain parasites. It has been proposed that the parasites should be reintroduced to the endangered population. A few cases of conservation-induced extinction have occurred in parasitic lice.
Examples
Parasite: Colpocephalum californici – host: California condorThe parasite most probably became extinct when the last individuals of its only host species were captured for a captive-breeding program; all parasites found were deliberately killed in an attempt to assist the host's survival.
Parasite: Felicola isidoroi – host: Iberian lynx.As with Colpocephalum californici, likely became extinct when the last individuals of its host species were taken into captivity and deloused to assist survival.
Parasite: Linognathus petasmatus – host: scimitar-horned oryxThe host specificity of this parasite is uncertain. Either it was specific to the scimitar-horned oryx and became extinct during captive breeding of the host, or – alternatively – it may be specific to the addax and possibly still surviving in the wild.
Parasite: Rallicola guami – host: Guam railThe only known host species of this parasite exists exclusively in captivity and kept under veterinary control. Little information about the fate of the parasite, likely extinct.
Parasite: Rallicola pilgrimi – host: little spotted kiwiThe parasite most probably became extinct when the last individuals of its only host species were captured and, after routine veter |
https://en.wikipedia.org/wiki/Soboleva%20modified%20hyperbolic%20tangent | The Soboleva modified hyperbolic tangent, also known as (parametric) Soboleva modified hyperbolic tangent activation function ([P]SMHTAF), is a special S-shaped function based on the hyperbolic tangent, given by
History
This function was originally proposed as "modified hyperbolic tangent" by Ukrainian scientist Elena V. Soboleva () as a utility function for multi-objective optimization and choice modelling in decision-making.
Practical usage
The function has since been introduced into neural network theory and practice.
It was also used in economics for modelling consumption and investment, to approximate current-voltage characteristics of field-effect transistors and light-emitting diodes, to design antenna feeders, and analyze plasma temperatures and densities in the divertor region of fusion reactors.
Sensitivity to parameters
Derivative of the function is defined by the formula:
The following conditions are keeping the function limited on y-axes: a ≤ c, b ≤ d.
A family of recurrence-generated parametric Soboleva modified hyperbolic tangent activation functions (NPSMHTAF, FPSMHTAF) was studied with parameters a = c and b = d. It is worth noting that in this case, the function is not sensitive to flipping the left and right-sides parameters:
The function is sensitive to ratio of the denominator coefficients and often is used without coefficients in the numerator:
With parameters a = b = c = d = 1 the modified hyperbolic tangent function reduces to the conventional tanh(x) function, whereas for a = b = 1 and c = d = 0, the term becomes equal to sinh(x).
See also
Activation function
e (mathematical constant)
Equal incircles theorem, based on sinh
Hausdorff distance
Inverse hyperbolic functions
List of integrals of hyperbolic functions
Poinsot's spirals
Sigmoid function
Notes
References
Further reading
(20 pages)
Elementary special functions
Exponentials
Hyperbolic functions
Analytic functions
Functions and mappings
Artificial neural netwo |
https://en.wikipedia.org/wiki/TCP%20Stealth | In computer networking, TCP Stealth is a proposed modification of the Transmission Control Protocol (TCP) to hide open ports of some TCP services from the public, in order to impede port scans. It is somewhat similar to the port knocking technique. it is an IETF Internet Draft specification.
The proposal modifies the TCP three-way handshake by only accepting connections from clients that transmit a proof of knowledge of a shared secret. If the connection attempt does not use TCP Stealth, or if authentication fails, the server acts as if no service was listening on the port number.
The project and initial Internet Draft specification was announced on 15 August 2014, following the revelations about the GCHQ project HACIENDA, which uses port scanning to find vulnerable systems for Five Eyes intelligence agencies. The draft was written by researchers from the Technische Universität München, Jacob Appelbaum of the Tor Project and Holger Kenn from Microsoft.
References
External links
Knock - patches for adding TCP Stealth to Linux and FreeBSD kernels and various applications
Master's thesis about TCP Stealth
TCP extensions
Computer network security |
https://en.wikipedia.org/wiki/Cerebral%20polyopia | Cerebral diplopia or polyopia describes seeing two or more images arranged in ordered rows, columns, or diagonals after fixation on a stimulus. The polyopic images occur monocular bilaterally (one eye open on both sides) and binocularly (both eyes open), differentiating it from ocular diplopia or polyopia. The number of duplicated images can range from one to hundreds. Some patients report difficulty in distinguishing the replicated images from the real images, while others report that the false images differ in size, intensity, or color. Cerebral polyopia is sometimes confused with palinopsia (visual trailing), in which multiple images appear while watching an object. However, in cerebral polyopia, the duplicated images are of a stationary object which are perceived even after the object is removed from the visual field. Movement of the original object causes all of the duplicated images to move, or the polyopic images disappear during motion. In palinoptic polyopia, movement causes each polyopic image to leave an image in its wake, creating hundreds of persistent images (entomopia).
Infarctions, tumors, multiple sclerosis, trauma, encephalitis, migraines, and seizures have been reported to cause cerebral polyopia. Cerebral polyopia has been reported in extrastriate visual cortex lesions, which is important for detecting motion, orientation, and direction. Cerebral polyopia often occurs in homonymous field deficits, suggesting deafferentation hyperexcitability could be a possible mechanism, similar to visual release hallucinations (Charles Bonnet syndrome).
Presentation
Cerebral polyopia is most often associated with occipital or temporal lobe lesions, as well as occipital lobe epilepsy. This condition is relatively uncommon, thus further research regarding its causes and mechanism has not been performed. Polyopia can be experienced as partial second or multiple images to either side (or in any eccentricity) of an object at fixation. Polyopia occurs when both ey |
https://en.wikipedia.org/wiki/LED%20strip%20light | An LED strip, tape, or ribbon light is a flexible circuit board populated by surface-mount light-emitting diodes (SMD LEDs) and other components that usually comes with an adhesive backing. Traditionally, strip lights had been used solely in accent lighting, backlighting, task lighting, and decorative lighting applications, such as cove lighting.
Increased luminous efficacy and higher-power SMDs have allowed LED strip lights to be used in applications such as high brightness task lighting, fluorescent and halogen lighting fixture replacements, indirect lighting applications, ultraviolet inspection during manufacturing processes, set and costume design, and growing plants.
Design
Variables in strip lighting consist of water resistance, color, adhesives, choice of SMD, driving voltage, control type, and whether it is constant current or constant voltage layout.
Uncoated LED tape is not considered to have any resistance to water ingress, but may be rated with an ingress protection code as IP20 for some physical ingress resistance. Such tapes are generally low voltage and safe for skin to touch but can be shorted by fine metal objects. Water resistant strip lighting is covered in a heat conducting epoxy or silicone to protect the circuitry from direct contact with water, and can be rated IP65, IP67, or with suitable sealed connections IP68. Both coated and uncoated LED tapes have a two sided adhesive backing to stick to walls, desks, doors, etc.
The most common design differences are in how individual LEDs are controlled, specifically differences in color and whether or not each LED is addressable.
Single color, non-addressable: Every LED on the strand is a single white colour, typically ranging from 2200K to 6500K in color temperature, or any of several monochrome colors covering the range of the visible spectrum (generally from 400-700 nanometers in wavelength).
Dynamic tunable white (often described with CCT), non-addressable: Allows the user to adjust the colo |
https://en.wikipedia.org/wiki/Massachusetts%20Division%20of%20Ecological%20Restoration | The Division of Ecological Restoration (DER) is a Division of the Massachusetts Department of Fish and Game within the Executive Office of Energy and Environmental Affairs. DER was created in 2009 with the merger of the Riverways and Wetlands Restoration Programs (formally within the Massachusetts Office of Coastal Zone Management). DER coordinates ecological restoration to improve ecological condition and to restore important ecosystem services that improve the quality of life for all Massachusetts citizens.
The Riverways Program (MGL Chapter 21A Section 8) has been maintained within DER and coordinates outreach and technical assistance to support watershed conservation and protection.
The Division and partners facilitate capital-based projects including (but not limited to) dam removal and culvert replacement with the goal of restoring aquatic habitats (e.g. salt marshes)
and ecosystems across the state. These projects support commercial and recreational fisheries and provide many other benefits such as reduced flooding, improved water quality, carbon sequestration and increased public safety.
Ecological restoration is a core component of the Commonwealth of Massachusetts efforts to build habitat resiliency to better allow fish and wildlife to adapt to climate change – including sea level rise, elevated water temperatures, and increased floods and periods of drought.
Beth Lambert currently serves as the Director of the Division of Ecological Restoration. The founding director was Tim Purinton, who served in this role from 2009-2017.
Philosophy
DER espouses a comprehensive, holistic approach to restoration, that prioritizes ecosystems over habitats.
DER places an emphasis on restoring ecological processes, rather than species specific habitats, focusing on implementation of restoration actions at a watershed scale.
DER staff identifies the most detrimental stressors in the larger watershed context and coordinates restoration actions to improve conditions in pa |
https://en.wikipedia.org/wiki/Something%20%28concept%29 | Something and anything are concepts of existence in ontology, contrasting with the concept of nothing. Both are used to describe the understanding that what exists is not nothing without needing to address the existence of everything. The philosopher, David Lewis, has pointed out that these are necessarily vague terms, asserting that "ontological assertions of common sense are correct if the quantifiers—such words as "something" and "anything"—are restricted roughly to ordinary or familiar things."
The idea that "something" is the opposite of "nothing" has existed at least since it was proposed by the Neoplatonist philosopher Porphyry in the 3rd century. One of the most basic questions of both science and philosophy is: why is there something rather than nothing at all? A question that follows from this is whether it is ever actually possible for there to be nothing at all, or whether there must always be something.
Grammatically, "something and anything are commonly classified as pronouns, although they do not stand for another noun so clearly as does thing itself, a word always classified as a noun".
In predicate logic
In predicate logic, what is described in layman's terms as "something" can more specifically be regarded as existential quantification, that is, the predication of a property or relation to at least one member of the domain. It is a type of quantifier, a logical constant which is interpreted as "there exists," "there is at least one," or "for some." It expresses that a propositional function can be satisfied by at least one member of a domain of discourse. In other terms, it is the predication of a property or relation to at least one member of the domain. It asserts that a predicate within the scope of an existential quantifier is true of at least one value of a predicate variable.
Nothing and something
Though considered to exclude one another, the concepts of something and nothing can coincide under special circumstances. One example would |
https://en.wikipedia.org/wiki/Electro-switchable%20biosurface | An electro-switchable biosurface is a biosensor that is based on an electrode (often gold) to which a layer of biomolecules (often DNA molecules) has been tethered. An alternating or fixed electrical potential is applied to the electrode which causes changes in the structure and position (movement) of the charged biomolecules. The biosensor is used in science, e.g. biomedical and biophysical research or drug discovery, to assess interactions between biomolecules and binding kinetics as well as changes in size or conformation of biomolecules.
Technological background
The general principle of a biosurface is a solid surface with an additional layer of biological macromolecules. Because this molecular layer will reversibly respond to changes in the environment of the surface, it is also called “stimuli-responsive monolayer”. The external stimuli can be for example changes in temperature, changes in magnetic fields, mechanical forces or changes in electric fields.
Different strategies can be used to attach a monolayer of biomolecules to a surface, for example atomic layer deposition or layer-by-layer deposition. Another option is the fabrication of self-assembled monolayers (SAM). The surface used most often with this strategy is a gold electrode. SAM form by spontaneous organization of the molecules, for example alkanethiolates, on the substrate. SAM can be used as surface layers for nanoparticles, e.g. in MRI contrast agents, they can protect metal films from corrosion, and have many other applications in electrochemistry and nanoscience. For their application as a biosensor, one of the most often used molecules self-assembling on gold electrodes is DNA. Due to its molecular structure, double stranded DNA molecules are negatively charged and rigid. By applying an alternating potential to the biosurface, the attached DNA strands can be moved systematically because they will switch between an upright position and a flat position. This enables the usage of the biosur |
https://en.wikipedia.org/wiki/Normal%20form%20%28dynamical%20systems%29 | In mathematics, the normal form of a dynamical system is a simplified form that can be useful in determining the system's behavior.
Normal forms are often used for determining local bifurcations in a system. All systems exhibiting a certain type of bifurcation are locally (around the equilibrium) topologically equivalent to the normal form of the bifurcation. For example, the normal form of a saddle-node bifurcation is
where is the bifurcation parameter. The transcritical bifurcation
near can be converted to the normal form
with the transformation .
See also canonical form for use of the terms canonical form, normal form, or standard form more generally in mathematics.
References
Further reading
Bifurcation theory
Dynamical systems |
https://en.wikipedia.org/wiki/Domain%20Based%20Security | "Domain Based Security", abbreviated to "DBSy", is a model-based approach to help analyze information security risks in a business context and provide a clear and direct mapping between the risks and the security controls needed to manage them. A variant of the approach is used by the UK government's HMG Infosec Standard No.1 technical risk-assessment method. DBSy is a registered trade mark of QinetiQ Ltd.
DBSy was developed in the late 1990s by the Defence Evaluation and Research Agency (DERA). It is a model-based approach to information assurance that describes the requirements for security in an organisation, taking account of the business that needs to be supported. The model is based around the concept of a security domain, which represents a logical place where people work with information using a computer system, and which has connections with other security domains where this is necessary to support business activity. Hence the focus is on the information that needs protection, the people that work with it and the people they exchange information with. The model can also describe the physical environments where people work and the system boundaries where major system security measures are placed. A systematic method is then applied to the model to identify and describe the risks to which valuable information assets are exposed and specify security measures that are effective in managing the risks.
History
DBSy has its origins in the late 1990s, having been developed by the Defence Evaluation and Research Agency (DERA) for the Ministry of Defence (MOD). Initially called the Domain Based Approach, it was developed alongside Purple Penelope to support the MOD's increasing need for interconnections between systems operating at different security levels,
It was recognised that the risks associated with such connections were directly related to the nature of the information exchange that was needed, and that an effective model for understanding and managing the |
https://en.wikipedia.org/wiki/Cyclopentadecanolide | Cyclopentadecanolide is a natural macrolide lactone and a synthetic musk.
Natural occurrence
Cyclopentadecanolide occurs in small quantities in angelica root essential oil and is responsible for its musklike odor.
Production
Cyclopentadecanolide is produced synthetically by ring expansion of cyclotetradecanone. Another synthesis route is the depolymerization of polyesters of 15-hydroxypentadecanoic acid.
Uses
Cyclopentadecanolide is used as a musklike perfume fixative in fine fragrances and as a flavoring agent. It is a substitute for the extremely expensive animal musk.
References
Flavors
Lactones
Perfume ingredients |
https://en.wikipedia.org/wiki/Cartesian%20monoidal%20category | In mathematics, specifically in the field known as category theory, a monoidal category where the monoidal ("tensor") product is the categorical product is called a cartesian monoidal category. Any category with finite products (a "finite product category") can be thought of as a cartesian monoidal category. In any cartesian monoidal category, the terminal object is the monoidal unit. Dually, a monoidal finite coproduct category with the monoidal structure given by the coproduct and unit the initial object is called a cocartesian monoidal category, and any finite coproduct category can be thought of as a cocartesian monoidal category.
Cartesian categories with an internal Hom functor that is an adjoint functor to the product are called Cartesian closed categories.
Properties
Cartesian monoidal categories have a number of special and important properties, such as the existence of diagonal maps Δx : x → x ⊗ x and augmentations ex : x → I for any object x. In applications to computer science we can think of Δ as "duplicating data" and e as "deleting data". These maps make any object into a comonoid. In fact, any object in a cartesian monoidal category becomes a comonoid in a unique way.
Examples
Cartesian monoidal categories:
Set, the category of sets with the singleton set serving as the unit.
Cat, the bicategory of small categories with the product category, where the category with one object and only its identity map is the unit.
Cocartesian monoidal categories:
Vect, the category of vector spaces over a given field, can be made cocartesian monoidal with the monoidal product given by the direct sum of vector spaces and the trivial vector space as unit.
Ab, the category of abelian groups, with the direct sum of abelian groups as monoidal product and the trivial group as unit.
More generally, the category R-Mod of (left) modules over a ring R (commutative or not) becomes a cocartesian monoidal category with the direct sum of modules as tensor product and th |
https://en.wikipedia.org/wiki/I.%20I.%20Rabi%20Award | The I. I. Rabi Award, founded in 1983, is awarded annually by IEEE.
"The Rabi Award is to recognize outstanding contributions related to the fields of atomic and molecular frequency standards, and time transfer and dissemination."
The award is named after Isidor Isaac Rabi, Nobel Prize winner in 1944. He was the first recipient of the award, for his experimental and theoretical work on atomic beam resonance spectroscopy.
Recipients
1983 - I. I. Rabi
1984 - David W. Allan
1985 - Norman Ramsey, Nobel Prize in 1989
1986 - Jerrold R. Zacharias
1987 - Louis Essen
1988 - Gernot M. R. Winkler
1989 - Leonard S. Cutler
1990 - Claude Audoin
1991 - Andrea De Marchi
1992 - James A. Barnes
1993 - Robert F. C. Vessot
1994 - Jacques Vanier
1995 - Fred L. Walls
1996 - Andre Clairon and Robert E. Drullinger
1997 - Harry E. Peters and Nikolai A. Demidov
1998 - David J. Wineland, Nobel Prize in 2012
1999 - Bernard Guinot
2000 - William J. Riley Jr.
2001 - Lute Maleki
2002 - Jon H. Shirley
2003 - Andreas Bauch
2005 - Theodor W. Hänsch, Nobel Prize in 2005
2004 - John L. Hall, Nobel Prize in 2005
2006 - James C. Bergquist
2007 - Patrick Gill and Leo Hollberg
2008 - Hidetoshi Katori
2009 - John D. Prestage
2010 - Long Sheng Ma
2011 - Fritz Riehle
2012 - James Camparo
2013 - Judah Levine
2014 - Harald R. Telle
2015 - Ulrich_L._Rohde
2016 - John Kitching
2017 - Scott Diddams
2018 - Jun Ye
2019 - Steven Jefferts
2020 - Robert Lutwak
2021 - Ekkehard Peik
See also
List of physics awards
References
Physics awards
Awards established in 1983
Atomic physics
Atomic, molecular, and optical physics
IEEE awards |
https://en.wikipedia.org/wiki/Node%20influence%20metric | In graph theory and network analysis, node influence metrics are measures that rank or quantify the influence of every node (also called vertex) within a graph. They are related to centrality indices. Applications include measuring the influence of each person in a social network, understanding the role of infrastructure nodes in transportation networks, the Internet, or urban networks, and the participation of a given node in disease dynamics.
Origin and development
The traditional approach to understanding node importance is via centrality indicators. Centrality indices are designed to produce a ranking which accurately identifies the most influential nodes. Since the mid 2000s, however, social scientists and network physicists have begun to question the suitability of centrality indices for understanding node influence. Centralities may indicate the most influential nodes, but they are rather less informative for the vast majority of nodes which are not highly influential.
Borgatti and Everett's 2006 review article
showed that the accuracy of centrality indices is highly dependent on network topology.
This finding has been repeatedly observed since then. (e.g.).
In 2012, Bauer and colleagues reminded us that centrality indices only rank nodes but do not quantify the difference between them.
In 2013, Sikic and colleagues presented strong evidence that centrality indices considerably underestimate the power of non-hub nodes.
The reason is quite clear. The accuracy of a centrality measure depends on network topology, but complex networks have heterogeneous topology. Hence a centrality measure which is appropriate for identifying highly influential nodes will most likely be inappropriate for the remainder of the network.
This has inspired the development of novel methods designed to measure the influence of all network nodes. The most general of these are
the accessibility, which uses the diversity of random walks to measure how accessible the rest of the net |
https://en.wikipedia.org/wiki/The%20Karlskrona%20Manifesto | The Karlskrona Manifesto for sustainability design in software was created as an output of the Third International Workshop on Requirements Engineering for Sustainable Systems (RE4SuSy) held in Karlskrona, Sweden, co-located with the 22nd IEEE International Requirements Engineering Conference (RE'14). The manifesto arose from a suggestion in the paper by Christoph Becker, "Sustainability and Longevity: Two Sides of the Same Quality?" that sustainability is a common ground for several disciplines related to software, but that this commonality had not been mapped out and made explicit and that a focal point of reference would be beneficial.
External links
Third International Workshop on Requirements Engineering for Sustainable Systems (RE4SuSy)
22nd IEEE International Requirements Engineering Conference (RE'14)
Proceedings of the Third International Workshop on Requirements Engineering for Sustainable Systems
Christoph Becker, Sustainability and Longevity: Two Sides of the Same Quality?
Christoph Becker, Ruzanna Chitchyan, Leticia Duboc, Steve Easterbrook, Martin Mahaux, Birgit Penzenstadler, Guillermo Rodriguez-Navas, Camille Salinesi, Norbert Seyff, Colin Venters, Coral Calero, Sedef Akinli Kocak, Stefanie Betz, The Karlskrona manifesto for sustainability design
Karlskrona Manifesto for Sustainability Design Official site
References
Sustainability
Software design |
https://en.wikipedia.org/wiki/Distributed%20Processing%20Technology | Distributed Processing Technology (DPT) was founded in 1977, in Maitland, Florida. DPT was an early pioneer in computer storage technology, popularizing the use of disk caching in the 1980s and 1990s. DPT was the first company to design, manufacture and sell microprocessor-based intelligent caching disk controllers to the OEM computer market. Prior to DPT, disk caching technology had been implemented in proprietary hardware in mainframe computing to improve the speed of disk access.
DPT's products popularized the use of disk caching in the 1980s. According to Bill Brothers, Unix product manager at the Santa Cruz Operation (SCO), a computer operating system vendor, "The kind of performance those guys (DPT) produce is phenomenal. It's unlike any other product on the market."
DPT was founded by Steve Goldman, who served as the President and Chief Executive Officer until DPT was acquired by Adaptec in November 1999.
External links
Floppy controller speeds access with cache
Caching Disk Controller Relieves System Bottlenecks
Disk Controller Unburdens Real Time Applications
References
Computer companies established in 1977
Computer companies disestablished in 1999
Defunct computer companies of the United States
Defunct computer hardware companies
1977 establishments in Florida
1999 disestablishments in Florida |
https://en.wikipedia.org/wiki/Browser-based%20computing | Browser-based computing is the use of the web browsers to perform computing tasks. Opportunities for computing on the Web have been noted as far back as 1997. Computing over the web was described in 2000. Applications include distributed computing for web workers as illustrated by James (formerly CrowdProcess) and HASH, the use of the browser's stack in QMachine, the embedding of web applications as semantic hypermedia components and the Signaling Server in Peer-to-peer networks set via WebRTC. Browser-based computing complements cloud computing, because they reduce server-side computational load, often using cloud-hosted, RESTful web services.
References
Distributed computing architecture |
https://en.wikipedia.org/wiki/North%20Pacific%20Marine%20Science%20Organization | The North Pacific Marine Science Organization, also called PICES (referring to the organization's status as a Pacific version of the International Council for the Exploration of the Sea), is an intergovernmental organization that promotes and coordinates marine scientific research in the North Pacific Ocean and provides a mechanism for information and data exchange among scientists in its member countries.
Legal framework
PICES is an international intergovernmental organization established under a Convention for a North Pacific Marine Science Organization. The Convention entered into force on 1992-03-24 with an initial membership that included the governments of Canada, Japan, and the United States of America. The Convention was ratified by the People's Republic of China on 1992-08-31 to increase membership to four countries. Although the Soviet Union had participated in the development of the Convention, it was not ratified there until 1994-12-16, by the Russian Federation. The Republic of Korea acceded to the Convention on 1995-7-30. The Republic of Mexico and the Democratic People's Republic of Korea are located within the Convention Area (generally north of 30°N) but are not members.
Oversight
A Governing Council consisting of up to two delegates appointed by each member country is the primary decision-making body. Day-to-day operations of the organization are managed by the staff of the PICES Secretariat, located in Canada at the Institute of Ocean Sciences (located on Patricia Bay in British Columbia).
History
The idea to create a Pacific version of ICES (International Council for the Exploration of the Sea) was first discussed by scientists from Canada, Japan, the Soviet Union, and the United States who were attending a conference, sponsored by the United Nations Food and Agriculture Organization, in Vancouver, British Columbia, Canada in February 1973. ICES had provided a forum since 1902 for scientists bordering the Atlantic Ocean and its marginal seas |
https://en.wikipedia.org/wiki/Kaye%20and%20Laby | Tables of Physical and Chemical Constants and Some Mathematical Functions is a scientific reference work. First compiled and published in 1911 by the physicists G. W. C. Kaye and T. H. Laby, it is more commonly known as Kaye and Laby. It is a standard textbook for scientists and engineers.
The final print edition was the 16th in 1995, after which the entire content was made available online in association with the National Physical Laboratory.
The online version was removed on 21 May 2019, the day after the redefinition of the SI base units. An archived version is still available online, but is no longer maintained, and does not have the updated values of physical constants.
References
External links
Kaye and Laby online
About Kaye and Laby
History of Kaye and Laby
1911 non-fiction books
Chemistry reference works
Mathematics textbooks
2019 disestablishments in the United Kingdom
Physics textbooks |
https://en.wikipedia.org/wiki/LIVAC%20Synchronous%20Corpus | LIVAC is an uncommon language corpus dynamically maintained since 1995. Different from other existing corpora, LIVAC has adopted a rigorous and regular as well as "Windows" approach in processing and filtering massive media texts from representative Chinese speech communities such as Hong Kong, Macau, Taipei, Singapore, Shanghai, Beijing, as well as Guangzhou, and Shenzhen. The contents are thus deliberately repetitive in most cases, represented by textual samples drawn from editorials, local and international news, cross-Formosan Straits news, as well as news on finance, sports and entertainment. By 2020, 3 billion characters of news media texts have been filtered so far, of which 700 million characters have been processed and analyzed and have yielded an expanding Pan-Chinese dictionary of 2.5 million words from the Pan-Chinese printed media. Through rigorous analysis based on computational linguistic methodology, LIVAC has at the same time accumulated a large amount of accurate and meaningful statistical data on the Chinese language and their speech communities in the Pan-Chinese region, and the results show considerable and important variations.
The "Windows" approach is the most representative feature of LIVAC and has enabled Pan-Chinese media texts to be quantitatively analyzed according to various attributes such as locations, time and subject domains. Thus, various types of comparative studies and applications in information technology as well as development of often related innovative applications have been possible. Moreover, LIVAC has allowed longitudinal developments to be taken into account, facilitating Key Word in Context (KWIC) and comprehensive study of target words and their underlying concepts as well as linguistic structures over the past 20 years, based on variables such as region, duration and content. Results from the extensive and accumulative data analysis contained in LIVAC have enabled the cultivation of textual databases of proper names, |
https://en.wikipedia.org/wiki/NamePros | NamePros is an online community for domain name investors. Its services include forums and domain name auctions. The forums implement a freemium business model, whereby membership is free, but paid subscriptions offer additional features.
History
Ron "RJ" James publicly launched NamePros in February, 2003. He created NamePros to fill the hole left by Afternic, a popular online community that was quickly failing. NamePros started to see success around June, 2003, four months after its launch. Ron James dismissed the idea of adopting a subscription business model, favoring free services.
Acquisition by Bodis
Bodis, a domain parking company, acquired NamePros in January, 2012. Speculation circulating around blogs and other communities point to a sale price in the range of $200,000 to $300,000 USD. Rumors of the sale began as early as January 11. By January 19, Matt Wegrzyn, owner of Bodis, had publicly confirmed the acquisition. Matt Wegrzyn hinted that improvements to the website and its services would follow and stated that Bodis continue to keep NamePros an open community. Former owner Ron James noted that Bodis had better resources and would be capable of supporting NamePros' continued growth. In years leading up to the acquisition, the domain investor community noted some issues with website reliability.
On May 14, 2013, NamePros released a series of updates to its service, including a new layout and the addition of a domain sales history tool. The new tool, developed by a community member, scraped domain sale information from other websites and RSS feeds. The author claimed to be indexing over 80,000 sales that members could query.
Third administration
NamePros changed ownership again in July, 2013.
In May, 2014, NamePros released another series of updates. They migrated from vBulletin to XenForo, resulting in significant changes. DNForum, a competing website, made a similar update a month prior.
Services
NamePros' services center aro |
https://en.wikipedia.org/wiki/Northern%20celestial%20hemisphere | The northern celestial hemisphere, also called the Northern Sky, is the northern half of the celestial sphere; that is, it lies north of the celestial equator. This arbitrary sphere appears to rotate westward around a polar axis due to Earth's rotation.
At any given time, the entire Northern Sky is visible from the geographic North Pole, while less of the hemisphere is visible the further south the observer is located. The southern counterpart is the southern celestial hemisphere.
Astronomy
In the context of astronomical discussions or writing about celestial cartography, this celestial hemisphere may also simply then be referred to as the Northern Hemisphere.
For celestial mapping, astronomers may conceive the sky like the inside of a sphere divided into two halves by the celestial equator. The Northern Sky or Northern Hemisphere is therefore the half of the celestial sphere that is north of the celestial equator.
Even if this geocentric model is the ideal projection of the terrestrial equator onto the imaginary celestial sphere, the northern and southern celestial hemispheres are not to be confused with descriptions of the terrestrial hemispheres of Earth itself.
See also
Southern celestial hemisphere
Astronomical coordinate systems
Hemispheres |
https://en.wikipedia.org/wiki/Mechanical%20load | Mechanical load is the physical stress on a mechanical system or component. Loads can be static or dynamic. Some loads are specified as part of the design criteria of a mechanical system. Depending on the usage, some mechanical loads can be measured by an appropriate test method in a laboratory or in the field.
Vehicle
It can be the external mechanical resistance against which a machine (such as a motor or engine), acts. The load can often be expressed as a curve of force versus speed.
For instance, a given car traveling on a road of a given slope presents a load which the engine must act against. Because air resistance increases with speed, the motor must put out more torque at a higher speed in order to maintain the speed. By shifting to a higher gear, one may be able to meet the requirement with a higher torque and a lower engine speed, whereas shifting to a lower gear has the opposite effect. Accelerating increases the load, whereas decelerating decreases the load.
Pump
Similarly, the load on a pump depends on the head against which the pump is pumping, and on the size of the pump.
Fan
Similar considerations apply to a fan. See Affinity laws.
See also
Structural load
Physical test
References
Mechanical engineering
Physical quantities
Structural analysis |
https://en.wikipedia.org/wiki/ISO/IEC%2029119 | ISO/IEC/IEEE 29119 Software and systems engineering -- Software testing is a series of five international standards for software testing. First developed in 2007 and released in 2013, the standard "defines vocabulary, processes, documentation, techniques, and a process assessment model for testing that can be used within any software development lifecycle."
History and revisions
Development of the set of ISO/IEC/IEEE 29119 software testing standards began in May 2007, based on existing standards such as the Institute of Electrical and Electronics Engineers's IEEE 829 (test documentation), and IEEE 1008 (unit testing); and the BSI Group's BS 7925-1 (vocabulary) and -2 (software components).
At first the International Organization for Standardization (ISO) had no working group with significant software testing experience, so the ISO created WG26, which by 2011 was represented by more than 20 different countries. Initially four sections were developed for the standard: Concepts and definitions (1), Test processes (2), Test documentation (3), and Test techniques (4). A fifth part concerning process assessment was considered for addition, ultimately becoming ISO/IEC 33063:2015, which ties to 29119-2's test processes. The actual fifth part of 29119 was published in November 2016 concerning the concept of keyword-driven testing.
, no major revisions have occurred to the five parts of the standard. These parts are, from most recent to oldest:
ISO/IEC/IEEE 29119-5:2016, Part 5: Keyword-driven testing, published in November 2016
ISO/IEC/IEEE 29119-4:2015, Part 4: Test techniques, published in December 2015
ISO/IEC/IEEE 29119-3:2013, Part 3: Test documentation, published in September 2013
ISO/IEC/IEEE 29119-2:2013, Part 2: Test processes, published in September 2013
ISO/IEC/IEEE 29119-1:2013, Part 1: Concepts and definitions, published in September 2013
Structure and contents
ISO/IEC/IEEE 29119-1:2013, Part 1: Concepts and definitions
ISO/IEC/IEEE 29119 Part 1 fa |
https://en.wikipedia.org/wiki/Broadcast%20Pioneers%20of%20Philadelphia | The Broadcast Pioneers of Philadelphia is a state-chartered, federally recognized, 501(c)(3) non-profit organization, devoted to collecting information about and promoting cultural work related to broadcasting and communications in Philadelphia, Pennsylvania, US, and its metropolitan area, the Delaware Valley. The group was founded in 1962 as a local chapter of the national Broadcast Pioneers organization, but became an independent organization in 1995.
The organization is made up of over 500 members of the broadcast community. Full members must have 10 years or more experience in the business while associate members need less.
Broadcast Pioneers of Philadelphia is more often simply referred to as just Broadcast Pioneers. They run college student career nights on local college campuses and annual symposiums, also for college students. The symposia rotate between three television stations in the Philadelphia market that originate local newscasts. In a three-year time period, local college students can visit all three. They also award college scholarships. In April 2014, they awarded 20 scholarships to local students.
Broadcast Pioneers also maintains several local broadcasting archives ranging from video to audio to photographs and other historical documents. Their website was awarded "Honorable Mention for Best Info Site" by Time magazine.
In addition to collecting information and archives, they professionally record interviews with local broadcast legends. These productions are available on YouTube for public access.
References
External links
+
Broadcast engineering
Delaware Valley
501(c)(3) organizations
Non-profit organizations based in Philadelphia
Organizations established in 1962
1962 establishments in Pennsylvania |
https://en.wikipedia.org/wiki/Laptop%E2%80%93tablet%20convergence | Laptop–tablet convergence describes the tendency in recent years for laptops and tablet computers to converge technologically.
In 2011, Canonical announced Ubuntu Touch, an attempt to bring Ubuntu to mobile devices such as phones and tablets. It is still under development.
In 2012, Windows 8 was released, which included the Metro UI, a touch-style UI framework for desktop applications. Microsoft encouraged application developers to develop Metro versions of their applications.
In 2014, Google announced that by late 2014, ChromeOS would allow Android apps to be run, meaning that Chromebooks and Chromeboxes would be able to access ordinary web applications, packaged ChromeOS apps, and also Android apps.
See also
Laplet
Convertible tablet
References
Mobile computers
Laptops
Tablet computers |
https://en.wikipedia.org/wiki/The%20Drunkard%27s%20Walk | The Drunkard's Walk: How Randomness Rules Our Lives is a 2008 popular science book by American physicist and author Leonard Mlodinow, which became a New York Times bestseller and a New York Times notable book.
Overview
The Drunkard's Walk discusses the role of randomness in everyday events, and the cognitive biases that lead people to misinterpret random events and stochastic processes. The title refers to a certain type of random walk, a mathematical process in which one or more variables change value under a series of random steps. Mlodinow discusses the contributions of mathematical heavyweights Jacob Bernoulli, Pierre-Simon Laplace, and Blaise Pascal, and introduces basic statistical concepts such as regression toward the mean and the law of large numbers, while discussing the role of probability in examples from wine ratings and school grades to political polls.
Reception
In 2008 the Committee for Skeptical Inquiry (CSICOP) awarded Mlodinow the Robert P. Balles Prize for Critical Thinking for the book.
References
Probability books
Popular science books
Pantheon Books books |
https://en.wikipedia.org/wiki/Keeper%20%28password%20manager%29 | Keeper Security, Inc. (Keeper) is a provider of zero-knowledge security and encryption software covering password management, secrets management, connection management, privileged access management, dark web monitoring, digital file storage, and encrypted messaging, among other offerings.
Keeper Password Manager
Keeper password manager uses a freemium pricing model for individual consumers and a subscription-based model for households and businesses. The free individual version of Keeper provides storage for passwords, identity data, and financial information, and includes a password generator and two-factor authentication (2FA) on a single mobile device. The subscription-based model for individual consumers offers additional features such as unlimited password, identity data, and financial data storage on an unlimited number of devices, cross-device syncing, and record-sharing capabilities.
Keeper is available as a mobile app for Android and iOS, as well as a desktop application for Windows, Linux, and MacOS. It offers a desktop browser extension for Safari, Chrome, Firefox, Edge, Internet Explorer, Opera, and Brave.
Users secure their Keeper vaults with a “master password.” Users can further protect their vaults via a variety of multi-factor authentication methods, including Google Authenticator, Duo Security, FIDO U2F, and biometrics.
Customer vaults are secured using an AES-256 key, which is derived from the user’s master password using PBKDF2 with 1,000,000 iterations by default. Only encrypted ciphertext is stored on Keeper’s servers, and the company has no way of decrypting the data its customers store in their digital vaults, nor can it retrieve their master passwords.
Keeper users can directly share passwords, files, and other information “vault to vault” with other Keeper users and through One-Time Share for non-Keeper users; all shared content is secured with PKI encryption.
Keeper Security Government Cloud
Keeper Security is listed as Authorize |
https://en.wikipedia.org/wiki/SandCraft | SandCraft, Inc., was a Silicon Valley based fabless semiconductor design company designer of microprocessors that were used as computing engines in electronics products utilizing the MIPS architecture and a series of RISC CPU chips. The markets targeted were consumer electronics, office automation, and communications applications, including Nintendo game consoles.
On 29 July 2003, SandCraft, Inc. was acquired by Raza Microelectronics Inc, which, in turn, was acquired by NetLogic Microsystems in 2008, which itself was acquired by Broadcom in 2012.
Name
SandCraft was named as such because sand represented silicon, the critical substrate in microprocessors, and craft to denote a design house.
History
Norman Yeung founded SandCraft, along with Mayank Gupta as chief architect in 1996. Microprocessors with Floating point units were designed verified and contracted out to Asian manufacturing based on MIPS IV, with 64 bit instruction sets, and a later processor called SR71000, which was at the time the world's highest performance MIPS processor. They incorporated superscalar, multi-staged pipeline design, and big/little endian support modes.
References
https://old.hotchips.org/wp-content/uploads/hc_archives/hc10/2_Mon/HC10.S2/HC10.2.2.pdf
1996 establishments in California
2003 disestablishments in California
American companies established in 1996
American companies disestablished in 2003
Computer companies established in 1996
Computer companies disestablished in 2003
Defunct companies based in the San Francisco Bay Area
Defunct computer companies of the United States
Defunct computer hardware companies
Defunct technology companies based in California
MIPS architecture |
https://en.wikipedia.org/wiki/Pseudocomplement | In mathematics, particularly in order theory, a pseudocomplement is one generalization of the notion of complement. In a lattice L with bottom element 0, an element x ∈ L is said to have a pseudocomplement if there exists a greatest element x* ∈ L with the property that x ∧ x* = 0. More formally, x* = max{ y ∈ L | x ∧ y = 0 }. The lattice L itself is called a pseudocomplemented lattice if every element of L is pseudocomplemented. Every pseudocomplemented lattice is necessarily bounded, i.e. it has a 1 as well. Since the pseudocomplement is unique by definition (if it exists), a pseudocomplemented lattice can be endowed with a unary operation * mapping every element to its pseudocomplement; this structure is sometimes called a p-algebra. However this latter term may have other meanings in other areas of mathematics.
Properties
In a p-algebra L, for all
The map x ↦ x* is antitone. In particular, 0* = 1 and 1* = 0.
The map x ↦ x** is a closure.
x* = x***.
(x∨y)* = x* ∧ y*.
(x∧y)** = x** ∧ y**.
The set S(L) ≝ { x** | x ∈ L } is called the skeleton of L. S(L) is a ∧-subsemilattice of L and together with x ∪ y = (x∨y)** = (x* ∧ y*)* forms a Boolean algebra (the complement in this algebra is *). In general, S(L) is not a sublattice of L. In a distributive p-algebra, S(L) is the set of complemented elements of L.
Every element x with the property x* = 0 (or equivalently, x** = 1) is called dense. Every element of the form x ∨ x* is dense. D(L), the set of all the dense elements in L is a filter of L. A distributive p-algebra is Boolean if and only if D(L) = {1}.
Pseudocomplemented lattices form a variety; indeed, so do pseudocomplemented semilattices.
Examples
Every finite distributive lattice is pseudocomplemented.
Every Stone algebra is pseudocomplemented. In fact, a Stone algebra can be defined as a pseudocomplemented distributive lattice L in which any of the following equivalent statements hold for all
S(L) is a sublattice of L;
(x∧y)* = x* ∨ y*;
(x |
https://en.wikipedia.org/wiki/Gigamacro | A gigapixel macro image is a digital image bitmap composed of one billion (109) pixels (picture elements), or 1000 times the information captured by a 1 megapixel digital camera. Creating such high-resolution images involves making mosaics (image stitching) of a large number of high-resolution digital photographs which are then combined into a single image.
Gigapixel macro images are made by 'stacking' a number of photographs together in order to increase the depth of field and then stitching the resulting images together in a technique known as 'stack and stitch'. Such images are usually very large in size and cannot be easily viewed. To make such images accessible, they are converted using tiled image techniques so that they may be viewed in a web browser. Such techniques are familiar in everyday use in e.g. Google Maps.
References
Gigamacro Image Gallery
Geological gigamacro Examples
Image processing |
https://en.wikipedia.org/wiki/Linux%20Lite | Linux Lite is a Linux distribution based on Debian and Ubuntu created by a team of programmers led by Jerry Bezencon. Created in 2012, it uses a customized implementation of Xfce as its desktop environment, and runs on the main Linux kernel.
The distribution aims to appeal to Linux beginners and Windows users, by trying to make the transition from Windows to Linux as smooth as possible. To achieve this, the distribution tries to conserve many of the visual and functional elements of Windows, to create an experience that can be perceived as familiar by Windows users. Additionally, the distro sets out to "dispel the myth that Linux is hard to use", by trying to offer a simple and intuitive desktop experience.
See also
Debian
Lubuntu
Ubuntu
References
External links
2012 software
Operating system distributions bootable from read-only media
Ubuntu derivatives
X86-64 Linux distributions
Linux distributions |
https://en.wikipedia.org/wiki/Multiplex%20%28sensor%29 | Multiplex sensor is a hand-held multiparametric optical sensor developed by Force-A. The sensor is a result of 15 years of research on plant autofluorescence conducted by the CNRS (National Center for Scientific Research) and University of Paris-Sud Orsay. It provides accurate and complete information on the physiological state of the crop, allowing real-time and non-destructive measurements of chlorophyll and polyphenols contents in leaves and fruits.
Technology
Multiplex assesses the chlorophyll and polyphenols indices by making use of two attributes of plant fluorescence: the effect of fluorescence re-absorption by chlorophyll and screening effect of polyphenols.
The sensor is an optical head which contains:
Optical sources (UV, blue, green and red)
Detectors (blue-green or yellow, red and far-red (NIR))
Applications
Alongside with other data, Multiplex is designed to provide input for decision support systems (DSS) for a range of crops, including:
Fertilization applications
Crop quality assessments (nitrogen status, maturity, freshness and disease detection)
As a standalone sensor, Multiplex is a tool for rapid collection of information concerning chlorophyll and flavonoids contents of the plant to be applied on ecophysiological research.
References
Sensors |
https://en.wikipedia.org/wiki/2014%20celebrity%20nude%20photo%20leak | On August 31, 2014, a collection of nearly five hundred private pictures of various celebrities, mostly women, with many containing nudity, were posted on the imageboard 4chan, and swiftly disseminated by other users on websites and social networks such as Imgur and Reddit. The leak has been popularly dubbed "The Fappening" and also "Celebgate". The images were initially believed to have been obtained via a breach of Apple's cloud services suite iCloud, or a security issue in the iCloud API which allowed them to make unlimited attempts at guessing victims' passwords. Apple claimed in a press release that access was gained via spear phishing attacks.
The incident was met with varied reactions from the media and fellow celebrities. Critics argued the leak was a major invasion of privacy for the photos' subjects, while some of the alleged subjects denied the images' authenticity. The leak also prompted increased concern from analysts surrounding the privacy and security of cloud computing services such as iCloud—with a particular emphasis on their use to store sensitive, private information.
Origin of the term
"The Fappening" is a jocular portmanteau coined by combining the words "fap", an internet slang term for masturbation, and the title of the 2008 film The Happening. Though the term is a vulgarism originating either with the imageboards where the pictures were initially posted or Reddit, mainstream media outlets soon adopted the term themselves, such as the BBC. The term has received criticism from journalists like Radhika Sanghani of The Daily Telegraph and Toyin Owoseje of the International Business Times, who said that the term not only trivialized the leak, but also, according to Sanghani, "[made] light of a very severe situation." Both articles used the term extensively to describe the event, including in their headlines.
"Celebgate" is a reference to the Watergate scandal.
Procurement and distribution
The images were obtained via the online storage offe |
https://en.wikipedia.org/wiki/Circuit%20topology | The circuit topology of a folded linear polymer refers to the arrangement of its intra-molecular contacts. Examples of linear polymers with intra-molecular contacts are nucleic acids and proteins. Proteins fold via formation of contacts of various nature, including hydrogen bonds, disulfide bonds, and beta-beta interactions. RNA molecules fold by forming hydrogen bonds between nucleotides, forming nested or non-nested structures. Contacts in the genome are established via protein bridges including CTCF and cohesins and are measured by technologies including Hi-C. Circuit topology categorises the topological arrangement of these physical contacts, that are referred to as hard contacts (or h-contacts). Furthermore, chains can fold via knotting (or formation of "soft" contacts (s-contacts)). Circuit topology uses a similar language to categorise both "soft" and "hard" contacts, and provides a full description of a folded linear chain. In this framework, a "circuit" refers to a segment of the chain where each contact site within the segment forms connections with other contact sites within the same segment, and thus is not left unpaired. A folded chain can thus be studied based on its constituting circuits.
A simple example of a folded chain is a chain with two hard contacts. For a chain with two binary contacts, three arrangements are available: parallel (P), series (S) and crossed (X). For a chain with n contacts, the topology can be described by an n by n matrix in which each element illustrates the relation between a pair of contacts and may take one of the three states, P, S and X. Multivalent contacts can also be categorised in full or via decomposition into several binary contacts. Similarly, circuit topology allows for classification of the pairwise arrangements of chain crossings and tangles, thus providing a complete 3D description of folded chains. Furthermore, one can apply circuit topology operations to soft and hard contacts to generate complex folds, usi |
https://en.wikipedia.org/wiki/Mac%20OS%20Turkish%20encoding | Each character is shown with its equivalent Unicode code point. Only the second half of the table (code points 128–255) is shown, the first half (code points 0–127) being the same as ASCII.
IBM uses code page 1281 (CCSID 1281) for Mac OS Turkish.
The character 0xF0 is a solid Apple logo. Apple uses U+F8FF in the Corporate Private Use Area for this logo, but it is usually not supported on non-Apple platforms.
References
Character sets
Turkish |
https://en.wikipedia.org/wiki/Mac%20OS%20Romanian%20encoding | Mac OS Romanian is a character encoding used on Apple Macintosh computers to represent the Romanian language. It is a derivative of Mac OS Roman.
IBM uses code page 1285 (CCSID 1285) for Mac OS Romanian.
Character set
Each character is shown with its equivalent Unicode code point. Only the second half of the table (code points 128–255) is shown, the first half (code points 0–127) being the same as ASCII.
References
Character sets
Romanian |
https://en.wikipedia.org/wiki/Mac%20OS%20Croatian%20encoding | Mac OS Croatian is a character encoding used on Apple Macintosh computers to represent Gaj's Latin alphabet. It is a derivative of Mac OS Roman. The three digraphs, Dž, Lj, and Nj, are not encoded.
IBM uses code page 1284 (CCSID 1284) for Mac OS Croatian, while Microsoft uses code page 10082.
The Croatian letters are added at the same positions as in ISO 8859-2. Despite having several added letters in common with Mac OS Central European, these are not encoded in the same positions.
Layout
Each character is shown with its equivalent Unicode code point and its decimal code point. Only the second half of the table (code points 128–255) is shown, the first half (code points 0–127) being the same as ASCII.
References
Character sets
Croatian |
https://en.wikipedia.org/wiki/PC%20Master%20Race | The PC Master Race (abbreviated PCMR), sometimes referred to by its original phrasing as the Glorious PC Gaming Master Race, is an internet meme, subculture and a tongue-in-cheek term used within video game culture to describe the grandiosity and god complex associated with PC gamers when comparing themselves to console gamers.
In current parlance, the term is commonly used by computer enthusiasts both to proudly proclaim themselves as an elitist gamer group, as well as a humorous self-parody of their own firm belief in the technical supremacy of personal computers as a video gaming platform over video game consoles such as PlayStation and Xbox, often citing gaming PC features like high-end graphics, faster frame rates, more precise gameplay control (especially with first-person shooters), free online play, wider variety of downloadable games, backward compatibility, better modifiability, upgradability and customization, lower cost-over-time, open standards, multitasking and overall superior performance. Popular imagery, discussion and media referencing the term also commonly belittles gamers who prefer playing consoles as "dirty console peasants", and describes people who prefer playing PC as the "PC master race".
History
Creation
In 2008, comedic writer Ben "Yahtzee" Croshaw employed the comedically extreme term "Glorious PC Gaming Master Race" in a Zero Punctuation video-review for the role-playing game The Witcher for the online gaming magazine The Escapist. Croshaw explained that his initial intent in referencing Nazi Germany's master race ideology when he coined the intense term Glorious PC Gaming Master Race was to poke fun at an elitist attitude he perceived among some of The Witchers PC playerbase at the time of The Witchers release, who had complained about the PC release of the game being possibly negatively affected by the console port of the game:
"It was intended to be ironic, to illustrate what I perceived at the time to be an elitist attitude am |
https://en.wikipedia.org/wiki/Lists%20of%20pharmaceutical%20industry%20topics | These are Wikipedia lists about the pharmaceutical industry. The pharmaceutical industry develops, produces, and markets drugs or pharmaceuticals licensed for use as medications. Pharmaceutical companies are allowed to deal in generic or brand medications and medical devices. They are subject to a variety of laws and regulations regarding the production, testing, and marketing of drugs.
List of pharmaceutical companies
List of largest selling pharmaceutical products
List of largest pharmaceutical settlements
List of off-label promotion pharmaceutical settlements
List of pharmaceutical sciences journals
List of pharmaceutical compound number prefixes
List of pharmaceutical manufacturers in the United Kingdom
List of pharmaceutical companies in Hyderabad
List of pharmaceutical companies in China
List of GlaxoSmithKline products
See also
Drug-related lists
Medical lists
Industry-related lists |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.