source stringlengths 31 227 | text stringlengths 9 2k |
|---|---|
https://en.wikipedia.org/wiki/Paris%20type | Paris type is a pattern of mycorrhizal infection which is coil-like in morphology.
These have direct intracellular growth to new cells.
The mycoheterotrophic plants use this to their advantage, as well as in many tree species, such as acer.
See also
Arbuscular mycorrhizal fungi
Arum type |
https://en.wikipedia.org/wiki/132nd%20meridian%20west | The meridian 132° west of Greenwich is a line of longitude that extends from the North Pole across the Arctic Ocean, North America, the Pacific Ocean, the Southern Ocean, and Antarctica to the South Pole.
The 132nd meridian west forms a great circle with the 48th meridian east.
From Pole to Pole
Starting at the North Pole and heading south to the South Pole, the 132nd meridian west passes through:
{| class="wikitable plainrowheaders"
! scope="col" width="130" | Co-ordinates
! scope="col" | Country, territory or sea
! scope="col" | Notes
|-
| style="background:#b0e0e6;" |
! scope="row" style="background:#b0e0e6;" | Arctic Ocean
| style="background:#b0e0e6;" |
|-
| style="background:#b0e0e6;" |
! scope="row" style="background:#b0e0e6;" | Beaufort Sea
| style="background:#b0e0e6;" |
|-valign="top"
|
! scope="row" |
| Northwest Territories Yukon — from British Columbia — from
|-valign="top"
|
! scope="row" |
| Alaska — Alaska Panhandle (mainland), Wrangell Island, Deer Island and Cleveland Peninsula (mainland)
|-
| style="background:#b0e0e6;" |
! scope="row" style="background:#b0e0e6;" | Clarence Strait
| style="background:#b0e0e6;" |
|-
|
! scope="row" |
| Alaska — Prince of Wales Island
|-
| style="background:#b0e0e6;" |
! scope="row" style="background:#b0e0e6;" | Dixon Entrance
| style="background:#b0e0e6;" |
|-
|
! scope="row" |
| British Columbia — Graham Island and Moresby Island
|-valign="top"
| style="background:#b0e0e6;" |
! scope="row" style="background:#b0e0e6;" | Pacific Ocean
| style="background:#b0e0e6;" |
|-
| style="background:#b0e0e6;" |
! scope="row" style="background:#b0e0e6;" | Southern Ocean
| style="background:#b0e0e6;" |
|-
|
! scope="row" | Antarctica
| Unclaimed territory
|-
|}
See also
131st meridian west
133rd meridian west
w132 meridian west |
https://en.wikipedia.org/wiki/Responsive%20web%20design | Responsive web design (RWD) or responsive design is an approach to web design that aims to make web pages render well on a variety of devices and window or screen sizes from minimum to maximum display size to ensure usability and satisfaction.
A responsive design adapts the web-page layout to the viewing environment by using techniques such as fluid proportion-based grids, flexible images, and CSS3 media queries, an extension of the @media rule, in the following ways:
The fluid grid concept calls for page element sizing to be in relative units like percentages, rather than absolute units like pixels or points.
Flexible images are also sized in relative units, so as to prevent them from displaying outside their containing element.
Media queries allow the page to use different CSS style rules based on characteristics of the device the site is being displayed on, e.g. width of the rendering surface (browser window width or physical display size).
Responsive layouts automatically adjust and adapt to any device screen size, whether it is a desktop, a laptop, a tablet, or a mobile phone.
Responsive web design became more important as users of mobile devices came to account for the majority of website visitors. In 2015, for instance, Google announced Mobilegeddon and started to boost the page ranking of mobile-friendly sites when searching from a mobile device.
Responsive web design is an example of user interface plasticity.
Challenges, and other approaches
Luke Wroblewski has summarized some of the RWD and mobile design challenges and created a catalog of multi-device layout patterns. He suggested that, compared with a simple HWD approach, device experience or RESS (responsive web design with server-side components) approaches can provide a user experience that is better optimized for mobile devices. Server-side CSS generator implementation of stylesheet languages like Sass can be part of such an approach. Google has recommended responsive design for smartphone |
https://en.wikipedia.org/wiki/Mittag-Leffler%20function | In mathematics, the Mittag-Leffler function is a special function, a complex function which depends on two complex parameters and . It may be defined by the following series when the real part of is strictly positive:
where is the gamma function. When , it is abbreviated as .
For , the series above equals the Taylor expansion of the geometric series and consequently .
In the case and are real and positive, the series converges for all values of the argument , so the Mittag-Leffler function is an entire function. This function is named after Gösta Mittag-Leffler. This class of functions are important in the theory of the fractional calculus.
For , the Mittag-Leffler function is an entire function of order , and is in some sense the simplest entire function of its order.
The Mittag-Leffler function satisfies the recurrence property (Theorem 5.1 of )
from which the following Poincaré asymptotic expansion holds : for and real such that
then for all , we can show the following asymptotic expansions (Section 6. of ):
-as :
,
-and as :
,
where we used the notation .
Special cases
For we find: (Section 2 of )
Error function:
The sum of a geometric progression:
Exponential function:
Hyperbolic cosine:
For , we have
For , the integral
gives, respectively: , , .
Mittag-Leffler's integral representation
The integral representation of the Mittag-Leffler function is (Section 6 of )
where the contour starts and ends at and circles around the singularities and branch points of the integrand.
Related to the Laplace transform and Mittag-Leffler summation is the expression (Eq (7.5) of with )
Applications of Mittag-Leffler function
One of the applications of the Mittag-Leffler function is in modeling fractional order viscoelastic materials. Experimental investigations into the time-dependent relaxation behavior of viscoelastic materials are characterized by a very fast decrease of the stress at the beginning of the relaxation process and an extr |
https://en.wikipedia.org/wiki/Neuropil | Neuropil (or "neuropile") is any area in the nervous system composed of mostly unmyelinated axons, dendrites and glial cell processes that forms a synaptically dense region containing a relatively low number of cell bodies. The most prevalent anatomical region of neuropil is the brain which, although not completely composed of neuropil, does have the largest and highest synaptically concentrated areas of neuropil in the body. For example, the neocortex and olfactory bulb both contain neuropil.
White matter, which is mostly composed of myelinated axons (hence its white color) and glial cells, is generally not considered to be a part of the neuropil.
Neuropil (pl. neuropils) comes from the Greek: neuro, meaning "tendon, sinew; nerve" and pilos, meaning "felt". The term's origin can be traced back to the late 19th century.
Location
Neuropil has been found in the following regions: outer neocortex layer, barrel cortex, inner plexiform layer and outer plexiform layer, posterior pituitary, and glomeruli of the cerebellum. These are all found in humans, with the exception of the barrel cortex, but many species have counterparts similar to our own regions of neuropil. However, the degree of similarity depends upon the composition of neuropil being compared. The concentrations of neuropil within certain regions are important to determine because simply using the proportions of the different postsynaptic elements does not verify the necessary, conclusive evidence. Comparing the concentrations can determine whether or not proportions of different postsynaptic elements contacted a particular axonal pathway. Relative concentrations could signify a reflection of different postsynaptic elements in the neuropil or show that axons sought out and formed synapses only with specific postsynaptic elements.
Function
Since neuropils have a diverse role in the nervous system, it is difficult to define a certain overarching function for all neuropils. For instance, the olfactory glomeru |
https://en.wikipedia.org/wiki/Paschen%27s%20law | Paschen's law is an equation that gives the breakdown voltage, that is, the voltage necessary to start a discharge or electric arc, between two electrodes in a gas as a function of pressure and gap length. It is named after Friedrich Paschen who discovered it empirically in 1889.
Paschen studied the breakdown voltage of various gases between parallel metal plates as the gas pressure and gap distance were varied:
With a constant gap length, the voltage necessary to arc across the gap decreased as the pressure was reduced and then increased gradually, exceeding its original value.
With a constant pressure, the voltage needed to cause an arc reduced as the gap size was reduced but only to a point. As the gap was reduced further, the voltage required to cause an arc began to rise and again exceeded its original value.
For a given gas, the voltage is a function only of the product of the pressure and gap length. The curve he found of voltage versus the pressure-gap length product (right) is called Paschen's curve. He found an equation that fit these curves, which is now called Paschen's law.
At higher pressures and gap lengths, the breakdown voltage is approximately proportional to the product of pressure and gap length, and the term Paschen's law is sometimes used to refer to this simpler relation. However, this is only roughly true, over a limited range of the curve.
Paschen curve
Early vacuum experimenters found a rather surprising behavior. An arc would sometimes take place in a long irregular path rather than at the minimal distance between the electrodes. For example, in air, at a pressure of one atmosphere, the distance for minimal breakdown voltage is about 7.5 μm. The voltage required to arc this distance is 327 V, which is insufficient to ignite the arcs for gaps that are either wider or narrower. For a 3.5 μm gap, the required voltage is 533 V, nearly twice as much. If 500 V were applied, it would not be sufficient to arc at the 2.85 μm distance, but |
https://en.wikipedia.org/wiki/Amber | Amber is fossilized tree resin that has been appreciated for its color and natural beauty since Neolithic times. Much valued from antiquity to the present as a gemstone, amber is made into a variety of decorative objects. Amber is used in jewelry and has been used as a healing agent in folk medicine.
There are five classes of amber, defined on the basis of their chemical constituents. Because it originates as a soft, sticky tree resin, amber sometimes contains animal and plant material as inclusions. Amber occurring in coal seams is also called resinite, and the term ambrite is applied to that found specifically within New Zealand coal seams.
Etymology
The English word amber derives from Arabic (ultimately from Middle Persian ambar) via Middle Latin ambar and Middle French ambre. The word referred to what is now known as ambergris (ambre gris or "grey amber"), a solid waxy substance derived from the sperm whale. The word, in its sense of "ambergris," was adopted in Middle English in the 14th century.
In the Romance languages, the sense of the word was extended to Baltic amber (fossil resin) from as early as the late 13th century. At first called white or yellow amber (ambre jaune), this meaning was adopted in English by the early 15th century. As the use of ambergris waned, this became the main sense of the word.
The two substances ("yellow amber" and "grey amber") conceivably became associated or confused because they both were found washed up on beaches. Ambergris is less dense than water and floats, whereas amber is too dense to float, though less dense than stone.
The classical names for amber, Latin electrum and Ancient Greek (ēlektron), are connected to a term ἠλέκτωρ (ēlektōr) meaning "beaming Sun". According to myth, when Phaëton son of Helios (the Sun) was killed, his mourning sisters became poplar trees, and their tears became elektron, amber. The word elektron gave rise to the words electric, electricity, and their relatives because of amber's abil |
https://en.wikipedia.org/wiki/NDUFA4L2 | NADH dehydrogenase (ubiquinone) 1 alpha subcomplex, 4-like 2 is a protein that in humans is encoded by the NDUFA4L2 gene. The NDUFA4L2 protein is a subunit of NADH dehydrogenase (ubiquinone), which is located in the mitochondrial inner membrane and is the largest of the five complexes of the electron transport chain.
Structure
The NDUFA4L2 gene is located on the long q arm of chromosome 12 at position 13.3 and it spans 5,860 base pairs. NDUFA4L2 is a subunit of the enzyme NADH dehydrogenase (ubiquinone), the largest of the respiratory complexes. The structure is L-shaped with a long, hydrophobic transmembrane domain and a hydrophilic domain for the peripheral arm that includes all the known redox centers and the NADH binding site. It has been noted that the N-terminal hydrophobic domain has the potential to be folded into an alpha helix spanning the inner mitochondrial membrane with a C-terminal hydrophilic domain interacting with globular subunits of Complex I. The highly conserved two-domain structure suggests that this feature is critical for the protein function and that the hydrophobic domain acts as an anchor for the NADH dehydrogenase (ubiquinone) complex at the inner mitochondrial membrane.
Function
The human NDUFA4L2 gene codes for a subunit of Complex I of the respiratory chain, which transfers electrons from NADH to ubiquinone. Initially, NADH binds to Complex I and transfers two electrons to the isoalloxazine ring of the flavin mononucleotide (FMN) prosthetic arm to form FMNH2. The electrons are transferred through a series of iron-sulfur (Fe-S) clusters in the prosthetic arm and finally to coenzyme Q10 (CoQ), which is reduced to ubiquinol (CoQH2). The flow of electrons changes the redox state of the protein, resulting in a conformational change and pK shift of the ionizable side chain, which pumps four hydrogen ions out of the mitochondrial matrix. |
https://en.wikipedia.org/wiki/Shared%20Whois%20Project | The Shared Whois Project (SWIP) is the process used to submit, maintain and update information to ensure up-to-date and efficient maintenance of WHOIS records, as structured in RFC 1491. The process updates WHOIS to contain information regarding what organization is using a specific IP address, or a specific block of addresses. Additionally, it provides means to track the use of an organization's current allocations of IP addresses, so that additional allocation of IP addresses may be justified and usage reports or case studies may be done.
SWIP must be used within seven days of any reassignment of an IP address space to a downstream customer to:
Provide new reassignments of blocks of eight or more IP addresses
Delete existing reassignments, or
Modify data connected to existing reassignments.
In contrast, an organization is not required to submit a SWIP template for fewer than eight IP addresses.
Allocations, assignments or reassignments of blocks of 8,192 addresses and more (greater than a /19 block) must be approved by ARIN beforehand. Companies assigned a block of 65,536 addresses or more will be responsible for maintaining all IN-ADDR.ARPA domain records for their customers.
There are two types of registration. The first is an allocation, for blocks of IP addresses which will later be reallocated or reassigned to third parties. The second is an assignment, for blocks that will not be reassigned, or the recipient is an end-user. |
https://en.wikipedia.org/wiki/Bochs | Bochs (pronounced "box") is a portable x86-32 and x86-64 IBM PC compatible emulator and debugger mostly written in C++ and distributed as free software under the GNU Lesser General Public License. It supports emulation of the processor(s) (including protected mode), memory, disks, display, Ethernet, BIOS and common hardware peripherals of PCs.
Many guest operating systems can be run using the emulator including DOS, several versions of Microsoft Windows, BSDs, Linux, Xenix and Rhapsody OS (precursor of Mac OS X Public Beta). Bochs runs on many host operating systems, including Android OS, Linux, macOS, PlayStation 2, Windows, and Windows CE along with its derivatives.
Bochs is mostly used for operating system development (when an emulated operating system crashes, it does not crash the host operating system, so the emulated OS can be debugged) and to run other guest operating systems inside already running host operating systems. It can also be used to run older software — such as PC games — which will not run on non-compatible, or too fast computers.
History
Bochs started as a program with a commercial license, at the price of US$25, for use as-is. If a user needed to link it to other software, that user would have to negotiate a special license. That changed on 22 March 2000, when Mandrakesoft (later Mandriva) bought Bochs from lead developer Kevin Lawton and released it for Linux under the GNU Lesser General Public License. Support for hosting on Windows XP ended with version 2.6.10.
Use
Bochs emulates the hardware needed by PC operating systems, including hard drives, CD drives, and floppy drives. It doesn't utilize any host CPU virtualization features, therefore is slower than most virtualization (as opposed to emulation) software. It provides additional security by completely isolating the guest OS from the hardware. Bochs also has extensive debugging features. It is widely used for OS development, as it removes the need for constant system restarts (to |
https://en.wikipedia.org/wiki/Self-complementary%20adeno-associated%20virus | Self-complementary adeno-associated virus (scAAV) is a viral vector engineered from the naturally occurring adeno-associated virus (AAV) to be used as a tool for gene therapy. Use of recombinant AAV (rAAV) has been successful in clinical trials addressing a variety of diseases. This lab-made progeny of rAAV is termed "self-complementary" because the coding region has been designed to form an intra-molecular double-stranded DNA template. A rate-limiting step for the standard AAV genome involves the second-strand synthesis since the typical AAV genome is a single-stranded DNA template. However, this is not the case for scAAV genomes. Upon infection, rather than waiting for cell mediated synthesis of the second strand, the two complementary halves of scAAV will associate to form one double stranded DNA (dsDNA) unit that is ready for immediate replication and transcription. The caveat of this construct is that instead of the full coding capacity found in rAAV (4.7–6kb) scAAV can only hold about half of that amount (≈2.4kb).
In gene therapy application utilizing rAAV, the virus transduces the cell with a single stranded DNA (ssDNA) flanked by two inverted terminal repeats (ITRs). These ITRs form hairpins at the end of the sequence to serve as primers to initiate synthesis of the second strand before subsequent steps of infection can begin. The second strand synthesis is considered to be one of several blocks to efficient infection. Additional advantages of scAAV include increased and prolonged transgene expression in vitro and in vivo, as well as "higher in vivo DNA stability and more effective circularization."
In gene therapy
scAAV is an attractive vector for use in gene therapy for many reasons. Its parent vector, AAV, is already being used in clinical trials. Due to a variety of scAAV serotypes available, scientists can choose a serotype which has properties desirable for their therapy. Selecting only a subset of cells improves specificity and lowers the risk of b |
https://en.wikipedia.org/wiki/Kam%C4%81l%20al-D%C4%ABn%20al-F%C4%81ris%C4%AB | Kamal al-Din Hasan ibn Ali ibn Hasan al-Farisi or Abu Hasan Muhammad ibn Hasan (1267– 12 January 1319, long assumed to be 1320)) () was a Persian Muslim scientist. He made two major contributions to science, one on optics, the other on number theory. Farisi was a pupil of the astronomer and mathematician Qutb al-Din al-Shirazi, who in turn was a pupil of Nasir al-Din Tusi.
According to Encyclopædia Iranica, Kamal al-Din was the most prominent Persian author on optics.
Optics
His work on optics was prompted by a question put to him concerning the refraction of light. Shirazi advised him to consult the Book of Optics of Ibn al-Haytham (Alhacen), and Farisi made such a deep study of this treatise that Shirazi suggested that he write what is essentially a revision of that major work, which came to be called the Tanqih. Qutb al-Din Al-Shirazi himself was writing a commentary on works of Avicenna at the time.
Farisi is known for giving the first mathematically satisfactory explanation of the rainbow, and an explication of the nature of colours that reformed the theory of Ibn al-Haytham Alhazen. Farisi also "proposed a model where the ray of light from the sun was refracted twice by a water droplet, one or more reflections occurring between the two refractions." He verified this through extensive experimentation using a transparent sphere filled with water and a camera obscura.
His research in this regard was based on theoretical investigations in dioptrics conducted on the so-called Burning Sphere (al-Kura al-muhriqa) in the tradition of Ibn Sahl (d. ca. 1000) and Ibn al-Haytham (d. ca. 1041) after him. As he noted in his Kitab Tanqih al-Manazir (The Revision of the Optics), Farisi used a large clear vessel of glass in the shape of a sphere, which was filled with water, in order to have an experimental large-scale model of a rain drop. He then placed this model within a camera obscura that has a controlled aperture for the introduction of light. He projected light |
https://en.wikipedia.org/wiki/Tdsoft | Tdsoft is an Israeli company, founded in 1994 by Yosi Albagli, Arie Shaham and Eytan Radian, together with Teledata Communications. Tdsoft developed V5.X based products, including V5Core Software, V5Proxi, and Tdgate family of TDM / ATM and Voice over IP access gateways.
Investors in Tdsoft included Apax, Gemini, HarbourVest Partners and Cisco Systems.
In 2001, Tdsoft acquired the Hunt 8110 product line from Cisco. In 2004, Tdsoft acquired the assets of Be-Connected from Telrad, and expanded its offering of access gateways.
In November 2005, Tdsoft took control over VocalTec (NASDAQ:VOCL), in a way of a reverse merger, and continued the combined business, by providing Voice over IP solutions to Telecommunications service providers, under the public entity and brand name of VocalTec.
See also
List of VoIP companies |
https://en.wikipedia.org/wiki/Sasaki%20%28company%29 | Sasaki is a design firm specializing in Architecture, Interior Design, Urban Design, Space Planning, Landscape Architecture, Ecology, Civil Engineering, and Place Branding. The firm is headquartered in Boston, Massachusetts, but practices on an international scale, with offices in Shanghai, and Denver, Colorado, and clients and projects globally.
History
Sasaki was founded in 1953 by landscape architect Hideo Sasaki while he served as a professor and landscape architecture chair at the Harvard Graduate School of Design. Sasaki was founded upon collaborative, interdisciplinary design, unprecedented in design practice at the time, and an emphasis on the integration of land, buildings, people, and their contexts.
Through the mid to late 1900s, Sasaki designed plazas (including Copley Square), corporate parks, college campuses, and master plans, among other projects.
The firm includes a team of in house designers, software developers, and data analysts who support the practice. Today, Sasaki has over 300 employees across its diverse practice areas and between its two offices. The firm engages in a wide variety of project types, across its many disciplines.
Milestones
In 2000, in honor of the passing of the firm's founder, the family of Hideo Sasaki together with Sasaki and other financial supporters, established the Sasaki Foundation. The foundation, which is a separate entity from Sasaki, gives yearly grants, supporting community-led research at Sasaki.
In 2012, Sasaki opened an office in Shanghai to support the firm's work in China and the larger Asia Pacific region.
In 2018, Sasaki opened the Incubator, a coworking space designed by and located within the Sasaki campus, which houses the Sasaki Foundation as curator of programming. The 5,000 square-foot space is home to several like-minded non-profits, organizations, and individuals.
In 2020, Sasaki established a new office in Denver, Colorado, marking the firm's third physical studio location. Opening an offi |
https://en.wikipedia.org/wiki/Reticulocyte%20binding%20protein%20homologs | Reticulocyte binding protein homologs (RHs) are a superfamily of proteins found in Plasmodium responsible for cell invasion. Together with the family of erythrocyte binding-like proteins (EBLs) they make up the two families of invasion proteins universal to Plasmodium. The two families function cooperatively.
This family is named after the reticulocyte binding proteins in P. vivax, a parasite that only infects reticulocytes (immature red blood cells) expressing the Duffy antigen. Homologs have since been identified in P. yoelii and P. reichenowi.
A P. falciparum protein complex called PfRH5-PfCyRPA-PfRipr (RCR) is known to bind basigin via the tip of RH5. The trimeric complex forms an elongated structure with RH5 and Ripr on distal ends and CyRPA in the middle. The RCR complex has been identified as a promising malaria vaccine target with each individual component capable of inducing strain transcending immunity in in vitro assays of parasite growth. Of the entire family of RHs, only RH5 appears to be essential for invasion and functions downstream of the other RHs during invasion.
PfRH4 is known to bind complement receptor 1.
RHs do not express any significant sequence feature for specific domains, except for a set of transmembrane helices at the C-terminal. From experimentation on partial proteins, RHs are known to contain enterocyte-binding and nucleotide-sensing domains (EBD and NBD) that may partially overlap. The structure of the EBD has been experimentally observed in 2011 by small angle X-ray scattering. A much better crystal structure for an N-terminal receptor-binding domain (presumably the same as EBD) was published in 2014. |
https://en.wikipedia.org/wiki/Addison%27s%20disease | Addison's disease, also known as primary adrenal insufficiency, is a rare long-term endocrine disorder characterized by inadequate production of the steroid hormones cortisol and aldosterone by the two outer layers of the cells of the adrenal glands (adrenal cortex), causing adrenal insufficiency. Symptoms generally come on slowly and insidiously and may include abdominal pain and gastrointestinal abnormalities, weakness, and weight loss. Darkening of the skin in certain areas may also occur. Under certain circumstances, an adrenal crisis may occur with low blood pressure, vomiting, lower back pain, and loss of consciousness. Mood changes may also occur. Rapid onset of symptoms indicates acute adrenal failure, which is a clinical emergency. An adrenal crisis can be triggered by stress, such as from an injury, surgery, or infection.
Addison's disease arises from problems with the adrenal gland such that not enough of the steroid hormone cortisol and possibly aldosterone are produced. It is an autoimmune disease which affects some genetically predisposed people in whom the body's own immune system has started to target the adrenal gland. While it can follow tuberculosis, in many adult cases it is unclear what has triggered onset of the disease. Causes can include certain medications, sepsis, and bleeding into both adrenal glands. Addison's disease is generally diagnosed by blood tests, urine tests, and medical imaging.
Treatment involves replacing the absent hormones. This involves taking a synthetic corticosteroid, such as hydrocortisone or fludrocortisone. These medications are usually taken by mouth. Lifelong, continuous steroid replacement therapy is required, with regular follow-up treatment and monitoring for other health problems. A high-salt diet may also be useful in some people. If symptoms worsen, an injection of corticosteroid is recommended (people need to carry a dose with them). Often, large amounts of intravenous fluids with the sugar dextrose are |
https://en.wikipedia.org/wiki/Free-standing%20Mathematics%20Qualifications | Free-standing Mathematics Qualifications (FSMQ) are a suite of mathematical qualifications available at levels 1 to 3 in the National Qualifications Framework – Foundation, Intermediate and Advanced.
Educational standard
They bridge a gap between GCSE and A-Level Mathematics. The advanced course is especially ideal for pupils who do not find GCSE maths particularly challenging and who often have extra time in their second year of GCSEs, having taken their Maths GCSE a year early. The qualification is commonly offered in private schools and is useful in allowing pupils to determine whether or not to pursue maths in subsequent stages of their schooling.
The highest grade achievable is an A. An FSMQ Unit at Advanced level is roughly equivalent to a single AS module with candidates receiving 10 UCAS points for an A grade. Intermediate level is equivalent to a GCSE in Mathematics. Coursework is often a key part of the FSMQ, but is sometimes omitted depending on the examining board.
Exam boards
The only examining board currently offering FSMQs is OCR.
Edexcel withdrew the qualification, the last exam being held in June 2004. AQA also withdrew the pilot advanced level FSMQ, the last exam being in June 2018, and a final re-sit opportunity in June 2019.
Examples
Additional Mathematics/AdMaths (OCR) (No coursework) |
https://en.wikipedia.org/wiki/Top-nodes%20algorithm | The top-nodes algorithm is an algorithm for managing a resource reservation calendar. The algorithm has been first published in 2003, and has been improved in 2009. It is used when a resource is shared among many users (for example bandwidth in a telecommunication link, or disk capacity in a large data center).
The algorithm allows users to:
check if an amount of resource is available during a specific period of time,
reserve an amount of resource for a specific period of time,
delete a previous reservation,
move the calendar forward (the calendar covers a defined duration, and it must be moved forward as time goes by).
Principle
The calendar is stored as a binary tree where leaves represent elementary time periods. Other nodes represent the period of time covered by all their descendants.
The period of time covered by a reservation is represented by a set of "top-nodes". This set is the minimal set of nodes that exactly cover the reservation period of time.
A node of the binary tree is a "top-node" for a given reservation if
all its descendants are inside the reservation period of time, and
it is the root node, or at least one descendant of the parent node is outside of the reservation period of time.
The following value is stored in each node:
q(node) = max(q(left child), q(right child))
+ total amount of reserved resource for all reservations having this node as a "top-node"
(for code optimization, the two parts of this sum are usually stored separately.)
Performance
The advantage of this algorithm is that the time to register a new resource reservation depends only on the calendar size (it does not depend on the total number of reservations).
Let be the number of elementary periods in the calendar.
The maximal number of "top-nodes" for a given reservation is 2.log n.
to check if an amount of resource is available during a specific period of time : O(log n)
to reserve an amount of resource for a specific period of time : O(log n)
to |
https://en.wikipedia.org/wiki/Stigma%20%28botany%29 | The stigma (: stigmas or stigmata) is the receptive tip of a carpel, or of several fused carpels, in the gynoecium of a flower.
Description
The stigma, together with the style and ovary (typically called the stigma-style-ovary system) comprises the pistil, which is part of the gynoecium or female reproductive organ of a plant. The stigma itself forms the distal portion of the style, or stylodia, and is composed of , the cells of which are receptive to pollen. These may be restricted to the apex of the style or, especially in wind pollinated species, cover a wide surface.
The stigma receives pollen and it is on the stigma that the pollen grain germinates. Often sticky, the stigma is adapted in various ways to catch and trap pollen with various hairs, flaps, or sculpturings. The pollen may be captured from the air (wind-borne pollen, anemophily), from visiting insects or other animals (biotic pollination), or in rare cases from surrounding water (hydrophily). Stigma can vary from long and slender to globe shaped to feathery.
Pollen is typically highly desiccated when it leaves an anther. Stigma have been shown to assist in the rehydration of pollen and in promoting germination of the pollen tube. Stigma also ensure proper adhesion of the correct species of pollen. Stigma can play an active role in pollen discrimination and some self-incompatibility reactions, that reject pollen from the same or genetically similar plants, involve interaction between the stigma and the surface of the pollen grain.
Shape
The stigma is often split into lobes, e.g. trifid (three lobed), and may resemble the head of a pin (capitate), or come to a point (punctiform). The shape of the stigma may vary considerably:
Style
Structure
The style is a narrow upward extension of the ovary, connecting it to the stigmatic papillae. Occasionally it may be absent, in which case the stigma is described as sessile. Styles are generally tube-like—either long or short. The style can be open (cont |
https://en.wikipedia.org/wiki/Diprosopus | Diprosopus (, "two-faced", from , , "two" and , [neuter], "face", "person"; with Latin ending), also known as craniofacial duplication (cranio- from Greek , "skull", the other parts Latin), is an extremely rare congenital disorder whereby parts (accessories) or all of the face are duplicated on the head.
Development
Although classically considered conjoined twinning (which it resembles), diprosopus is not normally due to the fusion or incomplete separation of two embryos. It is the result of abnormal activity by the protein SHH (sonic hedgehog). (The name of this protein was inspired by the Sonic the Hedgehog video game character and is part of an idiosyncratic naming tradition in molecular biology research.)
SHH and its corresponding genes have been found to play an important role in signaling craniofacial patterning during embryonic development. Among other things, SHH governs the width of facial features. In excess it leads to widening of facial features and to duplication of facial structures. The greater the widening, the more structures are duplicated, often in a mirror image form. This has been demonstrated in the laboratory by introducing pellets of the SHH protein into chicken embryos, resulting in chickens with duplicate beaks. Inadequate amounts of that protein lead to opposite conditions such as cyclopia where facial features are insufficiently developed.
Healthy brain development is also dependent on the signaling function of SHH. During embryonic development, SHH directs embryonic cells to organize in specific areas that later become specialized neural tissues, thus controlling the size and shape of brain structures.
Occurrences
Diprosopus often occurs in combination with other congenital disorders, particularly anencephaly, neural tube defect and cardiac malformations. When present, the brain may show abnormalities ranging from partial to complete duplication of brain structures, and/or underdevelopment of brain tissues.
Humans
Most human infa |
https://en.wikipedia.org/wiki/Mir-765%20microRNA%20precursor%20family | In molecular biology mir-765 microRNA is a short RNA molecule. MicroRNAs function to regulate the expression levels of other genes by several mechanisms.
Traumatic Brain Injury
miR-765 has been identified as a diagnostic marker for traumatic brain injury (TBI), due to significantly elevated levels in the plasma of TBI patients. Furthermore, combination with either miR-16 or miR-92a allows extremely accurate distinction between TBI patients and healthy controls.
miR-765 and pheochromocytomas
Increased miR-765 expression is observed in Von Hippel-Lindau disease compared with the sporadic benign forms of pheochromocytomas. It has been linked to neuronal plasticity through the neurotrophic tyrosine kinase receptor type 3 (NTRK3) gene. A NTRK3 3'UTR variant, ss102661458, is located in a functional target site of miR-765 and is able to significantly affect miRNA-mediated NTRK3 regulation. Through this there is restoration of the altered gene expression by miR-765. NTRK3 is underexpressed with high or overexpressed levels of miR-765, thus pointing towards a possible involvement of this relationship in pheochromocytoma pathogenesis.
See also
MicroRNA |
https://en.wikipedia.org/wiki/Ellen%20Hamaker | Ellen Louise "E.L." Hamaker (born February 13, 1974) is a Dutch-American psychologist, and statistician. Since 2018 she has been a full professor at Utrecht University, holding the chair Longitudinal Data Analysis at the Department of Methodology and Statistics. Her work focuses on the development of statistical models for the analysis of intensive longitudinal data in psychology, mainly within the frameworks of structural equation modeling and time series analysis.
Biography
Education
Hamaker received a bachelor's and master's degree in psychology from Utrecht University, in 1997 and 1999 respectively. She earned her PhD in psychological methods in 2004 at the University of Amsterdam, under the supervision of Peter Molenaar and Conor Dolan, for her thesis: "Time series analysis and the individual as the unit of psychological research"
Academic career
In 2005 Hamaker became a postdoctoral fellow at the Department of Psychology, University of Virginia. Hamaker returned to the Netherlands in 2006, being appointed assistant professor at the Department of Methodology and Statistics, Utrecht University. She was promoted to associate professor in 2011, before being promoted to full professor with the chair Longitudinal Data Analysis in 2018. Since 2014 she has also been a research fellow at the Research Group of Quantitative Psychology and Individual Differences, KU Leuven, in Belgium.
Research focus
Hamaker's early research focused on the application of time series techniques within psychology, to study process that unfold within individuals over time. Increased availability of data collected using the experience sampling method in psychology led to increased interest in this methodology in applied psychological research. Most notable of her papers is her critique of the cross-lagged panel model, in which she proposes to better account for stable trait-like differences between individuals through the use of multilevel models with random intercepts. She is also an |
https://en.wikipedia.org/wiki/Anarchist%20Black%20Cross | The Anarchist Black Cross (ABC), formerly the Anarchist Red Cross, is an anarchist support organization. The group is notable for its efforts at providing prisoners with political literature, but it also organizes material and legal support for class struggle prisoners worldwide. It commonly contrasts itself with Amnesty International, which is concerned mainly with prisoners of conscience and refuses to defend those accused of encouraging violence. The ABC openly supports those who have committed illegal activity in furtherance of revolutionary aims that anarchists accept as legitimate.
History
The Anarchist Black Cross offers aid to political prisoners.
In the early 20th century Russian Empire, dissidents including anarchists and socialists were jailed, exiled, or killed for their resistance to monarchy. Different political groups and organizations got together under Political Red Cross umbrella to provide material support for those repressed. Political Red Cross split when Social Democrats began filtering the group's support towards people with ideological alignment, thus creating the Anarchist Red Cross to help all social revolutionaries without regard to their political affiliation. By 1907, the Red Cross had expanded to Russia, Europe, and the United States, particularly as Russians fled persecution but from exile, continued to support imprisoned political dissidents. The Russian empire fell in 1917 and by releasing its political prisoners, obviated any need for the Red Cross, but as the Bolshevik communists rose and adopted the tsar's tactics, anarchists once again returned to prisoner aid. The group later changed its name from Red to Black Cross to not invoke the international humanitarian Red Cross.
Black Cross chapters in the early 20th century focused on helping anarchist dissidents during the Spanish Civil War, World War II, and in Francoist Spain. In the 1970s, the Black Cross turned away from international aid issues and towards local political is |
https://en.wikipedia.org/wiki/Human%20Genetics%20Commission | The Human Genetics Commission (HGC) was an advisory non-departmental public body that advised the UK government on the ethical and social aspects of genetics. This included genetic testing, cloning and other aspects of molecular medicine. The Commission was created after a review of the UK government biotechnology advisory framework in 1999. It was chaired initially by the lawyer, Baroness Helena Kennedy QC and, from 2007 to 2009, the acting chair was Sir John Sulston. From 2009, the Commission was chaired by Professor Jonathan Montgomery and comprised 21 members whose backgrounds include the law, medicine, consumer affairs, philosophy and ethics, scientific research, and clinical practice. Representatives of the Chief Medical Officers of England, Scotland, Wales, and Northern Ireland also sat on the Commission.
The Commission was abolished when quangos were reviewed by the newly elected government in October 2010. The Commission published its final paper in May 2012. |
https://en.wikipedia.org/wiki/Stieltjes%20moment%20problem | In mathematics, the Stieltjes moment problem, named after Thomas Joannes Stieltjes, seeks necessary and sufficient conditions for a sequence (m0, m1, m2, ...) to be of the form
for some measure μ. If such a function μ exists, one asks whether it is unique.
The essential difference between this and other well-known moment problems is that this is on a half-line [0, ∞), whereas in the Hausdorff moment problem one considers a bounded interval [0, 1], and in the Hamburger moment problem one considers the whole line (−∞, ∞).
Existence
Let
and
Then { mn : n = 1, 2, 3, ... } is a moment sequence of some measure on with infinite support if and only if for all n, both
{ mn : n = 1, 2, 3, ... } is a moment sequence of some measure on with finite support of size m if and only if for all , both
and for all larger
Uniqueness
There are several sufficient conditions for uniqueness, for example, Carleman's condition, which states that the solution is unique if |
https://en.wikipedia.org/wiki/Statistica%20%28journal%29 | Statistica is a quarterly peer-reviewed open access scientific journal dealing with methodological and technical aspects of statistics and statistical analyses in the various scientific fields.
It was established in 1931 as the Supplemento statistico ai nuovi problemi di Politica, Storia ed Economia (English: Statistical Supplement to the New Problems of Politics, History and Economics) and obtained its current title in 1941. It is published by the University of Bologna and is an historical Italian journal in the field of statistics. The founding editor-in-chief was Paolo Fortunati and Italo Scardovi was editor from 1981 till 2004. The current editor-in-chief is Christian Hennig (University of Bologna).
Famous scholars collaborated to Statistica as members of Scientific Board or simply as authors of papers. Among them there are Corrado Gini, Bruno De Finetti, Carlo E. Bonferroni, Marcel Fréchet, Samuel Kotz, Camilo Dagum, Estelle Bee Dagum, Italo Scardovi.
Abstracting and indexing
The journal is abstracted and indexed in Web of Science Core Collection – Emerging Sources Citation Index, a Clarivate Analytics database and Repec. |
https://en.wikipedia.org/wiki/Compulsion%20loop | A compulsion loop or core loop is a habitual chain of activities that will be repeated by the user to cause them to continue the activity. Typically, this loop is designed to create a neurochemical reward in the user such as the release of dopamine.
Compulsion loops are deliberately used in video game design as an extrinsic motivation for players, but may also result from other activities that create such loops, intentionally or not, such as gambling addiction and Internet addiction disorder.
Basis
The understanding of the motivations of compulsion loops came out of experiments performed on laboratory animals in operant conditioning chamber or a "Skinner box", where the animals are given both positive and negative stimuli for performing certain actions, such as providing food by pressing a lever. Besides demonstrating that animals would prefer positive rewards and thus learned to trigger the correct lever, B. F. Skinner found that the effects of random rewards and variable time between awards also became a factor towards how quickly the animals learned the rules of the positive reinforcement system. Ongoing research has shown that dopamine, synthesized in the animal brain, is a key neurotransmitter involved in this process; disabling the ability for receptors to react to dopamine in animal studies can impact how rapidly the animals can be conditioned.
Applying these principles to gaming, a compulsion loop creates a three-part cycle: the anticipation of receiving some reward, the activity that must be completed to receive that reward, and the act of finally obtaining the reward. From a neuroscience aspect, it is believed that the anticipation phase is where dopamine is created by the human brain, while it is released upon obtaining the reward. Dopamine creates feelings of pleasure in the brain and drives motivation, and while the neurotransmitter itself is not addictive, can lead to addictive behavior as the user desires to experience the further dopamine release. |
https://en.wikipedia.org/wiki/Mathematical%20morphology | Mathematical morphology (MM) is a theory and technique for the analysis and processing of geometrical structures, based on set theory, lattice theory, topology, and random functions. MM is most commonly applied to digital images, but it can be employed as well on graphs, surface meshes, solids, and many other spatial structures.
Topological and geometrical continuous-space concepts such as size, shape, convexity, connectivity, and geodesic distance, were introduced by MM on both continuous and discrete spaces. MM is also the foundation of morphological image processing, which consists of a set of operators that transform images according to the above characterizations.
The basic morphological operators are erosion, dilation, opening and closing.
MM was originally developed for binary images, and was later extended to grayscale functions and images. The subsequent generalization to complete lattices is widely accepted today as MM's theoretical foundation.
History
Mathematical Morphology was developed in 1964 by the collaborative work of Georges Matheron and Jean Serra, at the École des Mines de Paris, France. Matheron supervised the PhD thesis of Serra, devoted to the quantification of mineral characteristics from thin cross sections, and this work resulted in a novel practical approach, as well as theoretical advancements in integral geometry and topology.
In 1968, the Centre de Morphologie Mathématique was founded by the École des Mines de Paris in Fontainebleau, France, led by Matheron and Serra.
During the rest of the 1960s and most of the 1970s, MM dealt essentially with binary images, treated as sets, and generated a large number of binary operators and techniques: Hit-or-miss transform, dilation, erosion, opening, closing, granulometry, thinning, skeletonization, ultimate erosion, conditional bisector, and others. A random approach was also developed, based on novel image models. Most of the work in that period was developed in Fontainebleau.
From the |
https://en.wikipedia.org/wiki/Secure%20end%20node | A Secure End Node is a trusted, individual computer that temporarily becomes part of a trusted, sensitive, well-managed network and later connects to many other (un)trusted networks/clouds. SEN's cannot communicate good or evil data between the various networks (e.g. exfiltrate sensitive information, ingest malware, etc.). SENs often connect through an untrusted medium (e.g. the Internet) and thus require a secure connection and strong authentication (of the device, software, user, environment, etc.). The amount of trust required (and thus operational, physical, personnel, network, and system security applied) is commensurate with the risk of piracy, tampering, and reverse engineering (within a given threat environment). An essential characteristic of SENs is they cannot persist information as they change between networks (or domains).
The remote, private, and secure network might be organization's in-house network or a cloud service. A Secure End Node typically involves authentication of (i.e. establishing trust in) the remote computer's hardware, firmware, software, and/or user. In the future, the device-user's environment (location, activity, other people, etc.) as communicated by means of its (or the network's) trusted sensors (camera, microphone, GPS, radio, etc.) could provide another factor of authentication.
A Secure End Node solves/mitigates end node problem.
The common, but expensive, technique to deploy SENs is for the network owner to issue known, trusted, unchangeable hardware to users. For example, and assuming apriori access, a laptop's TPM chip can authenticate the hardware (likewise a user's smartcard authenticates the user). A different example is the DoD Software Protection Initiative's Cross Fabric Internet Browsing System that provides browser-only, immutable, anti-tamper thin clients to users Internet browsing. Another example is a non-persistent, remote client that boots over the network.
A less secure but very low cost approach i |
https://en.wikipedia.org/wiki/Hertwig%20rule | Hertwig's rule, or the long axis rule states that a cell divides along its long axis. Introduced by the German zoologist Oscar Hertwig in 1884, the rule emphasizes the cell shape as a default mechanism of spindle apparatus orientation. Hertwig's rule predicts cell division orientation, which is important for tissue architecture, cell fate and morphogenesis.
Discovery
Hertwig's experiments studied the orientation of frog egg divisions. The frog egg has a round shape and the first division occurs in a random orientation. Hertwig compressed the egg between two parallel plates. The compression forced the egg to change its shape from round to elongated. Hertwig noticed that elongated egg divides not randomly, but orthogonally to its long axis. The new daughter cells were formed along the longest axis of the cell. This observation thus became known as 'Hertwig's rule' or 'long axis rule'.
Confirmation and mechanism
Recent studies in animal and plant systems support the 'long axis rule'. The studied systems include the mouse embryo, Drosophila epithelium, Xenopus blastomeres (Strauss 2006), MDCK cell monolayers and plants (Gibson et al., 2011). The mechanism of the 'long axis rule' relies on interphase cell long axis sensing. However, during division many animal cell types undergo cell rounding, causing the long axis to disappear as the cell becomes round. It is at this rounding stage that the decision on the orientation of the cell division is made by the spindle apparatus. The spindle apparatus rotates in the round cell and after several minutes the spindle position is stabilised preferentially along the interphase cell long axis. The cell then divides along the spindle apparatus orientation. The first insights into how cells could remember their long axis came from studies on the Drosophila epithelium. The study indicated the participation of tricellular junctions (TCJs) in determining the spindle orientation. TCJs localized at the regions where three or more cells |
https://en.wikipedia.org/wiki/Strabismus%20%28protein%29 | Strabismus was originally identified as a Drosophila protein involved in planar cell polarity. Flies with mutated strabismus genes have altered development of ommatidia in their eyes. Vertebrates have two Strabismus-related proteins, VANGL1 and VANGL2 (an alternate name for the Drosophila "Strabismus" protein is "Van Gogh").
The amino acid sequence and localization studies for Strabismus indicate that it is a membrane protein. Prickle is another protein in the planar cell polarity signaling pathway. Prickle is recruited to the cell surface membrane by strabismus. In cells of the developing Drosophila wing, Prickle and Strabismus are concentrated at the cell surface membrane on the most proximal side of cells.
Vertebrate cell movement
VANGL2 is involved in the migration of groups of cells during vertebrate embryogenesis.
Humans
In humans, mutations in VANGL1 have been associated with neural tube defects including spina bifida, and with some forms of cancer including hepatocellular carcinoma. |
https://en.wikipedia.org/wiki/John%20Vandermeer | John Harry Vandermeer (born 1940) is an American ecologist, a mathematical ecologist, tropical ecologist and agroecologist. He is the Asa Gray Distinguished University Professor of Ecology and Evolutionary Biology and the Arthur F. Thurnau Professor at the University of Michigan, where he has taught since 1971. His research focuses on the ecology of agricultural systems, and he has operated a plot of coffee plants in Mexico for his research for more than fifteen years. In 2016, the symposium "Science with Passion and a Moral Compass" was held to honor his career as a scientist and activist. The symposium, also known as VandyFest, was held in Ann Arbor, Michigan from May 6 to May 8.
Early life and education
Vandermeer was born in 1940 in Chicago, Illinois. He was educated at the University of Illinois, the University of Kansas, and the University of Michigan.
Vandermeer has conducted field research mainly in Mexico, Puerto Rico, Costa Rica, Nicaragua and Guatemala. His research has focused on the dynamics of spatially explicit biological interactions in coffee farms in Mexico.
His long-term collaboration with a multi-national team of scientists focused on tropical rainforest dynamics after major hurricane disturbance in Nicaragua. Their research provides strong evidence in favor of the assertion that it is the chance to reach a recruitment space into the forest canopy that governs the maintenance of hundreds of tree species and to some lesser extent the multiple tree species competition for nutrients and light. This diverges from tropical tree species niche identity notion thus proposing that the tree species assemblage are to some extent the result of random dispersal and recruitment events.
Vandermeer and his colleagues Dr. Ivette Perfecto, Dr. Douglas Boucher and Dr. Inigo Granzow de la Cerda contributed to the groundwork that evolved into the university system in the Autonomous Regions of the Atlantic Coast of Nicaragua. |
https://en.wikipedia.org/wiki/Quantum%20reflection | Quantum reflection is a uniquely quantum phenomenon in which an object, such as a neutron or a small molecule, reflects smoothly and in a wavelike fashion from a much larger surface, such as a pool of mercury. A classically behaving neutron or molecule will strike the same surface much like a thrown ball, hitting only at one atomic-scale location where it is either absorbed or scattered. Quantum reflection provides a powerful experimental demonstration of particle-wave duality, since it is the extended quantum wave packet of the particle, rather than the particle itself, that reflects from the larger surface. It is similar to reflection high-energy electron diffraction, where electrons reflect and diffraction from surfaces, and grazing incidence atom scattering, where the fact that atoms (and ions) can also be waves is used to diffract from surfaces.
Definition
In a workshop about quantum reflection, the following definition of quantum reflection was suggested:
Quantum reflection is a classically counterintuitive phenomenon whereby the motion of particles is reverted "against the force" acting on them. This effect manifests the wave nature of particles and influences collisions of ultracold atoms and interaction of atoms with solid surfaces.
Observation of quantum reflection has become possible thanks to recent advances in trapping and cooling atoms.
Reflection of slow atoms
Although the principles of quantum mechanics apply to any particles, usually the term "quantum reflection" means reflection of atoms from a surface of condensed matter (liquid or solid). The full potential experienced by the incident atom does become repulsive at a very small distance from the surface (of order of size of atoms). This is when the atom becomes aware of the discrete character of material. This repulsion is responsible for the classical scattering one would expect for particles incident on a surface. Such scattering can be diffuse rather than specular, so this component of the |
https://en.wikipedia.org/wiki/Dental%20bonding | Adhesive dentistry is a branch of dentistry which deals with adhesion or bonding to the natural substance of teeth, enamel and dentin. It studies the nature and strength of adhesion to dental hard tissues, properties of adhesive materials, causes and mechanisms of failure of the bonds, clinical techniques for bonding and newer applications for bonding such as bonding to the soft tissue. There is also direct composite bonding which uses tooth-colored direct dental composites to repair various tooth damages such as cracks or gaps.
Dental bonding is a dental procedure in which a dentist applies a tooth-colored resin material (a durable plastic material) and cures it with visible, blue light. This ultimately "bonds" the material to the tooth and improves the overall appearance of teeth. Tooth bonding techniques have various clinical applications including operative dentistry and preventive dentistry as well as cosmetic and pediatric dentistry, prosthodontics, and orthodontics.
History
Adhesive dentistry began in 1955 with a paper by Dr. Michael Buonocore on the benefits of acid etching. Technologies have changed multiple times since then, with generally recognized generations established in the literature. Dental bonding agents have evolved from no-etch to total-etch (4th- and 5th-generation) to self-etch (6th- and 7th-generation) systems. improved convenience and reduced sensitivity to operator errors. However, the best bonding and longevity was achieved with 4th generation agents (having separate etch, prime, and bond steps).
Irwin Smigel founder and current president of the American Society for Dental Aesthetics and diplomate of the American Board of Aesthetic Dentistry, was one of the first to broaden the usage of bonding by using it to close gaps between teeth, lengthen teeth as well as to re-contour the entire mouth rather than using crowns. Having done more extensive work on the process than any other dentist, Dr. Smigel lectures worldwide on aesthetic dentist |
https://en.wikipedia.org/wiki/Mathematical%20Models%20%28Cundy%20and%20Rollett%29 | Mathematical Models is a book on the construction of physical models of mathematical objects for educational purposes. It was written by Martyn Cundy and A. P. Rollett, and published by the Clarendon Press in 1951, with a second edition in 1961. Tarquin Publications published a third edition in 1981.
The vertex configuration of a uniform polyhedron, a generalization of the Schläfli symbol that describes the pattern of polygons surrounding each vertex, was devised in this book as a way to name the Archimedean solids, and has sometimes been called the Cundy–Rollett symbol as a nod to this origin.
Topics
The first edition of the book had five chapters, including its introduction which discusses model-making in general and the different media and tools with which one can construct models. The media used for the constructions described in the book include "paper, cardboard, plywood, plastics, wire, string, and sheet metal".
The second chapter concerns plane geometry, and includes material on the golden ratio, the Pythagorean theorem, dissection problems, the mathematics of paper folding, tessellations, and plane curves, which are constructed by stitching, by graphical methods, and by mechanical devices.
The third chapter, and the largest part of the book, concerns polyhedron models, made from cardboard or plexiglass. It includes information about the Platonic solids, Archimedean solids, their stellations and duals, uniform polyhedron compounds, and deltahedra.
The fourth chapter is on additional topics in solid geometry and curved surfaces, particularly quadrics but also including topological manifolds such as the torus, Möbius strip and Klein bottle, and physical models helping to visualize the map coloring problem on these surfaces. Also included are sphere packings. The models in this chapter are constructed as the boundaries of solid objects, via two-dimensional paper cross-sections, and by string figures.
The fifth chapter, and the final one of the first editi |
https://en.wikipedia.org/wiki/Leased%20line | A leased line is a private telecommunications circuit between two or more locations provided according to a commercial contract. It is sometimes also known as a private circuit, and as a data line in the UK. Typically, leased lines are used by businesses to connect geographically distant offices.
Unlike traditional telephone lines in the public switched telephone network (PSTN) leased lines are generally not switched circuits, and therefore do not have an associated telephone number. Each side of the line is permanently connected, always active and dedicated to the other. Leased lines can be used for telephone, Internet, or other data communication services. Some are ringdown services, and some connect to a private branch exchange (PBX) or network router.
The primary factors affecting the recurring lease fees are the distance between end stations and the bandwidth of the circuit. Since the connection does not carry third-party communications, the carrier can assure a specified level of quality.
An Internet leased line is a premium Internet connectivity product, normally delivered over fiber, which provides uncontended, symmetrical bandwidth with full-duplex traffic. It is also known as an Ethernet leased line, dedicated line, data circuit or private line.
History
Leased line services (or private line services) became digital in the 1970s with the conversion of the Bell backbone network from analog to digital circuits. This allowed AT&T to offer Dataphone Digital Services (later re-branded digital data services) that started the deployment of ISDN and T1 lines to customer premises to connect.
Leased lines were used to connect mainframe computers with terminals and remote sites, via IBM's Systems Network Architecture (created in 1974) or DEC's DECnet (created in 1975).
With the extension of digital services in the 1980s, leased lines were used to connect customer premises to Frame Relay or ATM networks. Access data rates increased from the original T1 option wit |
https://en.wikipedia.org/wiki/Mixotrophic%20dinoflagellate | Dinoflagellates are eukaryotic plankton, existing in marine and freshwater environments. Previously, dinoflagellates had been grouped into two categories, phagotrophs and phototrophs. Mixotrophs, however include a combination of phagotrophy and phototrophy. Mixotrophic dinoflagellates are a sub-type of planktonic dinoflagellates and are part of the phylum Dinoflagellata. They are flagellated eukaryotes that combine photoautotrophy when light is available, and heterotrophy via phagocytosis. Dinoflagellates are one of the most diverse and numerous species of phytoplankton, second to diatoms.
Dinoflagellates have long whip-like structures called flagella that allow them to move freely throughout the water column. They are mainly marine but can also be found in freshwater environments. Combinations of phototrophy and phagotrophy allow organisms to supplement their inorganic nutrient uptake This means an increased trophic transfer to higher levels in food web compared to the traditional food web.
Mixotrophic dinoflagellates have the ability to thrive in changing ocean environments, resulting in shifts in red tide phenomenon and paralytic shellfish poisoning. It is unknown as to how many species of dinoflagellates have mixotrophic capabilities, as this is a relatively new feeding-mechanism discovery.
Species
Some dinoflagellates that live as parasites are probably mixotrophic.
Karenia, Karlodinium, and Lepidodinium are some of the dinoflagellate genera which are thought to contain peridinin, a carotenoid pigment necessary for photosynthesis in dinoflagellates; however, chlorophyll b has been found in these genera as an accessory pigment. This discovery has led scientists to assume that the pigment chlorophyll b actually came from prey which had been ingested by the dinoflagellates.
Some species of mixotrophic dinoflagellate are able to feed on toxic prey such as toxic algae and other toxic organisms. For example, Lingulodinium polyedra and Akashiwo sanguinea are two |
https://en.wikipedia.org/wiki/List%20of%20algebraic%20geometry%20topics | This is a list of algebraic geometry topics, by Wikipedia page.
Classical topics in projective geometry
Affine space
Projective space
Projective line, cross-ratio
Projective plane
Line at infinity
Complex projective plane
Complex projective space
Plane at infinity, hyperplane at infinity
Projective frame
Projective transformation
Fundamental theorem of projective geometry
Duality (projective geometry)
Real projective plane
Real projective space
Segre embedding of a product of projective spaces
Rational normal curve
Algebraic curves
Conics, Pascal's theorem, Brianchon's theorem
Twisted cubic
Elliptic curve, cubic curve
Elliptic function, Jacobi's elliptic functions, Weierstrass's elliptic functions
Elliptic integral
Complex multiplication
Weil pairing
Hyperelliptic curve
Klein quartic
Modular curve
Modular equation
Modular function
Modular group
Supersingular primes
Fermat curve
Bézout's theorem
Brill–Noether theory
Genus (mathematics)
Riemann surface
Riemann–Hurwitz formula
Riemann–Roch theorem
Abelian integral
Differential of the first kind
Jacobian variety
Generalized Jacobian
Moduli of algebraic curves
Hurwitz's theorem on automorphisms of a curve
Clifford's theorem on special divisors
Gonality of an algebraic curve
Weil reciprocity law
Algebraic geometry codes
Algebraic surfaces
Enriques–Kodaira classification
List of algebraic surfaces
Ruled surface
Cubic surface
Veronese surface
Del Pezzo surface
Rational surface
Enriques surface
K3 surface
Hodge index theorem
Elliptic surface
Surface of general type
Zariski surface
Algebraic geometry: classical approach
Algebraic variety
Hypersurface
Quadric (algebraic geometry)
Dimension of an algebraic variety
Hilbert's Nullstellensatz
Complete variety
Elimination theory
Gröbner basis
Projective variety
Quasiprojective variety
Canonical bundle
Complete intersection
Serre duality
Spaltenstein variety
Arithmetic genus, geometric genus, irregularity
Tangent space, Zariski tangent space
Function field of an algebraic variet |
https://en.wikipedia.org/wiki/Egbert%20Brieskorn | Egbert Valentin Brieskorn (7 July 1936, in Rostock – 11 July 2013, in Bonn) was a German mathematician who introduced Brieskorn spheres and the Brieskorn–Grothendieck resolution.
Education
Brieskorn was born in 1936 as the son of a mill construction engineer in East Prussia. He grew up in Freudenberg (Siegerland) and studied mathematics and physics at the Ludwig-Maximilians-Universität München and the Rheinische Friedrich-Wilhelms-Universität Bonn. In 1963 he received his doctorate at Bonn under Friedrich Hirzebruch with thesis Zur differentialtopologischen und analytischen Klassifizierung gewisser algebraischer Mannigfaltigkeiten, followed by his habilitation in 1968.
Career
From 1969 until 1973 he was professor ordinarius at Georg-August-Universität Göttingen and from 1973 to 1975 at the Sonderforschungsbereich Theoretische Mathematik in Bonn (since 1980 called the Max-Planck-Institut für Mathematik). From 1975 until his retirement as professor emeritus in 2001 he was a professor ordinarius at Bonn. He held temporary academic positions at the Massachusetts Institute of Technology (MIT) (where in 1965 he was Moore Instructor), the Institut des Hautes Études Scientifiques (IHES), the Eidgenössische Technische Hochschule Zürich (ETH Zürich), the University of Leiden, the University of Oxford, the University of Warwick, the University of Liverpool, and the University of Nice.
Contributions
Brieskorn was one of the editors of the collected works of Felix Hausdorff. In 1970 he was an invited speaker at the International Congress of Mathematicians in Nice (Singular elements of semi-simple algebraic groups). His doctoral students include Horst Knörrer, Peter Slodowy, Kyoji Saito, and Erhard Scholz.
Notes |
https://en.wikipedia.org/wiki/Boilery | A boilery or boiling house is a place of boiling, much as a bakery is a place of baking. Boilery can also mean the process and equipment for boiling. Although they are now generally confined to factories, and usually boil industrial products rather than food, historically they were more common in daily life. Boileries are typically for boiling large quantities of fluid.
In the 17th to 19th centuries, boileries were used to convert sugarcane juice into raw sugar. These boileries were usually sturdy places, built from stone, and contained several copper kettles, each with a furnace beneath it., Sugarcane juice was treated with lime in large clarifying vats, before it was heated in copper kettles over individual furnaces. Due to their importance, many Western sugar plantations had their own boileries on site.
Soap would also be made in a boiling house.
Another use for a boilery is to make salt through the evaporation of brine water. |
https://en.wikipedia.org/wiki/Interactive%20specialization | Interactive Specialization is a theory of brain development proposed by the British developmental cognitive neuroscientist Mark Johnson, formerly head of the Centre for Brain and Cognitive Development at Birkbeck, University of London, London and who is now Head of Psychology at the University of Cambridge.
In his book Developmental Cognitive Neuroscience
, Johnson contrasts two views of development. According to the first, the maturational hypothesis, the relationship between structure and function (i.e. which parts of the brain perform a particular task) is static, and specific cognitive skills come “on-line” as the cortical circuitry intrinsic to a particular task matures. Johnson likens this to a "mosaic" view of development.
According to the second, the Interactive Specialization (IS)
hypothesis, development is
not a unidirectional maturational process, but rather a set of complex, dynamic and back-propagated interactions between genetics, brain, body and environment. Development is not a simple question of a brain being built according to a pre-specified genetic blueprint - rather, the components of the brain are interacting with each other constantly - even prenatally, when patterns of spontaneous firing of cells in the eyes (before they have opened) transmit signals that appear to help develop the layered structure of the lateral geniculate nucleus
.
The hypothesis has attracted increasing attention in recent years as a number of neuroimaging studies on younger children have provided data that appears to fit specific predictions made by Johnson's model
.
Influences
In 1996, Johnson co-authored (with Jeffrey Elman, Annette Karmiloff-Smith, Elizabeth Bates, Domenico Parisi, and Kim Plunkett), the book Rethinking Innateness
, which argues against a strong nativist (innate) view on development. Other key influences include Gilbert Gottlieb's theory of Probabilistic Epigenesis
, a framework that emphasizes the reciprocity and ubiquity of gene-envir |
https://en.wikipedia.org/wiki/TIS-100 | TIS-100 is a programming/puzzle video game developed by Zachtronics Industries. The game has the player develop mock assembly language code to perform certain tasks on a fictional, virtualized 1970s computer that has been corrupted. The game was released for Microsoft Windows, OS X, and Linux personal computers in July 2015.
Gameplay
Within TIS-100, the player is said to have come across a malfunctioning TIS-100 computer ("Tessellated Intelligence System") and its manual, based on early computers of the 1980s. The computer is presented to the player as twelve separate processing nodes laid out in a four-by-three grid. Each node has a single processor register to store a numerical value as well as a backup register. Nodes also hold their own assembly language program as entered by the user. The assembly language, a simplified version of real-world assembly languages, allows the node to accept external input or a numerical value sent from an adjacent node, perform basic math and logic operations, store and backup the current data value, and then send results to an adjacent node or to the program's output. Later puzzles introduce stack nodes and an output to a simple 5-color graphics display.
The player is presented with a series of puzzles that require them to program the nodes to perform specific actions on a set of numbers from one or more input terminals to produce pre-determined output at other terminals. For example, one task requires the player to double the value of the input at the output terminal. The game presents the list of inputs and the target output values that it is expecting and requires the players to develop the code for each node to match this; if during execution the output nodes receive unexpected outputs, the player will have to rework their solution. Not all nodes are available in certain puzzles, so the player will need to route around these nodes. The game offers the player the ability to step through the execution of the code and insert de |
https://en.wikipedia.org/wiki/Clock%20signal | In electronics and especially synchronous digital circuits, a clock signal (historically also known as logic beat) is an electronic logic signal (voltage or current) which oscillates between a high and a low state at a constant frequency and is used like a metronome to synchronize actions of digital circuits. In a synchronous logic circuit, the most common type of digital circuit, the clock signal is applied to all storage devices, flip-flops and latches, and causes them all to change state simultaneously, preventing race conditions.
A clock signal is produced by an electronic oscillator called a clock generator. The most common clock signal is in the form of a square wave with a 50% duty cycle. Circuits using the clock signal for synchronization may become active at either the rising edge, falling edge, or, in the case of double data rate, both in the rising and in the falling edges of the clock cycle.
Digital circuits
Most integrated circuits (ICs) of sufficient complexity use a clock signal in order to synchronize different parts of the circuit, cycling at a rate slower than the worst-case internal propagation delays. In some cases, more than one clock cycle is required to perform a predictable action. As ICs become more complex, the problem of supplying accurate and synchronized clocks to all the circuits becomes increasingly difficult. The preeminent example of such complex chips is the microprocessor, the central component of modern computers, which relies on a clock from a crystal oscillator. The only exceptions are asynchronous circuits such as asynchronous CPUs.
A clock signal might also be gated, that is, combined with a controlling signal that enables or disables the clock signal for a certain part of a circuit. This technique is often used to save power by effectively shutting down portions of a digital circuit when they are not in use, but comes at a cost of increased complexity in timing analysis.
Single-phase clock
Most modern synchronous |
https://en.wikipedia.org/wiki/Lerner%20symmetry%20theorem | The Lerner symmetry theorem is a result used in international trade theory, which states that an ad valorem import tariff (a percentage of value or an amount per unit) will have the same effects as an export tax. The theorem is based on the observation that the effect on relative prices is the same regardless of which policy (ad valorem tariffs or export taxes) is applied.
The theorem was developed by economist Abba P. Lerner in 1936.
Notes |
https://en.wikipedia.org/wiki/PITPNM3 | Nir1 or Membrane-associated phosphatidylinositol transfer protein 3 (PITPNM3) is a mammalian protein that localizes to endoplasmic reticulum (ER) and plasma membrane (PM) membrane contact sites (MCS) and aids the transfer of phosphatidylinositol between these two membranes, potentially by recruiting additional proteins to the ER-PM MCS.
Classification
Nir1 has been classically categorized as a class IIA phosphatidylinositol transfer protein (PITP) that transfers phosphatidylinositol (PI) and phosphatidic acid (PA) between membranes. Class IIA PITPs are the multi-domain proteins PITPNM1/Nir2 (Drosophila homolog RdgBaI), PITPNM2/Nir3 (Drosophila homolog RdgBaII).. Nir1 shares high sequence similarity with Nir2 and Nir3, which led to its original categorization as a PITP. However, it was determined that Nir1 is not directly responsible for PI transfer, as it lacks the functional PITP domain seen within Nir2 and Nir3
Localization
Recently, Nir1 has been shown to localize to ER-PM MCS, both under basal conditions and upon phospholipase C (PLC) activation. Notably, PLC activation has previously been shown to regulate the localization of Nir2 and Nir3 at ER-PM MCS well.. The MCS-targeting by Nir1 is achieved by the N-terminus of Nir1 localizing to the ER and the C-terminus of Nir1 localizing to the PM. The domains responsible for binding these membranes are discussed below.
Structure
Nir1 contains three main structural elements that are shared with Nir2 and Nir3: an N-terminal FFAT motif, a DDHD domain, and a C-terminal Lipin/Ndel/Smp2 (LNS2) domain.
FFAT motif
The FFAT motif is made up of double phenylalanines (FF) in an Acidic Tract. This motif, made of residues EFFDA in Nir1, has been shown to be necessary for the Nir proteins to associate with the ER proteins VAPA and VAPB. Mutation of the phenylalanine residues in this motif or knockout of the VAPA and VAPB proteins results in a loss of ER-PM MCS localization and causes Nir1 to become fully localized to the PM |
https://en.wikipedia.org/wiki/Robert%20Bryant%20%28mathematician%29 | Robert Leamon Bryant (born August 30, 1953, Kipling) is an American mathematician. He works at Duke University and specializes in differential geometry.
Education and career
Bryant grew up in a farming family in Harnett County and was a first-generation college student. He obtained a bachelor's degree at North Caroline State University at Raleigh in 1974 and a PhD at University of North Carolina at Chapel Hill in 1979. His thesis was entitled "Some Aspects of the Local and Global Theory of Pfaffian Systems" and was written under the supervision of Robert Gardner.
He worked at Rice University for seven years, as assistant professor (1979–1981), associate professor (1981–1982) and full professor (1982–1986). He then moved to Duke University, where he worked for twenty years as J. M. Kreps Professor.
Between 2007 and 2013 he worked as full professor at University of California, Berkeley, where he served as the director of the Mathematical Sciences Research Institute (MSRI). In 2013 he returned to Duke University as Phillip Griffiths Professor of Mathematics.
Bryant was awarded in 1982 a Sloan Research Fellowship. In 1986 he was invited speaker at the International Congress of Mathematicians in Berkeley.
He was elected in 2002 a fellow of the American Academy of Arts and Sciences, in 2007 a member of the National Academy of Sciences, in 2013 a fellow of the American Mathematical Society and in 2022 a fellow of the American Association for the Advancement of Science. He is also a member of the Association for Women in Mathematics, the National Association of Mathematicians and the Mathematical Association of America.
He served as the president of the American Mathematical Society for the 2-years term 2015–2016, for which he was the first openly gay president.
Bryant is on the board of directors of EDGE, a transition program for women entering graduate studies in the mathematical sciences. He is also a board member of Spectra, an association for LGBT mathematicians |
https://en.wikipedia.org/wiki/William%20Shadish | William Raymond Shadish Jr. (November 3, 1949 – March 27, 2016) was an American psychologist and statistician who was a distinguished professor and founding faculty member at the University of California, Merced. He was known for his work in the field of behavioral science, especially on the topics of program evaluation, causal inference, meta-analysis, and the study of methodology.
Early life and education
Shadish was born on November 3, 1949 in Brooklyn, New York, to William Shadish Sr. and Maryjane Cartmell. Shadish Jr.'s father was a Korean War veteran who spent 30 months as a prisoner of war in North Korea. Shadish Jr. was raised in Redding, California, and received his bachelor's degree in sociology from Santa Clara University in 1972. He went on to attend Purdue University, where he received his master's degree and Ph.D in Clinical Psychology in 1975 and 1978, respectively.
Career
After teaching at the University of Memphis for many years, Shadish joined the faculty of the University of California, Merced (UC-Merced) in 2003. There, he designed the psychology major at the then-brand-new university, which did not have any students or buildings when he first joined its faculty. He also led the process that created the Ph.D. in psychology program at UC-Merced. He became the second ever distinguished professor at UC-Merced, and received the Distinguished Research Award from their Academic Senate in 2011.
Affiliations with learned societies
Shadish was the founding secretary-treasurer of the Society for Research Synthesis Methodology, later serving as its president from 2013 to 2014. He was also elected president of the American Evaluation Association in 1996, and of the Society of Multivariate Experimental Psychology in 2014. He was a fellow of the American Psychological Association.
Awards
Shadish received the 1994 Paul F. Lazarsfeld Award for Evaluation Theory and the 2000 Robert Ingle Award from the American Evaluation Association. He also received two Out |
https://en.wikipedia.org/wiki/DRE-i%20with%20enhanced%20privacy | Direct Recording Electronic with Integrity and Enforced Privacy (DRE-ip) is an End-to-End (E2E) verifiable e-voting system without involving any tallying authorities, proposed by Siamak Shahandashti and Feng Hao in 2016. It improves a previous DRE-i system by using a real-time computation strategy and providing enhanced privacy. A touch-screen based prototype of the system was trialed in the Gateshead Civic Centre polling station on 2 May 2019 during the 2019 United Kingdom local elections with positive voter feedback. A proposal that includes DRE-ip as a solution for large-scale elections was ranked 3rd place in the 2016 Economist Cybersecurity Challenge jointly organized by The Economist and Kaspersky Lab.
Protocol
The DRE-ip protocol is applicable to both onsite polling station voting and remote Internet voting implementations. In the specification below, it is described for polling station voting. The protocol consists of three stages: setup, voting and tallying.
Setup
Let and be two large primes, where . is a subgroup of of prime order . Let and be two random generators of , whose discrete logarithm relationship is unknown. This can be realized by choosing a non-identity element in as and computing based on applying a one-way hash function with the inclusion of election specific information such as the date, election title and questions as the input. All modulo operations are performed with respect to the modulus . Alternatively, the protocol can be implemented using an elliptic curve, while the protocol specification remains unchanged.
Voting
For simplicity, the voting process is described for a single-candidate (Yes/No) election held in a polling station using a touch-screen DRE machine. There are standard ways to extend a single candidate election to support multiple candidates, e.g., providing a Yes/No selection for each of the candidates or using different encoded values for different candidates as described by Baudron et al.
After being au |
https://en.wikipedia.org/wiki/Archimedean%20circle | In geometry, an Archimedean circle is any circle constructed from an arbelos that has the same radius as each of Archimedes' twin circles. If the arbelos is normed such that the diameter of its outer (largest) half circle has a length of 1 and r denotes the radiius of any of the inner half circles, then the radius ρ of such an Archimedean circle is given by
There are over fifty different known ways to construct Archimedean circles.
Origin
An Archimedean circle was first constructed by Archimedes in his Book of Lemmas. In his book, he constructed what is now known as Archimedes' twin circles.
Radius
If and are the radii of the small semicircles of the arbelos, the radius of an Archimedean circle is equal to
This radius is thus .
The Archimedean circle with center (as in the figure at right) is tangent to the tangents from the centers of the small semicircles to the other small semicircle.
Other Archimedean circles finders
Leon Bankoff
Leon Bankoff constructed other Archimedean circles called Bankoff's triplet circle and Bankoff's quadruplet circle.
Thomas Schoch
In 1978 Thomas Schoch found a dozen more Archimedean circles (the Schoch circles) that have been published in 1998. He also constructed what is known as the Schoch line.
Peter Y. Woo
Peter Y. Woo considered the Schoch line, and with it, he was able to create a family of infinitely many Archimedean circles known as the Woo circles.
Frank Power
In the summer of 1998, Frank Power introduced four more Archimedes circles known as Archimedes' quadruplets.
Archimedean circles in Wasan geometry (Japanese geometry)
In 1831, Nagata (永田岩三郎遵道) proposed a sangaku problem involving two Archimedean circles, which are denoted by W6 and W7 in [3].
In 1853, Ootoba (大鳥羽源吉守敬) proposed a sangaku problem involving an Archimedean circle. |
https://en.wikipedia.org/wiki/Phi-hiding%20assumption | The phi-hiding assumption or Φ-hiding assumption is an assumption about the difficulty of finding small factors of φ(m) where m is a number whose factorization is unknown, and φ is Euler's totient function. The security of many modern cryptosystems comes from the perceived difficulty of certain problems. Since P vs. NP problem is still unresolved, cryptographers cannot be sure computationally intractable problems exist. Cryptographers thus make assumptions as to which problems are hard. It is commonly believed that if m is the product of two large primes, then calculating φ(m) is currently computationally infeasible; this assumption is required for the security of the RSA Cryptosystem. The Φ-Hiding assumption is a stronger assumption, namely that if p1 and p2 are small primes exactly one of which divides φ(m), there is no polynomial-time algorithm which can distinguish which of the primes p1 and p2 divides φ(m) with probability significantly greater than one-half.
This assumption was first stated in the 1999 paper Computationally Private Information Retrieval with Polylogarithmic Communication, where it was used in a Private Information Retrieval scheme.
Applications
The Phi-hiding assumption has found applications in the construction of a few cryptographic primitives. Some of the constructions include:
Computationally Private Information Retrieval with Polylogarithmic Communication (1999)
Efficient Private Bidding and Auctions with an Oblivious Third Party (1999)
Single-Database Private Information Retrieval with Constant Communication Rate (2005)
Password authenticated key exchange using hidden smooth subgroups (2005) |
https://en.wikipedia.org/wiki/Inhomogeneous%20electromagnetic%20wave%20equation | In electromagnetism and applications, an inhomogeneous electromagnetic wave equation, or nonhomogeneous electromagnetic wave equation, is one of a set of wave equations describing the propagation of electromagnetic waves generated by nonzero source charges and currents. The source terms in the wave equations make the partial differential equations inhomogeneous, if the source terms are zero the equations reduce to the homogeneous electromagnetic wave equations. The equations follow from Maxwell's equations.
Maxwell's equations
For reference, Maxwell's equations are summarized below in SI units and Gaussian units. They govern the electric field E and magnetic field B due to a source charge density ρ and current density J:
{| class="wikitable" style="text-align: center;"
|-
! scope="col" style="width: 15em;" | Name
! scope="col" | SI units
! scope="col" | Gaussian units
|-
! scope="row" | Gauss's law
|
|
|-
! scope="row" | Gauss's law for magnetism
|
|
|-
! scope="row" | Maxwell–Faraday equation (Faraday's law of induction)
|
|
|-
! scope="row" | Ampère's circuital law (with Maxwell's addition)
|
|
|-
|}
where ε0 is the vacuum permittivity and μ0 is the vacuum permeability. Throughout, the relation
is also used.
SI units
E and B fields
Maxwell's equations can directly give inhomogeneous wave equations for the electric field E and magnetic field B. Substituting Gauss' law for electricity and Ampère's Law into the curl of Faraday's law of induction, and using the curl of the curl identity (The last term in the right side is the vector Laplacian, not Laplacian applied on scalar functions.) gives the wave equation for the electric field E:
Similarly substituting Gauss's law for magnetism into the curl of Ampère's circuital law (with Maxwell's additional time-dependent term), and using the curl of the curl identity, gives the wave equation for the magnetic field B:
The left hand sides of each equation correspond to wave motion (the D'Alembert operator act |
https://en.wikipedia.org/wiki/Burkholderiales-2%20RNA%20motif | The Burkholderiales-2 RNA motif is a conserved RNA structure that was discovered by bioinformatics.
Burkholderiales-2 motifs are found in Betaproteobacteria. Although one example is predicted in the phylum Bacteroidota, this is likely to be the result of a recent horizontal gene transfer or sequence contamination.
Burkholderiales-2 RNAs likely function in trans as sRNAs. There is weak evidence of an association with S24 peptidases. |
https://en.wikipedia.org/wiki/Cannon%20Lake%20%28microprocessor%29 | Cannon Lake (formerly Skymont) is Intel's codename for the 10 nm die shrink of the Kaby Lake microarchitecture. As a die shrink, Cannon Lake is a new process in Intel's process-architecture-optimization execution plan as the next step in semiconductor fabrication. Cannon Lake CPUs are the first mainstream CPUs to include the AVX-512 instruction set.
Prior to Cannon Lake's launch, Intel launched another 14 nm process refinement with the codename Coffee Lake.
The successor of Cannon Lake is Ice Lake, powered by the Sunny Cove microarchitecture, which represents the architecture phase in the process-architecture-optimization model.
Design history and features
Cannon Lake was initially expected to be released in 2015/2016, but the release was pushed back to 2018. Intel demonstrated a laptop with an unknown Cannon Lake CPU at CES 2017 and announced that Cannon Lake based products would be available in 2018 at the earliest.
At CES 2018 Intel announced that it had started shipping mobile Cannon Lake CPUs at the end of 2017 and would ramp up production in 2018.
On April 26, 2018 in its report on first-quarter 2018 financial results, Intel stated it was currently shipping low-volume 10 nm product and expects 10 nm volume production to shift to 2019. In July 2018, Intel announced that volume production of Cannon Lake would be delayed yet again, to late Q2 2019.
The first laptop featuring a Cannon Lake CPU, namely Intel Core i3-8121U, a dual core CPU with Hyper-Threading and Turbo Boost but without an integrated GPU, was released in May 2018 in very limited quantities.
On August 16, 2018 Intel announced two new models of NUCs would use the 10 nm Cannon Lake-U i3-8121U CPU. These models later became more readily available at retail in late November 2018.
On October 28, 2019, Intel announced that it will be discontinuing the i3-8121U and the Cannon Lake-powered Crimson Canyon NUC, with orders being taken till December 27, and shipping till February 28, 2020, making Can |
https://en.wikipedia.org/wiki/Croatian%20Operational%20Research%20Society | The Croatian Operational Research Society (CRORS) is the professional non-profit society for the scientific field of Operations Research in Croatia. Its main mission is to promote operational research in Croatia and worldwide for the benefit of science and society. Since 1994, CRORS has been recognized by the International Federation of Operational Research Societies and its subgrouping, the Association of European Operational Research Societies, as the main national society for Operations Research in Croatia, and CRORS actively participates in international promotion of operational research.
History
The Croatian Operational Research Society was established in 1992. as the only scientific society in Croatia specialized in operational research. Two groups of OR people, one from Zagreb and the other from Split, were active in organizing the OR Conference KOI ‘91 and establishing the CRORS, together with colleagues from Varaždin, Rijeka, Ljubljana and Maribor. In March 1992, The Founding meeting of CRORS was held at the Faculty of economics in Zagreb. The Statutes of CRORS were acquired at that meeting, and the Board of directors with a Supervisory board was elected. The first president was Prof. Luka Neralić from the University of Zagreb, Faculty of Economics Zagreb.
Governance
The main board of the society consists of a President, Vice President, Secretary and Treasurer. The society is managed by an Executive Committee consisting of 17 members; the Supervisory Board and Board of Ethics consists of 3 members. All members of the boards are elected by the membership.
Membership
, the society has 150 members - individuals and institutions from academia, industry and administration.
Publications
Scientific journal
Croatian Operational Research Review (CRORR), (Print), (Online) is published by Croatian Operational Research Society and Co-publishers: Faculty of Economics in Osijek (University of Osijek), Department of Mathematics (University of Osijek), Faculty of |
https://en.wikipedia.org/wiki/Antioxidant | Antioxidants are compounds that inhibit oxidation (usually occurring as autoxidation), a chemical reaction that can produce free radicals. Autoxidation leads to degradation of organic compounds, including living matter. Antioxidants are frequently added to industrial products, such as polymers, fuels, and lubricants, to extend their usable lifetimes. Food are also treated with antioxidants to forestall spoilage, in particular the rancidification of oils and fats. In cells, antioxidants such as glutathione, mycothiol or bacillithiol, and enzyme systems like superoxide dismutase, can prevent damage from oxidative stress.
Known dietary antioxidants are vitamins A, C, and E, but the term antioxidant has also been applied to numerous other dietary compounds that only have antioxidant properties in vitro, with little evidence for antioxidant properties in vivo. Dietary supplements marketed as antioxidants have not been shown to maintain health or prevent disease in humans.
History
As part of their adaptation from marine life, terrestrial plants began producing non-marine antioxidants such as ascorbic acid (vitamin C), polyphenols and tocopherols. The evolution of angiosperm plants between 50 and 200 million years ago resulted in the development of many antioxidant pigments – particularly during the Jurassic period – as chemical defences against reactive oxygen species that are byproducts of photosynthesis. Originally, the term antioxidant specifically referred to a chemical that prevented the consumption of oxygen. In the late 19th and early 20th centuries, extensive study concentrated on the use of antioxidants in important industrial processes, such as the prevention of metal corrosion, the vulcanization of rubber, and the polymerization of fuels in the fouling of internal combustion engines.
Early research on the role of antioxidants in biology focused on their use in preventing the oxidation of unsaturated fats, which is the cause of rancidity. Antioxidant activit |
https://en.wikipedia.org/wiki/Campbell%27s%20theorem%20%28geometry%29 | Campbell's theorem, named after John Edward Campbell, also known as Campbell’s embedding theorem and the Campbell-Magaard theorem, is a mathematical theorem guaranteeing that any n-dimensional Riemannian manifold can be locally embedded in an (n + 1)-dimensional Ricci-flat Riemannian manifold.
Statement
Campbell's theorem states that any n-dimensional Riemannian manifold can be embedded locally in an (n + 1)-manifold with a Ricci curvature of R'a b = 0. The theorem also states, in similar form, that an n-dimensional pseudo-Riemannian manifold can be both locally and isometrically embedded in an n(n + 1)/2-pseudo-Euclidean space.
Applications
Campbell’s theorem can be used to produce the embedding of numerous 4-dimensional spacetimes in 5-dimensional Ricci-flat spaces. It is also used to embed a class of n-dimensional Einstein spaces. |
https://en.wikipedia.org/wiki/Dorset%20Biological%20Warfare%20Experiments | The Dorset Biological Warfare Experiments were a series of experiments conducted between 1953 and 1975 to determine the extent to which a single ship or aircraft could dispense biological warfare agents over the United Kingdom. The tests between 1971 and 1975 were known as the DICE trials. The tests were conducted by scientists from Porton Down, initially using zinc cadmium sulfide (ZnCds) as a simulated agent. Early results clearly showed that one aircraft flying along the coast while spraying its agent could contaminate a target over 100 miles away, over an area of 10,000 square miles. This method of biological warfare attack and the test program to study it was known as the Large Area Coverage (LAC) concept.
In the early 1960s, Porton Down was asked to expand the scope of their tests to determine if using a live bacterium instead of ZnCds would significantly alter the results. Scientists from Microbiological Research Establishment at Porton Down selected South Dorset as the site for this next phase of testing, with Bacillus subtilis (also known as Bacillus globigii or BG) selected as the test agent.
This bacterium was sprayed across South Dorset without the knowledge or consent of the inhabitants.
Similar tests
In Operation Sea-Spray, unsuspecting inhabitants of the San Francisco Bay Area were sprayed with Serratia marcescens and Bacillus globigii, pathogens that were then implicated in some unusual outbreaks of illness, including pneumonia, and urinary tract infections, and even some deaths.
In Senate subcommittee hearings in 1977, the US Army revealed:
Between 1949 and 1969 open-air tests of biological agents were conducted 239 times. In 80 of those experiments, the Army said it used live bacteria that its researchers at the time thought were harmless. In the others, it used inert chemicals to simulate bacteria.
In the 1950s army researchers dispersed Serratia on Panama City and Key West Florida with no known illnesses resulting.
In the 1950s army re |
https://en.wikipedia.org/wiki/Echo%20state%20network | An echo state network (ESN) is a type of reservoir computer that uses a recurrent neural network with a sparsely connected hidden layer (with typically 1% connectivity). The connectivity and weights of hidden neurons are fixed and randomly assigned. The weights of output neurons can be learned so that the network can produce or reproduce specific temporal patterns. The main interest of this network is that although its behaviour is non-linear, the only weights that are modified during training are for the synapses that connect the hidden neurons to output neurons. Thus, the error function is quadratic with respect to the parameter vector and can be differentiated easily to a linear system.
Alternatively, one may consider a nonparametric Bayesian formulation of the output layer, under which: (i) a prior distribution is imposed over the output weights; and (ii) the output weights are marginalized out in the context of prediction generation, given the training data. This idea has been demonstrated in by using Gaussian priors, whereby a Gaussian process model with ESN-driven kernel function is obtained. Such a solution was shown to outperform ESNs with trainable (finite) sets of weights in several benchmarks.
Some publicly available implementations of ESNs are: (i) aureservoir: an efficient C++ library for various kinds of echo state networks with python/numpy bindings; (ii) Matlab code: an efficient matlab for an echo state network; (iii) ReservoirComputing.jl: an efficient Julia-based implementation of various types of echo state networks; and (iv) pyESN: simple echo state networks in Python.
Background
The Echo State Network (ESN) belongs to the Recurrent Neural Network (RNN) family and provide their architecture and supervised learning principle. Unlike Feedforward Neural Networks, Recurrent Neural Networks are dynamic systems and not functions. Recurrent Neural Networks are typically used for: Learn dynamical process: signal treatment in engineering and teleco |
https://en.wikipedia.org/wiki/Cray%20Time%20Sharing%20System | The Cray Time Sharing System, also known in the Cray user community as CTSS, was developed as an operating system for the Cray-1 or Cray X-MP line of supercomputers in 1978. CTSS was developed by the Los Alamos Scientific Laboratory (LASL now LANL) in conjunction with the Lawrence Livermore Laboratory (LLL now LLNL). CTSS was popular with Cray sites in the United States Department of Energy (DOE), but was used by several other Cray sites, such as the San Diego Supercomputing Center.
Overview
The predecessor of CTSS was the Livermore Time Sharing System (LTSS) which ran on Control Data CDC 7600 line of supercomputers. The first compiler was known as LRLTRAN, for Lawrence Radiation Laboratory forTRAN, a FORTRAN 66 programming language but with dynamic memory and other features. The Cray version, including automatic vectorization, was known as CVC, pronounced "Civic" like the Honda car of the period, for Cray Vector Compiler.
Some controversy existed at LASL with the first attempt to develop an operating system for the Cray-1 named DEIMOS, a message-passing, Unix-like operating system, by Forrest Basket. DEIMOS had initial "teething" problems common to the performance of all early operating systems. This left a bad taste for Unix-like systems at the National Laboratories and with the manufacturer, Cray Research, Inc., of the hardware who went on to develop their own batch oriented operating system, COS (Cray Operating System) and their own vectorizing Fortran compiler named "CFT" (Cray ForTran) both written in the Cray Assembly Language (CAL).
CTSS had the misfortune to have certain constants, structures, and lacking certain networking facilities (TCP/IP) which were optimized to be Cray-1 architecture-dependent without extensive rework when larger memory supercomputers like the Cray-2 and the Cray Y-MP came into use. CTSS has its final breaths running on Cray instruction-set-compatible hardware developed by Scientific Computer Systems (SCS-40 and SCS-30) and Supert |
https://en.wikipedia.org/wiki/Murid%20betaherpesvirus%203 | Murid betaherpesvirus 3 (MuHV-3) is a species of virus in the genus Roseolovirus, subfamily Betaherpesvirinae, family Herpesviridae, and order Herpesvirales. |
https://en.wikipedia.org/wiki/Ethanethiol | Ethanethiol, commonly known as ethyl mercaptan, is an organosulfur compound with the formula CH3CH2SH. is a colorless liquid with a distinct odor. Abbreviated EtSH, it consists of an ethyl group (Et), CH3CH2, attached to a thiol group, SH. Its structure parallels that of ethanol, but with sulfur in place of oxygen. The odor of EtSH is infamous. Ethanethiol is more volatile than ethanol due to a diminished ability to engage in hydrogen bonding. Ethanethiol is toxic in high concentrations. It occurs naturally as a minor component of petroleum, and may be added to otherwise odorless gaseous products such as liquefied petroleum gas (LPG) to help warn of gas leaks. At these concentrations, ethanethiol is not harmful.
Preparation
Ethanethiol is prepared by the reaction of ethylene with hydrogen sulfide in the presence of various catalysts. It is also prepared commercially by the reaction of ethanol with hydrogen sulfide gas over an acidic solid catalyst, such as alumina.
Historic methods
Ethanethiol was originally reported by Zeise in 1834. Zeise treated calcium ethyl sulfate with a suspension of barium sulfide saturated with hydrogen sulfide. He is credited with naming the C2H5S- group as mercaptum.
Ethanethiol can also be prepared by a halide displacement reaction, where ethyl halide is treated with aqueous sodium bisulfide. This conversion was demonstrated as early as 1840 by Henri Victor Regnault.
Odor
Ethanethiol has a strongly disagreeable odor that humans can detect in minute concentrations. The threshold for human detection is as low as one part in 2.8 billion parts of air (0.36 parts per billion). Its odor resembles that of leeks, onions, durian or cooked cabbage.
Employees of the Union Oil Company of California reported first in 1938 that turkey vultures would gather at the site of any gas leak. After finding that this was caused by traces of ethanethiol in the gas it was decided to boost the amount of ethanethiol in the gas, to make detection of leaks |
https://en.wikipedia.org/wiki/Injective%20sheaf | In mathematics, injective sheaves of abelian groups are used to construct the resolutions needed to define sheaf cohomology (and other derived functors, such as sheaf Ext).
There is a further group of related concepts applied to sheaves: flabby (flasque in French), fine, soft (mou in French), acyclic. In the history of the subject they were introduced before the 1957 "Tohoku paper" of Alexander Grothendieck, which showed that the abelian category notion of injective object sufficed to found the theory. The other classes of sheaves are historically older notions. The abstract framework for defining cohomology and derived functors does not need them. However, in most concrete situations, resolutions by acyclic sheaves are often easier to construct. Acyclic sheaves therefore serve for computational purposes, for example the Leray spectral sequence.
Injective sheaves
An injective sheaf is a sheaf that is an injective object of the category of abelian sheaves; in other words, homomorphisms from to can always be extended to any sheaf containing
The category of abelian sheaves has enough injective objects: this means that any sheaf is a subsheaf of an injective sheaf. This result of Grothendieck follows from the existence of a generator of the category (it can be written down explicitly, and is related to the subobject classifier). This is enough to show that right derived functors of any left exact functor exist and are unique up to canonical isomorphism.
For technical purposes, injective sheaves are usually superior to the other classes of sheaves mentioned above: they can do almost anything the other classes can do, and their theory is simpler and more general. In fact, injective sheaves are flabby (flasque), soft, and acyclic. However, there are situations where the other classes of sheaves occur naturally, and this is especially true in concrete computational situations.
The dual concept, projective sheaves, is not used much, because in a general category |
https://en.wikipedia.org/wiki/Bernoulli%20polynomials%20of%20the%20second%20kind | The Bernoulli polynomials of the second kind , also known as the Fontana-Bessel polynomials, are the polynomials defined by the following generating function:
The first five polynomials are:
Some authors define these polynomials slightly differently
so that
and may also use a different notation for them (the most used alternative notation is ). Under this convention, the polynomials form a Sheffer sequence.
The Bernoulli polynomials of the second kind were largely studied by the Hungarian mathematician Charles Jordan, but their history may also be traced back to the much earlier works.
Integral representations
The Bernoulli polynomials of the second kind may be represented via these integrals
as well as
These polynomials are, therefore, up to a constant, the antiderivative of the binomial coefficient and also that of the falling factorial.
Explicit formula
For an arbitrary , these polynomials may be computed explicitly via the following summation formula
where are the signed Stirling numbers of the first kind and are the Gregory coefficients.
The expansion of the Bernoulli polynomials of the second kind into a Newton series reads
It can be shown using the second integral representation and Vandermonde's identity.
Recurrence formula
The Bernoulli polynomials of the second kind satisfy the recurrence relation
or equivalently
The repeated difference produces
Symmetry property
The main property of the symmetry reads
Some further properties and particular values
Some properties and particular values of these polynomials include
where are the Cauchy numbers of the second kind and are the central difference coefficients.
Some series involving the Bernoulli polynomials of the second kind
The digamma function may be expanded into a series with the Bernoulli polynomials of the second kind
in the following way
and hence
and
where is Euler's constant. Furthermore, we also have
where is the gamma function. The Hurwitz and Riemann zeta f |
https://en.wikipedia.org/wiki/WikiVet | WikiVet is a wiki of veterinary content based on the MediaWiki platform. The website is a collaborative initiative between various veterinary schools, and its content covers the entire veterinary curriculum. WikiVet is part of the WikiVet Educational Foundation (UK registered charity number 1160546).
Full access to WikiVet requires a free registration, which is available to veterinarians, veterinary students and veterinary technicians. Except for content relating specifically to the veterinary curriculum, articles are authored by students or veterinarians, and subsequently peer reviewed by subject specialists.
History
WikiVet was established in 2007 to provide online access to a comprehensive veterinary undergraduate curriculum. The consortium was initially formed by three UK veterinary schools, London's Royal Veterinary College, the University of Edinburgh's Royal (Dick) School of Veterinary Studies, and the University of Cambridge's Department of Veterinary Medicine, and was subsequently joined by the University of Nottingham's School of Veterinary Medicine and Science.
WikiVet was initially funded by the Higher Education Academy (HEA) and Joint Information Systems Committee (JISC). Subsequent commercial sponsors include Mars Petcare, RCVS Trust, Pfizer Global Alliances, Ceva, and the Donkey Sanctuary.
In 2010, WikiVet had 6,000 registered users. A Spanish language version was launched in May 2011. A survey of first year veterinary students in Spain, conducted when only the English-language version of WikiVet was available, found that 9% had used WikiVet. In survey of veterinarians and veterinary students in Germany, reported in 2013, 8% of respondents used WikiVet, compared to 96% who used Wikipedia. In 2016, WikiVet had 50,000 registered users from more than 130 countries. |
https://en.wikipedia.org/wiki/Brown%E2%80%93Rho%20scaling | In quantum chromodynamics (QCD), Brown–Rho (BR) scaling is an approximate scaling law for hadrons in an ultra-hot, ultra-dense medium, such as hadrons in the quark epoch during the first microsecond of the Big Bang or within neutron stars.
According to Gerald E. Brown and Mannque Rho in their 1991 publication in Physical Review Letters:
refers to the pole mass of the ρ meson, whereas refers to the in-medium mass (or running mass in the medium) of the ρ meson according to QCD sum rules. The omega meson, sigma meson, and neutron are denoted by
ω, σ, and N, respectively. The symbol denotes the free-space pion decay constant. (Decay constants have a "running time" and a "pole time" similar to the "running mass" and "pole mass" concepts, according to special relativity.) The symbol is also used to denote the pion decay constant.
The hypothesis of Brown–Rho scaling is supported by experimental evidence on beta decay of 14C to the 14N ground state.
See also
Quantum chromodynamics
QCD matter
Pion decay constant |
https://en.wikipedia.org/wiki/Bi-wiring | Bi-wiring is a means of connecting a :loudspeaker to an audio :amplifier, primarily used in hi-fi systems. Normally, there is one pair of connectors on a loudspeaker and a single cable (two conductors) runs from the amplifier output to the terminals at the loudspeaker housing. From this point, connections are made to the loudspeaker drivers – usually through audio crossover networks.
In bi-wiring, each loudspeaker has two pairs of connectors and two cables are run from the same amplifier output to the speaker cabinet: one for the high frequency or tweeter driver and one for the low-frequency driver (through two separated crossover filters). The purported advantage of this split is that it "reduces magnetic interaction in the cable, resulting in better sound". However, technical analysis suggests that while bi-wired arrangements may be expected to have differences from single wired ones, these differences would normally be so small as to have little significance.
Some audiophiles feel that bi-wiring produces an audible improvement over standard single cabling. For example, John Atkinson, writing in Stereophile, states that he observes "subtle but important" differences, particularly in reduction of treble hardness and improvement in bass control in one review.
Critics of bi-wiring believe that both ways of making speaker connections are electrically equivalent (assuming no difference in speaker cable resistance), and thus cynically refer to the practice as "buy-wiring", implying it is nothing more than a marketing gimmick for buying more pairs of speaker wires.
Bi-wiring should not be confused with the hi-fi practice of bi-amping: the use of a separate amplifier for each driver, which brings improved separation of signal frequencies and removes the need for passive crossovers and the degraded efficiency, linearity, and cost that comes with them. |
https://en.wikipedia.org/wiki/Intensity%20%28physics%29 | In physics, the intensity or flux of radiant energy is the power transferred per unit area, where the area is measured on the plane perpendicular to the direction of propagation of the energy. In the SI system, it has units watts per square metre (W/m2), or kg⋅s−3 in base units. Intensity is used most frequently with waves such as acoustic waves (sound) or electromagnetic waves such as light or radio waves, in which case the average power transfer over one period of the wave is used. Intensity can be applied to other circumstances where energy is transferred. For example, one could calculate the intensity of the kinetic energy carried by drops of water from a garden sprinkler.
The word "intensity" as used here is not synonymous with "strength", "amplitude", "magnitude", or "level", as it sometimes is in colloquial speech.
Intensity can be found by taking the energy density (energy per unit volume) at a point in space and multiplying it by the velocity at which the energy is moving. The resulting vector has the units of power divided by area (i.e., surface power density). The intensity of a wave is proportional to the square of its amplitude. For example, the intensity of an electromagnetic wave is proportional to the square of the wave's electric field amplitude.
Mathematical description
If a point source is radiating energy in all directions (producing a spherical wave), and no energy is absorbed or scattered by the medium, then the intensity decreases in proportion to the distance from the object squared. This is an example of the inverse-square law.
Applying the law of conservation of energy, if the net power emanating is constant,
where
is the net power radiated;
is the intensity vector as a function of position;
the magnitude is the intensity as a function of position;
is a differential element of a closed surface that contains the source.
If one integrates a uniform intensity, , over a surface that is perpendicular to the intensity vector, for insta |
https://en.wikipedia.org/wiki/List%20of%20office%20suites | In computing, an office suite is a collection of productivity software usually containing at least a word processor, spreadsheet and a presentation program. There are many different brands and types of office suites. This wikipedia article is unique for its list of discontinued office suites.
Office suites
Free and open source suites
AndrOpen Office - available for Android
Apache OpenOffice - available for Linux, macOS and Windows
Calligra Suite - available for FreeBSD, Linux, macOS and Windows
Collabora Online - available for Android, ChromeOS, iOS, iPadOS, Linux, macOS, online and Windows
LibreOffice - available for Linux, macOS and Windows, and unofficial: Android, ChromeOS, FreeBSD, Haiku, iOS, iPadOS, OpenBSD, NetBSD and Solaris
NeoOffice - available for macOS
Freeware and proprietary suites
Ability Office - available for Windows
Google Workspace - available for Android, ChromeOS, iOS, iPadOS, Linux, macOS, online and Windows
Hancom Office - available for Windows
iWork - available for iOS, iPadOS, macOS and online
Ichitaro - a Japanese-language suite available for Windows
Microsoft 365 - available for Android, iOS, iPadOS, macOS, online and Windows
MobiSystems OfficeSuite - available for Android, iOS and Windows
ONLYOFFICE - available for Android, iOS, Linux, macOS, online and Windows
Polaris Office - available for iOS, macOS and Windows
SoftMaker Office - available for Android, iOS, iPadOS, Linux, macOS and Windows
Tiki Wiki CMS Groupware - online content management
WordPerfect Office - available for Windows
WPS Office - available for Android, iOS, macOS and Windows
Discontinued office suites
Aster*x
AUIS - an office suite developed by Carnegie Mellon University and named after Andrew Carnegie
Breadbox Office - DOS software
EasyOffice
AppleWorks
Breadbox Office
Corel WordPerfect for DOS
Hancom Office Suite (formerly ThinkFree Office)
IBM Lotus SmartSuite
IBM Lotus Symphony
IBM Works – an office suite for the IBM OS/2 operating |
https://en.wikipedia.org/wiki/Process | A process is a series or set of activities that interact to produce a result; it may occur once-only or be recurrent or periodic.
Things called a process include:
Business and management
Business process, activities that produce a specific service or product for customers
Business process modeling, activity of representing processes of an enterprise in order to deliver improvements
Manufacturing process management, a collection of technologies and methods used to define how products are to be manufactured.
Process architecture, structural design of processes, applies to fields such as computers, business processes, logistics, project management
Process area, related processes within an area which together satisfies an important goal for improvements within that area
Process costing, a cost allocation procedure of managerial accounting
Process management (project management), a systematic series of activities directed towards planning, monitoring the performance and causing an end result in engineering activities, business process, manufacturing processes or project management
Process-based management, is a management approach that views a business as a collection of processes
Law
Due process, the concept that governments must respect the rule of law
Legal process, the proceedings and records of a legal case
Service of process, the procedure of giving official notice of a legal proceeding
Science and technology
The general concept of the scientific process, see scientific method
Process theory, the scientific study of processes
Industrial processes, consists of the purposeful sequencing of tasks that combine resources to produce a desired output
Biology and psychology
Process (anatomy), a projection or outgrowth of tissue from a larger body
Biological process, a process of a living organism
Cognitive process, such as attention, memory, language use, reasoning, and problem solving
Mental process, a function or processes of the mind
Neuronal process, also neurite |
https://en.wikipedia.org/wiki/Quantum%20dilogarithm | In mathematics, the quantum dilogarithm is a special function defined by the formula
It is the same as the q-exponential function .
Let be "q-commuting variables", that is elements of a suitable noncommutative algebra satisfying Weyl's relation . Then, the quantum dilogarithm satisfies Schützenberger's identity
Faddeev-Volkov's identity
and Faddeev-Kashaev's identity
The latter is known to be a quantum generalization of Rogers' five term dilogarithm identity.
Faddeev's quantum dilogarithm is defined by the following formula:
where the contour of integration goes along the real axis outside a small neighborhood of the origin and deviates into the upper half-plane near the origin. The same function can be described by the integral formula of Woronowicz:
Ludvig Faddeev discovered the quantum pentagon identity:
where and are self-adjoint (normalized) quantum mechanical momentum and position operators satisfying Heisenberg's commutation relation
and the inversion relation
The quantum dilogarithm finds applications in mathematical physics, quantum topology, cluster algebra theory.
The precise relationship between the q-exponential and is expressed by the equality
valid for . |
https://en.wikipedia.org/wiki/Negation | In logic, negation, also called the logical not or logical complement, is an operation that takes a proposition to another proposition "not ", standing for " is not true", written , or . It is interpreted intuitively as being true when is false, and false when is true. Negation is thus a unary logical connective. It may be applied as an operation on notions, propositions, truth values, or semantic values more generally. In classical logic, negation is normally identified with the truth function that takes truth to falsity (and vice versa). In intuitionistic logic, according to the Brouwer–Heyting–Kolmogorov interpretation, the negation of a proposition is the proposition whose proofs are the refutations of .
Definition
Classical negation is an operation on one logical value, typically the value of a proposition, that produces a value of true when its operand is false, and a value of false when its operand is true. Thus if statement is true, then (pronounced "not P") would then be false; and conversely, if is true, then would be false.
The truth table of is as follows:
{| class="wikitable" style="text-align:center; background-color: #ddffdd;"
|- bgcolor="#ddeeff"
| ||
|-
| ||
|-
| ||
|}
Negation can be defined in terms of other logical operations. For example, can be defined as (where is logical consequence and is absolute falsehood). Conversely, one can define as for any proposition (where is logical conjunction). The idea here is that any contradiction is false, and while these ideas work in both classical and intuitionistic logic, they do not work in paraconsistent logic, where contradictions are not necessarily false. In classical logic, we also get a further identity, can be defined as , where is logical disjunction.
Algebraically, classical negation corresponds to complementation in a Boolean algebra, and intuitionistic negation to pseudocomplementation in a Heyting algebra. These algebras provide a semantics for classical and i |
https://en.wikipedia.org/wiki/C%C3%A0dl%C3%A0g | In mathematics, a càdlàg (French: "continue à droite, limite à gauche"), RCLL ("right continuous with left limits"), or corlol ("continuous on (the) right, limit on (the) left") function is a function defined on the real numbers (or a subset of them) that is everywhere right-continuous and has left limits everywhere. Càdlàg functions are important in the study of stochastic processes that admit (or even require) jumps, unlike Brownian motion, which has continuous sample paths. The collection of càdlàg functions on a given domain is known as Skorokhod space.
Two related terms are càglàd, standing for "continue à gauche, limite à droite", the left-right reversal of càdlàg, and càllàl for "continue à l'un, limite à l’autre" (continuous on one side, limit on the other side), for a function which at each point of the domain is either càdlàg or càglàd.
Definition
Let be a metric space, and let . A function is called a càdlàg function if, for every ,
the left limit exists; and
the right limit exists and equals .
That is, is right-continuous with left limits.
Examples
All functions continuous on a subset of the real numbers are càdlàg functions on that subset.
As a consequence of their definition, all cumulative distribution functions are càdlàg functions. For instance the cumulative at point correspond to the probability of being lower or equal than , namely . In other words, the semi-open interval of concern for a two-tailed distribution is right-closed.
The right derivative of any convex function defined on an open interval, is an increasing cadlag function.
Skorokhod space
The set of all càdlàg functions from to is often denoted by (or simply ) and is called Skorokhod space after the Ukrainian mathematician Anatoliy Skorokhod. Skorokhod space can be assigned a topology that, intuitively allows us to "wiggle space and time a bit" (whereas the traditional topology of uniform convergence only allows us to "wiggle space a bit"). For simplicity, take |
https://en.wikipedia.org/wiki/Jaikoz | Jaikoz is a Java program used for editing and mass tagging music file tags.
Jaikoz generates acoustic fingerprints from music files using the AcoustId service, it can then look up the metadata from MusicBrainz using the AcoustId, additionally it can match based on metadata to MusicBrainz or Discogs. Matching is first applied at album level, falling back to track level where a match at album level could not be made. This allows Jaikoz to automatically fix most of a users song collection.
Jaikoz uses a relatively unusual spreadsheet metaphor for both viewing and editing data, and allows editing of over fifty fields using this spreadsheet interface, the underlying jaudiotagger tag library is released under LGPL and is used by various Java applications.
Jaikoz is commercially licensed software, written in Java 1.5 by Paul Taylor. A shareware version, in which changes can only be saved to 20 files during one use, is also available as a 30-day free trial. 10% of every sale is paid to the MetaBrainz Foundation to support MusicBrainz development.
History
Originally released in 2006 as a standalone music tagger without any MusicBrainz support, but support for MusicBrainz was soon added. Changes in Jaikoz have always reflected changes in MusicBrainz, for example Jaikoz was the first application to make use of the new web service released as part of the MusicBrainz NGS release in 2011, and the first application to use the MusicBrainz seeding mechanism for adding new releases.
Summary of features
Acoustic matching using MusicBrainz and AcoustId to match songs based on the actual music
MetaData matching using MusicBrainz and Discogs to match tracks from the metadata in your files either automatically or manually
Fixes artwork.
Supports Multiple Audio formats and different audio formats can be edited the same easy way
Export/Import metadata to/from a spreadsheet.
Delete Duplicates based on Musicbrainz Id and/or Acoustic Fingerprint
Find And Replace feature that can |
https://en.wikipedia.org/wiki/Glossary%20of%20scientific%20naming | This is a list of terms and symbols used in scientific names for organisms, and in describing the names. For proper parts of the names themselves, see List of Latin and Greek words commonly used in systematic names. Note that many of the abbreviations are used with or without a stop.
Naming standards and taxonomic organizations and their codes and taxonomies
ICTV – International Committee on Taxonomy of Viruses
ICSP – International Committee on Systematics of Prokaryotes
formerly the ICSB – International Committee on Systematic Bacteriology
publishes the ICNP – International Code of Nomenclature of Prokaryotes
formerly the International Code of Nomenclature of Bacteria (ICNB) or Bacteriological Code (BC)
ICZN – International Commission on Zoological Nomenclature
publishes ICZN the International Code of Zoological Nomenclature or "ICZN Code"
IBC – International Botanical Congress
publishes ICN the International Code of Nomenclature for algae, fungi, and plants
formerly ICBN or the International Code of Botanical Nomenclature (current version the Shenzhen Code)
also publishes ICNCP or the International Code of Nomenclature for Cultivated Plants
IAPT – International Association for Plant Taxonomy
publishes Taxon
also publishes Regnum Vegetabile which contains the IBC's ICN, the Index Nominum Genericorum, and the Index Herbariorum
General terms
clade, cladistics
phylum, phylogeny
taxon, taxonomy; Taxon is a journal of the IAPT, where proposals are made
synonym: a name for a taxon different from the currently accepted name
pro parte (abbreviation p. p.; "for part" in Latin)
senior synonym, (zoology): the earliest (correctly published) name
junior synonym, (zoology): any later name
homotypic synonym (botany)
heterotypic synonym (botany): (or taxonomic synonym) a synonym that comes into being when a taxon is reduced in status ("reduced to synonymy") and becomes part of a different taxon; the zoological equivalent is "subjective synonym"
objecti |
https://en.wikipedia.org/wiki/Takeda%20Oncology | Takeda Oncology (originally Millennium Pharmaceuticals) is a biopharmaceutical company based in Cambridge, Massachusetts. It is a fully owned subsidiary of Takeda Pharmaceutical.
Takeda Oncology's research, development and commercialization activities focused in two therapeutic areas: oncology and inflammation to develop a line of new product candidates. It was one of the first companies to systematically search for genes linked to disease, although none of the drugs which it is marketing or has in clinical trial, with one partial exception, have been the results of that research.
It is particularly known for bringing bortezomib (marketed as Velcade) through clinical trials to approval for treatment of patients with multiple myeloma by the U.S. FDA, but has a growing clinical development pipeline of other product candidates.
On May 14, 2008, Japanese company Takeda Pharmaceutical announced the completion of its acquisition of Millennium for US$25.00 per share in cash—a deal worth $8.8 billion. Takeda completed the acquisition through a tender offer and subsequent merger as a wholly owned subsidiary Millennium: The Takeda Oncology Company - with the name being simplified to Takeda Oncology in 2013.
History
Millennium was founded by Mark J. Levin, CEO, in Cambridge, Massachusetts in 1993. In its early years, Millennium focused on building science and business teams.
Beginning in 1994, Millennium created more than 20 strategic alliances with pharmaceutical and biotechnology companies. These alliances provided Millennium with close to $2 billion of committed funding that was used to develop and enhance its pipeline.
A merger with Leukosite in 1999 brought the company its first drug close-to-market, Campath (alemtuzumab) Injection, and additional investigational drugs in clinical trials. In 2000, a merger with Cambridge Discovery Chemistry gave Millennium a strong presence in the United Kingdom and added to the organization more than 100 scientists with expertis |
https://en.wikipedia.org/wiki/Temporal%20bone | The temporal bones are situated at the sides and base of the skull, and lateral to the temporal lobes of the cerebral cortex.
The temporal bones are overlaid by the sides of the head known as the temples, and house the structures of the ears. The lower seven cranial nerves and the major vessels to and from the brain traverse the temporal bone.
Structure
The temporal bone consists of four parts— the squamous, mastoid, petrous and tympanic parts. The squamous part is the largest and most superiorly positioned relative to the rest of the bone. The zygomatic process is a long, arched process projecting from the lower region of the squamous part and it articulates with the zygomatic bone. Posteroinferior to the squamous is the mastoid part. Fused with the squamous and mastoid parts and between the sphenoid and occipital bones lies the petrous part, which is shaped like a pyramid. The tympanic part is relatively small and lies inferior to the squamous part, anterior to the mastoid part, and superior to the styloid process. The styloid, from the Greek stylos, is a phallic shaped pillar directed inferiorly and anteromedially between the parotid gland and internal jugular vein. An elongated or deviated styloid process can result from calcification of the stylohyoid ligament in a condition known as Eagle syndrome.
Borders
Development
The temporal bone is ossified from eight centers, exclusive of those for the internal ear and the tympanic ossicles: one for the squama including the zygomatic process, one for the tympanic part, four for the petrous and mastoid parts, and two for the styloid process. Just before the end of prenatal development [Fig. 6] the temporal bone consists of three principal parts:
The squama is ossified in membrane from a single nucleus, which appears near the root of the zygomatic process about the second month.
The petromastoid part is developed from four centers, which make their appearance in the cartilaginous ear capsule about the fifth or si |
https://en.wikipedia.org/wiki/Quickselect | In computer science, quickselect is a selection algorithm to find the kth smallest element in an unordered list, also known as the kth order statistic. Like the related quicksort sorting algorithm, it was developed by Tony Hoare, and thus is also known as Hoare's selection algorithm. Like quicksort, it is efficient in practice and has good average-case performance, but has poor worst-case performance. Quickselect and its variants are the selection algorithms most often used in efficient real-world implementations.
Quickselect uses the same overall approach as quicksort, choosing one element as a pivot and partitioning the data in two based on the pivot, accordingly as less than or greater than the pivot. However, instead of recursing into both sides, as in quicksort, quickselect only recurses into one side – the side with the element it is searching for. This reduces the average complexity from to , with a worst case of .
As with quicksort, quickselect is generally implemented as an in-place algorithm, and beyond selecting the th element, it also partially sorts the data. See selection algorithm for further discussion of the connection with sorting.
Algorithm
In quicksort, there is a subprocedure called partition that can, in linear time, group a list (ranging from indices left to right) into two parts: those less than a certain element, and those greater than or equal to the element. Here is pseudocode that performs a partition about the element list[pivotIndex]:
function partition(list, left, right, pivotIndex) is
pivotValue := list[pivotIndex]
swap list[pivotIndex] and list[right] // Move pivot to end
storeIndex := left
for i from left to right − 1 do
if list[i] < pivotValue then
swap list[storeIndex] and list[i]
increment storeIndex
swap list[right] and list[storeIndex] // Move pivot to its final place
return storeIndex
This is known as the Lomuto partition scheme, which is simpler but le |
https://en.wikipedia.org/wiki/Science%20Writing%20Award | The American Institute of Physics (AIP) instituted their Science Writing Award to "promote effective science communication in print and broadcast media in order to improve the general public's appreciation of physics, astronomy, and allied science fields." The winner receives $3000, and an engraved Windsor chair. The award is given in three broad categories: 1) science writing, 2) work intended for children, and 3) work done in new media. The AIP stopped issuing awards to three categories: 1) work by a professional journalist (last awarded in 2011) 2) work by a scientist (last awarded in 2009), and 3) broadcast media (last awarded in 2009)
Winners of this Science Writing Award include Nobel Prize winners Charles Townes, Steven Weinberg, and Kip Thorne; other notable winners include Simon Singh, Neil DeGrasse Tyson, Lawrence Krauss, John Wheeler, Leonard Susskind, Clifford Martin Will, Abraham Pais, Heinz Pagels, Banesh Hoffmann, and Martin Gardner. Marcia Bartusiak has won the award three times, twice for her books (in 2019 and 2001) and once for her journalism (in 1982).
Winners: New Media
2012: - Anna Rothschild for Nova, "The Amazing Atomic Clock"
Past Winners: Books
2020: Susan Hockfield for The Age of Living Machines (W.W. Norton & Company).
2019: Marcia Bartusiak for Dispatches from Planet 3 (Yale University Press).
2019: David Hu for How to Walk on Water and Climb Up Walls (Yale University Press).
2018: David Baron for American Eclipse: A Nation's Epic Race to Catch the Shadow of the Moon and Win the Glory of the World (Liveright Publishing Corporation/W. W. Norton & Company).
2017: Timothy Jorgensen for Strange Glow: The Story of Radiation (Princeton University Press).
2016: Chris Woodford for Atoms Under the Floorboards: The Surprising Science Hidden in Your Home (Bloomsbury).
2015: Charles Adler for Wizards, Aliens, and Starships: Physics and Math in Fantasy and Science Fiction (Princeton University Press).
2014: Lee Billings for Five Billion Years of |
https://en.wikipedia.org/wiki/H3K79me2 | H3K79me2 is an epigenetic modification to the DNA packaging protein Histone H3. It is a mark that indicates the di-methylation at the 79th lysine residue of the histone H3 protein. H3K79me2 is detected in the transcribed regions of active genes.
Nomenclature
H3K79me2 indicates dimethylation of lysine 79 on histone H3 protein subunit:
Lysine Methylation
This diagram shows the progressive methylation of a lysine residue. The di-methylation denotes the methylation present in H3K79me2.
Histone modifications
The genomic DNA of eukaryotic cells is wrapped around special protein molecules known as Histones. The complexes formed by the looping of the DNA are known as chromatin. The basic structural unit of chromatin is the nucleosome: this consists of the core octamer of histones (H2A, H2B, H3 and H4) as well as a linker histone and about 180 base pairs of DNA. These core histones are rich in lysine and arginine residues. The carboxyl (C) terminal end of these histones contribute to histone-histone interactions, as well as histone-DNA interactions. The amino (N) terminal charged tails are the site of the post-translational modifications, such as the one seen in H3K36me3.
Epigenetic implications
The post-translational modification of histone tails by either histone modifying complexes or chromatin remodelling complexes are interpreted by the cell and lead to complex, combinatorial transcriptional output. It is thought that a Histone code dictates the expression of genes by a complex interaction between the histones in a particular region. The current understanding and interpretation of histones comes from two large scale projects: ENCODE and the Epigenomic roadmap. The purpose of the epigenomic study was to investigate epigenetic changes across the entire genome. This led to chromatin states which define genomic regions by grouping the interactions of different proteins and/or histone modifications together.
Chromatin states were investigated in Drosophila cells by l |
https://en.wikipedia.org/wiki/Amanda%20Bradford | Dr. Amanda Bradford is a marine mammal biologist who is currently researching cetacean population dynamics for the National Marine Fisheries Service of the National Oceanic and Atmospheric Administration. Bradford is currently a Research Ecologist with the Pacific Islands Fisheries Science Center's Cetacean Research Program. Her research primarily focuses on assessing populations of cetaceans, including evaluating population size, health, and impacts of human-caused threats, such as fisheries interactions. Bradford is a cofounder and organizer of the Women in Marine Mammal Science (WIMMS) Initiative.
Education
Undergraduate education
Bradford received her Bachelor of Science in Marine Biology from Texas A&M University in Galveston, Texas in 1998. She worked in the lab of Bernd Würsig.
While Bradford was an undergraduate, she was a volunteer at the Texas Marine Mammal Stranding Network from 1994 to 1998. Bradford, monitored live stranded delphinids and performed basic husbandry and life-support for bottlenose dolphins and false killer whales. Bradford also participated in marine mammal necropsies.
During her senior year, Bradford began analyzing photo-identification data from the western North Pacific population of gray whales. Shortly after graduation, Bradford traveled to northeastern Sakhalin Island in the Russian Far East to join a collaborative Russia-U.S. field study of these whales on their primary feeding ground. Once Bradford returned from the field, she spent a year as a research assistant for this project based at the Southwest Fisheries Science Center in La Jolla, California.
Graduate education
Bradford attended the University of Washington, School of Aquatic and Fishery Sciences (SAFS) in Seattle, Washington, receiving her Masters of Science in 2003 and then Doctorate of Philosophy (PhD) in 2011. Bradford studied under the late Glenn VanBlaricom for both degrees.
During her time at SAFS, Bradford spent 10 summers in the Russian Far East studying |
https://en.wikipedia.org/wiki/Inverse%20demand%20function | In economics, an inverse demand function is the mathematical relationship that expresses price as a function of quantity demanded (is is therefore also known as a price function).
Historically, the economists first expressed the price of a good as a function of demand (holding the other economic variables, like income, constant), and plotted the price-demand relationship with demand on the x (horizontal) axis (the demand curve). Later the additional variables, like prices of other goods, came into analysis, and it became more convenient to express the demand as a multivariate function (the demand function):
, so the original demand curve now depicts the inverse demand function with extra variables fixed.
Definition
In mathematical terms, if the demand function is , then the inverse demand function is . The value of the inverse demand function is the highest price that could be charged and still generate the quantity demanded. This is useful because economists typically place price (P) on the vertical axis and quantity (demand, Q) on the horizontal axis in supply-and-demand diagrams, so it is the inverse demand function that depicts the graphed demand curve in the way the reader expects to see.
The inverse demand function is the same as the average revenue function, since P = AR.
To compute the inverse demand function, simply solve for P from the demand function. For example, if the demand function has the form then the inverse demand function would be . Note that although price is the dependent variable in the inverse demand function, it is still the case that the equation represents how the price determines the quantity demanded, not the reverse.
Relation to marginal revenue
There is a close relationship between any inverse demand function for a linear demand equation and the marginal revenue function. For any linear demand function with an inverse demand equation of the form P = a - bQ, the marginal revenue function has the form MR = a - 2bQ. The invers |
https://en.wikipedia.org/wiki/Trackbuster | Trackbuster is an online service used to identify and remove email trackers.
Email tracking was mainly used in the beginning by marketers to monitor the delivery of emails. However, new tools now allow individuals to track emails easily, and this practice is on the rise. In addition to revealing when and whether an email has been opened or not, email trackers can also reveal one’s software configuration and geographic location (through IP geolocation). Trackbuster helps email users get rid of these trackers, by detecting and disabling tracking images and tracked links.
At launch, Trackbuster received praise and positive press from well-known figures in the tech world, including Robert Scoble, Om Malik and Chris Messina.
Trackbuster is currently compatible with Gmail and Google Apps. |
https://en.wikipedia.org/wiki/Polytrope | In astrophysics, a polytrope refers to a solution of the Lane–Emden equation in which the pressure depends upon the density in the form
where is pressure, is density and is a constant of proportionality. The constant is known as the polytropic index; note however that the polytropic index has an alternative definition as with n as the exponent.
This relation need not be interpreted as an equation of state, which states P as a function of both ρ and T (the temperature); however in the particular case described by the polytrope equation there are other additional relations between these three quantities, which together determine the equation. Thus, this is simply a relation that expresses an assumption about the change of pressure with radius in terms of the change of density with radius, yielding a solution to the Lane–Emden equation.
Sometimes the word polytrope may refer to an equation of state that looks similar to the thermodynamic relation above, although this is potentially confusing and is to be avoided. It is preferable to refer to the fluid itself (as opposed to the solution of the Lane–Emden equation) as a polytropic fluid. The equation of state of a polytropic fluid is general enough that such idealized fluids find wide use outside of the limited problem of polytropes.
The polytropic exponent (of a polytrope) has been shown to be equivalent to the pressure derivative of the bulk modulus where its relation to the Murnaghan equation of state has also been demonstrated. The polytrope relation is therefore best suited for relatively low-pressure (below 107 Pa) and high-pressure (over 1014 Pa) conditions when the pressure derivative of the bulk modulus, which is equivalent to the polytrope index, is near constant.
Example models by polytropic index
An index polytrope is often used to model rocky planets. The reason is that polytrope has constant density, i.e., incompressible interior. This is a zero order approximation for rocky (solid/liquid) plane |
https://en.wikipedia.org/wiki/Methods%20of%20successive%20approximation | Mathematical methods relating to successive approximation include the following:
Babylonian method, for finding square roots of numbers
Fixed-point iteration
Means of finding zeros of functions:
Halley's method
Newton's method
Differential-equation matters:
Picard–Lindelöf theorem, on existence of solutions of differential equations
Runge–Kutta methods, for numerical solution of differential equations
Approximation algorithms
Root-finding algorithms |
https://en.wikipedia.org/wiki/Excitable%20medium | An excitable medium is a nonlinear dynamical system which has the capacity to propagate a wave of some description, and which cannot support the passing of another wave until a certain amount of time has passed (known as the refractory time).
A forest is an example of an excitable medium: if a wildfire burns through the forest, no fire can return to a burnt spot until the vegetation has gone through its refractory period and regrown. In chemistry, oscillating reactions are excitable media, for example the Belousov–Zhabotinsky reaction and the Briggs–Rauscher reaction. Cell excitability is the change in membrane potential that is necessary for cellular responses in various tissues. The resting potential forms the basis of cell excitability and these processes are fundamental for the generation of graded and action potentials. Normal and pathological activities in the heart and brain can be modelled as excitable media. A group of spectators at a sporting event are an excitable medium, as can be observed in a Mexican wave (so-called from its initial appearance in the 1986 World Cup in Mexico).
Modelling excitable media
Excitable media can be modelled using both partial differential equations and cellular automata.
With cellular automata
Cellular automata provide a simple model to aid in the understanding of excitable media. Perhaps the simplest such model is in. See Greenberg-Hastings cellular automaton for this model.
Each cell of the automaton is made to represent some section of the medium being modelled (for example, a patch of trees in a forest, or a segment of heart tissue). Each cell can be in one of the three following states:
Quiescent or excitable — the cell is unexcited, but can be excited. In the forest fire example, this corresponds to the trees being unburnt.
Excited — the cell is excited. The trees are on fire.
Refractory — the cell has recently been excited and is temporarily not excitable. This corresponds to a patch of land where the |
https://en.wikipedia.org/wiki/Awwwards | Awwwards (Awwwards Online SL) is a professional web design and development competition body. It aims to recognize and promote the best of innovative web design. It is a website competition that developers can submit to. The best year-round submissions are awarded at the Awwwards conference and prize-giving ceremony, which take place in various cities across the United States and Europe.
Nomination process
In the first phase of the nomination process, web designers submit their work through Awwwards' website for consideration for Site of the Day. The work is then judged by the Awwwards community and an international jury panel composed of designers, developers, and agencies. The best daily sites are also published in Awwward's year-end "The 365 Best Websites Around the World" book.
In the second voting phase, the Site of the Month is chosen. Site of the Month winners proceed to the final judging phase of the competition: Site of the Year. This award is given out at the Awwwards conference and prize-giving ceremony.
Jury
The jury consists of multidisciplinary designers, developers, journalists, and agencies from across the globe. The panel assesses the talent, effort, technicality, and insight that goes into the web projects submitted for consideration.
Awards granted
Members of the Awwwards Jury score nominees on four separate criteria: design, creativity, usability and content. High scoring sites can win several different types of awards.
Daily
Honorable Mention
All websites awarded 6.5 or above receive an Honorable Mention.
Site of the Day
Site of the Day recognizes aesthetic, usability, and technical achievements in web design innovation. Sites remain in competition for Site of the Day for three months from the submission date. Only the highest scoring sites are awarded Site of the Day.
Developer Award
The developer's award was created in partnership with Microsoft, and it awards developers who have achieved significant web development programming |
https://en.wikipedia.org/wiki/Alternated%20hypercubic%20honeycomb | In geometry, the alternated hypercube honeycomb (or demicubic honeycomb) is a dimensional infinite series of honeycombs, based on the hypercube honeycomb with an alternation operation. It is given a Schläfli symbol h{4,3...3,4} representing the regular form with half the vertices removed and containing the symmetry of Coxeter group for n ≥ 4. A lower symmetry form can be created by removing another mirror on an order-4 peak.
The alternated hypercube facets become demihypercubes, and the deleted vertices create new orthoplex facets. The vertex figure for honeycombs of this family are rectified orthoplexes.
These are also named as hδn for an (n-1)-dimensional honeycomb. |
https://en.wikipedia.org/wiki/Inertial%20response | Inertial response is a property of large synchronous generators, which contain large synchronous rotating masses, and which acts to overcome any immediate imbalance between power supply and demand for electric power systems, typically the electrical grid. Due to the ever existing power imbalance between mechanical power supply and electric power demand the rotational frequency of the rotating masses in all synchronous generators in the grid either speed up and thus absorb the extra power in case of an excess power supply, or slow down and provide additional power in case of an excess power demand. This response in case of a synchronous generator is built-in into the design and happens without any external intervention or coordination, providing the automatic generation control and the grid operator with valuable time (few seconds) to rebalance the system The grid frequency is the combined result of the detailed motions of all individual synchronous rotors in the grid, which are modeled by a general equation of motion called the swing equation.
In the US power systems, the grid operator is mandated to keep the frequency within a tight range, and can be financially responsible if the monitoring by the North American Electric Reliability Corporation detects a non-compliance. Furthermore, in order to protect the equipment, a portion of the load will be disconnected ("underfrequency load shedding", UFLS) if the frequency drops below a limit (59.5 Hz in most of the US, 59.3 Hz in Texas). When an unexpected supply disruption occurs (for example, a generator failure), the primary frequency response kicks in automatically - a sensor detects the lower frequency and adjusts the power of the prime mover accordingly. For a typical synchronous generator, this adjustment involves manipulation of the mechanical devices (valves, etc.) and thus takes time. During this time, the power grid has to rely on the accumulated inertia to slow down the decrease in frequency.
Synchronous g |
https://en.wikipedia.org/wiki/Kelly%27s%20lemma | In probability theory, Kelly's lemma states that for a stationary continuous-time Markov chain, a process defined as the time-reversed process has the same stationary distribution as the forward-time process. The theorem is named after Frank Kelly.
Statement
For a continuous time Markov chain with state space S and transition-rate matrix Q (with elements qij) if we can find a set of non-negative numbers q'ij and a positive measure π that satisfy the following conditions:
then q'''ij are the rates for the reversed process and π is proportional to the stationary distribution for both processes.
Proof
Given the assumptions made on the qij and π we have
so the global balance equations are satisfied and the measure π is proportional to the stationary distribution of the original process.
By symmetry, the same argument shows that π'' is also proportional to the stationary distribution of the reversed process. |
https://en.wikipedia.org/wiki/Starlight%20Information%20Visualization%20System | Starlight is a software product originally developed at Pacific Northwest National Laboratory and now by Future Point Systems. It is an advanced visual analysis environment. In addition to using information visualization to show the importance of individual pieces of data by showing how they relate to one another, it also contains a small suite of tools useful for collaboration and data sharing, as well as data conversion, processing, augmentation and loading.
The software, originally developed for the intelligence community, allows users to load data from XML files, databases, RSS feeds, web services, HTML files, Microsoft Word, PowerPoint, Excel, CSV, Adobe PDF, TXT files, etc. and analyze it with a variety of visualizations and tools. The system integrates structured, unstructured, geospatial, and multimedia data, offering comparisons of information at multiple levels of abstraction, simultaneously and in near real-time. In addition Starlight allows users to build their own named entity-extractors using a combination of algorithms, targeted normalization lists and regular expressions in the Starlight Data Engineer (SDE).
As an example, Starlight might be used to look for correlations in a database containing records about chemical spills. An analyst could begin by grouping records according to the cause of the spill to reveal general trends. Sorting the data a second time, they could apply different colors based on related details such as the company responsible, age of equipment or geographic location. Maps and photographs could be integrated into the display, making it even easier to recognize connections among multiple variables.
Starlight has been deployed to both the Iraq and Afghanistan wars and used on a number of large-scale projects.
PNNL began developing Starlight in the mid-1990s, with funding from the Land Information Warfare Agency, a part of the Army Intelligence and Security Command and continued developed at the laboratory with funding from th |
https://en.wikipedia.org/wiki/Manbang | Manbang () is a series of state-owned digital media players issued by North Korea's Korean Central Broadcasting Committee, providing over-the-top content in the form of channels. It was created in response to streaming platforms like Netflix and Roku in the west, and the popularity of Chinese-made Notel players in North Korea.
Manbang, which ironically translates to "everywhere," is only available to citizens in Pyongynag, Siniju and Sariwon. Due to North Korea's isolationism, users connect to the service not by internet but via the state-controlled intranet using the IPTV protocol. It's hard to tell if the technology at play is IPTV or VOD but according to description it is a mixture of both.
Like Apple TV or Roku, the device is an Internet Protocol Television that works through a separate box. The system comes as a set-top box which first has to be connected to a modem and after that to the phone line. The box can be connected to a television through a HDMI cable.
History
Exact release date of Manbang system is unclear. One of the first set-top box appears to be manufactured in 2015. An intranet sites listing in 2015 included a site named “Manbang” with the operator being Korean Central Television.
On 16 August 2016, Manbang for the first time appeared on a report “망 TV다매체열람기‘만방'” by Korean Central Television. It reported that the implementation of “Intranet” Protocol Television (IPTV) which run on North Korea’s Kwangmyong intranet has begun. It also showcased a set-top box which was developed by Manbang IT company on which Manbang is based. KCTV also stated that the new service already has "several hundred users" and is "making the lives of citizens and children flourish" .
On 25 August 2016, Netflix took a light-hearted jab at Manbang by changing its Twitter bio description to read: "Manbang knockoff."
During the following years it appears that the North Korean government working towards making the service widely accessible . It was reported that the de |
https://en.wikipedia.org/wiki/Mois%C3%A9s%20Exp%C3%B3sito-Alonso | Moisés Expósito-Alonso (born March, 1990, Alicante) is a Spanish scientist principal investigator in the Departments of Plant Biology and Global Ecology from the Carnegie Institution for Science and assistant professor at Stanford University. His research includes the study of plants and how climate change affects their evolution. In 2023, Exposito-Alonso was selected as one of the inaugural Freeman Hrabowski Scholars from the Howard Hughes Medical Institute.
Education
Exposito-Alonso graduated with a Bachelor of Science degree in biology from the University of Seville, Spain, in 2013. He completed his Master of Science in quantitative and population genetics from the University of Edinburgh in 2014. He conducted his doctorate in 2018 at the Max Planck Institute of Biology in Tübingen, Germany under Detlef Weigel. After his PhD, Exposito-Alonso conducted a postdoctoral position in 2019 at the University of California, Berkeley in the Integrative Biology department with Rasmus Nielsen.
Career and research
Exposito-Alonso's research includes the study of genetic and phenotypic basis of plant local adaptation to different climates using Arabidopsis thaliana as model system. His research combines large-scale common garden experiments and genome sequencing and CRISPR/Cas9 to study the consequence of gene edits in Arabidopsis. His research has been published in journals such as Nature (journal), Science (journal), Cell (journal), Evolution (journal), the Proceedings of the National Academy of Sciences of the United States of America and independent news media such as The Scientist (magazine) and El País.
In January 2024, Expósito Alonso will be joining the University of California, Berkeley as an assistant professor of global change biology, member of the Innovative Genomics Institute and Howard Hughes Medical Institute.
Selected publications
Exposito-Alonso M., Booker T.R., Czech .L, Gillespie L., Hateley S., Kyriazis, C.C., Lang, P., Leventhal, L., Nogues-Bravo, |
https://en.wikipedia.org/wiki/Land%C3%A9%20interval%20rule | In atomic physics, the Landé interval rule states that, due to weak angular momentum coupling (either spin-orbit or spin-spin coupling), the energy splitting between successive sub-levels are proportional to the total angular momentum quantum number (J or F) of the sub-level with the larger of their total angular momentum value (J or F).
Background
The rule assumes the Russell–Saunders coupling and that interactions between spin magnetic moments can be ignored. The latter is an incorrect assumption for light atoms. As a result of this, the rule is optimally followed by atoms with medium atomic numbers.
The rule was first stated in 1923 by German-American physicist Alfred Landé.
Derivation
As an example, consider an atom with two valence electrons and their fine structures in the LS-coupling scheme. We will derive heuristically the interval rule for the LS-coupling scheme and will remark on the similarity that leads to the interval rule for the hyperfine structure.
The interactions between electrons couple their orbital and spin angular momentums. Let's denote the spin and orbital angular momentum as and for each electrons. Thus, the total orbital angular momentum is and total spin momentum is . Then the coupling in the LS-scheme gives rise to a Hamiltonian:
where and encode the strength of the coupling. The Hamiltonian acts as a perturbation to the state . The coupling would cause the total orbital and spin angular momentums to change directions, but the total angular momentum would remain constant. Its z-component would also remain constant, since there is no external torque acting on the system. Therefore, we shall change the state to , which is a linear combination of various . The exact linear combination, however, is unnecessary to determine the energy shift.
To study this perturbation, we consider the vector model where we treat each as a vector. and precesses around the total orbital angular momentum . Con |
https://en.wikipedia.org/wiki/Biodiversity%20of%20Wales | The biodiversity of Wales (Welsh: Bioamrywiaeth Cymru) is the wide variety of ecosystems, living organisms, and the genetic makeups found in Wales.
Wales is a predominantly mountainous peninsula located between England and the Irish Sea, covering 8,023 square miles. It has terrestrial habitats and many protected areas rich in biodiversity, including three national parks and five Areas of Outstanding Natural Beauty (AONB). The national parks being: Snowdonia, Pembrokeshire Coast, and Brecon Beacons, and the AONBs of: Anglesey, the Clwydian Range and Dee Valley, Gower Peninsula, Llŷn Peninsula, and Wye Valley (partially in England). Wales also has many locations categorised as Site of Special Scientific Interest, Special Area of Conservation, Special Protection Area and local nature reserve. There are many zoos and gardens, including the National Botanic Garden of Wales.
On the coast, a great diversity of species such as seals, dolphins, sharks, jellyfish, crabs and lobsters can be found. There are also seabird colonies on the islands near the coast. Species which can only be found in Wales are the Radnor lily and a type of fish, the gwyniad, only found in Lake Bala. The rare fen orchid (Liparis loeselii) is one of the most threatened species in northwestern Europe and has vanished from many places in Wales. The Welsh Government funds Natural Resources Wales (NRW), Plantlife, Bridgend County Borough Council and the Wales Biodiversity Partnership coastal ecosystem group to help reconstruct its natural habitat and secure the future of this threatened species.
The Welsh Government works closely with the Wales Biodiversity Partnership (WBP) which promotes and monitors the Wales biodiversity action plan. In 2010 the Welsh government launched a Natural Environment Framework, "A Living Wales", which focuses on sustainable land and marine management in Wales. The Environment (Wales) Act 2016 put into place a range of powers and duties designed to enable the natural resourc |
https://en.wikipedia.org/wiki/Axial%20precession | In astronomy, axial precession is a gravity-induced, slow, and continuous change in the orientation of an astronomical body's rotational axis. In the absence of precession, the astronomical body's orbit would show axial parallelism. In particular, axial precession can refer to the gradual shift in the orientation of Earth's axis of rotation in a cycle of approximately 26,000 years. This is similar to the precession of a spinning top, with the axis tracing out a pair of cones joined at their apices. The term "precession" typically refers only to this largest part of the motion; other changes in the alignment of Earth's axis—nutation and polar motion—are much smaller in magnitude.
Earth's precession was historically called the precession of the equinoxes, because the equinoxes moved westward along the ecliptic relative to the fixed stars, opposite to the yearly motion of the Sun along the ecliptic. Historically,
the discovery of the precession of the equinoxes is usually attributed in the West to the 2nd-century-BC astronomer Hipparchus. With improvements in the ability to calculate the gravitational force between planets during the first half of the nineteenth century, it was recognized that the ecliptic itself moved slightly, which was named planetary precession, as early as 1863, while the dominant component was named lunisolar precession. Their combination was named general precession, instead of precession of the equinoxes.
Lunisolar precession is caused by the gravitational forces of the Moon and Sun on Earth's equatorial bulge, causing Earth's axis to move with respect to inertial space. Planetary precession (an advance) is due to the small angle between the gravitational force of the other planets on Earth and its orbital plane (the ecliptic), causing the plane of the ecliptic to shift slightly relative to inertial space. Lunisolar precession is about 500 times greater than planetary precession. In addition to the Moon and Sun, the other planets also cause |
https://en.wikipedia.org/wiki/Flag%20of%20Israel | The national flag of the State of Israel ( ; ) was adopted on 28 October 1948, five months after the establishment of the state. It depicts a blue hexagram on a white background, between two horizontal blue stripes. The Israeli flag legislation states that the official measurements are 160 × 220 cm. Therefore, the official proportions are 8:11. Variants can be found at a wide range of proportions, with 2:3 being common.
The blue color is described as "dark sky-blue", and varies from flag to flag, ranging from a hue of pure blue, sometimes shaded almost as dark as navy blue, to hues about 75% toward pure cyan and shades as light as very light blue. An early version of the flag was displayed in 1885 at a procession marking the third anniversary of Rishon LeZion. A similar version was designed for the Zionist Movement in 1891. The basic design recalls the (), the Jewish prayer shawl, which is white with black or blue stripes. The symbol in the center represents the Star of David (, ), a Jewish symbol dating from late medieval Prague, which was adopted by the First Zionist Congress in 1897.
Origin of the flag
In the Middle Ages, mystical powers were attributed to the pentagram and hexagram, which were used in talismans against evil spirits. Both were called the "Seal of Solomon", but eventually the name became exclusive to the pentagram, while the hexagram became known as a "Magen David", or "Shield of David". Later the star began to appear in Jewish art. In 1648, Ferdinand II permitted the Jews of Prague to fly a "Jewish flag" over their synagogue. This flag was red with a yellow Magen David in the middle.
The idea that the blue and white colors were the national color of the Jewish people was voiced early on by Ludwig August von Frankl (1810–94), an Austrian Jewish poet. In his poem, "Judah's Colours", he writes:
In 1885, the agricultural village of used a blue and white flag incorporating a blue Star of David, designed by Israel Belkind and Fanny Abramovitch, |
https://en.wikipedia.org/wiki/Cladding%20%28fiber%20optics%29 | Cladding in optical fibers is one or more layers of materials of lower refractive index in intimate contact with a core material of higher refractive index.
The cladding causes light to be confined to the core of the fiber by total internal reflection at the boundary between the core and cladding. Light propagation within the cladding is typically suppressed for most fibers. However, some fibers can support cladding modes in which light propagates through the cladding as well as the core. Depending upon the quantity of modes that are supported, they are referred to as multi-mode fibers and single-mode fibers. Improving transmission through fibers by applying a cladding was discovered in 1953 by Dutch scientist Bram van Heel.
History
The fact that transmission through fibers could be improved by applying a cladding was discovered in 1953 by van Heel, who used it to demonstrate image transmission through a bundle of optical fibers. Early cladding materials included oils, waxes, and polymers. Lawrence E. Curtiss at the University of Michigan developed the first glass cladding in 1956, by inserting a glass rod into a tube of glass with a lower refractive index, fusing the two together, and drawing the composite structure into an optical fiber.
Modes
A cladding mode is a mode that is confined to the cladding of an optical fiber by virtue of the fact that the cladding has a higher refractive index than the surrounding medium, which is either air or the primary polymer overcoat. These modes are generally undesired. Modern fibers have a primary polymer overcoat with a refractive index that is slightly higher than that of the cladding, so that light propagating in the cladding is rapidly attenuated and disappears after only a few centimeters of propagation. An exception to this is double-clad fiber, which is designed to support a mode in its inner cladding, as well as one in its core.
Advantages
In the production of glass fibers, there will inevitably be surface irre |
https://en.wikipedia.org/wiki/HistoryWorld | HistoryWorld is an interactive online history encyclopaedia that seeks to make world history more easily accessible through interactive narratives and timelines. It was established by Bamber Gascoigne who started developing it in 1994. It went online in June 2001 and in 2002 it won the New Statesman New Media award for the best educational website. In 2007 Gascoigne launched a related site, at TimeSearch , using timelines as a way of searching the internet.
HistoryWorld currently consists of about 300 narratives and some 10,000 events on searchable timelines. All the content (apart from "The Wellcome History of Medicine", by Dr Carole Reeves) has been written by Gascoigne.
The HistoryWorld website, which is free to use, also contains more than 5000 entries from Gascoigne's Encyclopedia of Britain, originally published by Macmillan in 1993, and a pilot project, Places in History for Richmond-upon-Thames, which uses placemarks in Google Maps to identify the exact position of a building, street or other feature, with a satellite view of the location. The maps then link to pages in HistoryWorld for historical details, images and timelines.
Harvey McGavin, writing in the TES, said that the history website "is remarkably easy to navigate" and "should help teachers and pupils find all the answers". |
https://en.wikipedia.org/wiki/Content%20as%20a%20service | Content as a service (CaaS) or managed content as a service (MCaaS) is a service-oriented model, where the service provider delivers the content on demand to the service consumer via web services that are licensed under subscription. The content is hosted by the service provider centrally in the cloud and offered to a number of consumers that need the content delivered into any applications or system, hence content can be demanded by the consumers as and when required.
Content as a Service is a way to provide raw content (in other words, without the need for a specific human compatible representation, such as HTML) in a way that other systems can make use of it. Content as a Service is not meant for direct human consumption, but rather for other platforms to consume and make use of the content according to their particular needs. This happens usually on the cloud, with a centralized platform which can be globally accessible and provides a standard format for your content. With Content as a Service, you centralize your content into a single repository, where you can manage it, categorize it, make it available to others, search for it, or do whatever you wish with it.
Overview
The content delivered typically could be one or more of the following
The technical terminology related to equipments or spares that is required to procure or design the materials
The industrial terminology of the equipments or spares
Technical values pertaining to various types, specifications, applications, characteristics of equipments or spares
Sourcing information which will help in procurement or supply-chain management of equipments or spares
Descriptive specifications of equipments or spares based on the product reference number or identifier
UNSPSC codes or industry practiced classifications
ISO, IEC compliant terminology
Ontology or Technical Dictionary of products & services
Predefined content for specific business needs
The term "Content as a service" (CaaS) is considered |
https://en.wikipedia.org/wiki/Dedekind%20eta%20function | In mathematics, the Dedekind eta function, named after Richard Dedekind, is a modular form of weight 1/2 and is a function defined on the upper half-plane of complex numbers, where the imaginary part is positive. It also occurs in bosonic string theory.
Definition
For any complex number with , let ; then the eta function is defined by,
Raising the eta equation to the 24th power and multiplying by gives
where is the modular discriminant. The presence of 24 can be understood by connection with other occurrences, such as in the 24-dimensional Leech lattice.
The eta function is holomorphic on the upper half-plane but cannot be continued analytically beyond it.
The eta function satisfies the functional equations
In the second equation the branch of the square root is chosen such that when .
More generally, suppose are integers with , so that
is a transformation belonging to the modular group. We may assume that either , or and . Then
where
Here is the Dedekind sum
Because of these functional equations the eta function is a modular form of weight and level 1 for a certain character of order 24 of the metaplectic double cover of the modular group, and can be used to define other modular forms. In particular the modular discriminant of Weierstrass can be defined as
and is a modular form of weight 12. Some authors omit the factor of , so that the series expansion has integral coefficients.
The Jacobi triple product implies that the eta is (up to a factor) a Jacobi theta function for special values of the arguments:
where is "the" Dirichlet character modulo 12 with and . Explicitly,
The Euler function
has a power series by the Euler identity:
Because the eta function is easy to compute numerically from either power series, it is often helpful in computation to express other functions in terms of it when possible, and products and quotients of eta functions, called eta quotients, can be used to express a great variety of modular forms.
The pictur |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.