source
stringlengths 31
227
| text
stringlengths 9
2k
|
|---|---|
https://en.wikipedia.org/wiki/General%20Mission%20Analysis%20Tool
|
General Mission Analysis Tool (GMAT) is open-source space mission analysis software developed by NASA and private industry.
It has been used for several missions, including LCROSS, the Lunar Reconnaissance Orbiter, OSIRIS-REx, the Magnetospheric Multiscale Mission, and the Transiting Exoplanet Survey Satellite (TESS) mission.
GMAT is an open-source alternative to software like Systems Tool Kit and FreeFlyer.
|
https://en.wikipedia.org/wiki/Nvidia%20NVDEC
|
Nvidia NVDEC (formerly known as NVCUVID) is a feature in its graphics cards that performs video decoding, offloading this compute-intensive task from the CPU.
It is accompanied by NVENC for video encoding in Nvidia's Video Codec SDK.
Technology
NVDEC can offload video decoding to full fixed-function decoding hardware (Nvidia PureVideo), or (partially) decode via CUDA software running on the GPU, if fixed-function hardware is not available.
Depending on the GPU architecture, the following codecs are supported:
MPEG-2
VC-1
H.264 (AVC)
H.265 (HEVC)
VP8
VP9
AV1
Versions
NVCUVID was originally distributed as part of the Nvidia CUDA Toolkit. Later, it was renamed to NVDEC and moved to the Nvidia Video Codec SDK.
Operating system support
NVDEC is available for Windows and Linux operating systems. As NVDEC is a proprietary API (as opposed to the open-source VDPAU API), it is only supported by the proprietary Nvidia driver on Linux.
Application and library support
Gstreamer has supported NVDEC since 2017.
FFmpeg has supported NVDEC since 2017.
mpv has supported NVDEC since 2017 by the use of FFmpeg.
GPU support
HW accelerated decode and encode are supported on Nvidia GeForce, Quadro, Tesla, and GRID products with Fermi or newer generation GPUs.
See also
AMD Video Core Next, AMD's equivalent SIP core since 2018
AMD Unified Video Decoder, AMD's equivalent SIP core up to 2017
Intel Quick Sync Video, Intel's equivalent SIP core
List of Nvidia graphics processing units
Qualcomm Hexagon
Nvidia NVENC
|
https://en.wikipedia.org/wiki/Apache%20Axis2
|
Apache Axis2 is a web service engine. It is a redesign and re-write of the widely used Apache Axis SOAP stack. Implementations of Axis2 are available in Java and C.
Axis2 provides the capability to add Web services interfaces to Web applications. It can also function as a standalone application server.
Why Apache Axis2
A new architecture for Axis2 was introduced during the August 2004 Axis2 Summit in Colombo, Sri Lanka. Some concepts from Axis 1.x, like handlers etc., have been preserved in the new architecture.
Apache Axis2 supports SOAP 1.1 and SOAP 1.2, and it has integrated support for the REST style of Web services. The same business-logic implementation can offer both a WS-* style interface as well as a REST/POX style interface simultaneously.
Axis2/Java has support for Spring Framework.
Axis2/C is a high-performance Web services implementation that has been implemented with portability and ability to be embedded or hosted in Apache Httpd, Microsoft IIS or Axis Http Server. See article about Apache Axis2/C Performance (2008) (latest release occurred in 2009).
Axis2 came with new features, enhancements and industry specification implementations. Key features include:
Axis2 Features
Apache Axis2 includes support for following standards:
WS-ReliableMessaging
WS-Coordination Via Apache Kandula2
WS-AtomicTransaction Via Apache Kandula2
WS-SecurityPolicy Via Apache Rampart
WS-Security Via Apache Rampart
WS-Trust Via Apache Rampart
WS-SecureConversation Via Apache Rampart
SAML 1.1 Via Apache Rampart
SAML 2.0 Via Apache Rampart
WS-Addressing Module included as part of Axis2 core
Below a list of features and selling points cited from the Apache axis site:
Speed Axis2 uses its own object model and StAX (Streaming API for XML) .
Low memory foot print Axis2 was designed to consume a low amount of memory.
AXIOM Axis2 comes with its own light-weight object model, AXIOM
Hot Deployment Axis2 can deploy Web services and handlers while the system is runnin
|
https://en.wikipedia.org/wiki/Ureaplasma%20felinum
|
Ureaplasma felinum is a species of Ureaplasma, a genus of bacteria belonging to the family Mycoplasmataceae. It has been isolated from cats. It possesses the sequence accession no. (16S rRNA gene) for the type strain: D78651.
|
https://en.wikipedia.org/wiki/File%20URI%20scheme
|
In programming, a file uniform resource identifier (URI) scheme is a specific format of URI, used to specifically identify a file on a host computer. While URIs can be used to identify anything, there is specific syntax associated with identifying files.
Format
A file URI has the format
file://host/path
where host is the fully qualified domain name of the system on which the path is accessible, and path is a hierarchical directory path of the form directory/directory/.../name. If host is omitted, it is taken to be "localhost", the machine from which the URL is being interpreted. Note that when omitting host, the slash is not omitted (while "file:///foo.txt" is valid, "file://foo.txt" is not, although some interpreters manage to handle the latter).
RFC 3986 includes additional information about the treatment of ".." and "." segments in URIs.
Number of slash characters
The character sequence of two slash characters (//) after the string file: denotes that either a hostname or the literal term localhost follows, although this part may be omitted entirely, or may contain an empty hostname.
The single slash between host and path denotes the start of the local-path part of the URI and must be present.
A valid file URI must therefore begin with either file:/path (no hostname), file:///path (empty hostname), or file://hostname/path.
file://path (i.e. two slashes, without a hostname) is never correct, but is often used.
Further slashes in path separate directory names in a hierarchical system of directories and subdirectories. In this usage, the slash is a general, system-independent way of separating the parts, and in a particular host system it might be used as such in any pathname (as in Unix systems).
There are two ways that Windows UNC filenames (such as \\server\folder\data.xml) can be represented. These are both described in RFC 8089, Appendix E as "non-standard". The first way (called here the 2-slash format) is to represent the server name using the Autho
|
https://en.wikipedia.org/wiki/Asplenia
|
Asplenia refers to the absence of normal spleen function and is associated with some serious infection risks. Hyposplenism is used to describe reduced ('hypo-') splenic functioning, but not as severely affected as with asplenism.
Functional asplenia occurs when splenic tissue is present but does not work well (e.g. sickle-cell disease, polysplenia) -such patients are managed as if asplenic-, while in anatomic asplenia, the spleen itself is absent.
Causes
Congenital
Congenital asplenia is rare. There are two distinct types of genetic disorders: heterotaxy syndrome and isolated congenital asplenia.
polysplenia
Acquired
Acquired asplenia occurs for several reasons:
Following splenectomy due to splenic rupture from trauma or because of tumor
After splenectomy with the goal of interfering with splenic function, as a treatment for diseases (e.g. idiopathic thrombocytopenic purpura, thalassemia, spherocytosis), in which the spleen's usual activity exacerbates the disease
After splenectomy with the goal of arresting the progression of cancers (Chronic lymphocytic leukemia, Hodgkin's disease (starting in the 1970s), non-Hodgkin lymphoma)
Due to underlying diseases that destroy the spleen (autosplenectomy), e.g. sickle-cell disease.
Celiac disease: unknown physiopathology. In a 1970 study, it was the second most common cause of abnormalities of red blood cells linked to hyposplenism, after surgical splenectomy.
Functional asplenia
Functional asplenia can occur when patients with metabolic or haematological disorders have their splenic tissue organisation altered. This can lead to results similar to those seen in patients who have undergone a splenectomy e.g. becoming infected with encapsulated bacteria such as Haemophilus influenzae, Streptococcus pneumoniae and Neisseria meningitidis. Patients who have some form of asplenia have an increased susceptibility to these encapsulated bacterial infections mainly because they lack IgM memory B cells and their non-adhere
|
https://en.wikipedia.org/wiki/Neuroimaging%20Informatics%20Tools%20and%20Resources%20Clearinghouse
|
The Neuroimaging Tools and Resources Collaboratory (NITRC) is a neuroimaging informatics knowledge environment for MR, PET/SPECT, CT, EEG/MEG, optical imaging, clinical neuroinformatics, imaging genomics, and computational neuroscience tools and resources.
Description
Initiated in 2006 and currently funded by NIH Grant number: 1R24EB029173 , NITRC's mission is to provide a user-friendly knowledge environment that enables the distribution, enhancement, and adoption of neuroimaging tools and resources and has expanded from MR to Imaging Genomics, EEG/MEG, PET/SPECT, CT, optical imaging, clinical neuroinformatics, and computational neuroscience. Supporting 143,000 page views per month, NITRC's 1,000+ tools and resources have been downloaded over 11.4 million times by 1.4 million users.
NITRC's goal is to support researchers dedicated to enhancing, adopting, distributing, and contributing to the evolution of previously funded neuroimaging analysis tools and resources for broader community use. Promoting software tools, workflows, resources, vocabularies, test data, and pre-processed, community-generated images through its Image Repository (NITRC-IR), NITRC gives researchers greater and more efficient access to the tools and resources they need; better categorizing and organizing existing tools and resources via a controlled vocabulary; facilitating interactions between researchers and developers through forums, direct email contact, ratings and reviews; and promoting better use through enhanced documentation.
Approach
To meet the disparate needs of neuroimaging informatics developers and researchers, NITRC offers collaborative functionality like that found in platforms such as GitHub and SourceForge. To provide such functionality, we customized the open-source GForge project. Thus, within NITRC, each tool or resource has the option to offer descriptive content as well as use MediaWiki, CVS/SVN, bug tracking, news, and forums to distribute information and downloads
|
https://en.wikipedia.org/wiki/CASA%20ratio
|
CASA ratio stands for current and savings account ratio. CASA ratio of a bank is the ratio of deposits in current, and saving accounts to total deposits. A higher CASA ratio indicates a lower cost of funds, because banks do not usually give any interests on current account deposits and the interest on saving accounts is usually very low 3-4%. If a large part of a bank's deposits comes from these funds, it means that the bank is getting those funds at a relative lower cost. It is generally understood that a higher CASA ratio leads to higher net interest margin. In India, it is used as one of the metrics to assess the profitability of a bank.
Formula
See also
Banking in India
|
https://en.wikipedia.org/wiki/Westcott%20Rule%20Company
|
The history of school and office products brand Westcott dates back to 1872, when Henry Westcott, together with his two sons Charles and Frank Westcott, started manufacturing wooden furniture used by printers and trellis units in Seneca Falls, New York. The company expanded rapidly and became one of the largest manufacturers of desk and school rulers in the world. Throughout its history it remained in Westcott family hands until it was purchased by Acme Shear Co. in 1968, which later changed its name to Acme United Corporation.
Because Westcott was such a strong brand, Acme United started selling dozens of other school and office items, like scissors, pencil sharpeners and paper trimmers under the Westcott name. Nowadays, Westcott sells 60 to 80 million scissors and 15 to 18 million rulers worldwide annually.
History
1808–1871
Henry Westcott was born in Newport, Rhode Island in 1808 where he learned the sash and blind makers' trade. Later, he started his own company together with his brother Edwin and an associate called Hiram Miller to manufacture sash, doors and blinds. After retiring from this business, he moved to Seneca Falls in 1847 to embark in the manufacture of an improved churn and butter pail on which he had secured patents. Also this business proved to be a success, mainly because Henry Westcott invented and built machinery to make the wooden products he sold, while competitors were still making them by hand. In 1868, at the age of 60, he sold his interest in this company.
1872–1893
In 1872 Henry, along with his two sons Charles and Frank, established the firm Westcott Brothers to manufacture a variety of wood specialties. While Charles concentrated on operating the factory and Frank on selling the products, here too, Henry's experience and inventive genius were responsible for the company's growth. When Henry retired in 1890, the company was renamed to Westcott Bros. Co. and expanded into numerous products like toy blocks and other games.
1894–192
|
https://en.wikipedia.org/wiki/Plug%20compatible
|
Plug compatible refers to "hardware that is designed to perform exactly like another vendor's product." The term PCM was originally applied to manufacturers who made replacements for IBM peripherals. Later this term was used to refer to IBM-compatible computers.
PCM and peripherals
Before the rise of the PCM peripheral industry, computing systems were either configured with peripherals designed and built by the CPU vendor, or designed to use vendor-selected rebadged devices.
The first example of plug-compatible IBM subsystems were tape drives and controls offered by Telex beginning 1965. Memorex in 1968 was first to enter the IBM plug-compatible disk followed shortly thereafter by a number of suppliers such as CDC, Itel, and Storage Technology Corporation. This was boosted by the world's largest user of computing equipment in both directions.
Ultimately plug-compatible products were offered for most peripherals and system main memory.
PCM and computer systems
A plug-compatible machine is one that has been designed to be backward compatible with a prior machine. In particular, a new computer system that is plug-compatible has not only the same connectors and protocol interfaces to peripherals, but also binary-code compatibility—it runs the same software as the old system. A plug compatible manufacturer or PCM is a company that makes such products.
One recurring theme in plug-compatible systems is the ability to be bug compatible as well. That is, if the forerunner system had software or interface problems, then the successor must have (or simulate) the same problems. Otherwise, the new system may generate unpredictable results, defeating the full compatibility objective. Thus, it is important for customers to understand the difference between a "bug" and a "feature", where the latter is defined as an intentional modification to the previous system (e.g. higher speed, lighter weight, smaller package, better operator controls, etc.).
PCM and IBM mainframes
The or
|
https://en.wikipedia.org/wiki/NASP%20%28gene%29
|
Nuclear autoantigenic sperm protein is a protein that in humans is encoded by the NASP gene. Multiple isoforms are encoded by transcript variants of this gene.
Function
This gene encodes a histone H1 binding protein that is involved in transporting histones into the nucleus of dividing cells. The somatic form is expressed in all mitotic cells, is localized to the nucleus, and is coupled to the cell cycle. The testicular form is expressed in embryonic tissues, tumor cells, and the testis. In male germ cells, this protein is localized to the cytoplasm of primary spermatocytes, the nucleus of spermatids, and the periacrosomal region of mature spermatozoa.
|
https://en.wikipedia.org/wiki/Instant%20soup
|
Instant soup is a type of soup designed for fast and simple preparation. Some are homemade, and some are mass-produced on an industrial scale and treated in various ways to preserve them. A wide variety of types, styles and flavors of instant soups exist. Commercial instant soups are usually dried or dehydrated, canned, or treated by freezing.
Types
Commercial instant soups are manufactured in several types. Some consist of a packet of dry soup stock. These do not contain water, and are prepared by adding water and then heating the product for a short time, or by adding hot water directly to the dry soup mix. Instant soup can also be produced in a dry powder form, such as Unilever's Cup-a-Soup
Canned (tinned) instant soups contain liquid soup that is prepared by heating their contents. Some canned soups are condensed, and require additional water to bring them to their intended strength, while others are canned in a ready-to-eat, single-strength form. Dr. John T. Dorrance, an employee with the Campbell Soup Company, invented condensed soup in 1897. Consumers sometimes use condensed soups (without diluting them), as a sauce base. Some instant liquid soups are manufactured in microwaveable containers. Additionally, some instant soups, such as Knorr's Erbswurst, are prepared in a concentrated paste form. Knorr ceased production of Erbswurst on December 31, 2018.
Instant noodle soups such as Cup Noodles contain dried instant ramen noodles, dehydrated vegetable and meat products, and seasonings, and are prepared by adding hot water. Packaged instant ramen noodle soup is typically formed as a cake, and often includes a seasoning packet that is added to the noodles and water during preparation. Some also include separate packets of oil and garnishes used to season the product. Momofuku Ando, the founder of Nissin Foods, developed packaged ramen noodle soup in 1958.
Varieties
A multitude of instant soup varieties exist. For example, there are several Lipton and Knorr-b
|
https://en.wikipedia.org/wiki/189%20%28number%29
|
189 (one hundred [and] eighty-nine) is the natural number following 188 and preceding 190.
In mathematics
189 is a centered cube number and a heptagonal number.
The centered cube numbers are the sums of two consecutive cubes, and 189 can be written as sum of two cubes in two ways: and The smallest number that can be written as the sum of two positive cubes in two ways is 1729.
There are 189 zeros among the decimal digits of the positive integers with at most three digits.
The largest prime number that can be represented in 256-bit arithmetic is the "ultra-useful prime" used in quasi-Monte Carlo methods and in some cryptographic systems.
See also
The year AD 189 or 189 BC
List of highways numbered 189
|
https://en.wikipedia.org/wiki/Adhesive%20category
|
In mathematics, an adhesive category is a category where pushouts of monomorphisms exist and work more or less as they do in the category of sets. An example of an adhesive category is the category of directed multigraphs, or quivers, and the theory of adhesive categories is important in the theory of graph rewriting.
More precisely, an adhesive category is one where any of the following equivalent conditions hold:
C has all pullbacks, it has pushouts along monomorphisms, and pushout squares of monomorphisms are also pullback squares and are stable under pullback.
C has all pullbacks, it has pushouts along monomorphisms, and the latter are also (bicategorical) pushouts in the bicategory of spans in C.
If C is small, we may equivalently say that C has all pullbacks, has pushouts along monomorphisms, and admits a full embedding into a Grothendieck topos preserving pullbacks and preserving pushouts of monomorphisms.
|
https://en.wikipedia.org/wiki/Semipermeable%20membrane
|
Semipermeable membrane is a type of biological or synthetic, polymeric membrane that will allow certain molecules or ions to pass through it by osmosis. The rate of passage depends on the pressure, concentration, and temperature of the molecules or solutes on either side, as well as the permeability of the membrane to each solute. Depending on the membrane and the solute, permeability may depend on solute size, solubility, properties, or chemistry. How the membrane is constructed to be selective in its permeability will determine the rate and the permeability. Many natural and synthetic materials which are rather thick are also semipermeable. One example of this is the thin film on the inside of the egg.
Biological membranes are selectively permeable, with the passage of molecules controlled by facilitated diffusion, passive transport or active transport regulated by proteins embedded in the membrane.
Biological membranes
An example of a biological semi-permeable membrane is the lipid bilayer, on which is based the plasma membrane that surrounds all biological cells. A group of phospholipids (consisting of a phosphate head and two fatty acid tails) arranged into a double layer, the phospholipid bilayer is a semipermeable membrane that is very specific in its permeability. The hydrophilic phosphate heads are in the outside layer and exposed to the water content outside and within the cell. The hydrophobic tails are the layer hidden in the inside of the membrane. Cholesterol molecules are also found throughout the plasma membrane and act as a buffer of membrane fluidity. The phospholipid bilayer is most permeable to small, uncharged solutes. Protein channels are embedded in or through phospholipids, and, collectively, this model is known as the fluid mosaic model. Aquaporins are protein channel pores permeable to water.
Cellular communication
Information can also pass through the plasma membrane when signaling molecules bind to receptors in the cell membrane. Th
|
https://en.wikipedia.org/wiki/Common%20Weakness%20Enumeration
|
The Common Weakness Enumeration (CWE) is a category system for hardware and software weaknesses and vulnerabilities. It is sustained by a community project with the goals of understanding flaws in software and hardware and creating automated tools that can be used to identify, fix, and prevent those flaws. The project is sponsored by the office of the U.S. Department of Homeland Security (DHS) Cybersecurity and Infrastructure Security Agency (CISA), which is operated by The MITRE Corporation, with support from US-CERT and the National Cyber Security Division of the U.S. Department of Homeland Security.
Version 4.10 of the CWE standard was released in July 2021.
CWE has over 600 categories, including classes for buffer overflows, path/directory tree traversal errors, race conditions, cross-site scripting, hard-coded passwords, and insecure random numbers.
Examples
CWE category 121 is for stack-based buffer overflows.
CWE compatibility
Common Weakness Enumeration (CWE) Compatibility program allows a service or a product to be reviewed and registered as officially "CWE-Compatible" and "CWE-Effective". The program assists organizations in selecting the right software tools and learning about possible weaknesses and their possible impact.
In order to obtain CWE Compatible status a product or a service must meet 4 out of 6 requirements, shown below:
There are 56 organizations as of September 2019 that develop and maintain products and services that achieved CWE Compatible status.
Research, critiques, and new developments
Some researchers think that ambiguities in CWE can be avoided or reduced.
See also
Common Vulnerabilities and Exposures (CVE)
Common Vulnerability Scoring System (CVSS)
National Vulnerability Database
|
https://en.wikipedia.org/wiki/Abstraction-Filtration-Comparison%20test
|
The Abstraction-Filtration-Comparison test (AFC) is a method of identifying substantial similarity for the purposes of applying copyright law. In particular, the AFC test is used to determine whether non-literal elements of a computer program have been copied by comparing the protectable elements of two programs. The AFC test was developed by the United States Court of Appeals for the Second Circuit in 1992 in its opinion for
Computer Associates Int. Inc. v. Altai Inc. It has been widely adopted by United States courts and recognized by courts outside the United States as well.
Background
The Second Circuit Court of Appeals developed the AFC test for use in Computer Associates Int'l, Inc. v. Altai Inc. In that case, Computer Associates sued Altai for copyright infringement of a computer job scheduler program that was designed to be easily ported between operating systems. Proving copyright infringement requires proving both ownership of the copyright and that copying took place. This second requirement can be met either by direct proof, or as is more usually done, by demonstrating the following: 1) the defendant had access to the copyright material and 2) there is substantial similarity between the copyrighted work and the defendant's work. Demonstrating substantial similarity can be difficult when the two works are not exact replicas, either in full or in part. The Second Circuit court had found there was little previous guidance on how best to do this.
One notable previous treatment of substantial similarity of the structure, sequence and organization (SSO) of software was adopted by the Third Circuit in Whelan v. Jaslow. The court there suggested identifying the main function of a program as the idea and everything that is not strictly necessary for the purposes of the idea can be considered expression. The Altai court declined to follow this method, noting that the Whelan method "did not place enough emphasis on practical considerations". The AFC test was de
|
https://en.wikipedia.org/wiki/Fibre%20Channel
|
Fibre Channel (FC) is a high-speed data transfer protocol providing in-order, lossless delivery of raw block data. Fibre Channel is primarily used to connect computer data storage to servers in storage area networks (SAN) in commercial data centers.
Fibre Channel networks form a switched fabric because the switches in a network operate in unison as one big switch. Fibre Channel typically runs on optical fiber cables within and between data centers, but can also run on copper cabling. Supported data rates include 1, 2, 4, 8, 16, 32, 64, and 128 gigabit per second resulting from improvements in successive technology generations. The industry now notates this as Gigabit Fibre Channel (GFC).
There are various upper-level protocols for Fibre Channel, including two for block storage. Fibre Channel Protocol (FCP) is a protocol that transports SCSI commands over Fibre Channel networks. FICON is a protocol that transports ESCON commands, used by IBM mainframe computers, over Fibre Channel. Fibre Channel can be used to transport data from storage systems that use solid-state flash memory storage medium by transporting NVMe protocol commands.
Etymology
When the technology was originally devised, it ran over optical fiber cables only and, as such, was called "Fiber Channel". Later, the ability to run over copper cabling was added to the specification. In order to avoid confusion and to create a unique name, the industry decided to change the spelling and use the British English fibre for the name of the standard.
History
Fibre Channel is standardized in the T11 Technical Committee of the International Committee for Information Technology Standards (INCITS), an American National Standards Institute (ANSI)-accredited standards committee. Fibre Channel started in 1988, with ANSI standard approval in 1994, to merge the benefits of multiple physical layer implementations including SCSI, HIPPI and ESCON.
Fibre Channel was designed as a serial interface to overcome limitati
|
https://en.wikipedia.org/wiki/MPEG%20program%20stream
|
Program stream (PS or MPEG-PS) is a container format for multiplexing digital audio, video and more. The PS format is specified in MPEG-1 Part 1 (ISO/IEC 11172-1) and MPEG-2 Part 1, Systems (ISO/IEC standard 13818-1/ITU-T H.222.0). The MPEG-2 Program Stream is analogous and similar to ISO/IEC 11172 Systems layer and it is forward compatible.
Program streams are used on DVD-Video discs and HD DVD video discs, but with some restrictions and extensions. The filename extensions are VOB and EVO respectively.
Coding structure
Program streams are created by combining one or more Packetized Elementary Streams (PES), which have a common time base, into a single stream. It is designed for reasonably reliable media such as disks, in contrast to MPEG transport stream which is for data transmission in which loss of data is likely. Program streams have variable size records and minimal use of start codes which would make over the air reception difficult, but has less overhead. Program stream coding layer allows only one program of one or more elementary streams to be packaged into a single stream, in contrast to transport stream, which allows multiple programs.
MPEG-2 Program stream can contain MPEG-1 Part 2 video, MPEG-2 Part 2 video, MPEG-1 Part 3 audio (MP3, MP2, MP1) or MPEG-2 Part 3 audio. It can also contain MPEG-4 Part 2 video, MPEG-2 Part 7 audio (AAC) or MPEG-4 Part 3 (AAC) audio, but they are rarely used. The MPEG-2 Program stream has provisions for non-standard data (e.g. AC-3 audio or subtitles) in the form of so-called private streams. International Organization for Standardization authorized SMPTE Registration Authority, LLC as the registration authority for MPEG-2 format identifiers. It publishes a list of compression formats which can be encapsulated in MPEG-2 transport stream and program stream.
Coding details
See also
Elementary stream
MPEG transport stream
|
https://en.wikipedia.org/wiki/Za%27atar
|
Za'atar ( ; , ) is a culinary herb or family of herbs. It is also the name of a spice mixture that includes the herb along with toasted sesame seeds, dried sumac, often salt, as well as other spices. As a family of related Middle Eastern herbs, it contains plants from the genera Origanum (oregano), Calamintha (basil thyme), Thymus (typically Thymus vulgaris, i.e., thyme), and Satureja (savory) plants. The name za'atar alone most properly applies to Origanum syriacum, considered in biblical scholarship to be the ezov of the Hebrew Bible, often translated as hyssop but distinct from modern Hyssopus officinalis.
Used in Levantine cuisine, both the herb and spice mixture are popular throughout the Mediterranean region of the Middle East.
Etymology
According to Ignace J. Gelb, an Akkadian language word that can be read sarsar may refer to a spice plant. This word could be attested in the Syriac satre, and Arabic za'atar (or sa'tar), possibly the source of Latin Satureia. Satureia (Satureja) is a common name for Satureja thymbra, a species of savory whose other common and ethnic names include, "Persian za'atar", "za'atar rumi" (Roman hyssop), and "za'atar franji" (European hyssop). In the Modern Hebrew language, za'atar is used as an Arabic loanword.
Thymus capitatus (also called Satureja capitata) is a species of wild thyme found throughout the hills of the Levant and Mediterranean Middle East. Thyme is said to be a plant "powerfully associated with Palestine", and the spice mixture za'atar is common fare there. Thymbra spicata, a plant native to Greece and to Palestine and has been cultivated in North America by Syrian, Palestinian, and Lebanese immigrants for use in their za'atar preparations since the 1940s.
Another species identified as "wild za'atar" (Arabic: za'atar barri) is Origanum vulgare, commonly known as European oregano, oregano, pot marjoram, wild marjoram, winter marjoram, or wintersweet. This species is also extremely common in Lebanon, Syria, Is
|
https://en.wikipedia.org/wiki/Rifleman%27s%20rule
|
Rifleman's rule is a "rule of thumb" that allows a rifleman to accurately fire a rifle that has been calibrated for horizontal targets at uphill or downhill targets. The rule says that only the horizontal range should be considered when adjusting a sight or performing hold-over in order to account for bullet drop. Typically, the range of an elevated target is considered in terms of the slant range, incorporating both the horizontal distance and the elevation distance (possibly negative, i.e. downhill), as when a rangefinder is used to determine the distance to target. The slant range is not compatible with standard ballistics tables for estimating bullet drop.
The Rifleman's rule provides an estimate of the horizontal range for engaging a target at a known slant range (the uphill or downhill distance from the rifle). For a bullet to strike a target at a slant range of and an incline of , the rifle sight must be adjusted as if the shooter were aiming at a horizontal target at a range of . Figure 1 illustrates the shooting scenario. The rule holds for inclined and declined shooting (all angles measured with respect to horizontal). Very precise computer modeling and empirical evidence suggests that the rule does appear to work with reasonable accuracy in air and with both bullets and arrows.
Background
Definitions
There is a device that is mounted on the rifle called a sight. While there are many forms of rifle sight, they all permit the shooter to set the angle between the bore of the rifle and the line of sight (LOS) to the target. Figure 2 illustrates the relationship between the LOS and bore angle.
This relationship between the LOS to the target and the bore angle is determined through a process called "zeroing." The bore angle is set to ensure that a bullet on a parabolic trajectory will intersect the LOS to the target at a specific range. A properly adjusted rifle barrel and sight are said to be "zeroed." Figure 3 illustrates how the LOS, bullet trajecto
|
https://en.wikipedia.org/wiki/Indian%20Institute%20of%20Science%20Education%20and%20Research%2C%20Thiruvananthapuram
|
Indian Institute of Science Education and Research, Thiruvananthapuram (IISER Thiruvananthapuram or IISER-TVM) is an autonomous public university located in Thiruvananthapuram, Kerala, India. The institute is one of the seven IISERs established by the Ministry of Human Resource Development, Government of India, to bridge the gap between research and basic sciences' education at the undergraduate level. All IISERs are declared as Institutes of National Importance by the Parliament of India in 2012 through the NIT Amendment Act.
IISER Thiruvananthapuram awards Bachelor of Science and Master of Science (BS-MS) dual-degrees, and Doctor of Philosophy (Ph.D.) degrees in Mathematics, Physics, Chemistry, and Biology. IISER TVM is to offer five new BS-MS programmes (i2 Sciences) and new two-year M.Sc programmes in 2021.
History
Indian Institutes of Science Education and Research (IISERs) were established in 2006 through a proclamation of the Ministry of Education, Government of India to promote quality collegiate education and research in basic sciences. Each IISER is a degree-granting autonomous institution with a prime focus to integrate science education and research. IISERs receive generous funding from the Government of India.
Soon after MHRD's announcement in 2006, two of these institutes were established at Pune and Kolkata in 2006. This was followed by another institute at Mohali in 2007, and two more at Bhopal and Thiruvananthapuram in 2008. More recently, two more IISERs were established at Tirupati and Berhampur in 2015 and 2016 respectively.
Campus
IISER TVM campus is located in the foothills of the Western Ghats mountain. Spanning around 200 acres, it is located 40 km from Thiruvananthapuram, Kerala.
Organization and administration
Governance
The institute is administered by a Board of Governors with Madhavan Nair Rajeevan as Chairperson, S. Murty Srinivasula as Deputy Director, a 13-member Senate, and the Director, Prof. Jarugu Narasimha Moorthy. The
|
https://en.wikipedia.org/wiki/Galactic%20quadrant
|
A galactic quadrant, or quadrant of the Galaxy, is one of four circular sectors in the division of the Milky Way Galaxy.
Quadrants in the galactic coordinate system
In actual astronomical practice, the delineation of the galactic quadrants is based upon the galactic coordinate system, which places the Sun as the pole of the mapping system. The Sun is used instead of the Galactic Center for practical reasons since all astronomical observations (by humans) to date have been based on Earth or within the Solar System.
Delineation
Quadrants are described using ordinals—for example, "1st galactic quadrant", "second galactic quadrant", or "third quadrant of the Galaxy". Viewing from the north galactic pole with 0 degrees (°) as the ray that runs starting from the Sun and through the galactic center, the quadrants are as follows (where is galactic longitude):
1st galactic quadrant – 0° ≤ ≤ 90°
2nd galactic quadrant – 90° ≤ ≤ 180°
3rd galactic quadrant – 180° ≤ ≤ 270°
4th galactic quadrant – 270° ≤ ≤ 360°
Constellations grouped by galactic quadrants
Visibility of each quadrant
Due to the orientation of the Earth with respect to the rest of the galaxy, the 2nd galactic quadrant is primarily only visible from the northern hemisphere while the 4th galactic quadrant is mostly only visible from the southern hemisphere. Thus, it is usually more practical for amateur stargazers to use the celestial quadrants. Nonetheless, cooperating or international astronomical organizations are not so bound by the Earth's horizon.
Based on a view from Earth, one may look towards major constellations for a rough sense of where the borders of the quadrants are: (Note: by drawing a line through the following, one can also approximate the galactic equator.)
For 0°, look towards the Sagittarius constellation. (The galactic center)
For 90°, look towards the Cygnus constellation.
For 180°, look towards the Auriga constellation. (The galactic anticenter)
For 270°, look towards the Ve
|
https://en.wikipedia.org/wiki/Covariance%20intersection
|
Covariance intersection (CI) is an algorithm for combining two or more estimates of state variables in a Kalman filter when the correlation between them is unknown.
Formulation
Items of information a and b are known and are to be fused into information item c. We know a and b have mean/covariance , and , , but the cross correlation is not known. The covariance intersection update gives mean and covariance for c as
where ω is computed to minimize a selected norm, e.g., the trace, or the logarithm of the determinant. While it is necessary to solve an optimization problem for higher dimensions, closed-form solutions exist for lower dimensions.
Application
CI can be used in place of the conventional Kalman update equations to ensure that the resulting estimate is conservative, regardless of the correlation between the two estimates, with covariance strictly non-increasing according to the chosen measure. The use of a fixed measure is necessary for rigor to ensure that a sequence of updates does not cause the filtered covariance to increase.
Advantages
According to a recent survey paper and, the covariance intersection has the following advantages:
The identification and computation of the cross covariances are completely avoided.
It yields a consistent fused estimate, and thus a non-divergent filter is obtained.
The accuracy of the fused estimate outperforms each local one.
It gives a common upper bound of actual estimation error variances, which has robustness with respect to unknown correlations.
These advantages have been demonstrated in the case of simultaneous localization and mapping (SLAM) involving over a million map features/beacons.
Motivation
It is widely believed that unknown correlations exist in a diverse range of multi-sensor fusion problems. Neglecting the effects of unknown correlations can result in severe performance degradation, and even divergence. As such, it has attracted and sustained the attention of researchers for decades
|
https://en.wikipedia.org/wiki/Novell%20ZENworks%20Application%20VIrtualization
|
Novell ZENworks Application Virtualization is an application virtualization and portable application creation console by Novell that allows a user to build applications that run like an executable file.
This product is now known as Micro Focus Desktop Containers.
History
ZENworks Application Virtualization was first made available on September 2, 2008. Since its initial release, it has undergone additional major releases, including versions 6.x, 7.x and the current version 8.x.
Architecture & Technology
ZENworks Application Virtualization is an application virtualization solution that allows users to build highly portable virtual Windows applications that run like common executable files. Virtual applications are self-contained and include an isolated sandbox that interfaces with the operating system registry indirectly. Virtualized applications run without an agent/server architecture and without a local client install.
The technology also supports application streaming, a form of on-demand application distribution. This type of application streaming uses a predictive algorithm that maps common user behavior in applications to anticipate what portions of application code should be sent first, based upon changing user behavior, in an effort to improve performance.
Building virtual applications
The process for creating a virtual application with ZENworks Application Virtualization follows one of three paths:
Auto-configured applications: Use wizards to build pre-configured applications. Included applications are limited to popular or common applications.
Snap-Shot: Take "snapshots" of an example system's state before and after the target application is installed.
Manual configuration: Manually configure the registry, sandbox, and other portions of the target application. This approach typically requires an experienced programmer.
Users can then define other application behavior, such as setting an application to expire on a certain date, or to make the appl
|
https://en.wikipedia.org/wiki/Istv%C3%A1n%20Ruszny%C3%A1k
|
István Rusznyák (Budapest, 22 January 1889 – Budapest, 15 October 1974), was a Hungarian physician. He was the President of the Hungarian Academy of Sciences between 1949 and 1970.
Biography
Rusznyák came from a family of Jewish intellectuals. In 1911 he got a diploma in medicine from the Budapest University of Medical Sciences. He worked in the Pathology Department. He fought in the First World War. In 1926, private professor lecturers, he worked as a lecturer on.
Between 1931 and 1944 he was the director of the Department of Medicine i Medical Faculty of University of Szeged. In 1937/38 school year he was elected dean. With start of Second World War, in 1944, he and his family were deported but was returned from Austria.( eyewitness) After the end of the Second World War, he worked as the Head of the Department of Internal Medicine. In 1946, he was elected a full member of the Hungarian Academy of Sciences and was elected its president in 1949. In 1963, he retired as a university professor but he continued as the President of the Hungarian Academy of Sciences until 1970. He remained an academic advisor from 1971 until his death in 1974.
Contributions
Along with Albert Szent-Györgyi he discovered Vitamin P and proved that chemically it belongs to the flavones.
|
https://en.wikipedia.org/wiki/Dataram
|
Dataram (formerly Dataram Corporation) is a manufacturer of computer memory and software products headquartered in Princeton, New Jersey. Dataram Memory was founded in 1967. It provided core memory for many early Digital Equipment computer systems. Dataram products include memory and storage, and related technical products and services for desktops, laptops, workstations and servers. The company sells worldwide to OEMs, distributors, value-added resellers, embedded manufacturers, enterprise customers, and end users. Dataram provides compatible server memory for companies including HP, IBM, Sun Microsystems, Oracle, Dell, Cisco, Lenovo, Intel and AMD. The company's manufacturing facility is in Montgomeryville, Pennsylvania, United States, and has sales offices in the United States, Europe, China, and Japan.
History
Dataram was founded in 1967 as a manufacturer of computer memory for Digital Equipment Corporation (DEC), selling 16 KB core memory.
In 1968, Dataram completed its initial public offering, and began trading on the American Stock Exchange under the symbol DTM.
In 1974, the company developed memory for Digital PDP-11.
In 1976 the company produced the first solid-state drive called BULK CORE for DEC and Data General computers.
In 1987, Dataram produced memory for Sun Microsystems and Apollo workstations.
In 1989, they produced memory for HP/Apollo systems.
In 1999, the company moved from the American Stock Exchange to NASDAQ, trading under the symbol DRAM. They earned the highest number of Intel Advanced Memory Module Qualifications that year.
In 2001, the company acquired Memory Card Technology assets, and again earned the highest number of Intel Advanced Memory Module Qualifications for the year.
In 2002, the company's DDR memory modules for the Intel market received validation from Advanced Validation Labs.
In 2008, the company signed a service and support agreement with IBM, on 9 October 2018 acquired Cenatek Inc., a privately owned company loc
|
https://en.wikipedia.org/wiki/Learnability
|
Learnability is a quality of products and interfaces that allows users to quickly become familiar with them and able to make good use of all their features and capabilities.
Software testing
In software testing learnability, according to ISO/IEC 9126, is the capability of a software product to enable the user to learn how to use it. Learnability may be considered as an aspect of usability, and is of major concern in the design of complex software applications.
Learnability is defined in the Standard glossary of terms used in software testing published by the International Software Testing Qualifications Board.
Computational learning theory
In computational learning theory, learnability is the mathematical analysis of machine learning. It is also employed in language acquisition in arguments within linguistics.
Frameworks include:
Language identification in the limit proposed in 1967 by E. Mark Gold. Subsequently known as Algorithmic learning theory.
Probably approximately correct learning (PAC learning) proposed in 1984 by Leslie Valiant
|
https://en.wikipedia.org/wiki/Myrmecotrophy
|
Myrmecotrophy is the ability of plants to obtain nutrients from ants, a form of mutualism. Due to this behaviour the invasion of vegetation into harsh environments is promoted. The dead remains of insects thrown out by the ants are absorbed by the lenticular warts in myrmecophytes like Hydnophytum and Myrmecodia. Myrmecodia uses its lenticular warts to suck nutrients from the insects thrown out by the ants. The ants in turn benefit with a secure location to form their colony. The pitcher plant Nepenthes bicalcarata obtains an estimated 42% of its total foliar nitrogen from ant waste.
|
https://en.wikipedia.org/wiki/TVE%20test%20card
|
The TVE colour test card (Spanish: Carta de ajuste en color de TVE) was an electronic analogue TV test card adopted by Televisión Española with the introduction of PAL colour broadcasts in 1975. It is notable for its unique design, created by the Danish engineer Finn Hendil in 1973, under the supervision of Erik Helmer Nielsen at the Philips TV & Test Equipment laboratory in Brøndby Municipality near Copenhagen, the same team that developed the popular Philips PM5544 test pattern. It replaced a previous black and white version developed by Eduardo Gavilán.
The test card was considered to be part of the regular TV schedule, figuring among daily program listings published in newspapers and magazines. It was said to be the most viewed program in some days due to people watching the test card while waiting for broadcasts to start in the afternoon. It was also relevant in the context of general work strikes, where the test card was sometimes broadcast in place of regular programming, marking it a visible sign of the strike's success.
It was used on several TVE channels, like La Primera, TVE 2, Canal Clásico, Teledeporte or TVE Internacional.
With the start of continuous 24-hour broadcasting on TVE's channels, the test card was phased out. It stopped being broadcast on La Primera in 1996 and on La 2 in the early morning hours of 6 January 2001, although it continued to be broadcast sporadically on Teledeporte and TVE Internacional until 2005.
Operation and features
As Televisión Española adopted the PAL colour system in 1975, the test card has specific elements that allow proper colour adjustments. Being a creation of the same team behind the Philips PM5544 test card, it has many elements in common with it (like colour and grey bars or castellations), but introduces some differences (for example, different resolution gratings and coloured background rectangle and circle).
There were two generations of the TVE test card. The original was generated by a heavily modif
|
https://en.wikipedia.org/wiki/TinyMUCK
|
TinyMUCK or, more broadly, a MUCK, is a type of user-extendable online text-based role-playing game, designed for role playing and social interaction. Backronyms like "Multi-User Chat/Created/Computer/Character/Carnal Kingdom" and "Multi-User Construction Kit" are sometimes cited, but are not the actual origin of the term; "muck" is simply a play on the term MUD.
History
The original TinyMUCK 1.0 server was written by Stephen White from University of Waterloo in winter of 1990, based on TinyMUD 1.5.2 codebase. This version improved building capabilities for the users.
TinyMUCK 2.0 was released in June 1990 by Piaw "Lachesis" Na from Berkeley, who added the programming language MUF for in-game server extensions.
TinyMUCK 2.1 and 2.2 were released in July 1990 and April 1991 by Robert "ChupChup" Earl of San Diego, California. These were mostly bugfix releases as the code was cleaned up and ported to new operating systems and architectures.
FuzzBall MUCK server was built on TinyMUCK 2.2 codebase by Belfry Webworks and, as of version 5, released in 1995, includes the alternative programming language MPI. version 6, available at SourceForge project fbmuck also supports MCP and MCP-GUI.
Characteristics
MUCKs are extensible by design, players can create and modify ("build") all internal objects of the game environment, including rooms, exits, and even the system commands, for which the MUCKs use the MUF (Multi-User Forth) language. Fuzzball MUCKs also use Message Parsing Interpreter (MPI) which can be used to embed executable code into descriptions of all in-game objects. Unlike many other virtual worlds, however, TinyMUCK and its descendants do not usually have computer-controlled monsters for players to kill.
Usage
TinyMUCKs are popular among members of furry fandom; examples of active, large TinyMUCKs include FurryMUCK and Tapestries MUCK, both of which run the Fuzzball version of MUCK server code.
See also
MUD
MUSH
MOO
Online text-based role-playing game
|
https://en.wikipedia.org/wiki/Facility%20for%20Antiproton%20and%20Ion%20Research
|
The Facility for Antiproton and Ion Research (FAIR) is an international accelerator facility under construction which will use antiprotons and ions to perform research in the fields of: nuclear, hadron and particle physics, atomic and anti-matter physics, high density plasma physics, and applications in condensed matter physics, biology and the bio-medical sciences. It is situated in Darmstadt in Germany.
FAIR will be based upon an expansion of the GSI Helmholtz Centre for Heavy Ion Research, the details of which have been laid out in the FAIR Baseline Technical Report 2006. On October 4, 2010 the Facility for Antiproton and Ion Research in Europe limited liability company (German GmbH), abbreviated as FAIR GmbH, was founded which coordinates the construction of the new accelerators and experiments.
The construction begun in the summer of 2017. Commissioning is planned for 2025. The original budget had been estimated at 1262 million euro (2005 price level), in 2018 the German Federal Court of Auditors stated that the cost of FAIR had increased to at least 1669 million euro (2005 prices), in 2019 a further report by the Federal Court of Auditors quoted results from a review board that an additional 850 million euro (2019 prices) would be needed to finish building the facility.
FAIR science case
The four scientific pillars of FAIR are:
Atomic, Plasma Physics and Applications – APPA,
Compressed Baryonic Matter – CBM,
Nuclear Structure, Astrophysics and Reactions – NUSTAR,
antiProton ANnihilation at DArmstadt – PANDA.
Those are described on the web pages of FAIR (see and links therein).
FAIR accelerators
Beams of protons will be prepared in the proton linear accelerator, p-LINAC, while heavy ions will be prepared in the UNILAC. Both of them will be fed into the SIS18. From there they will be directed into SIS100 (and SIS300 if applicable). Protons will be used either to produce antiproton beams by directing them on a dedicated production target or directly u
|
https://en.wikipedia.org/wiki/Power%20Balance
|
Power Balance is the original brand of hologram bracelets claimed by its manufacturers and vendors to use "holographic technology" to "resonate with and respond to the natural energy field of the body" to increase athletic performance. Numerous independent studies of the device have found it to be no more effective than placebo for enhancing athletic performance. As a result, in 2010, the Australian distributor, Power Balance Australia Pty Ltd, was forced by the Australian Competition & Consumer Commission (ACCC) to retract any previous claims.
The product was originally promoted at trade shows in the beginning of 2006 utilizing applied kinesiology as its most effective sales tool. The bracelets went on sale in 2007 and had several celebrity endorsements. The bracelets became a trend among high school, collegiate, and professional sports teams between 2008 and 2012. This sustained prevalence compelled journalist Darren Rovell to remark that "a growing number of professional sportsmen and their attendants are starting to sound like New Age crystal healers." CNBC Sports named Power Balance Product of the Year in 2010 for its strong sales and celebrity endorsements.
Power Balance headquarters, which was located in Laguna Niguel, California at the time, denied that they made any medical or scientific claims about their products. However, the company had been the focus of significant criticism, particularly for false advertising. The Power Balance bracelet has been described as "like the tooth fairy" and a "very successful marketing scam". Dylan Evans, a lecturer in behavioral science at Cork University's School of Medicine, stated the marketing of Power Balance was impressive: "They have managed to get away without deceiving anyone in the sense of an overt lie. There are no claims on the packaging itself. They don't make any reference at all to any health outcomes. They leave that as an inference that most people will draw."
By the end of 2011 the company was reporte
|
https://en.wikipedia.org/wiki/Porina%20chlorotica
|
Porina chlorotica is a species of lichen belonging to the family Porinaceae.
Synonym:
Verrucaria chlorotica Ach., 1810 (= basionym)
|
https://en.wikipedia.org/wiki/Combinatorics%20of%20Finite%20Geometries
|
Combinatorics of Finite Geometries is an undergraduate mathematics textbook on finite geometry by Lynn Batten. It was published by Cambridge University Press in 1986 with a second edition in 1997 ().
Topics
The types of finite geometry covered by the book include partial linear spaces, linear spaces, affine spaces and affine planes, projective spaces and projective planes, polar spaces, generalized quadrangles, and partial geometries. A central connecting concept is the "connection number" of a point and a line not containing it, equal to the number of lines that meet the given point and intersect the given line.
The second edition adds a final chapter on blocking sets.
Beyond the basic theorems and proofs of this subject, the book includes many examples and exercises, and some history and information about current research.
Audience and reception
The book is aimed at advanced undergraduates, assuming only an introductory-level of abstract algebra and some knowledge of linear algebra. Its coverage of recent research also makes it useful as background reading for researchers in this area.
Reviewer Michael J. Kallaher cites as a "serious shortcoming" of the first edition its lack of coverage of applications of this subject, for instance to the design of experiments and to coding theory. The second edition has a section on applications but reviewer Tamás Szőnyi writes that it needs additional expansion.
Because of the many types of geometry covered in the book, the coverage of each of them is, at times, shallow; for instance, reviewer Theodore G. Ostrom complains that there is only half a page on non-Desarguesian planes. Additionally, Kallaher feels that block designs should have been included in place of some of the more esoteric geometries described by Batten. Reviewer Thomas Brylawski criticizes the book for "glossing over or ignoring" important results, for overcomplicated proofs, and for missed cases in some of its case analysis.
On the other hand, reviewer
|
https://en.wikipedia.org/wiki/Saliva%20testing
|
Saliva testing or Salivaomics is a diagnostic technique that involves laboratory analysis of saliva to identify markers of endocrine, immunologic, inflammatory, infectious, and other types of conditions. Saliva is a useful biological fluid for assaying steroid hormones such as cortisol, genetic material like RNA, proteins such as enzymes and antibodies, and a variety of other substances, including natural metabolites, including saliva nitrite, a biomarker for nitric oxide status (see below for Cardiovascular Disease, Nitric Oxide: a salivary biomarker for cardio-protection).
Saliva testing is used to screen for or diagnose numerous conditions and disease states, including Cushing's disease, anovulation, HIV, cancer, parasites, hypogonadism, and allergies. Salivary testing has even been used by the U.S. government to assess circadian rhythm shifts in astronauts before flight and to evaluate hormonal profiles of soldiers undergoing military survival training.
Proponents of saliva testing cite its ease of collection, safety, non-invasiveness, affordability, accuracy, and capacity to circumvent venipuncture as the primary advantages when compared to blood testing and other types of diagnostic testing. Additionally, since multiple samples can be readily obtained, saliva testing is particularly useful for performing chronobiological assessments that span hours, days, or weeks. Collecting whole saliva by passive drool has a myriad of advantages. Passive drool collection facilitates large sample size collection. Consequently, this allows the sample to be tested for more than one biomarker. It also gives the researcher the ability to freeze the left over specimen to be used at a later time. Additionally, it lessens the possibility of contamination by eliminating extra collection devices and the need to induce saliva flow.
The testing of salivation by the use of mercury was performed at least as early as 1685. Testing the acidity of saliva occurred at least as early as
|
https://en.wikipedia.org/wiki/Statistical%20genetics
|
Statistical genetics is a scientific field concerned with the development and application of statistical methods for drawing inferences from genetic data. The term is most commonly used in the context of human genetics. Research in statistical genetics generally involves developing theory or methodology to support research in one of three related areas:
population genetics - Study of evolutionary processes affecting genetic variation between organisms
genetic epidemiology - Studying effects of genes on diseases
quantitative genetics - Studying the effects of genes on 'normal' phenotypes
Statistical geneticists tend to collaborate closely with geneticists, molecular biologists, clinicians and bioinformaticians. Statistical genetics is a type of computational biology.
|
https://en.wikipedia.org/wiki/NEDD4L
|
Neural precursor cell expressed developmentally downregulated gene 4-like (NEDD4L) or NEDD4-2 is an enzyme (ubiquitin ligase) of the NEDD4 family.
In human the protein is encoded by the NEDD4L gene. In mouse the protein is commonly known as NEDD4-2 and the gene Nedd4-2.
NEDD4-2 has been shown to ubiquitinate and therefore down regulate the epithelial sodium channel (ENaC) in the collecting ducts of the kidneys, therefore opposing the actions of aldosterone and increasing salt excretion. In Liddle's Syndrome NEDD4 is unable to bind to the ENaC and lead to salt retention and hypertension occur.
NEDD4L belongs to the NEDD4 family of E3 HECT domain ubiquitin ligases. It is the closest homologue of NEDD4, the prototypic member of the family and probably arose as a result of gene duplication. While NEDD4 orthologues are present in all eukaryotes, NEDD4L proteins are limited to vertebrates. NEDD4L proteins are known to be involved in regulating many membrane proteins via ubiquitination and endocytosis.
NEDD4L protein is expressed widely. The primary targets of NEDD4-2 include the epithelial sodium channel (ENaC), the Na+-Cl- co-transporter (NCC), and the voltage gated sodium channels (Navs), although additional targets are predicted from in vitro studies. NEDD4-2 gene in mice is essential for animal survival and the polymorphisms in NEDD4L are associated with human hypertension.
Protein architecture
The NEDD4-2 protein consists of an amino-terminal Ca2+-phospholipid binding domain (C2), 4 WW domains (protein-protein interaction domains) and the carboxyl-terminal HECT domain (ubiquitin ligase domain). The WW domains in the protein are responsible for binding the substrates, regulatory proteins and adaptors. These domains generally recognize PPxY (or similar) motifs in the target proteins.
Expression
Human NEDD4L gene is located on chromosome 18q12.31 with 38 exons that transcribe multiple splice variants of NEDD4L. The protein expressed in the brain, lung, heart and
|
https://en.wikipedia.org/wiki/Constant-Q%20transform
|
In mathematics and signal processing, the constant-Q transform and variable-Q transform, simply known as CQT and VQT, transforms a data series to the frequency domain. It is related to the Fourier transform and very closely related to the complex Morlet wavelet transform. Its design is suited for musical representation.
The transform can be thought of as a series of filters fk, logarithmically spaced in frequency, with the k-th filter having a spectral width δfk equal to a multiple of the previous filter's width:
where δfk is the bandwidth of the k-th filter, fmin is the central frequency of the lowest filter, and n is the number of filters per octave.
Calculation
The short-time Fourier transform of x[n] for a frame shifted to sample m is calculated as follows:
Given a data series at sampling frequency fs = 1/T, T being the sampling period of our data, for each frequency bin we can define the following:
Filter width, δfk.
Q, the "quality factor":
This is shown below to be the integer number of cycles processed at a center frequency fk. As such, this somewhat defines the time complexity of the transform.
Window length for the k-th bin:
Since fs/fk is the number of samples processed per cycle at frequency fk, Q is the number of integer cycles processed at this central frequency.
The equivalent transform kernel can be found by using the following substitutions:
The window length of each bin is now a function of the bin number:
The relative power of each bin will decrease at higher frequencies, as these sum over fewer terms. To compensate for this, we normalize by N[k].
Any windowing function will be a function of window length, and likewise a function of window number. For example, the equivalent Hamming window would be
Our digital frequency, , becomes .
After these modifications, we are left with
Variable-Q bandwidth calculation
The variable-Q transform is the same as constant-Q transform, but the only difference is the filter Q is variable,
|
https://en.wikipedia.org/wiki/Hadamard%27s%20gamma%20function
|
In mathematics, Hadamard's gamma function, named after Jacques Hadamard, is an extension of the factorial function, different from the classical gamma function (it is an instance of a pseudogamma function.) This function, with its argument shifted down by 1, interpolates the factorial and extends it to real and complex numbers in a different way than Euler's gamma function. It is defined as:
where denotes the classical gamma function. If is a positive integer, then:
Properties
Unlike the classical gamma function, Hadamard's gamma function is an entire function, i.e. it has no poles in its domain. It satisfies the functional equation
with the understanding that is taken to be for positive integer values of .
Representations
Hadamard's gamma can also be expressed as
where is the Lerch zeta function, and as
where denotes the digamma function.
|
https://en.wikipedia.org/wiki/Reduced%20dimensions%20form
|
In biophysics and related fields, reduced dimension forms (RDFs) are unique on-off mechanisms for random walks that generate two-state trajectories (see Fig. 1 for an example of a RDF and Fig. 2 for an example of a two-state trajectory). It has been shown that RDFs solve two-state trajectories, since only one RDF can be constructed from the data, where this property does not hold for on-off kinetic schemes, where many kinetic schemes can be constructed from a particular two-state trajectory (even from an ideal on-off trajectory). Two-state time trajectories are very common in measurements in chemistry, physics, and the biophysics of individual molecules (e.g. measurements of protein dynamics and DNA and RNA dynamics, activity of ion channels, enzyme activity, quantum dots ), thus making RDFs an important tool in the analysis of data in these fields.
Since RDFs are uniquely obtained from the data, they have many advantages over other mathematical and statistical methods that were developed for solving two-state trajectories.
Description of RDF
A RDF is a lattice of substates, each substate represents either the on state or the off state, and has a particular number (see Figure 1). The connections are only among substates of different states.
A simulation of an on-off trajectory from a RDF is made with a generalized Gillespie algorithm, where here a random jumping time is first taken from density functions that are (usually) not exponential using the rejection method, and then the specific next substate is chosen according to the jumping probabilities that are determined from the jumping time probability density functions.
A RDF can have irreversible connections, yet, it generates an on-off trajectory that has the property of microscopic reversibility, meaning that the physical system fluctuates around equilibrium.
Two-state trajectories
A two-state trajectory is a fluctuating signal made of on periods and off periods; an on period, and then an off period, and s
|
https://en.wikipedia.org/wiki/Trauma-sensitive%20yoga
|
Trauma-sensitive yoga is yoga as exercise, adapted from 2002 onwards for work with individuals affected by psychological trauma. Its goal is to help trauma survivors to develop a greater sense of mind-body connection, to ease their physiological experiences of trauma, to gain a greater sense of ownership over their bodies, and to augment their overall well-being. However, a 2019 systematic review found that the studies to date were not sufficiently robustly designed to provide strong evidence of yoga's effectiveness as a therapy; it called for further research.
Background
Psychological trauma and the body
Psychological trauma occurs when an individual has experienced a traumatic event which becomes lived and relived in the body and the mind. Trauma can trigger a chronic stress response in the body, which may manifest as an uncontrollable and constant state of heightened arousal and fear. Those with a history of trauma may also interpret this chronic stress response as a threat to their sense of self and relation to the world. Traumatized individuals often have difficulty soothing their overactive internal sensations without relying on external stimuli, such as food, substances, or self-harm. Therefore, psychological trauma is not only associated with psychological disorders such as post-traumatic stress disorder (PTSD), depression, and anxiety, but also with somatic disorders.
Though most evidence-based treatments focus on the psychological effects of trauma first and foremost, attrition rates are still high, possibly due to heightened physiological arousal during the initial stages of exposure therapy. Mind-body approaches offer a complementary method to traditional psychotherapy, allowing traumatized individuals to reconnect with and identify their own physical sensations. Mind-body approaches allow participants to work through their somatic trauma memories and feel safe enough to emotionally and verbally process their traumatic memories. Such approaches invo
|
https://en.wikipedia.org/wiki/Quantum%20logic%20clock
|
A quantum clock is a type of atomic clock with laser cooled single ions confined together in an electromagnetic ion trap. Developed in 2010 by physicists at the U.S. National Institute of Standards and Technology, the clock was 37 times more precise than the then-existing international standard. The quantum logic clock is based on an aluminium spectroscopy ion with a logic atom.
Both the aluminum-based quantum clock and the mercury-based optical atomic clock track time by the ion vibration at an optical frequency using a UV laser, that is 100,000 times higher than the microwave frequencies used in NIST-F1 and other similar time standards around the world. Quantum clocks like this are able to be far more precise than microwave standards.
Accuracy
The NIST team are not able to measure clock ticks per second because the definition of a second is based on the standard NIST-F1, which cannot measure a machine more precise than itself. However, the aluminum ion clock's measured frequency to the current standard is . NIST have attributed the clock's accuracy to the fact that it is insensitive to background magnetic and electric fields, and unaffected by temperature.
In March 2008, physicists at NIST described an experimental quantum logic clock based on individual ions of beryllium and aluminum. This clock was compared to NIST's mercury ion clock. These were the most accurate clocks that had been constructed, with neither clock gaining nor losing time at a rate that would exceed a second in over a billion years.
In February 2010, NIST physicists described a second, enhanced version of the quantum logic clock based on individual ions of magnesium and aluminium. Considered the world's most precise clock in 2010 with a fractional frequency inaccuracy of , it offers more than twice the precision of the original.
In terms of standard deviation, the quantum logic clock deviates one second every 3.68 billion () years, while the then current international standard NIST-F1 C
|
https://en.wikipedia.org/wiki/Orthogonal%20Procrustes%20problem
|
The orthogonal Procrustes problem is a matrix approximation problem in linear algebra. In its classical form, one is given two matrices and and asked to find an orthogonal matrix which most closely maps to . Specifically, the orthogonal Procrustes problem is an optimization problem given by
where denotes the Frobenius norm. This is a special case of Wahba's problem (with identical weights; instead of considering two matrices, in Wahba's problem the columns of the matrices are considered as individual vectors). Another difference is, that Wahba's problem tries to find a proper rotation matrix, instead of just an orthogonal one.
The name Procrustes refers to a bandit from Greek mythology who made his victims fit his bed by either stretching their limbs or cutting them off.
Solution
This problem was originally solved by Peter Schönemann in a 1964 thesis, and shortly after appeared in the journal Psychometrika.
This problem is equivalent to finding the nearest orthogonal matrix to a given matrix , i.e. solving the closest orthogonal approximation problem
.
To find matrix , one uses the singular value decomposition (for which the entries of are non-negative)
to write
Proof of Solution
One proof depends on basic properties of the Frobenius inner product that induces the Frobenius norm:
This quantity is an orthogonal matrix (as it is a product of orthogonal matrices) and thus the expression is maximised when equals the identity matrix . Thus
where is the solution for the optimal value of that minimizes the norm squared .
Generalized/constrained Procrustes problems
There are a number of related problems to the classical orthogonal Procrustes problem. One might generalize it by seeking the closest matrix in which the columns are orthogonal, but not necessarily orthonormal.
Alternately, one might constrain it by only allowing rotation matrices (i.e. orthogonal matrices with determinant 1, also known as special orthogonal matrices). In this case, o
|
https://en.wikipedia.org/wiki/Simple%20homotopy%20theory
|
In mathematics, simple homotopy theory is a homotopy theory (a branch of algebraic topology) that concerns with the simple-homotopy type of a space. It was originated by Whitehead in his 1950 paper "Simple homotopy types".
See also
Whitehead torsion
|
https://en.wikipedia.org/wiki/NCOA4
|
Nuclear receptor coactivator 4, also known as Androgen Receptor Activator (ARA70), is a protein that in humans is encoded by the NCOA4 gene. It plays an important role in ferritinophagy, acting as a cargo receptor, binding to the ferritin heavy chain and latching on to ATG8 on the surface of the autophagosome.
Interactions
NCOA4 has been shown to interact with:
Androgen receptor, and
Peroxisome proliferator-activated receptor gamma
Ferritin
ATG8
See also
Transcription coregulator
|
https://en.wikipedia.org/wiki/When%20Topology%20Meets%20Chemistry
|
When Topology Meets Chemistry: A Topological Look At Molecular Chirality is a book in chemical graph theory on the graph-theoretic analysis of chirality in molecular structures. It was written by Erica Flapan, based on a series of lectures she gave in 1996 at the Institut Henri Poincaré, and was published in 2000 by the Cambridge University Press and Mathematical Association of America as the first volume in their shared Outlooks book series.
Topics
A chiral molecule is a molecular structure that is different from its mirror image. This property, while seemingly abstract, can have big consequences in biochemistry, where the shape of molecules is essential to their chemical function, and where a chiral molecule can have very different biological activities from its mirror-image molecule. When Topology Meets Chemistry concerns the mathematical analysis of molecular chirality.
The book has seven chapters, beginning with an introductory overview and ending with a chapter on the chirality of DNA molecules.
Other topics covered through the book include the rigid geometric chirality of tree-like molecular structures such as tartaric acid, and the stronger topological chirality of molecules that cannot be deformed into their mirror image without breaking and re-forming some of their molecular bonds. It discusses results of Flapan and Jonathan Simon on molecules with the molecular structure of Möbius ladders, according to which every embedding of a Möbius ladder with an odd number of rungs is chiral while Möbius ladders with an even number of rungs have achiral embeddings. It uses the symmetries of graphs, in a result that the symmetries of certain graphs can always be extended to topological symmetries of three-dimensional space, from which it follows that non-planar graphs with no self-inverse symmetry are always chiral. It discusses graphs for which every embedding is topologically knotted or linked. And it includes material on the use of knot invariants to detect topol
|
https://en.wikipedia.org/wiki/Transition-minimized%20differential%20signaling
|
Transition-minimized differential signaling (TMDS) is a technology for transmitting high-speed serial data used by the DVI and HDMI video interfaces, as well as by other digital communication interfaces.
The transmitter incorporates an advanced coding algorithm which reduces electromagnetic interference over copper cables and enables robust clock recovery at the receiver to achieve high skew tolerance for driving longer cables as well as shorter low-cost cables.
Coding
The method is a form of 8b/10b encoding but using a code-set that differs from the original IBM form. A two-stage process converts an input of 8 bits into a 10 bit code with particular desirable properties. In the first stage, the first bit is untransformed and each subsequent bit is either XOR or XNOR transformed against the previous bit. The encoder chooses between XOR and XNOR by determining which will result in the fewest transitions; the ninth bit encodes which operation was used. In the second stage, the first eight bits are optionally inverted to even out the balance of ones and zeros and therefore the sustained average DC level; the tenth bit encodes whether this inversion took place.
The 10-bit TMDS symbol can represent either an 8-bit data value during normal data transmission, or 2 bits of control signals during screen blanking. Of the 1,024 possible combinations of the 10 transmitted bits:
460 combinations are used to represent an 8-bit data value, as most of the 256 possible values have two encoded variants (some values have only one),
4 combinations are used to represent 2 bits of control signals (C0 and C1 in the table below); unlike the data symbols these have such properties that they can be reliably recognized even if sync is lost and are therefore also used for synchronizing the decoder,
2 combinations are used as a guard band before HDMI data,
558 remaining combinations are reserved and forbidden.
Control data is encoded using the values in the table below. Control data c
|
https://en.wikipedia.org/wiki/Alpha%20particle
|
Alpha particles, also called alpha rays or alpha radiation, consist of two protons and two neutrons bound together into a particle identical to a helium-4 nucleus. They are generally produced in the process of alpha decay, but may also be produced in other ways. Alpha particles are named after the first letter in the Greek alphabet, α. The symbol for the alpha particle is α or α2+. Because they are identical to helium nuclei, they are also sometimes written as or indicating a helium ion with a +2 charge (missing its two electrons). Once the ion gains electrons from its environment, the alpha particle becomes a normal (electrically neutral) helium atom .
Alpha particles have a net spin of zero. Due to the mechanism of their production in standard alpha radioactive decay, alpha particles generally have a kinetic energy of about 5 MeV, and a velocity in the vicinity of 4% of the speed of light. (See discussion below for the limits of these figures in alpha decay.) They are a highly ionizing form of particle radiation, and (when resulting from radioactive alpha decay) usually have low penetration depth (stopped by a few centimetres of air, or by the skin).
However, so-called long range alpha particles from ternary fission are three times as energetic, and penetrate three times as far. The helium nuclei that form 10–12% of cosmic rays are also usually of much higher energy than those produced by nuclear decay processes, and thus may be highly penetrating and able to traverse the human body and also many metres of dense solid shielding, depending on their energy. To a lesser extent, this is also true of very high-energy helium nuclei produced by particle accelerators.
Name
Some science authors use doubly ionized helium nuclei () and alpha particles as interchangeable terms. The nomenclature is not well defined, and thus not all high-velocity helium nuclei are considered by all authors to be alpha particles. As with beta and gamma particles/rays, the name used for t
|
https://en.wikipedia.org/wiki/Monovalent%20cation%3Aproton%20antiporter-1
|
The Monovalent Cation:Proton Antiporter-1 (CPA1) Family (TC# 2.A.36) is a large family of proteins derived from Gram-positive and Gram-negative bacteria, blue-green bacteria, archaea, yeast, plants and animals. The CPA1 family belongs to the VIC superfamily. Transporters from eukaryotes have been functionally characterized to catalyze Na+:H+ exchange. Their primary physiological functions are thought to be in (1) cytoplasmic pH regulation, extruding the H+ generated during metabolism, and (2) salt tolerance (in plants), due to Na+ uptake into vacuoles. Bacterial homologues have also been found to facilitate Na+:H+ antiport, but some also catalyze Li+:H+ antiport or Ca2+:H+ antiport under certain conditions.
Phylogeny
The phylogenetic tree for the CPA1 family shows three principal clusters. The first cluster includes proteins derived exclusively from animals, and all of the functionally characterized members of the family belong to this cluster. Of the two remaining clusters, one includes all bacterial homologues while the other includes one from Arabidopsis thaliana, one from Homo sapiens and two from yeast (S. cerevisiae and S. pombe). Several organisms possess multiple paralogues; for example, seven paralogues are found in C. elegans, and five are known in humans. Most of these paralogues are very similar in sequence, and they belong to the animal specific cluster.
A representative list of proteins belonging to the CPA1 family can be found in the Transporter Classification Database.
Structure
Numerous members of the CPA1 family have been sequenced, and these proteins vary substantially in size. The bacterial proteins have 520-550 amino acyl residues (aas) while eukaryotic proteins are generally larger, varying in size from 540-900 residues. They exhibit 10-12 putative transmembrane α-helical spanners (TMSs). A proposed topological model suggests that in addition to 12 TMSs, a region between TMSs 9 and 10 dips into the membrane to line the pore. However, one h
|
https://en.wikipedia.org/wiki/Hydrophobic%20effect
|
The hydrophobic effect is the observed tendency of nonpolar substances to aggregate in an aqueous solution and exclude water molecules. The word hydrophobic literally means "water-fearing", and it describes the segregation of water and nonpolar substances, which maximizes hydrogen bonding between molecules of water and minimizes the area of contact between water and nonpolar molecules. In terms of thermodynamics, the hydrophobic effect is the free energy change of water surrounding a solute. A positive free energy change of the surrounding solvent indicates hydrophobicity, whereas a negative free energy change implies hydrophilicity.
The hydrophobic effect is responsible for the separation of a mixture of oil and water into its two components. It is also responsible for effects related to biology, including: cell membrane and vesicle formation, protein folding, insertion of membrane proteins into the nonpolar lipid environment and protein-small molecule associations. Hence the hydrophobic effect is essential to life. Substances for which this effect is observed are known as hydrophobes.
Amphiphiles
Amphiphiles are molecules that have both hydrophobic and hydrophilic domains. Detergents are composed of amphiphiles that allow hydrophobic molecules to be solubilized in water by forming micelles and bilayers (as in soap bubbles). They are also important to cell membranes composed of amphiphilic phospholipids that prevent the internal aqueous environment of a cell from mixing with external water.
Folding of macromolecules
In the case of protein folding, the hydrophobic effect is important to understanding the structure of proteins that have hydrophobic amino acids (such as glycine, alanine, valine, leucine, isoleucine, phenylalanine, tryptophan and methionine) clustered together within the protein. Structures of water-soluble proteins have a hydrophobic core in which side chains are buried from water, which stabilizes the folded state. Charged and polar side ch
|
https://en.wikipedia.org/wiki/Robert%20Connelly
|
Robert Connelly (born July 15, 1942) is a mathematician specializing in discrete geometry and rigidity theory. Connelly received his Ph.D. from University of Michigan in 1969. He is currently a professor at Cornell University.
Connelly is best known for discovering embedded flexible polyhedra. One such polyhedron is in the National Museum of American History. His recent interests include tensegrities and the carpenter's rule problem. In 2012 he became a fellow of the American Mathematical Society.
Asteroid 4816 Connelly, discovered by Edward Bowell at Lowell Observatory 1981, was named after Robert Connelly. The official was published by the Minor Planet Center on 18 February 1992 ().
Author
Connelly has authored or co-authored several articles on mathematics, including Conjectures and open questions in rigidity; A flexible sphere; and A counterexample to the rigidity conjecture for polyhedra.
|
https://en.wikipedia.org/wiki/Periodic%20instantons
|
Periodic instantons are finite energy solutions of Euclidean-time field equations which communicate (in the sense of quantum tunneling) between two turning points in the barrier of a potential and are therefore also known as bounces. Vacuum instantons, normally simply called instantons, are the corresponding zero energy configurations in the limit of infinite Euclidean time. For completeness we add that ``sphalerons´´ are the field configurations at the very top of a potential barrier. Vacuum instantons carry a winding (or topological) number, the other configurations do not. Periodic instantons werde discovered with the explicit solution of Euclidean-time field equations for double-well potentials and the cosine potential with non-vanishing energy and are explicitly expressible in terms of Jacobian elliptic functions (the generalization of trigonometrical functions). Periodic instantons describe the oscillations between two endpoints of a potential barrier between two potential wells. The frequency of these oscillations or the tunneling between the two wells is related to the bifurcation or level splitting of the energies of states or wave functions related to the wells on either side of the barrier, i.e. . One can also interpret this energy change as the energy contribution to the well energy on either side originating from the integral describing the overlap of the wave functions on either side in the domain of the barrier.
Evaluation of by the path integral method requires summation over an infinite number of widely separated pairs of periodic instantons -- this calculation is therefore said to be that in the ``dilute gas approximation´´.
Periodic instantons have meanwhile been found to occur in numerous theories and at various levels of complication. In particular they arise in investigations of the following topics.
(1) Quantum mechanics and path integral treatment of periodic and anharmonic potentials.
(2) Macroscopic spin systems (like ferromagneti
|
https://en.wikipedia.org/wiki/LIBOX
|
LIBOX was a free platform that allowed users to access and share their high definition media collections, including video, photos and music, across various devices and with friends. LIBOX offered this service for free thanks to a patent pending combination of peer-to-peer, grid and distributed computing technologies. LIBOX consisted of a downloadable desktop application that works on both Windows PCs and Macs, and a web-based interface. The service was accessed by any Web browser and placed no limitations on the amount of media that can be added or the number of people with which it can be shared.
History
LIBOX was founded in 2008 by Erez Pilosof, who previously founded Walla!, the first major web portal in Israel. Pilosof created LIBOX to allow users to manage and share media across all devices and keep its original high quality. He saw that as a consumer, trying to store your media on several different devices and in many different partial areas online was becoming an annoyance; it "seemed very limited and tedious and problematic” Pilosof created LIBOX as a way to provide a smooth and dependable way for people to enjoy his/her media anywhere.
The company started working on the patent-pending technology to power LIBOX in the Fall of 2008, released an Alpha version in October 2009 and launched a Beta version on June 22, 2010. The company has received funding from investors such as Evergreen Venture Partners and Rhodium to help grow the platform.
LIBOX closed down in 2011.
Technology
The distributed LIBOX platform effectively creates private clouds that communicate between devices and Web browsers through a combination of algorithms, grids and peer-to-peer networking technologies. Files are not uploaded to an external server but streamed straight from the computer of the user that holds the file. The mixture of technologies allows LIBOX to never limit how much media can be added to the platform, while keeping the service free for users. The LIBOX platform uses
|
https://en.wikipedia.org/wiki/Autonomously%20replicating%20sequence
|
An autonomously replicating sequence (ARS) contains the origin of replication in the yeast genome. It contains four regions (A, B1, B2, and B3), named in order of their effect on plasmid stability. The A-Domain is highly conserved, any mutation abolishes origin function. Mutations on B1, B2, and B3 will diminish, but not prevent functioning of the origin.
Element A is highly conserved, consisting of the consensus sequence:
(where Y is either pyrimidine and R is either purine). When this element is mutated, the ARS loses all activity.
As seen above the ARS are considerably A-T rich which makes it easy for replicative proteins to disrupt the H-bonding in that area. ORC protein complex (origin recognition complex) is bound at the ARS throughout the cell cycle, allowing replicative proteins access to the ARS.
Mutational analysis for the yeast ARS elements have shown that any mutation in the B1, B2 and B3 regions result in a reduction of function of the ARS element. A mutation in the A region results in a complete loss of function.
Melting of DNA occurs within domain B2, induced by attachment of ARS binding factor 1 to B3.
A1 and B1 domain binds with origin recognition complex.
To identify these sequences, yeast mutants unable to synthesize histidine were transformed with plasmids containing the His gene and random fragments of the yeast genome. If the genome fragment contained an origin of replication, cells were able to grow in a medium lacking histidine. These sequences were termed autonomously replicating sequences, because they were replicated and inherited by progeny without integrating into the host chromosome.
|
https://en.wikipedia.org/wiki/Trinomial%20tree
|
The trinomial tree is a lattice-based computational model used in financial mathematics to price options. It was developed by Phelim Boyle in 1986. It is an extension of the binomial options pricing model, and is conceptually similar. It can also be shown that the approach is equivalent to the explicit finite difference method for option pricing. For fixed income and interest rate derivatives see Lattice model (finance)#Interest rate derivatives.
Formula
Under the trinomial method, the underlying stock price is modeled as a recombining tree, where, at each node the price has three possible paths: an up, down and stable or middle path. These values are found by multiplying the value at the current node by the appropriate factor , or where
(the structure is recombining)
and the corresponding probabilities are:
.
In the above formulae: is the length of time per step in the tree and is simply time to maturity divided by the number of time steps; is the risk-free interest rate over this maturity; is the corresponding volatility of the underlying; is its corresponding dividend yield.
As with the binomial model, these factors and probabilities are specified so as to ensure that the price of the underlying evolves as a martingale, while the moments considering node spacing and probabilities are matched to those of the log-normal distribution (and with increasing accuracy for smaller time-steps). Note that for , , and to be in the interval the following condition on has to be satisfied .
Once the tree of prices has been calculated, the option price is found at each node largely as for the binomial model, by working backwards from the final nodes to the present node (). The difference being that the option value at each non-final node is determined based on the threeas opposed to two later nodes and their corresponding probabilities.
If the length of time-steps is taken as an exponentially distributed random variable and interpreted as the waiting time
|
https://en.wikipedia.org/wiki/Rustock%20botnet
|
The Rustock botnet was a botnet that operated from around 2006 until March 2011.
It consisted of computers running Microsoft Windows, and was capable of sending up to 25,000 spam messages per hour from an infected PC. At the height of its activities, it sent an average of 192 spam messages per compromised machine per minute. Reported estimates on its size vary greatly across different sources, with claims that the botnet may have comprised anywhere between 150,000 and 2,400,000 machines. The size of the botnet was increased and maintained mostly through self-propagation, where the botnet sent many malicious e-mails intended to infect machines opening them with a trojan which would incorporate the machine into the botnet.
The botnet took a hit after the 2008 takedown of McColo, an ISP which was responsible for hosting most of the botnet's command and control servers. McColo regained Internet connectivity for several hours, and in those hours up to 15 Mbit a second of traffic was observed, likely indicating a transfer of command and control to Russia. While these actions temporarily reduced global spam levels by around 75%, the effect did not last long: spam levels increased by 60% between January and June 2009, 40% of which was attributed to the Rustock botnet.
On March 16, 2011, the botnet was taken down through what was initially reported as a coordinated effort by Internet service providers and software vendors. It was revealed the next day that the take-down, called Operation b107, was the action of Microsoft, U.S. federal law enforcement agents, FireEye, and the University of Washington.
To capture the individuals involved with the Rustock botnet, on July 18, 2011, Microsoft is offering "a monetary reward in the amount of US$250,000 for new information that results in the identification, arrest and criminal conviction of such individual(s)."
Operations
Botnets are composed of infected computers used by unwitting Internet users. In order to hide its presen
|
https://en.wikipedia.org/wiki/Surface%20modification
|
Surface modification is the act of modifying the surface of a material by bringing physical, chemical or biological characteristics different from the ones originally found on the surface of a material.
This modification is usually made to solid materials, but it is possible to find examples of the modification to the surface of specific liquids.
The modification can be done by different methods with a view to altering a wide range of characteristics of the surface, such as: roughness, hydrophilicity, surface charge, surface energy, biocompatibility and reactivity.
Surface engineering
Surface engineering is the sub-discipline of materials science which deals with the surface of solid matter. It has applications to chemistry, mechanical engineering, and electrical engineering (particularly in relation to semiconductor manufacturing).
Solids are composed of a bulk material covered by a surface. The surface which bounds the bulk material is called the Surface phase. It acts as an interface to the surrounding environment. The bulk material in a solid is called the Bulk phase.
The surface phase of a solid interacts with the surrounding environment. This interaction can degrade the surface phase over time. Environmental degradation of the surface phase over time can be caused by wear, corrosion, fatigue and creep.
Surface engineering involves altering the properties of the Surface Phase in order to reduce the degradation over time. This is accomplished by making the surface robust to the environment in which it will be used.
Applications and Future of Surface Engineering
Surface engineering techniques are being used in the automotive, aerospace, missile, power, electronic, biomedical, textile, petroleum, petrochemical, chemical, steel, power, cement, machine tools, construction industries. Surface engineering techniques can be used to develop a wide range of functional properties, including physical, chemical, electrical, electronic, magnetic, mechanical, wear-r
|
https://en.wikipedia.org/wiki/Cartan%20matrix
|
In mathematics, the term Cartan matrix has three meanings. All of these are named after the French mathematician Élie Cartan. Amusingly, the Cartan matrices in the context of Lie algebras were first investigated by Wilhelm Killing, whereas the Killing form is due to Cartan.
Lie algebras
A (symmetrizable) generalized Cartan matrix is a square matrix with integer entries such that
For diagonal entries, .
For non-diagonal entries, .
if and only if
can be written as , where is a diagonal matrix, and is a symmetric matrix.
For example, the Cartan matrix for G2 can be decomposed as such:
The third condition is not independent but is really a consequence of the first and fourth conditions.
We can always choose a D with positive diagonal entries. In that case, if S in the above decomposition is positive definite, then A is said to be a Cartan matrix.
The Cartan matrix of a simple Lie algebra is the matrix whose elements are the scalar products
(sometimes called the Cartan integers) where ri are the simple roots of the algebra. The entries are integral from one of the properties of roots. The first condition follows from the definition, the second from the fact that for is a root which is a linear combination of the simple roots ri and rj with a positive coefficient for rj and so, the coefficient for ri has to be nonnegative. The third is true because orthogonality is a symmetric relation. And lastly, let and . Because the simple roots span a Euclidean space, S is positive definite.
Conversely, given a generalized Cartan matrix, one can recover its corresponding Lie algebra. (See Kac–Moody algebra for more details).
Classification
An matrix A is decomposable if there exists a nonempty proper subset such that whenever and . A is indecomposable if it is not decomposable.
Let A be an indecomposable generalized Cartan matrix. We say that A is of finite type if all of its principal minors are positive, that A is of affine type if its proper princi
|
https://en.wikipedia.org/wiki/Gingerbreadman%20map
|
In dynamical systems theory, the Gingerbreadman map is a chaotic two-dimensional map. It is given by the piecewise linear transformation:
See also
List of chaotic maps
|
https://en.wikipedia.org/wiki/Telmac%201800
|
The Telmac 1800 was an early microcomputer delivered in kit form. It was introduced in 1977 by Telercas Oy, the Finnish importer of RCA microchips. Most of the 2,000 kits manufactured over four years were bought by electronics enthusiasts in Finland, Sweden and Norway.
An expansion board, OSCOM, later became available, and included an alphanumeric video display, and up to of memory. A Tiny BASIC could be run on this configuration.
The first-ever commercial video game to be developed in Finland, Chesmac (fi), was developed by Raimo Suonio on a Telmac 1800 computer in 1979.
The Telmac 1800 was followed by the Oscom Nano and the Telmac 2000.
Major features
RCA 1802 (COSMAC) microprocessor CPU @ 1.75 MHz
Cassette tape interface
2 kB RAM, expandable to 4 kB
RCA CDP1861 'Pixie' video chip, 64×128 pixels display resolution
Sound limited to a fixed frequency tone
Able to run a CHIP-8 interpreter
|
https://en.wikipedia.org/wiki/Home%20screen
|
A home screen, homescreen, or start screen, is the main screen on a device or computer program. Home screens are not identical because users rearrange icons as they please, and home screens often differ across mobile operating systems. Almost every smartphone has some form of home screen, which typically displays links to applications, settings, and notifications.
Common features
Home screens usually consist of a grid of application links or shortcuts that can often be arranged over multiple pages, and serve as the user's main method of accessing phone functions. Home screens also tend to include a dock along an edge of the screen, where application links can be stored and accessed from any page on the home screen. Most operating systems allow users to add folders to the home screen in order to further organize application links. Some home screens may also include a panel where push notifications are displayed or select system settings can be accessed.
In addition to applications links, many home screens are also capable of displaying ambient information, such as live tiles on Windows Phone or widgets on Android. Such tiles or widgets may link to applications, however they differ from traditional links in that they show current, dynamic information instead of a static icon. However, the increased relevance of the information can come at the cost of device battery life, bandwidth, and the ease of recognition afforded by static application icons.
Alternative home screens
Although most home screens have a similar structure, not all are designed in common. Two notable examples of less-common home screens paradigms include Siri and WebOS. The former is Apple's natural language user interface, which performs functions similar to more traditional home screens such as opening applications, displaying relevant data, and managing phone settings. The latter is notable for its use of entirely dynamic application icons that mimic the current state of the application, similar
|
https://en.wikipedia.org/wiki/Human%20male%20sexuality
|
Human male sexuality encompasses a wide variety of feelings and behaviors. Men's feelings of attraction may be caused by various physical and social traits of their potential partner. Men's sexual behavior can be affected by many factors, including evolved predispositions, individual personality, upbringing, and culture. While most men are heterosexual, significant minorities are homosexual or varying degrees of bisexual.
Sexual attraction
Physical factors
Research indicates that men tend to be attracted to young women with bodily symmetry. Facial symmetry, femininity, and averageness are also linked with attractiveness. Men typically find female breasts attractive and this holds true for a variety of cultures. A preference for lighter-skinned women has been documented across many cultures.
Women with a relatively low waist-to-hip ratio (WHR) are considered more attractive. The exact ratio varies among cultures, depending on the WHR of the women in the local culture. In Western cultures, a WHR of 0.70 is preferred. Other possible physical factors of attraction include low body mass index, low waist circumference, longer legs, and greater lower back curvature. Preference for a slim or a plump body build is culturally variable, but in a predictable manner. In cultures where food is scarce, plumpness is associated with higher status and is more attractive, but the reverse is true in wealthy cultures.
Men generally prefer their wives to be younger than they are, but by how much exactly varies between cultures. Older men prefer greater age differences, while teenage males prefer females slightly older than they are.
The exact degree to which physical appearance is considered important in selecting a long-term mate varies between cultures.
Non-physical factors
When choosing long-term partners, both men and women desire those who are intelligent, kind, understanding, and healthy. They also show a preference for partners who have similar values, attitudes, personal
|
https://en.wikipedia.org/wiki/Peaucellier%E2%80%93Lipkin%20linkage
|
The Peaucellier–Lipkin linkage (or Peaucellier–Lipkin cell, or Peaucellier–Lipkin inversor), invented in 1864, was the first true planar straight line mechanism – the first planar linkage capable of transforming rotary motion into perfect straight-line motion, and vice versa. It is named after Charles-Nicolas Peaucellier (1832–1913), a French army officer, and Yom Tov Lipman Lipkin (1846–1876), a Lithuanian Jew and son of the famed Rabbi Israel Salanter.
Until this invention, no planar method existed of converting exact straight-line motion to circular motion, without reference guideways. In 1864, all power came from steam engines, which had a piston moving in a straight-line up and down a cylinder. This piston needed to keep a good seal with the cylinder in order to retain the driving medium, and not lose energy efficiency due to leaks. The piston does this by remaining perpendicular to the axis of the cylinder, retaining its straight-line motion. Converting the straight-line motion of the piston into circular motion was of critical importance. Most, if not all, applications of these steam engines, were rotary.
The mathematics of the Peaucellier–Lipkin linkage is directly related to the inversion of a circle.
Earlier Sarrus linkage
There is an earlier straight-line mechanism, whose history is not well known, called the Sarrus linkage. This linkage predates the Peaucellier–Lipkin linkage by 11 years and consists of a series of hinged rectangular plates, two of which remain parallel but can be moved normally to each other. Sarrus' linkage is of a three-dimensional class sometimes known as a space crank, unlike the Peaucellier–Lipkin linkage which is a planar mechanism.
Geometry
In the geometric diagram of the apparatus, six bars of fixed length can be seen: , , , , , . The length of is equal to the length of , and the lengths of , , , and are all equal forming a rhombus. Also, point is fixed. Then, if point is constrained to move along a circle (for exampl
|
https://en.wikipedia.org/wiki/Reverse%20proxy
|
In computer networks, a reverse proxy is an application that sits in front of back-end applications and forwards client (e.g. browser) requests to those applications. Reverse proxies help increase scalability, performance, resilience and security. The resources returned to the client appear as if they originated from the web server itself.
Large websites and content delivery networks use reverse proxies, together with other techniques, to balance the load between internal servers. Reverse proxies can keep a cache of static content, which further reduces the load on these internal servers and the internal network. It is also common for reverse proxies to add features such as compression or TLS encryption to the communication channel between the client and the reverse proxy.
Reverse proxies are typically owned or managed by the web service, and they are accessed by clients from the public Internet. In contrast, a forward proxy is typically managed by a client (or their company) who is restricted to a private, internal network, except that the client can ask the forward proxy to retrieve resources from the public Internet on behalf of the client.
Reverse proxy servers are implemented in popular open-source web servers such as Apache, Nginx, and Caddy. This software can inspect HTTP headers, which, for example, allows it to present a single IP address to the Internet while relaying requests to different internal servers based on the URL of the HTTP request. Dedicated reverse proxy servers such as the open source software HAProxy and Squid are used by some of the biggest websites on the Internet.
Uses
Reverse proxies can hide the existence and characteristics of origin servers. This can make it more difficult to determine the actual location of the origin server / website and, for instance, more challenging to initiate legal action such as takedowns or block access to the website, as the IP address of the website may not be immediately apparent. Additionally, th
|
https://en.wikipedia.org/wiki/50th%20meridian%20west
|
The meridian 50° west of Greenwich is a line of longitude that extends from the North Pole across the Arctic Ocean, Greenland, the Atlantic Ocean, South America, the Southern Ocean, and Antarctica to the South Pole.
The 50th meridian west forms a great circle with the 130th meridian east.
From Pole to Pole
Starting at the North Pole and heading south to the South Pole, the 50th meridian west passes through:
{| class="wikitable plainrowheaders"
! scope="col" width="120" | Co-ordinates
! scope="col" | Country, territory or sea
! scope="col" | Notes
|-
| style="background:#b0e0e6;" |
! scope="row" style="background:#b0e0e6;" | Arctic Ocean
| style="background:#b0e0e6;" |
|-
| style="background:#b0e0e6;" |
! scope="row" style="background:#b0e0e6;" | Lincoln Sea
| style="background:#b0e0e6;" |
|-
|
! scope="row" |
| Beaumont Island
|-
| style="background:#b0e0e6;" |
! scope="row" style="background:#b0e0e6;" | Lincoln Sea
| style="background:#b0e0e6;" |
|-
|
! scope="row" |
|Wulff Land
|-
| style="background:#b0e0e6;" |
! scope="row" style="background:#b0e0e6;" | Sherard Osborn Fjord
| style="background:#b0e0e6;" |
|-
|
! scope="row" |
|
|-
| style="background:#b0e0e6;" |
! scope="row" style="background:#b0e0e6;" | Atlantic Ocean
| style="background:#b0e0e6;" |
|-
|
! scope="row" |
| Amapá — mainland and the island of Bailique
|-
| style="background:#b0e0e6;" |
! scope="row" style="background:#b0e0e6;" | Atlantic Ocean
| style="background:#b0e0e6;" | Mouth of the Amazon River
|-valign="top"
|
! scope="row" |
| Pará — islands of Caviana and Marajó, and the mainland Tocantins — from Goiás — from Minas Gerais — from São Paulo — from Paraná — from Santa Catarina — from Rio Grande do Sul — from Santa Catarina — for about 11 km from Rio Grande do Sul — from
|-
| style="background:#b0e0e6;" |
! scope="row" style="background:#b0e0e6;" | Atlantic Ocean
| style="background:#b0e0e6;" |
|-
| style="background:#b0e0e6;" |
! scope="row" style="background:#
|
https://en.wikipedia.org/wiki/Broadcast%20reference%20monitor
|
A video reference monitor, also called a broadcast reference monitor or just reference monitor, is a specialized display device similar to a television set, used to monitor the output of a video-generating device, such as playout from a video server, IRD, video camera, VCR, or DVD player. It may or may not have professional audio monitoring capability. Unlike a television set, a video monitor has no tuner and, as such, is unable independently to tune into an over-the-air broadcast like a television receiver. One common use of video monitors is in television stations, television studios, production trucks and in outside broadcast vehicles, where broadcast engineers use them for confidence checking of analog signal and digital signals throughout the system. They can also be used for color grading if calibrated, during post-production.
Common display types for video monitors
Cathode ray tube
Liquid crystal display
Plasma display
OLED
Common monitoring formats for security
Composite video
S-Video
Broadcast reference monitor
Broadcast reference monitors must be used for video compliance at television or television studio facilities, because they do not perform any video enhancements and try to produce as accurate an image as possible. For quality control purposes, it is necessary for a broadcast reference monitor to produce (reasonably) consistent images from facility to facility, to reveal any flaws in the material, and also not to introduce any image artifacts (such as aliasing) that is not in the source material. Broadcast monitors will try to avoid post processing such as a video scaler, line doubling and any image enhancements such as dynamic contrast. However, display technologies with fixed pixel structures (e.g. LCD, plasma) must perform image scaling when displaying SD signals as the signal contains non-square pixels while the display has square pixels. LCDs and plasmas are also inherently progressive displays and may need to perform deinterlacing
|
https://en.wikipedia.org/wiki/Targeted%20alpha-particle%20therapy
|
Targeted alpha-particle therapy (or TAT) is an in-development method of targeted radionuclide therapy of various cancers. It employs radioactive substances which undergo alpha decay to treat diseased tissue at close proximity. It has the potential to provide highly targeted treatment, especially to microscopic tumour cells. Targets include leukemias, lymphomas, gliomas, melanoma, and peritoneal carcinomatosis. As in diagnostic nuclear medicine, appropriate radionuclides can be chemically bound to a targeting biomolecule which carries the combined radiopharmaceutical to a specific treatment point.
It has been said that "α-emitters are indispensable with regard to optimisation of strategies for tumour therapy".
Advantages of alpha emitters
The primary advantage of alpha particle (α) emitters over other types of radioactive sources is their very high linear energy transfer (LET) and relative biological effectiveness (RBE). Beta particle (β) emitters such as yttrium-90 can travel considerable distances beyond the immediate tissue before depositing their energy, while alpha particles deposit their energy in 70–100 μm long tracks.
Alpha particles are more likely than other types of radiation to cause double-strand breaks to DNA molecules, which is one of several effective causes of cell death.
Production
Some α emitting isotopes such as 225Ac and 213Bi are only available in limited quantities from 229Th decay, although cyclotron production is feasible. Among alpha-emitting radiometals according to availability, chelation chemistry, and half-life, 212Pb is also a promising candidate for targeted alpha-therapy.
The ARRONAX cyclotron can produce 211At by irradiation of 209Bi.
Applications
Though many α-emitters exist, useful isotopes would have a sufficient energy to cause damage to cancer cells, and a half-life that is long enough to provide a therapeutic dose without remaining long enough to damage healthy tissue.
Immunotherapy
Several radionuclides have been studi
|
https://en.wikipedia.org/wiki/Biochemistry%20of%20Alzheimer%27s%20disease
|
The biochemistry of Alzheimer's disease, the most common cause of dementia, is not yet very well understood. Alzheimer's disease (AD) has been identified as a proteopathy: a protein misfolding disease due to the accumulation of abnormally folded amyloid beta (Aβ) protein in the brain. Amyloid beta is a short peptide that is an abnormal proteolytic byproduct of the transmembrane protein amyloid-beta precursor protein (APP), whose function is unclear but thought to be involved in neuronal development. The presenilins are components of proteolytic complex involved in APP processing and degradation.
Amyloid beta monomers are soluble and contain short regions of beta sheet and polyproline II helix secondary structures in solution, though they are largely alpha helical in membranes; however, at sufficiently high concentration, they undergo a dramatic conformational change to form a beta sheet-rich tertiary structure that aggregates to form amyloid fibrils. These fibrils and oligomeric forms of Aβ deposit outside neurons in formations known as senile plaques. There are different types of plaques, including the diffuse, compact, cored or neuritic plaque types, as well as Aβ deposits in the walls of small blood vessel walls in the brain called cerebral amyloid angiopathy.
AD is also considered a tauopathy due to abnormal aggregation of the tau protein, a microtubule-associated protein expressed in neurons that normally acts to stabilize microtubules in the cell cytoskeleton. Like most microtubule-associated proteins, tau is normally regulated by phosphorylation; however, in Alzheimer's disease, hyperphosphorylated tau accumulates as paired helical filaments that in turn aggregate into masses inside nerve cell bodies known as neurofibrillary tangles and as dystrophic neurites associated with amyloid plaques. Although little is known about the process of filament assembly, depletion of a prolyl isomerase protein in the parvulin family has been shown to accelerate the accumu
|
https://en.wikipedia.org/wiki/Lasing%20without%20inversion
|
Lasing without inversion (LWI), or lasing without population inversion, is a technique used for light amplification by stimulated emission without the requirement of population inversion. A laser working under this scheme exploits the quantum interference between the probability amplitudes of atomic transitions in order to eliminate absorption without disturbing the stimulated emission. This phenomenon is also the essence of electromagnetically induced transparency.
The basic LWI concept was first predicted by Ali Javan in 1956. The first demonstration of LWI was carried out by Marlan Scully in an experiment in rubidium and sodium at Texas A&M University, and then at NIST in Boulder.
|
https://en.wikipedia.org/wiki/Jigglypuff
|
Jigglypuff ( ), known in Japan as , is a Pokémon species. Jigglypuff first appeared in the video games Pokémon Red and Blue and subsequent sequels, later appearing in various merchandise, spinoff titles and animated and printed adaptations of the franchise. Jigglypuff is voiced by Rachael Lillis in English and by Mika Kanai in Japanese. In the live-action musical Pokémon Live!, Jigglypuff is portrayed by Leah Smith. Jigglypuff is also very well known for singing a lullaby in the Pokémon anime series.
Known as the Balloon Pokémon, Jigglypuff evolves from Igglybuff when it reaches a certain point of happiness, and evolves into Wigglytuff when exposed to a Moon Stone. Its English name is a combination of the words "jiggly" and "puff", intended to relate to its jelly-like appearance. The character has been featured in a recurring role in the anime series and served as the focus for several printed adaptions of the franchise. Since it appeared in the Pokémon series, Jigglypuff has received generally positive reception. It has been featured in several forms of merchandise, including figurines, plush toys, and the Pokémon Trading Card Game. It also has appeared as a playable character in every entry of the Super Smash Bros. series.
Design and characteristics
Jigglypuff was one of 151 different Generation I Pokémon designs conceived by Game Freak's character development team and finalized by Ken Sugimori, for the first-generation of Pocket Monsters games Red and Green, which were localized outside Japan as Pokémon Red and Blue. Its Japanese name "Purin", derives from the Japanese loanword for custard or pudding. Nintendo decided to give the various Pokémon species "clever and descriptive names" related to their appearance or features when translating the game for western audiences as a means to make the characters more relatable to American children. Deciding to use a name better suited for its jelly-like appearance, the species was renamed "Jigglypuff", a combination of
|
https://en.wikipedia.org/wiki/Re-Pair
|
Re-Pair (short for recursive pairing) is a grammar-based compression algorithm that, given an input text, builds a straight-line program, i.e. a context-free grammar generating a single string: the input text. In order to perform the compression in linear time, it consumes the amount of memory that is approximately five times the size of its input.
The grammar is built by recursively replacing the most frequent pair of characters occurring in the text.
Once there is no pair of characters occurring twice, the resulting string is used as the axiom of the grammar.
Therefore, the output grammar is such that all rules but the axiom have two symbols on the right-hand side.
How it works
Re-Pair was first introduced by NJ. Larsson and A. Moffat in 1999.
In their paper the algorithm is presented together with a detailed description of the data structures required to implement it with linear time and space complexity.
The experiments showed that Re-Pair achieves high compression ratios and offers good performance for decompression.
However, the major drawback of the algorithm is its memory consumption, which is approximately 5 times the size of the input. Such memory usage is required in order to perform the compression in linear time but makes the algorithm impractical for compressing large files.
The image on the right shows how the algorithm works compresses the string .
During the first iteration, the pair , which occurs three times in , is replaced by a new symbol .
On the second iteration, the most frequent pair in the string , which is , is replaced by a new symbol .
Thus, at the end of the second iteration, the remaining string is .
In the next two iterations, the pairs and are replaced by symbols and respectively.
Finally, the string contains no repeated pair and therefore it is used as the axiom of the output grammar.
Data structures
In order to achieve linear time complexity, Re-Pair requires the following data structures
A sequence representing the
|
https://en.wikipedia.org/wiki/Outside%20broadcasting
|
Outside broadcasting (OB) is the electronic field production (EFP) of television or radio programmes (typically to cover television news and sports television events) from a mobile remote broadcast television studio. Professional video camera and microphone signals come into the production truck for processing, recording and possibly transmission.
Some outside broadcasts use a mobile production control room (PCR) inside a production truck.
History
Outside radio broadcasts have been taking place since the early 1920s and television ones since the late 1920s. The first large-scale outside broadcast was the televising of the Coronation of George VI and Elizabeth in May 1937, done by the BBC's first Outside Broadcast truck, MCR 1 (short for Mobile Control Room).
After the Second World War, the first notable outside broadcast was of the 1948 Summer Olympics. The Coronation of Elizabeth II followed in 1953, with 21 cameras being used to cover the event.
In December 1963 instant replays were used for the first time. Director Tony Verna used the technique on the Army-Navy game which aired on CBS Sports on December 7, 1963.
The 1968 Summer Olympics was the first with competitions televised in colour. The 1972 Olympic Games were the first where all competitions were captured by outside broadcast cameras.
During the 1970s, ITV franchise holder Southern Television was unique in having an outside broadcast boat, named Southener.
The wedding of Prince Charles and Lady Diana Spencer in July 1981 was the biggest outside broadcast at the time, with an estimated 750 million viewers.
New technology
In 2008, the first 3D outside broadcast took place with the transmission of a Calcutta Cup rugby match, but only to an audience of industry professionals who had been invited by BBC Sport.
In March 2010, the first public 3D outside broadcast took place with an NHL game between the New York Rangers and New York Islanders.
The first commercial ultra-high definition outside broad
|
https://en.wikipedia.org/wiki/NhaC%20family
|
The NhaC family (TC# 2.A.35) belongs to the Ion Transporter (IT) Superfamily. A representative list of proteins belonging to the NhaC family can be found in the Transporter Classification Database.
Two members of the NhaC family have been functionally characterized. One is believed to be a Na+:H+ antiporter; the other is a malate·H+:lactate·Na+ antiporter. Several paralogues are found in Vibrio cholerae, and two paralogues are found encoded in the completely sequenced genomes of bothHaemophilus influenzae and Bacillus subtilis. E. coli lacks such a homologue. Pyrococcus species also have at least one homologue each. Thus, members of the NhaC family are found in both Gram-negative bacteria and Gram-positive bacteria as well as archaea. NhaC of B. firmus is 462 amino acyl residues long and possesses 12 putative transmembrane α-helical segments. MleN of B. subtilis (468 aas; TC# 2.A.35.1.2) also exhibits 12 putative TMSs.
The transport reaction catalyzed by NhaC is probably:Na+ (in) + nH+ (out) ⇌ Na+ (out) + nH+ (in). (n > 1)That catalyzed by MleN is probably:Malate (out) + H+ (out) + Lactate (in) + Na+ (in) ⇌ Malate (in) + H+ (in) + Lactate (out) + Na+ (out)
See also
Sodium-Proton antiporter
Antiporter
Transporter Classification Database
Further reading
Ivey, D. M.; Guffanti, A. A.; Bossewitch, J. S.; Padan, E.; Krulwich, T. A. (1991-12-05). "Molecular cloning and sequencing of a gene from alkaliphilic Bacillus firmus OF4 that functionally complements an Escherichia coli strain carrying a deletion in the nhaA Na+/H+ antiporter gene". The Journal of Biological Chemistry 266 (34): 23483–23489. ISSN 0021-9258. PMID 1660475.
Liew, Chong Wai; Illias, Rosli Md; Mahadi, Nor Muhammad; Najimudin, Nazalan (2007-11-01)."Expression of the Na+/H+ antiporter gene (g1-nhaC) of alkaliphilic Bacillus sp. G1 in Escherichia coli". FEMS Microbiology Letters 276 (1): 114–122. doi:10.1111/j.1574-6968.2007.00925.x. ISSN 0378-1097. PMID 17937670.
Panina, Ekaterina M.; Vitreschak,
|
https://en.wikipedia.org/wiki/Critica%20Botanica
|
Critica Botanica ("Critique of botany", Leiden, July 1737) was written by Swedish botanist, physician, zoologist and naturalist Carl Linnaeus (1707–1778). The book was published in Germany when Linnaeus was 29 with a discursus by the botanist Johannes Browallius (1707–1755), bishop of Åbo. The first edition was published in July 1737 under the full title Critica botanica in qua nomina plantarum generica, specifica & variantia examini subjiciuntur, selectoria confirmantur, indigna rejiciuntur; simulque doctrina circa denominationem plantarum traditur. Seu Fundamentorum botanicorum pars IV Accedit Johannis Browallii De necessitate historiae naturalis discursus.
Linnaeus's principles of botanical nomenclature were first expounded in Fundamenta Botanica ("Foundations of botany") of 1736 chapters VII to X which contained the aphorisms (principles) 210 to 324 that outlined the rules for the acceptance and formation of names. These were later elaborated, with numerous examples, in his Critica Botanica of 1737. The practical application of these rules was soon seen in subsequent publications such as Flora Lapponica ("Flora of Lapland", 1737), Hortus Cliffortianus ("In honour of Clifford's garden", 1738), and Flora Svecica ("Flora of Sweden", 1746). Together the Fundamenta and Critica summarised Linnaeus's thoughts on plant nomenclature and classification which he later revised and elaborated in his Philosophia Botanica of 1751.
In the Critica Linnaeus presented a series of rules which guided him in his own publications, established standards of procedure for his followers, and led him to discard on a grand scale the names used by his predecessors. Many of his canons have long since been disregarded, but they ensured that modern botanical nomenclature at least began with a series of well-formed, euphonious and convenient names.
Binomial nomenclature
To understand the objectives of the Critica it is first necessary to appreciate the state of botanical nomenclature at th
|
https://en.wikipedia.org/wiki/Mihrab
|
Mihrab (, , pl. ) is a niche in the wall of a mosque that indicates the qibla, the direction of the Kaaba in Mecca towards which Muslims should face when praying. The wall in which a mihrab appears is thus the "qibla wall".
The minbar, which is the raised platform from which an imam (leader of prayer) addresses the congregation, is located to the right of the mihrab.
Etymology
The origin of the word miḥrāb is complicated and multiple explanations have been proposed by different sources and scholars. It may come from Old South Arabian (possibly Sabaic) mḥrb meaning a certain part of a palace, as well as "part of a temple where tḥrb (a certain type of visions) is obtained," from the root word ḥrb "to perform a certain religious ritual (which is compared to combat or fighting and described as an overnight retreat) in the mḥrb of the temple." It may also possibly be related to Ethiopic məkʷrab "temple, sanctuary," whose equivalent in Sabaic is mkrb of the same meaning, from the root word krb "to dedicate" (cognate with Akkadian karābu "to bless" and related to Hebrew kerūḇ "cherub (either of the heavenly creatures that bound the Ark in the inner sanctuary)").
Arab lexicographers traditionally derive the word from the Arabic root (Ḥ-R-B) relating to "war, fighting or anger," (which, though cognate with the South Arabian root, does not however carry any relation to religious rituals) thus leading some to interpret it to mean a "fortress", or "place of battle (with Satan)," the latter due to mihrabs being private prayer chambers. The latter interpretation though bears similarity to the nature of the ḥrb ritual.
The word mihrab originally had a non-religious meaning and simply denoted a special room in a house; a throne room in a palace, for example. The Fath al-Bari (p. 458), on the authority of others, suggests the mihrab is "the most honorable location of kings" and "the master of locations, the front and the most honorable." The Mosques in Islam (p. 13),
|
https://en.wikipedia.org/wiki/Nuclear%20power%20in%20Kenya
|
In 2017, the Kenya Nuclear Electrification Board (Kneb) estimated that a 1,000 MW nuclear plant could be operational by 2027 and cost Ksh500-600 billion ($5-$6 billion), to be located near a large body of water, such as the Indian Ocean, Lake Victoria or Lake Turkana.
Background
In September 2010 Former Energy and Petroleum Ministry PS Patrick Nyoike announced that Kenya aimed to build a 1,000 MW nuclear power plant between 2017 and 2022.
The projected cost using South Korean technology was US$3.5 billion.
Nuclear and renewable sources of energy such as wind, solar and geothermal plants could play a major role in helping Kenya achieve middle income status, as the reduction of carbon emissions becomes a higher priority.
Nuclear energy programme
Kenya has embarked on a programme to see the country generate 1 GW (1,000 MW) from nuclear sources between 2020 and 2022. By 2030 Kenya was slated to have installed a capacity of 4 GW of nuclear energy, generating about 19% of Kenya's energy needs, meaning that nuclear power would be the second largest source of energy in Kenya coming second after geothermal power which is a clean form of energy.
The Kenya Nuclear Electricity Board (NuPEA) is in charge of spearheading this sector in the country.
See also
Geothermal power in Kenya
Wind power in Kenya
|
https://en.wikipedia.org/wiki/Alexandre-Th%C3%A9ophile%20Vandermonde
|
Alexandre-Théophile Vandermonde (28 February 1735 – 1 January 1796) was a French mathematician, musician, and chemist who worked with Bézout and Lavoisier; his name is now principally associated with determinant theory in mathematics. He was born in Paris, and died there.
Biography
Vandermonde was a violinist, and became engaged with mathematics only around 1770. In Mémoire sur la résolution des équations (1771) he reported on symmetric functions and solution of cyclotomic polynomials; this paper anticipated later Galois theory (see also abstract algebra for the role of Vandermonde in the genesis of group theory). In Remarques sur des problèmes de situation (1771) he studied knight's tours, and presaged the development of knot theory by explicitly noting the importance of topological features when discussing the properties of knots:
"Whatever the twists and turns of a system of threads in space, one can always obtain an expression for the calculation of its dimensions, but this expression will be of little use in practice. The craftsman who fashions a braid, a net, or some knots will be concerned, not with questions of measurement, but with those of position: what he sees there is the manner in which the theads are interlaced"
The same year he was elected to the French Academy of Sciences. Mémoire sur des irrationnelles de différents ordres avec une application au cercle (1772) was on combinatorics, and Mémoire sur l'élimination (1772) on the foundations of determinant theory. These papers were presented to the Académie des Sciences, and constitute all his published mathematical work. The Vandermonde determinant does not make an explicit appearance.
He was professor at the École Normale Supérieure, member of the Conservatoire national des arts et métiers and examiner at the École polytechnique.
Honors
A special class of matrices, the Vandermonde matrices are named after him, as is an elementary fact of combinatorics, Vandermonde's identity.
Vandermonde is th
|
https://en.wikipedia.org/wiki/Regular%20tree%20grammar
|
In theoretical computer science and formal language theory, a regular tree grammar is a formal grammar that describes a set of directed trees, or terms. A regular word grammar can be seen as a special kind of regular tree grammar, describing a set of single-path trees.
Definition
A regular tree grammar G is defined by the tuple
G = (N, Σ, Z, P),
where
N is a finite set of nonterminals,
Σ is a ranked alphabet (i.e., an alphabet whose symbols have an associated arity) disjoint from N,
Z is the starting nonterminal, with , and
P is a finite set of productions of the form A → t, with , and , where TΣ(N) is the associated term algebra, i.e. the set of all trees composed from symbols in according to their arities, where nonterminals are considered nullary.
Derivation of trees
The grammar G implicitly defines a set of trees: any tree that can be derived from Z using the rule set P is said to be described by G.
This set of trees is known as the language of G.
More formally, the relation ⇒G on the set TΣ(N) is defined as follows:
A tree can be derived in a single step into a tree
(in short: t1 ⇒G t2), if there is a context S and a production such that:
t1 = S[A], and
t2 = S[t].
Here, a context means a tree with exactly one hole in it; if S is such a context, S[t] denotes the result of filling the tree t into the hole of S.
The tree language generated by G is the language .
Here, TΣ denotes the set of all trees composed from symbols of Σ, while ⇒G* denotes successive applications of ⇒G.
A language generated by some regular tree grammar is called a regular tree language.
Examples
Let G1 = (N1,Σ1,Z1,P1), where
N1 = {Bool, BList } is our set of nonterminals,
Σ1 = { true, false, nil, cons(.,.) } is our ranked alphabet, arities indicated by dummy arguments (i.e. the symbol cons has arity 2),
Z1 = BList is our starting nonterminal, and
the set P1 consists of the following productions:
Bool → false
Bool → true
BList → nil
BList → cons(Bool,BList)
A
|
https://en.wikipedia.org/wiki/FURPS
|
FURPS is an acronym representing a model for classifying software quality attributes (functional and non-functional requirements):
Functionality - Capability (Size & Generality of Feature Set), Reusability (Compatibility, Interoperability, Portability), Security (Safety & Exploitability)
Usability (UX) - Human Factors, Aesthetics, Consistency, Documentation, Responsiveness
Reliability - Availability (Failure Frequency (Robustness/Durability/Resilience), Failure Extent & Time-Length (Recoverability/Survivability)), Predictability (Stability), Accuracy (Frequency/Severity of Error)
Performance - Speed, Efficiency, Resource Consumption (power, ram, cache, etc.), Throughput, Capacity, Scalability
Supportability (Serviceability, Maintainability, Sustainability, Repair Speed) - Testability, Flexibility (Modifiability, Configurability, Adaptability, Extensibility, Modularity), Installability, Localizability
The model, developed at Hewlett-Packard was first publicly elaborated by Grady and Caswell. FURPS+ is now widely used in the software industry. The + was later added to the model after various campaigns at HP to extend the acronym to emphasize various attributes.
See also
Types of requirements
Expanded list of types of requirements
Further reading
External links
IBM on Furps+
Software requirements
Mnemonics
|
https://en.wikipedia.org/wiki/Long-term%20potentiation
|
In neuroscience, long-term potentiation (LTP) is a persistent strengthening of synapses based on recent patterns of activity. These are patterns of synaptic activity that produce a long-lasting increase in signal transmission between two neurons. The opposite of LTP is long-term depression, which produces a long-lasting decrease in synaptic strength.
It is one of several phenomena underlying synaptic plasticity, the ability of chemical synapses to change their strength. As memories are thought to be encoded by modification of synaptic strength, LTP is widely considered one of the major cellular mechanisms that underlies learning and memory.
LTP was discovered in the rabbit hippocampus by Terje Lømo in 1966 and has remained a popular subject of research since. Many modern LTP studies seek to better understand its basic biology, while others aim to draw a causal link between LTP and behavioral learning. Still, others try to develop methods, pharmacologic or otherwise, of enhancing LTP to improve learning and memory. LTP is also a subject of clinical research, for example, in the areas of Alzheimer's disease and addiction medicine.
History
Early theories of learning
At the end of the 19th century, scientists generally recognized that the number of neurons in the adult brain (roughly 100 billion) did not increase significantly with age, giving neurobiologists good reason to believe that memories were generally not the result of new neuron production. With this realization came the need to explain how memories could form in the absence of new neurons.
The Spanish neuroanatomist Santiago Ramón y Cajal was among the first to suggest a mechanism of learning that did not require the formation of new neurons. In his 1894 Croonian Lecture, he proposed that memories might instead be formed by strengthening the connections between existing neurons to improve the effectiveness of their communication. Hebbian theory, introduced by Donald Hebb in 1949, echoed Ramón y Cajal's
|
https://en.wikipedia.org/wiki/Office%20of%20Technology%20Assessment
|
The Office of Technology Assessment (OTA) was an office of the United States Congress that operated from 1974 to 1995. OTA's purpose was to provide congressional members and committees with objective and authoritative analysis of the complex scientific and technical issues of the late 20th century, i.e. technology assessment. It was a leader in practicing and encouraging delivery of public services in innovative and inexpensive ways, including early involvement in the distribution of government documents through electronic publishing. Its model was widely copied around the world.
The OTA was authorized in 1972 and received its first funding in fiscal year 1974. It was defunded at the end of 1995, following the 1994 mid-term elections which led to Republican control of the Senate and the House. House Republican legislators characterized the OTA as wasteful and hostile to GOP interests.
Princeton University hosts The OTA Legacy site, which holds "the complete collection of OTA publications along with additional materials that illuminate the history and impact of the agency". On July 23, 2008 the Federation of American Scientists launched a similar archive that includes interviews and additional documents about OTA.
History
Congress established the Office of Technology Assessment with the Technology Assessment Act of 1972. It was governed by a twelve-member board, comprising six members of Congress from each party—half from the Senate and half from the House of Representatives. During its twenty-four-year life it produced about 750 studies on a wide range of topics, including acid rain, health care, global climate change, and polygraphs.
Closure
Criticism of the agency was fueled by Fat City, a 1980 book by Washington Times journalist Donald Lambro that was regarded favorably by the Reagan administration; it called OTA an "unnecessary agency" that duplicated government work done elsewhere. OTA was abolished (technically "de-funded") in the "Contract with America"
|
https://en.wikipedia.org/wiki/Pterosaur%20size
|
Pterosaurs included the largest flying animals ever to have lived. They are a clade of prehistoric archosaurian reptiles closely related to dinosaurs. Species among pterosaurs occupied several types of environments, which ranged from aquatic to forested. Below are the lists that comprise the smallest and the largest pterosaurs known .
Smallest pterosaurs
The smallest known pterosaur is Nemicolopterus with a wingspan of about . The specimen found may be a juvenile or a subadult, however, and adults may have been larger. Anurognathus is another small pterosaur, with a wingspan of and in body mass.
Pterosaurs with largest wingspan
This is a list of pterosaurs with estimated maximum wingspan of more than 5 meters (16 feet):
Hatzegopteryx thambema
Quetzalcoatlus northropi
Cryodrakon boreas
Undescribed specimen from Mongolia
Thanatosdrakon amaru
Arambourgiania philadelphiae
Tropeognathus mesembrinus
Pteranodon longiceps
Thapunngaka shawi
Alanqa saharica
Santanadactylus araripensis
Cearadactylus atrox
The largest of non-pterodactyloid pterosaurs as well as the largest Jurassic pterosaur was Dearc, with an estimated wingspan between and . Only a fragmentary rhamphorhynchid specimen from Germany could be larger (184 % the size of the biggest Rhamphorhynchus). Other huge non-pterodactyloid pterosaurs are Sericipterus, Campylognathoides and Harpactognathus, with the wingspan of , , and , respectively. Middle Jurassic Angustinaripterus had a wingspan of .
Speculation about pterosaur size and flight
Some species of pterosaurs grew to very large sizes and this has implications for their capacity for flight. Many pterosaurs were small but the largest had wingspans which exceeded . The largest of these are estimated to have weighed . For comparison, the wandering albatross has the largest wingspan of living birds at up to but usually weighs less than . This indicates that the largest pterosaurs may have had higher wing loadings than modern birds
|
https://en.wikipedia.org/wiki/Obturator%20nerve
|
The obturator nerve in human anatomy arises from the ventral divisions of the second, third, and fourth lumbar nerves in the lumbar plexus; the branch from the third is the largest, while that from the second is often very small.
Structure
The obturator nerve originates from the anterior divisions of the L2, L3, and L4 spinal nerve roots. It descends through the fibers of the psoas major, and emerges from its medial border near the brim of the pelvis. It then passes behind the common iliac arteries, and on the lateral side of the internal iliac artery and vein, and runs along the lateral wall of the lesser pelvis, above and in front of the obturator vessels, to the upper part of the obturator foramen.
Here it enters the thigh, through the obturator canal, and divides into an anterior and a posterior branch, which are separated at first by some of the fibers of the obturator externus, and lower down by the adductor brevis.
An accessory obturator nerve may be present in approximately 8% to 29% of the general population.
Branches
Anterior branch of obturator nerve
Posterior branch of obturator nerve
Cutaneous branch of the obturator nerve
Function
The obturator nerve is responsible for the sensory innervation of the skin of the medial aspect of the thigh.
The nerve is also responsible for the motor innervation of the adductor muscles of the lower limb (external obturator, adductor longus, adductor brevis, adductor magnus, gracilis) and the pectineus (inconstant). It is, notably, not responsible for the innervation of the obturator internus, despite the similarity in name.
Clinical significance
An obturator nerve block may be used during knee surgery and urethral surgery in combination with other anaesthetics.
Additional images
|
https://en.wikipedia.org/wiki/Block%20reflector
|
"A block reflector is an orthogonal, symmetric matrix that reverses a subspace whose dimension may be greater than one."
It is built out of many elementary reflectors.
It is also referred to as a triangular factor, and is a triangular matrix and they are used in the Householder transformation.
A reflector belonging to can be written in the form :
where is the identity matrix for , is a scalar and belongs to .
LAPACK routines
Here are some of the LAPACK routines that apply to block reflectors
"*larft" forms the triangular vector T of a block reflector H=I-VTVH.
"*larzb" applies a block reflector or its transpose/conjugate transpose as returned by "*tzrzf" to a general matrix.
"*larzt" forms the triangular vector T of a block reflector H=I-VTVH as returned by "*tzrzf".
"*larfb" applies a block reflector or its transpose/conjugate transpose to a general rectangular matrix.
See also
Reflection (mathematics)
Householder transformation
Unitary matrix
Triangular matrix
|
https://en.wikipedia.org/wiki/The%20Lichenologist
|
The Lichenologist is a peer-reviewed scientific journal specializing in lichenology. It is published bimonthly by the British Lichen Society. According to the Journal Citation Reports, the 2020 impact factor of The Lichenologist is 1.514, ranking it 149 out of 235 in plant sciences and 26 of 29 in mycology.
More than 51,000 lichen-related articles were published up to 2019, about 4.7% (over 2400) of which were published in The Lichenologist; about half of these were published under the senior editorship of Peter Crittenden, who had a 20-year tenure at the journal, from 2000 to 2020.
|
https://en.wikipedia.org/wiki/Erd%C5%91s%20Lectures
|
Erdős Lectures in Discrete Mathematics and Theoretical Computer Science is a distinguished lecture series at Hebrew University of Jerusalem named after mathematician Paul Erdős. It is bringing an outstanding mathematician or computer scientist to Israel every year in the Spring. The subject of the lectures is Discrete Mathematics and Theoretical Computer Science.
The first lecture series took place in 1998.
List of Erdős Lecturers
1998: Alexander Razborov (Steklov Institute, Russia), Jeff Kahn (Rutgers University, U.S.)
1999: Richard Stanley (MIT, U.S.), Johan Håstad (Royal Institute of Technology, Sweden)
2001: Joel Spencer (NYU, U.S.)
2002: Madhu Sudan (NYU, U.S.)
2003: Maria Chudnovsky (Princeton University, U.S.)
2004: Imre Bárány (Alfréd Rényi Mathematical Institute, Hungary)
2005: János Pach (NYU, U.S.), Endre Szemerédi (Alfréd Rényi Institute of Mathematics, Hungary)
2006: József Beck (Princeton University, U.S.)
2007: Van H. Vu (Rutgers University, U.S.)
2008: Henry Cohn (Microsoft Research, U.S.)
2010: Éva Tardos (Cornell University, U.S.)
2011: Günter M. Ziegler (Freie Universität Berlin, Germany)
2012: Luca Trevisan (Stanford University, U.S.)
2013: Michael Saks (Rutgers University, U.S.)
2014: Daniel Spielman (Yale University, U.S.)
2015: Subhash Khot (NYU, U.S.)
2016: June Huh (IAS & Princeton University, U.S.)
2017: József Solymosi (UBC, Canada)
2018: Igor Pak (UCLA, U.S.)
2022: Shachar Lovett (UCSD, U.S.)
See also
List of things named after Paul Erdős
|
https://en.wikipedia.org/wiki/Lost%20Planet%3A%20Extreme%20Condition
|
Lost Planet: Extreme Condition is a third-person shooter video game developed and published by Capcom for Xbox 360, Microsoft Windows and PlayStation 3. The game was released in Japan in December 2006 and worldwide in January 2007. Originally intended to be an Xbox 360 exclusive, it was later ported and released for Microsoft Windows in June 2007 and PlayStation 3 in February 2008.
Gameplay
The game is played through a third person over-the-shoulder view. Players are allowed to switch between first-person and third-person at any moment. Players either travel on foot or ride various types of mechanized suits called Vital Suits (VSs). VSs carry heavy weapons such as chain guns and rocket launchers. They can pick up weapons lying on the ground and fire multiple weapons at once. On foot, players are able to use a grappling hook to pull themselves up to normally hard-to-reach places, or to hook onto a VS and hijack it. Driving VSs and using certain weapons requires thermal energy. Also, the planet's cold temperature causes the characters' thermal energy level to continually decrease. Players can replenish their thermal energy level by defeating enemies or activating data posts. Data posts also allow players to use their navigational radars to see incoming enemies. Each of the 11 levels is accompanied by a boss, which can be either a VS or a large Akrid.
Multiplayer
Online multiplayer versus also requires players to monitor their thermal energy level, but here, reaching zero does not cause death. Instead, the characters cannot use VSs or fire the weapons which require thermal energy. Online multiplayer versus consists of four modes, called Elimination, Team Elimination, Post Grab, and Fugitive. Players score points by killing other players and activating posts, and they lose points for being killed or committing suicide. Post grab is a mode where players on opposite teams compete to capture as many posts as possible before the set time runs out. Team Elimination is a 1
|
https://en.wikipedia.org/wiki/Cognitive%20network
|
In communication networks, cognitive network (CN) is a new type of data network that makes use of cutting edge technology from several research areas (i.e. machine learning, knowledge representation, computer network, network management) to solve some problems current networks are faced with. Cognitive network is different from cognitive radio (CR) as it covers all the layers of the OSI model (not only layers 1 and 2 as with CR ).
History
The first definition of the cognitive network was provided by Theo Kanter in his doctoral research at KTH, The Royal Institute of Technology, Stockholm, including a presentation in June 1998 of the cognitive network as the network with memory. Theo was a student of Chip Maguire who also was advising Joe Mitola, the originator of cognitive radio. Mitola focused on cognition in the nodes, while Kantor focused on cognition in the network. Mitola's Licentiate thesis, published in August, 1999 includes the following quote "Over time, the [Radio Knowledge Representation Language] RKRL-empowered network can learn to distinguish a feature of the natural environment that does not match the models. It could declare the errors to a cognitive network." This is the earliest publication of the concept cognitive network, since Kantor published a bit later.
IBM's autonomic networks challenge of 2001 instigated the introduction of a cognition cycle into networks. Cognitive radio, Kantor's cognitive networks, and IBM's autonomic networks provided the foundation for the parallel evolution of cognitive wireless networks and other cognitive networks. In 2004, Petri Mahonen, currently at RWTH, Aachen, and a member of Mitola's doctoral committee organized the first international workshop on cognitive wireless networks at Dagstuhl, Germany. In addition, the EU's E2R and E3 programs developed cognitive network theory under the rubric of self* - self organizing networks, self-aware networks, and so forth. One of the attempts to define the concep
|
https://en.wikipedia.org/wiki/GLRA2
|
Glycine receptor subunit alpha-2 is a protein that in humans is encoded by the GLRA2 gene.
See also
Glycine receptor
|
https://en.wikipedia.org/wiki/Food%20plot
|
A food plot is a planted area set aside to act as a supplementary food source for wildlife. The term was coined by the U.S. hunting and outdoor industries and food plots are most commonly planted for game species.
Food plot crops generally consist of but are not limited to legumes (clovers, alfalfa, beans, etc.), grains, or certain wildflowers. The plants may be annual or perennial and specific recommendations for local varieties that perform best in your area for specific wildlife species are often available. In agricultural areas, food plots may be planted in fields after the crop has been harvested, and left standing through the fall, winter, and early spring for the use of wildlife.
In some cases farmers and landowners may be reimbursed by the government for providing food plots that meet government specifications. Among other considerations, a food plot should be located close to a source of cover for the animals, and it should be only one part of a comprehensive habitat management program. Ideally a food plot should be small (2 to 4 acres), irregularly shaped, and located away from roads. A variety of cool-season grasses and forbs can be used to attract and enhance nutrition of multiple game species simultaneously when mixed according to recommendations.
Food plots differ from revegetation, which is intended to stabilize and rebuild the soil of disturbed land using naturally growing grasses, legumes, shrubs, and trees. Food plots are intended to feed wildlife rather than rebuild the soil, and generally use agricultural forages rather than native or naturally occurring plants. The oldest company to start developing products for food plots is the Whitetail Institute of North America in 1988.
Food plots can greatly increase the wildlife carrying capacity of a particular ecosystem, enhancing opportunities for hunting or wildlife viewing. Target animals include Whitetail deer, Bear, Moose, Rabbit, Woodchuck, Deer, Wild turkey, Grouse, and songbirds.
In
|
https://en.wikipedia.org/wiki/ICloud
|
iCloud is a cloud service developed by Apple Inc. Launched on October 12, 2011, iCloud enables users to store and sync data across devices, including Apple Mail, Apple Calendar, Apple Photos, Apple Notes, contacts, settings, backups, and files, to collaborate with other users, and track assets through Find My. It is built into iOS, iPadOS, watchOS, tvOS and macOS and may additionally be accessed through a limited web interface and Windows application.
iCloud offers users 5 GB of free storage which may be upgraded to 50 GB, 200 GB, or 2 TB through optional paid plans. Starting in September 2023, storage options for 6 TB and 12 TB have been added. All paid plans include iCloud+ which additionally provides Private Relay, Hide My Email and Custom Email Domain.
, the service had an estimated 850 million users, up from 782 million users in 2016.
In December 2022, Apple rolled out optional end-to-end encryption for all iCloud data (including iCloud Photos, iCloud Drive, and more), with the exception of Calendar, Contacts, and Mail, which rely on legacy sync technologies for compatibility with third-party apps (CalDav, CardDav, IMAP). The feature was released to U.S. customers on December 13, 2022, and rolled out worldwide on January 23, 2023.
History
iCloud was announced on May 31, 2011, in a press release. On June 6, 2011, during the WWDC 2011 keynote, Steve Jobs announced that iCloud would replace MobileMe, which had been widely seen as a "failure", a fact which Steve Jobs acknowledged during the announcement. iCloud was released on October 12, 2011, and MobileMe was discontinued on June 30, 2012. Previous MobileMe users could keep their @mac.com and @me.com email addresses as aliases to their new @icloud.com address. Earlier versions included Back to My Mac, which was previously part of MobileMe. This service allowed users to create point-to-point connections between computers. It was discontinued on July 1, 2019.
iCloud had 20 million users within a week after la
|
https://en.wikipedia.org/wiki/SAGE%20KE
|
The Science of Aging Knowledge Environment (SAGE KE) was an online scientific resource provided by the American Association for the Advancement of Science (AAAS).
History and organization
The American Association for the Advancement of Science established a collaboration with Stanford University Libraries and The Center for Resource Economics/Island Press (Island Press) in 1996 to find means to utilize internet-based technologies to enhance access to scientific information and improve the effectiveness of information transfer. The collaborative coined the term Knowledge Environment (KE) to describe the collection of electronic networking tools they were seeking to develop.
SAGE KE is the third in a series of Knowledge Environments developed by Science and AAAS, after the Signal Transduction Knowledge Environment (STKE) and AIDScience. Funding for SAGE KE comes from The Ellison Medical Foundation, founded and supported by Oracle Corporation CEO Larry Ellison.
SAGE KE published its final issue on 28 June 2006 due to lack of funding. The interactive content was discontinued during the summer of 2006, leaving the SAGE KE site as an archive by August 2006.
Activities
The focus of SAGE KE was to provide timely access to information about advances on basic mechanisms of aging and age-related diseases through the internet, to provide searchable databases of information on aging and to provide an active environment in which biogerontologists could share and debate their understandings.
Ouroboros
Ouroboros is a WordPress community weblog devoted to research in the biology of aging. It was established in July 2006 in reaction to the termination of the SAGE KE. The primary mission of the site is to provide timely commentary and review of recently published articles in the scholarly literature, either directly or indirectly related to aging. Articles on the site discuss a range of scientific topics, including Alzheimer's disease, bioinformatics, calorie restriction, regul
|
https://en.wikipedia.org/wiki/International%20Association%20for%20Statistical%20Education
|
The International Association for Statistical Education (IASE) is a section of the International Statistical Institute (ISI), a professional association of statisticians, devoted to statistics education. It was founded in 1991 as an outgrowth of the ISI Statistical Education Committee, which had operated since 1948.
Since 2002 the ISI and IASE have published the Statistics Education Research Journal. The IASE is also associated with the quadrennial International Conference on Teaching Statistics, with satellite conferences of the World Statistics Congress, and with smaller roundtable workshops.
The presidents of the IASE have included
David Vere-Jones (1991–1993),
David S. Moore (1993–1995),
Anne Hawkins (1995–1997),
Maria Gabriella Ottaviani (1997–1999),
Brian Phillips (1999–2001),
Carmen Batanero (2001–2003),
Chris Wild (2003–2005),
Gilberte Schuyten (2005–2007),
Allan Rossman (2007–2009),
Helen MacGillivray (2009–2011),
John Harraway (2011–2013),
Iddo Gal (2013–2015),
Andrej Blejec (2015–2017),
Gail F. Burrill (2017–2019),
and Joachim Engel (2019–2021).
|
https://en.wikipedia.org/wiki/Higher-order%20function
|
In mathematics and computer science, a higher-order function (HOF) is a function that does at least one of the following:
takes one or more functions as arguments (i.e. a procedural parameter, which is a parameter of a procedure that is itself a procedure),
returns a function as its result.
All other functions are first-order functions. In mathematics higher-order functions are also termed operators or functionals. The differential operator in calculus is a common example, since it maps a function to its derivative, also a function. Higher-order functions should not be confused with other uses of the word "functor" throughout mathematics, see Functor (disambiguation).
In the untyped lambda calculus, all functions are higher-order; in a typed lambda calculus, from which most functional programming languages are derived, higher-order functions that take one function as argument are values with types of the form .
General examples
map function, found in many functional programming languages, is one example of a higher-order function. It takes as arguments a function f and a collection of elements, and as the result, returns a new collection with f applied to each element from the collection.
Sorting functions, which take a comparison function as a parameter, allowing the programmer to separate the sorting algorithm from the comparisons of the items being sorted. The C standard function qsort is an example of this.
filter
fold
apply
Function composition
Integration
Callback
Tree traversal
Montague grammar, a semantic theory of natural language, uses higher-order functions
Support in programming languages
Direct support
The examples are not intended to compare and contrast programming languages, but to serve as examples of higher-order function syntax
In the following examples, the higher-order function takes a function, and applies the function to some value twice. If has to be applied several times for the same it preferably should return a fu
|
https://en.wikipedia.org/wiki/Imagined%20speech
|
Imagined speech (also called silent speech, covert speech, inner speech, or, in the original Latin terminology used by clinicians, endophasia) is thinking in the form of sound – "hearing" one's own voice silently to oneself, without the intentional movement of any extremities such as the lips, tongue, or hands. Logically, imagined speech has been possible since the emergence of language, however, the phenomenon is most associated with its investigation through signal processing and detection within electroencephalograph (EEG) data as well as data obtained using alternative non-invasive, brain–computer interface (BCI) devices.
History
In 2008, the US Defense Advanced Research Projects Agency (DARPA) provided a $4 million grant to the University of California (Irvine), with the intent of providing a foundation for synthetic telepathy. According to DARPA, the project "will allow user-to-user communication on the battlefield without the use of vocalized speech through neural signals analysis. The brain generates word-specific signals prior to sending electrical impulses to the vocal cords. These imagined speech signals would be analyzed and translated into distinct words allowing covert person-to-person communication." In his "Impossible languages" (2016) Andrea Moro discusses the "sound of thoughts" and the relationship between linguistics units and imagined speech, mainly capitalizing on Magrassi et al. (2015) "Sound representation in higher language areas during language production".
DARPA's program outline has three major goals:
To attempt to identify EEG patterns unique to individual words
To ensure these patterns are common to different users to avoid extensive device training
To construct a prototype that would decode the signals and transmit them over a limited range
Detection methods
The process for analyzing subjects' silent speech is composed of recording subjects' brain waves, and then using a computer to process the data and determine the content of t
|
https://en.wikipedia.org/wiki/Infrared%20cirrus
|
Infrared cirrus or galactic cirrus are galactic filamentary structures seen in space over most of the sky that emit far-infrared light. The name is given because the structures are cloud-like in appearance. These structures were first detected by the Infrared Astronomy Satellite at wavelengths of 60 and 100 micrometres.
See also
Cosmic infrared background
|
https://en.wikipedia.org/wiki/Infraorbital%20artery
|
The infraorbital artery is a small artery in the head that arises from the maxillary artery and passes through the inferior orbital fissure to enter the orbit, then passes forward along the floor of the orbit, finally exiting the orbit through the infraorbital foramen to reach the face.
Anatomy
Origin
The infraorbital artery arises from the maxillary artery; it often arises in conjunction with the posterior superior alveolar artery. It may be considered a continuation of the third part of the maxillary artery and continues the direction of the maxillary artery.
Course
It passes anterior-ward to enter the orbit through the inferior orbital fissure. In the orbit, it courses along the floor of the orbit with the infraorbital nerve first along the infraorbital groove and then the infraorbital canal. It exits the orbit (with the infraorbital nerve) through infraorbital foramen to reach the face, beneath the infraorbital head of the levator labii superioris muscle.
Branches
While in the canal, it gives off:
Orbital branches - assist in supplying the inferior rectus and inferior oblique and the lacrimal sac.
Anterior superior alveolar artery - supplies upper/maxillary canine and incisor teeth.
Middle superior alveolar artery - upper/maxillary canine and incisor teeth. May be absent.
On the face, some branches pass upward to the medial angle of the orbit and the lacrimal sac, anastomosing with the angular artery, a branch of the facial artery; others run toward the nose, anastomosing with the dorsal nasal branch of the ophthalmic artery; and others descend between the levator labii superioris and the levator anguli oris, and anastomose with the facial artery, transverse facial artery, and buccal artery.
The four remaining branches arise from that portion of the maxillary artery which is contained in the pterygopalatine fossa.
Additional images
|
https://en.wikipedia.org/wiki/TCL%20%28GTPase%29
|
TCL is a small (~21 kDa) signaling G protein (more specifically a GTPase), and is a member of the Rho family of GTPases.,.
TCL (TC10-like) shares 85% and 78% amino acid similarity to TC10 and Cdc42, respectively. TCL mRNA is 2.5 kb long and is mainly expressed in heart. In vitro, TCL shows rapid GDP/GTP exchange and displays higher GTP dissociation and hydrolysis rates than TC10. Like other Rac/Cdc42/RhoUV members, GTP-bound TCL interacts with CRIB domains, such as those found in PAK and WASP. TCL produces large and dynamic F-actin-rich ruffles on the dorsal cell membrane in REF-52 fibroblasts. TCL activity is blocked by dominant negative Rac1 and Cdc42 mutants, suggesting a cross-talk between these three Rho GTPases.
TCL is unrelated to TCL1A, a proto-oncogene implicated in the development of T-Cell Leukemias.
See also
TCL1A
|
https://en.wikipedia.org/wiki/Phototroph
|
Phototrophs () are organisms that carry out photon capture to produce complex organic compounds (e.g. carbohydrates) and acquire energy. They use the energy from light to carry out various cellular metabolic processes. It is a common misconception that phototrophs are obligatorily photosynthetic. Many, but not all, phototrophs often photosynthesize: they anabolically convert carbon dioxide into organic material to be utilized structurally, functionally, or as a source for later catabolic processes (e.g. in the form of starches, sugars and fats). All phototrophs either use electron transport chains or direct proton pumping to establish an electrochemical gradient which is utilized by ATP synthase, to provide the molecular energy currency for the cell. Phototrophs can be either autotrophs or heterotrophs. If their electron and hydrogen donors are inorganic compounds (e.g., , as in some purple sulfur bacteria, or , as in some green sulfur bacteria) they can be also called lithotrophs, and so, some photoautotrophs are also called photolithoautotrophs. Examples of phototroph organisms are Rhodobacter capsulatus, Chromatium, and Chlorobium.
History
Originally used with a different meaning, the term took its current definition after Lwoff and collaborators (1946).
Photoautotroph
Most of the well-recognized phototrophs are autotrophic, also known as photoautotrophs, and can fix carbon. They can be contrasted with chemotrophs that obtain their energy by the oxidation of electron donors in their environments. Photoautotrophs are capable of synthesizing their own food from inorganic substances using light as an energy source. Green plants and photosynthetic bacteria are photoautotrophs. Photoautotrophic organisms are sometimes referred to as holophytic.
Oxygenic photosynthetic organisms use chlorophyll for light-energy capture and oxidize water, "splitting" it into molecular oxygen.
Ecology
In an ecological context, phototrophs are often the food source for neighboring he
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.