source stringlengths 31 203 | text stringlengths 28 2k |
|---|---|
https://en.wikipedia.org/wiki/Lumus | Lumus is an Israeli-based augmented reality company headquartered in Ness Ziona, Israel. Founded in 2000, Lumus has developed technology for see-through wearable displays, via its patented Light-guide Optical Element (LOE) platform to market producers of smart glasses and augmented reality eyewear.
Technology
The LOE is a patented optical waveguide that makes use of multiple partial reflectors embedded in a single substrate to reflect a virtual image into the eye of the wearer. Specifically, the image is coupled into the LOE by a "Pod" (micro-display projector) that sits at the edge of the waveguide—in an eyeglass configuration, this is embedded in the temple of the glasses. The image travels through total internal reflection to the multiple array of partial reflectors and are reflected to the eye. While each partial reflector shows only a portion of the image, the optics are such that the wearer sees the combined array and perceives it as a single uniform image projected at infinity. The transparent display enables a virtual image to be seamlessly overlaid over the wearer's real world view. This is especially true when the source image comprises a black background with light color wording or symbology being displayed. Black is essentially see-through color, while lighter colored objects, symbols or characters appear to float in the wearer's line of sight. Conversely, full screen images like documents, internet pages, movies which are typically brighter colors can be displayed to look like a large virtual image floating a few meter's away from the wearer.
Lumus, with the LOE, has a single waveguide that works on all colors. The thickness of their one LOE is similar to the stack of multiple (one per red, green, and blue) thinner waveguides on HoloLens. They simply cut the waveguide's entrance at an angle to get the light to enter (rather than use a color specific diffraction grating), and then they use a series of very specially designed partial mirrors to cause t |
https://en.wikipedia.org/wiki/Castor%20%28framework%29 | Castor is a data binding framework for Java with some features like Java to Java-to-XML binding, Java-to-SQL persistence, paths between Java objects, XML documents, relational tables, etc. Castor is one of the oldest data binding projects.
Process flow
Basic process flows include class generation, marshalling, unmarshalling, etc. Marshalling framework includes a set of ClassDescriptors and FieldDescription to describe objects.
Class generation
Class generation is similar to JAXB and Zeus. Castor supports XML Schema instead of DTDs (DTDs are not supported by Castor).
Unmarshalling and marshalling
Unmarshalling and marshalling are dealt with marshall() and unmarshall() methods respectively. During marshalling, conversion process from Java to XML is carried out, and, during unmarshalling, conversion process from XML to Java is carried out. Mapping files are the equivalent of a binding schema, which allows to transforms names from XML to Java and vice versa.
Additional features
Castor offers some additional features which are not present in JAXB. Additional features include:
Database and directory server mappings - mapping between databases and directory servers to Java
JDO - Caster supports Java Data Objects.
Code samples
Code for marshalling may look like as follows:
package javajaxb;
import java.io.File;
import java.io.FileReader;
import java.io.FileWriter;
import java.io.IOException;
// Castor
import org.exolab.castor.xml.MarshalException;
import org.exolab.castor.xml.ValidationException;
// Generated hr.xml classes
import javajaxb.generated.hr.*;
public class EmployeeLister {
// Existing methods
public void modify()
throws IOException, MarshalException, ValidationException {
// Add a new employee
Employee employee = new Employee();
employee.setName("Ben Rochester");
Address address = new Address();
address.setStreet1("708 Teakwood Drive");
address.setCity("Flower Mound");
address.se |
https://en.wikipedia.org/wiki/Cavalier-Smith%27s%20system%20of%20classification | The initial version of a classification system of life by British zoologist Thomas Cavalier-Smith appeared in 1978. This initial system continued to be modified in subsequent versions that were published until he died in 2021. As with classifications of others, such as Carl Linnaeus, Ernst Haeckel, Robert Whittaker, and Carl Woese, Cavalier-Smith's classification attempts to incorporate the latest developments in taxonomy., Cavalier-Smith used his classifications to convey his opinions about the evolutionary relationships among various organisms, principally microbial. His classifications complemented his ideas communicated in scientific publications, talks, and diagrams. Different iterations might have a wider or narrow scope, include different groupings, provide greater or lesser detail, and place groups in different arrangements as his thinking changed. His classifications has been a major influence in the modern taxonomy, particularly of protists.
Cavalier-Smith has published extensively on the classification of protists. One of his major contributions to biology was his proposal of a new kingdom of life: the Chromista, although the usefulness of the grouping is questionable given that it is generally agreed to be an arbitrary (polyphyletic) grouping of taxa. He also proposed that all chromista and alveolata share the same common ancestor, a claim later refuted by studies of morphological and molecular evidence by other labs. He named this new group the Chromalveolates. He also proposed and named many other high-rank taxa, like Opisthokonta (1987), Rhizaria (2002), and Excavata (2002), though he himself consistently does not include Opisthonkonta as a formal taxon in his schemes. Together with Chromalveolata, Amoebozoa (he amended their description in 1998), and Archaeplastida (which he called Plantae since 1981) the six formed the basis of the taxonomy of eukaryotes in the middle 2000s. He has also published prodigiously on issues such as the origin of var |
https://en.wikipedia.org/wiki/Island%20Conservation | Island Conservation is a non-profit organization with the mission to prevent extinctions by removing invasive species from islands. Island Conservation has therefore focused its efforts on islands with species categorized as Critically Endangered and Endangered on the IUCN's Red List. Working in partnership with local communities, government management agencies, and conservation organizations, Island Conservation develops plans and implements the removal of invasive alien species, and conducts field research to document the benefits of the work and to inform future projects.
Island Conservation's approach is now being shown to have a wider beneficial effect on the marine systems surrounding its project areas. In addition, invasive vertebrate eradication has now been shown to have many benefits besides conservation of species. Specifically, the approach has been found to align with 13 UN Sustainable Development Goals and 42 associated targets encompassing marine and terrestrial biodiversity conservation, promotion of local and global partnerships, economic development, climate change mitigation, human health and sanitation and sustainable production and consumption.
To date Island Conservation has deployed teams to protect 1,195 populations of 487 species and subspecies on 64 islands.
The work of Island Conservation is not without controversy, This is documented in the book Battle at the End of Eden. Restoring islands requires removing whole populations of an invasive species. There is an ethical question of whether humankind has the right to remove one species to save others. However, a 2019 study suggests that if eradications of invasive animals were conducted on just 169 islands, the survival prospects of 9.4% of the Earth's most highly threatened terrestrial insular vertebrates would be improved.
History
Island Conservation was founded by Bernie Tershy and Don Croll, both Professors at UCSC's Long Marine Lab. These scientists learned about the story of Clipp |
https://en.wikipedia.org/wiki/Emerging%20Sources%20Citation%20Index | The Emerging Sources Citation Index (ESCI) is a citation index produced since 2015 by Thomson Reuters and now by Clarivate. According to the publisher, the index includes "peer-reviewed publications of regional importance and in emerging scientific fields".
The ESCI is accessible through the Web of Science, together with other Clarivate indexes. As of June 2021, all journals indexed in ESCI are also included in the Journal Citation Reports. While these journals still did not receive an impact factor until the next year, they did contribute citations to the calculation of other journals' impact factors. In July 2022, Clarivate announced that journals in the ESCI obtain an impact factor effective from JCR Year 2022 first released in June 2023.
Inclusion criteria
To be included in the ESCI, journals must be:
Peer reviewed
Follow ethical publishing practices
Meet technical requirements
Have English language bibliographic information
Be recommended or requested by a scholarly audience of Web of Science users
Criticism
Jeffrey Beall argued that among the databases produced by Clarivate, the ESCI is the easiest one to get into and that as a result it contains many predatory journals.
References
External links
Citation indices
Online databases
Clarivate |
https://en.wikipedia.org/wiki/Synthetic%20Biology%20Open%20Language | The Synthetic Biology Open Language (SBOL) is a proposed data standard for exchanging synthetic biology designs between software packages. It has been under development by the SBOL Developers Group since 2008. This group aims to develop the standard in a way that is open and democratic in order to include as many interests as possible and to avoid domination by a single company. The group also aims to develop and improve the design standard over time as the field of synthetic biology reflects this development.
A graphical modeling language called SBOL Visual has also been created to visualize SBOL designs.
Releases
References
Biocybernetics
Bioinformatics
Systems biology |
https://en.wikipedia.org/wiki/Metascape | Metascape is a free gene annotation and analysis resource that helps biologists make sense of one or multiple gene lists. Metascape provides automated meta-analysis tools to understand either common or unique pathways and protein networks within a group of orthogonal target-discovery studies.
History
In the "OMICs" age, it is important to gain biological insights into a list of genes. Although a number of bioinformatics sources exist for this purpose, such as DAVID, they are not all free, easy to use, and well maintained. To analyze multiple lists of genes originated from orthogonal but complementary "OMICs" studies, tools often require computational skills that are beyond the reach of many biologists. According to the Metascape blog, a team of scientists self-organized to address this challenge. The team includes core members Yingyao Zhou, Bin Zhou, Lars Pache, Max Chang, Christopher Benner, and Sumit Chanda, as well as other contributors over the time. Metascape was first released as a beta version on Oct 8, 2015. The first Metascape application was published on Dec 9, 2015. Metascape has gone through multiple releases since then. It currently supports key model organisms, pathway enrichment analysis, protein-protein interaction network and component analysis, automatic presentation of the results as publication-ready web report, Excel and PowerPoint presentations.
The paper titled "Metascape provides a biologist-oriented resource for the analysis of systems-level datasets" was published on Apr 3, 2019 in Nature Communications.
Analysis workflow
Metascape implements a CAME analysis workflow:
Conversion: Convert gene identifiers from popular types (such as Symbol, RefSeq, Ensembl, UniProt, UCSC) into human Entrez gene IDs and vice versa.
Annotation: Extract from dozens of function-relevant gene annotations, including protein families, transmembrane/secreted predictions, disease associations, compound associations, etc.
Membership: Flag gene membersh |
https://en.wikipedia.org/wiki/Riemannian%20metric%20and%20Lie%20bracket%20in%20computational%20anatomy | Computational anatomy (CA) is the study of shape and form in medical imaging. The study of deformable shapes in computational anatomy rely on high-dimensional diffeomorphism groups which generate orbits of the form . In CA, this orbit is in general considered a smooth Riemannian manifold
since at every point of the manifold there is an inner product inducing the norm on the tangent space
that varies smoothly from point to point in the manifold of shapes . This is generated by viewing the
group of diffeomorphisms as a Riemannian manifold with , associated to the tangent space at . This induces the norm and metric on the orbit under the action from the group of diffeomorphisms.
The diffeomorphisms group generated as Lagrangian and Eulerian flows
The diffeomorphisms in computational anatomy are generated to satisfy the Lagrangian and Eulerian specification of the flow fields, , generated via the ordinary differential equation
with the Eulerian vector fields in for , with the inverse for the flow given by
and the Jacobian matrix for flows in given as
To ensure smooth flows of diffeomorphisms with inverse, the vector fields must be at least 1-time continuously differentiable in space which are modelled as elements of the Hilbert space using the Sobolev embedding theorems so that each element has 3-square-integrable derivatives thusly implies embeds smoothly in 1-time continuously differentiable functions. The diffeomorphism group are flows with vector fields absolutely integrable in Sobolev norm:
The Riemannian orbit model
Shapes in Computational Anatomy (CA) are studied via the use of diffeomorphic mapping for establishing correspondences between anatomical coordinate systems. In this setting, 3-dimensional medical images are modelled as diffemorphic transformations of some exemplar, termed the template , resulting in the observed images to be elements of the random orbit model of CA. For images these are defined as , with for charts represen |
https://en.wikipedia.org/wiki/Expert360 | Expert360 is an online marketplace co-founded by Bridget Loudon and Emily Yue and headquartered in Sydney, Australia. Expert360 acts as a digital network for matching independent business consultants with clients (companies, organizations) for short or long-term project work. The company is best known for its innovative approach to the local and international freelance marketplace.
Description
Expert360's online platform is structured like a standard freelance marketplace as it provides the tools to potential employers to post and manage jobs of interest to independent consultants in the upper tier employment market. Its main consultant base is made of highly qualified workers with experience in senior or executive positions in large and stable corporations or investment firms. The marketplace also has junior representatives from management consulting and investment firms.
Expert360's clients range from small-medium businesses to enterprises such as QSHR, Woolworths and Telstra, as well as consulting and investment firms.
History
Expert360 launched on July 1, 2013. The company's co-founders, Bridget Loudon and Emily Yue, raised $1 million (AUD) from investors in their first funding round at the end of 2013
In 2015, Expert360 closed an oversubscribed capital raising round of $4.1 million (AUD), backed by Russian investment fund Frontier Ventures, Australian technology fund Rampersand and several Australian angel investors. Allan Moss AO is also a notable investor of Expert360. On March 30, 2016, the company announced that it was opening an office in New York City in order to expand its reach in the United States.
See also
Freelance marketplace
Freelancers Union
Independent contractor
Mercenary
Misclassification of employees as independent contractors
Recruitment advertising
Self-employment
References
Telecommuting
Freelance marketplace websites
Companies based in Sydney
Online marketplaces of Australia |
https://en.wikipedia.org/wiki/Mathematics%20and%20the%20Search%20for%20Knowledge | Mathematics and the Search for Knowledge is a 1985 book by Morris Kline about the role of mathematics when understanding of the physical world. It is preceded by Kline's work, Mathematics: The Loss of Certainty.
In the book, Kline gives an outline of the development of physics, from ancient Greek astronomy to modern physics. He explains that modern physics (consisting of theories such as electromagnetism, relativity and quantum mechanics) differs from previous theories such as Newtonian mechanics in being purely mathematical models without any intuitive ways of being visualized. Further, unlike sensory perception, modern theories have provided predictions that have been verified and are immune to sensory illusions. Thus, Kline argues that it is mathematics that provides a true understanding of physical reality, rather than our senses.
Bibliography
Notes
Books about mathematics
1985 non-fiction books |
https://en.wikipedia.org/wiki/Roots%20%282016%20miniseries%29 | Roots is a 2016 American miniseries and a remake of the 1977 miniseries with the same name, based on Alex Haley's 1976 novel, Roots: The Saga of an American Family, which follows an African man who is shipped to North America as a slave and his descendants. It first aired on May 30, 2016 and stars Malachi Kirby, Forest Whitaker, Anna Paquin, Laurence Fishburne, Jonathan Rhys Meyers, Anika Noni Rose, T.I. and South African actress Nokuthula Ledwaba. It was produced on a budget of $50 million.
Plot
Part 1
In 1767, Kunta Kinte (Malachi Kirby) is a young Mandinka man from Jufureh in the Gambia region of West Africa. One day, Kunta is taken into the jungle with other Mandinka youth as part of their training to become warriors. During a test where Kunta is made to run through the jungle in a certain period of time, he sees a man's dead body lying in a canoe. Kunta then sees the rival Koro family in another group of canoes, hunting for slaves to sell to European slave traders. After returning home, Kunta reports his discovery to his teachers, who decided in light of this information that Kunta's training must come to an end. After undergoing an initiation ceremony and becoming a warrior, Kunta tells his father that he wishes to attend a university in Timbuktu; in response to his father disagreeing with him, Kunta runs off into the jungle where he meets up with his lover, Jinna. The two are suddenly surrounded by the Koro family, who intend to ransom Kunta back to his relatives. However, Kunta manages to escape, killing one of the Koro family in the process before being recaptured. The Koro family subsequently decide to sell Kunta and Jinna to the crew of the Lord Ligonier, a British slave ship.
After Kunta and Jinna are brought onboard the Lord Ligonier, they discover that Kunta's uncle, Silla, has also been sold into slavery. Silla attempts to escape but is thwarted by a sailor who shoots him in the arm. The captives are then brought to the ship's hold and chained next |
https://en.wikipedia.org/wiki/Caribherp | Caribherp is an online database containing information on amphibians and reptiles of the Caribbean Islands. It was established in 1999 and serves as a resource for determining the species that occur on specific islands, viewing their distributions, and identifying them by images. Besides the primary search capability by regions and islands, the site features a global search functionality and the ability to refine lists by taxon and origin (endemic or introduced), and to sort by various features. Caribherp also includes common and scientific names, sightings, images, videos, audio of frog calls, distribution maps, geographic regions, and conservation status provided by the International Union for Conservation of Nature (IUCN).
The development and maintenance of Caribherp is accomplished through the work of S. Blair Hedges and his colleagues, and students from Penn State University and (since 2014) Temple University.
Contents
Caribherp database currently contains 1,022 reptile and amphibian species, maps for each species, and about 2000 professional images. This is 5% of the roughly 8,579 amphibian species and 11,940 reptiles species in the world. New species are continually being discovered and described.
References
External links
Official Website
Online databases
Biological databases
Herpetology
Biodiversity databases
Natural history of the Caribbean |
https://en.wikipedia.org/wiki/Large%20deformation%20diffeomorphic%20metric%20mapping | Large deformation diffeomorphic metric mapping (LDDMM) is a specific suite of algorithms used for diffeomorphic mapping and manipulating dense imagery based on diffeomorphic metric mapping within the academic discipline of computational anatomy, to be distinguished from its precursor based on diffeomorphic mapping. The distinction between the two is that diffeomorphic metric maps satisfy the property that the length associated to their flow away from the identity induces a metric on the group of diffeomorphisms, which in turn induces a metric on the orbit of shapes and forms within the field of Computational Anatomy. The study of shapes and forms with the metric of diffeomorphic metric mapping is called diffeomorphometry.
A diffeomorphic mapping system is a system designed to map, manipulate, and transfer information which is stored in many types of spatially distributed medical imagery.
Diffeomorphic mapping is the underlying technology for mapping and analyzing information measured in human anatomical coordinate systems which have been measured via Medical imaging. Diffeomorphic mapping is a broad term that actually refers to a number of different algorithms, processes, and methods. It is attached to many operations and has many applications for analysis and visualization. Diffeomorphic mapping can be used to relate various sources of information which are indexed as a function of spatial position as the key index variable. Diffeomorphisms are by their Latin root structure preserving transformations, which are in turn differentiable and therefore smooth, allowing for the calculation of metric based quantities such as arc length and surface areas. Spatial location and extents in human anatomical coordinate systems can be recorded via a variety of Medical imaging modalities, generally termed multi-modal medical imagery, providing either scalar and or vector quantities at each spatial location. Examples are scalar T1 or T2 magnetic resonance imagery, or as 3x3 di |
https://en.wikipedia.org/wiki/IBM%203090 | The IBM 3090 family is a family of mainframe computers that was a high-end successor to the IBM System/370 series, and thus indirectly the successor to the IBM System/360 launched 25 years earlier.
Announced on 12 February 1985, the press releases did not explicitly mention that the two models, Model 200 and Model 400, were backwardly compatible with the 370; instead, they were simply positioned as replacements for the IBM 3033. The IBM 3090/200 version was rated at 18 MIPS and 31,000 UNIX Dhrystones. This was true of the entire line, which expanded with the release of the Model 120E, 150, 150E, 180, 180E, 200, 200E, 300, 300E, 400, 400E, 600E, 600J, and 600S 3090 were described as using "ideas from the ... IBM 3033, extending them ... It also took ... from the ... IBM 308X."
The 400 and 600 were respectively two 200s or 300s coupled together as one system and could run in either single-system image mode or partitioned into two systems.
Models and features
Cooling
By the late 1970s and early 1980s, patented technology allowed Amdahl mainframes of this era to be completely air-cooled, unlike IBM systems that required chilled water and its supporting infrastructure. The eight largest of the 18 models of the ES/9000 systems introduced in 1990 were water-cooled; the other ten were air-cooled.
Enterprise Systems Architecture/370
On February 15, 1988, IBM announced Enterprise Systems Architecture/370 for 3090 enhanced ("E") models and for 4381 model groups 91E and 92E. In additional to the primary and secondary addressing modes that System/370 Extended Architecture (S/370-XA) supports, ESA has an AR mode in which each use of general register 1-15 as a base register uses an associated access register to select an address space. In addition to the normal address spaces that S/370-XA supports, ESA also allows data spaces, which contain no executable code.
Processor Resource/Systems Manager (PR/SM)
On February 15, 1988, IBM announced
Processor Resource/Systems Mana |
https://en.wikipedia.org/wiki/Content%20as%20a%20service | Content as a service (CaaS) or managed content as a service (MCaaS) is a service-oriented model, where the service provider delivers the content on demand to the service consumer via web services that are licensed under subscription. The content is hosted by the service provider centrally in the cloud and offered to a number of consumers that need the content delivered into any applications or system, hence content can be demanded by the consumers as and when required.
Content as a Service is a way to provide raw content (in other words, without the need for a specific human compatible representation, such as HTML) in a way that other systems can make use of it. Content as a Service is not meant for direct human consumption, but rather for other platforms to consume and make use of the content according to their particular needs. This happens usually on the cloud, with a centralized platform which can be globally accessible and provides a standard format for your content. With Content as a Service, you centralize your content into a single repository, where you can manage it, categorize it, make it available to others, search for it, or do whatever you wish with it.
Overview
The content delivered typically could be one or more of the following
The technical terminology related to equipments or spares that is required to procure or design the materials
The industrial terminology of the equipments or spares
Technical values pertaining to various types, specifications, applications, characteristics of equipments or spares
Sourcing information which will help in procurement or supply-chain management of equipments or spares
Descriptive specifications of equipments or spares based on the product reference number or identifier
UNSPSC codes or industry practiced classifications
ISO, IEC compliant terminology
Ontology or Technical Dictionary of products & services
Predefined content for specific business needs
The term "Content as a service" (CaaS) is considered |
https://en.wikipedia.org/wiki/Sensor%20Web%20Enablement | Sensor Web Enablement (SWE) is a suite of standards developed and maintained by Open Geospatial Consortium. SWE standards enable developers to make all types of sensors, transducers and sensor data repositories discoverable, accessible and usable via the Web.
SWE Standards include:
Sensor Observation Service
Sensor Planning Service
Observations and Measurements
Sensor Model Language
SensorThings API
References
Open Geospatial Consortium
Internet Standards |
https://en.wikipedia.org/wiki/Hadamard%27s%20gamma%20function | In mathematics, Hadamard's gamma function, named after Jacques Hadamard, is an extension of the factorial function, different from the classical gamma function (it is an instance of a pseudogamma function.) This function, with its argument shifted down by 1, interpolates the factorial and extends it to real and complex numbers in a different way than Euler's gamma function. It is defined as:
where denotes the classical gamma function. If is a positive integer, then:
Properties
Unlike the classical gamma function, Hadamard's gamma function is an entire function, i.e. it has no poles in its domain. It satisfies the functional equation
with the understanding that is taken to be for positive integer values of .
Representations
Hadamard's gamma can also be expressed as
where is the Lerch zeta function, and as
where denotes the digamma function.
References
Gamma and related functions
Analytic functions
Special functions |
https://en.wikipedia.org/wiki/Armed%20Forces%20Engineering%20Authority%20%28Egypt%29 | Armed Forces Engineering Authority is one of the Egyptian Ministry of Defense agencies.
Tasks
Contribute in the areas of development in Egypt and the establishment of major development projects
Contribute to remove the effects of local disasters and crises
Participation of in the aftermath of a major disaster to remove the effects (to reconstruct)
Participation in infrastructure and construction of bridges, roads and fields
Participation in education by building schools
Participation in the fields of psychology and clearing of land mines and explosive remnants of war
Participation in the areas of political development and the development of sports facilities
Supply the civil sector effective cadres
Participation in the area of economic housing
General planning of cities and the creation of Bedouin villages and Nubian houses
Remove the effects of disasters and floods and earthquakes, removing bombs and booby traps
References
Construction and civil engineering companies of Egypt
Defence agencies of Egypt
Military engineering |
https://en.wikipedia.org/wiki/NASA%27s%20Lunabotics%20Competition | NASA's Lunabotics Challenge
Since 2010, NASA’s Lunabotics competition has provided college students from around the country an opportunity to engage with the NASA Systems Engineering process to design and build a robotic Lunar excavator capable of mining regolith and icy regolith simulants.
Some of the deliverables include a Project Management Plan, a Public Outreach Report, Presentations and Demonstrations, and a Systems Engineering Paper. The rules and rubrics evolve each year to account for changes to the Artemis Program mission objectives and advances in commercially available technology. The competition allows NASA to gather and evaluate design and operational data for future robotic excavators and builders. The complexities the robots will have to master are the abrasive characteristics of the regolith simulant, the resources required to excavate and construct, the weight and size limitations of the Lunar robot and the ability to operate by remote control (tele-operate) or through autonomous operations.
For more than a decade, NASA has been able to gather valuable data about necessary excavation hardware and surface locomotion processes that can be implemented as the agency prepares to return to the Moon through the Artemis program. We plan to put in place sustainable infrastructure that will allow us to explore and study more of the Moon than ever before and all in preparation for the human exploration of Mars. To support these goals, participating teams will use the systems engineering process to design and construct a prototype Lunar robot to demonstrate the technologies required for a sustainable human presence on the Moon, and to build those items using “Infrastructure to Stay” technologies.
This is a two-semester, virtual challenge, designed to educate college students in the application of the NASA Systems Engineering process that may culminate with the design and build of a prototype Lunar robot. The events are as follows:
The Challenges
1. Pr |
https://en.wikipedia.org/wiki/List%20of%20genetically%20modified%20crops | Genetically modified crops are plants used in agriculture, the DNA of which has been modified using genetic engineering techniques. In most cases, the aim is to introduce a new trait to the plant which does not occur naturally in the species. As of 2015, 26 plant species have been genetically modified and approved for commercial release in at least one country. The majority of these species contain genes that make them either tolerant to herbicides or resistant to insects. Other common traits include virus resistance, delayed ripening, modified flower colour or altered composition. In 2014, 28 countries grew GM crops, and 39 countries imported but did not grow them.
Background
Regulations regarding the commercialisation of genetically modified crops are mostly conducted by individual countries. For cultivation, environmental approval determines whether a crop can be legally grown. Separate approval is generally required to use GM crops in food for human consumption or as animal feed.
GM crops were first planted commercially on a large scale in 1996, in the US, China, Argentina, Canada, Australia, and Mexico. Some countries have approved but not actually cultivated GM crops, due to public uncertainty or further government restrictions, while at the same time, they may import GM foods for consumption. For example, Japan is a leading GM food importer, and permits but has not grown GM food crops. The European Union regulates importation of GM foods, while individual member states determine cultivation. In the US, separate regulatory agencies handle approval for cultivation (USDA, EPA) and for human consumption (FDA).
Two genetically modified crops have been approved for food use in some countries, but have not obtained approval for cultivation. A GM Melon engineered for delayed senescence was approved in 1999 and a herbicide tolerant GM wheat was approved in 2004.
Genetically modified crops cultivated in 2014
In 2014, 181.5 million hectares of genetically modifie |
https://en.wikipedia.org/wiki/Complement-dependent%20cytotoxicity | Complement-dependent cytotoxicity (CDC) is an effector function of IgG and IgM antibodies. When they are bound to surface antigen on target cell (e.g. bacterial or viral infected cell), the classical complement pathway is triggered by bonding protein C1q to these antibodies, resulting in formation of a membrane attack complex (MAC) and target cell lysis.
Complement system is efficiently activated by human IgG1, IgG3 and IgM antibodies, weakly by IgG2 antibodies and it is not activated by IgG4 antibodies.
It is one mechanism of action by which therapeutic antibodies or antibody fragments can achieve an antitumor effect.
Use of CDC assays
Therapeutic antibodies
Development of antitumor therapeutic antibodies involves in vitro analysis of their effector functions including ability to trigger CDC to kill target cells. Classical approach is to incubate antibodies with target cells and source of complement (serum). Then cell death is determined with several approaches:
Radioactive method: target cells are labeled with 51Cr before CDC assay, chromium is released during cell lysis and amount of radioactivity is measured.
Measuring of the metabolic activity of live cells (live cells staining): after incubation of target cells with antibodies and complement, plasma membrane-permeable dye is added (e.g. calcein-AM or resazurin). Live cells metabolise it into impermeable fluorescent product that can be detected by flow cytometry. This product can’t be formed in metabolically inactive dead cells.
Measuring of the activity of released intracellular enzymes: dead cells release enzyme (e.g. LDH or GAPDH) and addition of its substrate leads to color change, that is usually quantified as change of absorbance or luminiscence.
Dead cells staining: a (fluorescent) dye gets inside the dead cells through their damaged plasma membrane. For instance propidium iodide binds to DNA of dead cells and fluorescent signal is measured by flow cytometry.
HLA typing and crossmatch test
CDC |
https://en.wikipedia.org/wiki/Mesentoblast | Mesentoblasts, also called 4d cells, are the cells from which the mesoderm originates. Mesentoblasts are found in the blastopore area between the endoderm and the ectoderm. In protostomes the embryos are mosaic, so mesentoblast removal will result in failure of formation of the mesoderm and other structures related to the mesoderm, which in turn will give abnormal embryos. The mesentoblast migrates to the blastocoel where it reproduces to form a mass of cells that becomes the mesoderm.
References
Developmental biology |
https://en.wikipedia.org/wiki/Offshore%20Structures%20%28Britain%29%20Ltd. | Offshore Structures (Britain) Ltd. is a marine offshore structure foundation manufacture based at Haverton Hill near Billingham on the north bank of the River Tees. The company was formed as a joint venture between Bladt Industries and EEW at the factory established by TAG Energy Solutions.
The factory was used briefly in the late 2000s by Tees Alliance Group for a prematurely terminated contract to manufacture oil rig substructures.
The works is based on the site of the Haverton Hill shipyard, initially operating as the Furness Shipbuilding Co. Ltd., established 1918. Shipbuilding at the yard ended in 1979.
History
The Haverton Hill shipyard opened in 1917 under a wartime construction programme. Shipbuilding took place at the yard from 1919 to 1969 as the Furness Shipbuilding Company under the shipping group Furness Withy. In 1968 the yard was acquired by Swan Hunter who operated the yard for a further decade, and finally by Smith's Dock Co. Ltd who built three ships at the yard in 1978. Shipbuilding ceased in 1979.
In 2008 the shipyard reopened under Tees Alliance Group (TAG) to construct the foundation structure of an oil rig 'SeaDragon 1', initially to be leased to Pemex (Mexico). In 2009 the contract between TAG and the rig development company (Sea Dragon Offshore) was terminated, with the financier (Lloyds TSB) citing financial risk, and the work transferred to a yard in Singapore with previous experience of rig construction.
TAG Energy Solutions was established in 2010. The company invested in a new production facility with machines for production of large diameter steel tubes suitable for monopoles, jacket foundations and other offshore structures. In 2013 the company was awarded a contract for monopoles and transition pieces for the Humber Gateway wind farm.
In September 2014 the company was reported entered insolvency with the loss of around 100 jobs, having run up a £61 million debt.
In late 2014 Bladt Industries (Denmark) and EEW (Germany) ac |
https://en.wikipedia.org/wiki/Bayesian-optimal%20mechanism | A Bayesian-optimal mechanism (BOM) is a mechanism in which the designer does not know the valuations of the agents for whom the mechanism is designed, but the designer knows that they are random variables and knows the probability distribution of these variables.
A typical application is a seller who wants to sell some items to potential buyers. The seller wants to price the items in a way that will maximize their profit. The optimal prices depend on the amount that each buyer is willing to pay for each item. The seller does not know these amounts, but assumes that they are drawn from a certain known probability distribution. The phrase "Bayesian optimal mechanism design" has the following meaning:
Bayesian means that we know the probability distribution from which the agents' valuations are drawn (in contrast to prior-free mechanism design, which do not assume any prior probability distribution).
Optimal means that we want to maximize the expected revenue of the auctioneer, where the expectation is over the randomness in the agents' valuations.
Mechanism means that we want to design rules that define a truthful mechanism, in which each agent has an incentive to report their true value.
Example
There is one item for sale. There are two potential buyers. The valuation of each buyer is drawn i.i.d. from the uniform distribution on [0,1].
The Vickrey auction is a truthful mechanism and its expected profit, in this case, is 1/3 (the first-price sealed-bid auction is a non-truthful mechanism and its expected profit is the same).
This auction is not optimal. It is possible to get a better profit by setting a reservation price. The Vickrey auction with a reservation price of 1/2 achieves an expected profit of 5/12, which in this case is optimal.
Notation
We assume that the agents have single-parameter utility functions, such as a single-item auction. Each agent has a value which represents the agent's "winning value" (e.g, the agent's valuation of the item). We |
https://en.wikipedia.org/wiki/Mobile%20industrial%20robots | Mobile industrial robots are pieces of machinery that are able to be programmed to perform tasks in an industrial setting. Typically these have been used in stationary and workbench applications; however, mobile industrial robots introduce a new method for lean manufacturing. With advances in controls and robotics, current technology has been improved allowing for mobile tasks such as product delivery. This additional flexibility in manufacturing can save a company time and money during the manufacturing process, and therefore results in a cheaper end product.
Mobile robot technology has potential to revolutionize many sectors of industry; however, it carries with it some disadvantages. The logistics of manufacturing will be streamlined by allowing robots to autonomously navigate to different areas for their work. The labour demands for employees will be lessened as robots will be able to work alongside humans, and robots will assist with medicine and surgery more and more. However, there are drawbacks to this technology. Coordinating the movement of robots around facilities and calibrating their position at their destination is tedious and far from perfect. A robot malfunctioning in a manufacturing setting will hold up production - and this robot could malfunction anywhere in a facility. Human safety must also be considered. Robots must prioritize the safety of human operators over their programmed task - which may complicate the coordination of multiple autonomous robots. Especially in a surgical setting, there is no room for error on the robot's part. Even though some challenges are present, mobile robot technology promises to streamline aspects across much of the industry.
History
Automation began in the automobile industry in the years surrounding WWII (1946) and the origin of the term itself belongs with D.S. Harder, the engineering manager at the Ford Motor Company. At first, the term was used to describe the increased presence of automatic devices in prod |
https://en.wikipedia.org/wiki/List%20of%20cluster%20management%20software | List of software for cluster management.
Free and open source
HA
Apache Mesos, from the Apache Software Foundation
Kubernetes, founded by Google Inc, from the Cloud Native Computing Foundation
Heartbeat, from Linux-HA
Docker Swarm
Red Hat cluster suite
OpenShift and OKD, from Red Hat
Nomad, from HashiCorp
Rancher, from Rancher Labs
TrinityX from ClusterVision Solutions
Corosync Cluster Engine
OpenSVC
K3s (“Lightweight Kubernetes“), from Rancher Labs
non-HA
oneSIS
OpenHPC
OpenSAF, founded by Motorola, from OpenSAF Foundation, implements Service Availability Forum
Rocks Cluster Distribution
Stacki, from StackIQ
YARN, distributed with Apache Hadoop
xCAT
Warewulf
Foreman
Proprietary
Amazon Elastic Container Service
Borg, used at Google
Bright Cluster Manager, from Bright Computing
CycleCloud, from Cycle Computing acquired By Microsoft
IBM Tivoli System Automation for Multiplatforms, from IBM
Microsoft Cluster Server, from Microsoft
Twine, from Facebook
HPE Performance Cluster Manager - HPCM, from Hewlett Packard Enterprise Company
Dell/EMC - Remote Cluster Manager (RCM)
Aspen Systems Inc - Aspen Cluster Management Environment (ACME)
Evidian SafeKit
IBM PowerHA system mirror
Veritas Cluster Server
See also
Comparison of cluster software
References
Cluster computing
Cluster management software |
https://en.wikipedia.org/wiki/Holland%20Computing%20Center | The Holland Computing Center, often abbreviated HCC, is the high-performance computing core for the University of Nebraska System. HCC has locations in both the University of Nebraska-Lincoln June and Paul Schorr III Center for Computer Science & Engineering and the University of Nebraska Omaha Peter Kiewit Institute.
The center was named after Omaha businessman Richard Holland who donated considerably to the university for the project.
Both locations provide various research computing services and hardware. The retrofitted facilities at the PKI location include the Crane Supercomputer which “is used by scientists and engineers to study topics such as nanoscale chemistry, subatomic physics, meteorology, crashworthiness, artificial intelligence and bioinformatics” and Anvil, the Holland Computing Center's "Cloud" based on the OpenStack Architecture. Other resources include "Rhino" for shared memory processing and "Red" for LHC grid computing.
Active resources
Crane
The Crane Supercomputer is HCC's most powerful supercomputer and is used as the primary computational resource for many researchers within the University of Nebraska system across a variety of disciplines. When it was implemented in 2013, Crane was ranked 474 in the TOP500. As of May 2019, Crane is composed of 548 nodes offering a total of 12,236 cores, 68,000 GB of memory, and 57 Nvidia GPU's. Crane has 1.5 PB of available Lustre storage (1 PB = 1 million gigabytes).
In 2017, Crane received a major upgrade, adding nodes with the Omnipath InfiniBand Architecture.
Rhino
Rhino is the latest addition to HCC's Resources, taking the place of the former Tusker super computer, using nodes from both Tusker and Sandhills. At its creation in June 2019, Rhino was composed of 112 nodes offering a total of 7,168 cores, 25,856 GB of memory. The cluster has 360 TB of Lustre storage available.
Red
Red is the resource for the University of Nebraska-Lincoln's US CMS Tier-2 site. Initially created in August 2005, th |
https://en.wikipedia.org/wiki/Babesia%20sp.%20%27North%20Carolina%20dog%27 | Babesia sp. 'North Carolina dog' is an unclassified species of Babesia, identified from a 18S ribosomal gene partial sequence performed in the investigation of dog piroplasms.
Unlike the other piroplasms investigated, the piroplasm is "in a distinct phylogenetic clade, closely related to babesial isolates from wildlife and humans from the Western US".
References
North Carolina dog
Undescribed species |
https://en.wikipedia.org/wiki/Scaled%20agile%20framework | The scaled agile framework (SAFe) is a set of organization and workflow patterns intended to guide enterprises in scaling lean and agile practices. Along with disciplined agile delivery (DAD), SAFe is one of a growing number of frameworks that seek to address the problems encountered when scaling beyond a single team.
SAFe promotes alignment, collaboration, and delivery across large numbers of agile teams. It was developed by and for practitioners, by leveraging three primary bodies of knowledge: agile software development, lean product development, and systems thinking.
The primary reference for the scaled agile framework was originally the development of a big picture view of how work flowed from product management (or other stakeholders), through governance, program, and development teams, out to customers. With the collaboration of others in the agile community, this was progressively refined and then first formally described in a 2007 book. The framework continues to be developed and shared publicly; with an academy and an accreditation scheme supporting those who seek to implement, support, or train others in the adoption of SAFe.
Starting at its first release in 2011, six major versions have been released while the latest edition, version 6.0, was released in March 2023.
While SAFe continues to be recognised as the most common approach to scaling agile practices (at 30 percent and growing),, it also has received criticism for being too hierarchical and inflexible. It also receives criticism for giving organizations the illusion of adopting Agile, while keeping familiar processes intact.
Challenges of scaling agile principles and practices
Coping with longer planning horizons
Development teams typically refine their backlog up to two to three iterations ahead, but in larger organizations the product marketing team needs to plan further ahead for their commitments to market and discussions with customers. They will often work with a very high level, 12 |
https://en.wikipedia.org/wiki/Zephyr%20%28operating%20system%29 | Zephyr () is a small real-time operating system (RTOS) for connected, resource-constrained and embedded devices (with an emphasis on microcontrollers) supporting multiple architectures and released under the Apache License 2.0. Zephyr includes a kernel, and all components and libraries, device drivers, protocol stacks, file systems, and firmware updates, needed to develop full application software.
History
Zephyr originated from Virtuoso RTOS for digital signal processors (DSPs). In 2001, Wind River Systems acquired Belgian software company Eonic Systems, the developer of Virtuoso. In November 2015, Wind River Systems renamed the operating system to Rocket, made it open-source and royalty-free. Compared to Wind River's other RTOS, VxWorks, Rocket had a much smaller memory needs, especially suitable for sensors and single-function embedded devices. Rocket could fit into as little as 4 KB of memory, while VxWorks needed 200 KB or more.
In February 2016, Rocket became a hosted collaborative project of the Linux Foundation under the name Zephyr. Wind River Systems contributed the Rocket kernel to Zephyr, but still provided Rocket to its clients, charging them for the cloud services. As a result, Rocket became "essentially the commercial version of Zephyr".
Since then, early members and supporters of Zephyr include Intel, NXP Semiconductors, Synopsys, Linaro, Texas Instruments, DeviceTone, Nordic Semiconductor, Oticon, and Bose.
, Zephyr had the largest number of contributors and commits compared to other RTOSes (including Mbed, RT-Thread, NuttX, and RIOT).
Features
Zephyr intends to provide all components needed to develop resource-constrained and embedded or microcontroller-based applications. This includes, but is not limited to:
A small kernel
A flexible configuration and build system for compile-time definition of required resources and modules
A set of protocol stacks (IPv4 and IPv6, Constrained Application Protocol (CoAP), LwM2M, MQTT, 802.15.4, Thread, Bl |
https://en.wikipedia.org/wiki/Mark%2034%20Gun%20Weapon%20System | The Mark 34 Gun Weapon System (GWS) is a component of the Aegis Combat System that is responsible for controlling and providing fire control to the 5" Mark 45 gun. It is used on the U.S. Navy Arleigh Burke-class destroyer and several later Ticonderoga-class cruisers. The Mk 34 GWS receives target data from the ship's sensors and off-ship sources, performs ballistic calculations, and produces gun control orders. The system is made up of the gun mount itself, the fire-control computer, and an optical sight.
The Mk 34 GWS was developed to improve the Arleigh Burke-class destroyer's ability to use the Mk 45 gun against a variety of threats. It is different from previous gun fire-control systems in that it was developed under a one-system concept and is a fully integrated subsystem of Aegis. The Aegis Command and Decision (C&D) system issues target engagement orders to the Mk 34 GWS.
Components
The Mark 34 Gun Weapon System consists of three components:
5" Mark 45 gun
Mark 160 Gun Computer System
Mark 46 Optical Sight System or Mark 20 Electro-Optical Sight System
5" Mark 45 gun
The 5"/54 caliber or 5"/62 caliber Mark 45 gun is fully-automated and capable of engaging surface targets in an anti-ship role, air targets in an anti-aircraft or defensive anti-missile role, and land targets in a naval gunfire support role. The gun has a range of 13 nmi (24.1 km) for 5"/54 variants or 20 nmi (37.0 km) for the 5"/62 variant. Its maximum rate of fire is 16–20 rounds per minute.
Mark 160 Gun Computer System (GCS)
The Mark 160 GCS is responsible for receiving target information from the ship's sensors and computing a ballistic solution based on the selected ammunition type. Using the solution, the Mk 160 GCS provides control and fire orders to the Mark 45 gun.
The Mark 160 Gun Computer System consists of:
Gun computer console (GCC) – Located in the ship's Combat Information Center (CIC), it is the primary interface between the Mk 34 GWS, the Aegis C&D system, and the s |
https://en.wikipedia.org/wiki/System%20U | In mathematical logic, System U and System U− are pure type systems, i.e. special forms of a typed lambda calculus with an arbitrary number of sorts, axioms and rules (or dependencies between the sorts). They were both proved inconsistent by Jean-Yves Girard in 1972. This result led to the realization that Martin-Löf's original 1971 type theory was inconsistent as it allowed the same "Type in Type" behaviour that Girard's paradox exploits.
Formal definition
System U is defined as a pure type system with
three sorts ;
two axioms ; and
five rules .
System U− is defined the same with the exception of the rule.
The sorts and are conventionally called “Type” and “Kind”, respectively; the sort doesn't have a specific name. The two axioms describe the containment of types in kinds () and kinds in (). Intuitively, the sorts describe a hierarchy in the nature of the terms.
All values have a type, such as a base type (e.g. is read as “ is a boolean”) or a (dependent) function type (e.g. is read as “ is a function from natural numbers to booleans”).
is the sort of all such types ( is read as “ is a type”). From we can build more terms, such as which is the kind of unary type-level operators (e.g. is read as “ is a function from types to types”, that is, a polymorphic type). The rules restrict how we can form new kinds.
is the sort of all such kinds ( is read as “ is a kind”). Similarly we can build related terms, according to what the rules allow.
is the sort of all such terms.
The rules govern the dependencies between the sorts: says that values may depend on values (functions), allows values to depend on types (polymorphism), allows types to depend on types (type operators), and so on.
Girard's paradox
The definitions of System U and U− allow the assignment of polymorphic kinds to generic constructors in analogy to polymorphic types of terms in classical polymorphic lambda calculi, such as System F. An example of such a generic construct |
https://en.wikipedia.org/wiki/Sam%20Labs | SAM Labs is a startup that makes app-enabled construction kits, designed for people of all ages to learn STEM, play, and create with technology and the Internet of Things. The company has been featured in WIRED and The Telegraph. Founded by Belgian born CEO Joachim Horn, the company works out of their headquarters in East London.
A month-long Kickstarter campaign initiated in October 2014 raised over $160,000, and helped launch the mass production of the kits. Backers included Jawbone co-founder Alexander Asseily, who pre-ordered the kit.
Kickstarter campaign
On 29 September 2014, SAM Labs launched a month-long Kickstarter campaign to crowdfund the mass manufacture of the SAM kits for the spring of 2015, primarily through pre-orders of the kits. At the conclusion of the campaign on 29 October, SAM Labs had raised over $160,000 from 817 backers.
Products
SAM Labs produces app-enabled construction toys.
The SAM Science Museum Inventor Kit was produced in collaboration with the London Science Museum. Using the SAM Blocks and the free SAM Space app inside of the kit, kids can build awesome projects, games, inventions and hacks. Super quick. Super smart.
The company's hero kit, SAM's Curious Cars, launched in October 2016. Available now at Barnes & Noble across the United States and John Lewis (department store) in the United Kingdom, the Curious Cars kit allows kids to build and program their own cars and games. Winning an IFA Markit Innovation Award, Consumer Electronics Show Award, and a KAPi award, the triple-award-winning kit also comes with its own app: Curious Cars.
Partnerships
SAM Labs has created a habit of working with local design talent. SAM Labs first partnered with MAP Project Office, the London-based creative consultancy that specialises in industrial design, to deliver the Kickstarter campaign. Map also featured SAM Labs in a video collaboration with Honda, where SAM led modules were featured in this promotional video, covering "Technology for Exp |
https://en.wikipedia.org/wiki/General%20Mission%20Analysis%20Tool | General Mission Analysis Tool (GMAT) is open-source space mission analysis software developed by NASA and private industry.
It has been used for several missions, including LCROSS, the Lunar Reconnaissance Orbiter, OSIRIS-REx, the Magnetospheric Multiscale Mission, and the Transiting Exoplanet Survey Satellite (TESS) mission.
GMAT is an open-source alternative to software like Systems Tool Kit and FreeFlyer.
References
External links
GMAT Wiki
GMAT Download (SourceForge)
GMAT channel on YouTube
Aerospace engineering
3D graphics software
Astronomy software
Mathematical software
Physics software |
https://en.wikipedia.org/wiki/Population%20proportion | In statistics, a population proportion, generally denoted by or the Greek letter , is a parameter that describes a percentage value associated with a population. For example, the 2010 United States Census showed that 83.7% of the American population was identified as not being Hispanic or Latino; the value of .837 is a population proportion. In general, the population proportion and other population parameters are unknown. A census can be conducted in order to determine the actual value of a population parameter, but often a census is not practical due to its costs and time consumption.
A population proportion is usually estimated through an unbiased sample statistic obtained from an observational study or experiment. For example, the National Technological Literacy Conference conducted a national survey of 2,000 adults to determine the percentage of adults who are economically illiterate. The study showed that 72% of the 2,000 adults sampled did not understand what a gross domestic product is. The value of 72% is a sample proportion. The sample proportion is generally denoted by and in some textbooks by .
Mathematical definition
A proportion is mathematically defined as being the ratio of the quantity of elements (a countable quantity) in a subset to the size of a set :
where is the count of successes in the population, and is the size of the population.
This mathematical definition can be generalized to provide the definition for the sample proportion:
where is the count of successes in the sample, and is the size of the sample obtained from the population.
Estimation
One of the main focuses of study in inferential statistics is determining the "true" value of a parameter. Generally, the actual value for a parameter will never be found, unless a census is conducted on the population of study. However, there are statistical methods that can be used to get a reasonable estimation for a parameter. These methods include confidence intervals and hypoth |
https://en.wikipedia.org/wiki/PiliPinas%20Debates%202016 | PiliPinas Debates 2016 is a debate series administered by the Commission on Elections (COMELEC) with the assistance of the Kapisanan ng mga Brodkaster ng Pilipinas in preparation for the May 9, 2016, general election. In January 2016, the Commission on Elections confirmed that they would hold three presidential debates and one vice presidential debate. This will be the first time that the COMELEC will host debates since the 1992 elections, with ABC-5 as the host network fresh from the launch of the recently restored broadcast network on February 21 of the same year, after a 20-year rest due to Martial Law. The debates were branded as PiliPinas Debates 2016. The first word is a portmanteau of Pili, Filipino for "choose" and Pinas, shortcut for the Philippines in Filipino.
Broadcast and coverage
The first leg of the debate was conducted on Mindanao and was held on February 21 at the Mini-Theater Building of the Capitol University in Cagayan de Oro. It aired on GMA Network and was simulcast on Super Radyo DZBB, RGMA stations and KBP-affiliated member radio stations nationwide (GMA was a KBP member from April 1973 until the said network withdrew from its membership in September 2003) and its livestream channels. The Mindanao leg was moderated by Mike Enriquez and Jessica Soho of GMA Network and John Nery of Philippine Daily Inquirer.
The second leg was conducted in Visayas and was held at the Performing Arts Hall of the University of the Philippines Cebu on March 20, 2016. It was aired simultaneously over TV5, AksyonTV, Bloomberg TV Philippines, Radyo5 92.3 News FM and KBP-affiliated member radio stations nationwide with live streaming broadcast also made available via news5.com.ph, bilangpilipino.com and YouTube. It was moderated by News5 chief Luchi Cruz-Valdez.
CNN Philippines hosted the Vice Presidential Debate on April 10, 2016. It was moderated by Pia Hontiveros and Pinky Webb. The debate was held at the Quadricentennial Pavillion of the University of Santo To |
https://en.wikipedia.org/wiki/Semasiography | Semasiography (from (semasia) "signification, meaning" and (graphia) "writing") is "writing with signs", a non-phonetic based technique to "communicate information without the necessary intercession of forms of speech." It means written symbols and languages that are not based on spoken words. It predated the advent of the creation of the language-based writing system and is used contemporarily in computer icons, musical notation, emoji, Blissymbols and mathematical notation. It is studied in semasiology within the field of linguistics.
References |
https://en.wikipedia.org/wiki/Jones%E2%80%93Dole%20equation | The Jones–Dole equation, or Jones–Dole expression, is an empirical expression that describes the relationship between the viscosity of a solution and the concentration of solute within the solution (at a fixed temperature and pressure). The Jones–Dole equation is written as
where
η is the viscosity of the solution (at a fixed temperature and pressure),
η0 is the viscosity of the solvent at the same temperature and pressure,
A is a coefficient that describes the impact of charge–charge interactions on the viscosity of a solution (it is usually positive) and can be calculated from Debye–Hückel theory,
B is a coefficient that characterises the solute–solvent interactions at a defined temperature and pressure,
C is the solute concentration.
The Jones–Dole B coefficient is often used to classify ions as either structure-makers (kosmotropes) or structure-breakers (chaotropes) according to their supposed strengthening or weakening of the hydrogen-bond network of water. The Jones–Dole expression works well up to about 1 M, but at higher concentrations breaks down, as the viscosity of all solutions increase rapidly at high concentrations.
The large increase in viscosity as a function of solute concentration seen in all solutions above about 1 M is the effect of a jamming transition at a high concentration. As a result, the viscosity increases exponentially as a function of concentration and then diverges at a critical concentration. This has been referred to as the "Mayonnaise effect", as the viscosity of mayonnaise (essentially a solution of oil in water) is extremely high because of the jamming of micrometer-scale droplets.
References
Equations |
https://en.wikipedia.org/wiki/Amazon%20Lumberyard | Amazon Lumberyard is a now-superseded freeware cross-platform game engine developed by Amazon and based on CryEngine (initially released in 2002), which was licensed from Crytek in 2015.
In July 2021, Amazon and the Linux Foundation announced that parts of the engine would be used to create a new open source game engine called Open 3D Engine, which would replace it. A new Open 3D Foundation, run by the Linux Foundation, will manage the new engine, which will be licensed under the open source Apache 2.0 license. The new engine is reportedly partially based on Lumberyard but with many parts rewritten, and is considered a new engine.
Overview
The Lumberyard engine features integration with Amazon Web Services to allow developers to build or host their games on Amazon's servers, as well as support for livestreaming via Twitch.
Additionally, the engine includes Twitch ChatPlay, allowing viewers of the Twitch stream to influence the game through the associated chat, a method of play inspired by the Twitch Plays Pokémon phenomenon.
The source code is available to end users with limitations: Users may not publicly release the Lumberyard engine source code or use it to release their own game engine.
Lumberyard launched on February 9, 2016 alongside GameLift, a fee-based managed service for deploying and hosting multiplayer games, intended to allow developers the easy development of games that attract "large and vibrant communities of fans."
, the software is currently in beta status and can be used to build games for Microsoft Windows, PlayStation 4, and Xbox One, with limited support for iOS and Android and support for macOS being planned for future releases. Virtual reality integration was added in Beta 1.3, allowing developers to build games supporting devices like Oculus Rift and HTC Vive.
Despite being based on the architecture of Crytek's CryEngine, the engine has been developed to use many of its own custom-developed systems, some of which are in a preview mode. A |
https://en.wikipedia.org/wiki/Misuse%20of%20p-values | Misuse of p-values is common in scientific research and scientific education. p-values are often used or interpreted incorrectly; the American Statistical Association states that p-values can indicate how incompatible the data are with a specified statistical model. From a Neyman–Pearson hypothesis testing approach to statistical inferences, the data obtained by comparing the p-value to a significance level will yield one of two results: either the null hypothesis is rejected (which however does not prove that the null hypothesis is false), or the null hypothesis cannot be rejected at that significance level (which however does not prove that the null hypothesis is true). From a Fisherian statistical testing approach to statistical inferences, a low p-value means either that the null hypothesis is true and a highly improbable event has occurred or that the null hypothesis is false.
Clarifications about p-values
The following list clarifies some issues that are commonly misunderstood regarding p-values:
The p-value is not the probability that the null hypothesis is true, or the probability that the alternative hypothesis is false. A p-value can indicate the degree of compatibility between a dataset and a particular hypothetical explanation (such as a null hypothesis). Specifically, the p-value can be taken as the probability of obtaining an effect that is at least as extreme as the observed effect, given that the null hypothesis is true. This should not be confused with the probability that the null hypothesis is true given the observed effect (see prosecutor's fallacy). In fact, frequentist statistics does not attach probabilities to hypotheses.
The p-value is not the probability that the observed effects were produced by random chance alone. The p-value is computed under the assumption that a certain model, usually the null hypothesis, is true. This means that the p-value is a statement about the relation of the data to that hypothesis.
The 0.05 significance lev |
https://en.wikipedia.org/wiki/Eddy-current%20sensor | Eddy current sensors uses the principle of eddy current formation to sense displacement. These sensors measure shaft displacement in rotating machinery and have been around for many years as they offer manufacturers high-linearity, high-speed measurements, and high resolution.
Eddy currents are formed when a moving or changing magnetic field intersects a conductor or vice versa. The relative motion causes a circulating flow of electrons, or currents, within the conductor. These circulating eddies of current create electromagnets with magnet fields that oppose the effect of applied magnetic field. The stronger the applied magnetic field, or greater the electrical conductivity of the conductor, or greater the relative velocity of motion, the greater the currents developed and greater the opposing field. Eddy current probes senses this formation of secondary fields to find out the distance between the probe and target material.
References
Position sensors |
https://en.wikipedia.org/wiki/Talehunt | Talehunt is an online community that allows people to write, read, and post very short stories. It is available as a mobile app (the TaleHunt App) on both Android and iPhone mobile platforms. In Talehunt, any user can create an account and post stories, fan fiction, and poems, which are under 250 characters in length. This allows anyone to write very short "tales" (also known as Stories) and build a fanbase of followers, giving people the chance to have their creative works available to a wider audience. Around 55% of Talehunt users are from the United States while the remainder are primarily from Europe and Asia.
History
The early beta version of Talehunt was available by the end of 2015. The app was officially released January 2016.
Design
The app has a minimalist design which is simple and is reflected in its logo: a quill and a scribble on a dark green background. The dark green theme is intended to make the TaleHunt App user-friendly and pleasant to read. Once in the app, even before logging in, the user can immediately read other users' short stories and follow the writer if the user liked the tale. Once followed, all stories will show up in the personal timeline, which is customized for each user based on the people he or she follows. Each author also has their own page, which includes their username, profile photo, follow button, number of followers, number of users followed, and a feed of all of the author's stories.
See also
Flash fiction
References
Ebook suppliers
Electronic paper technology
Android (operating system) software
IOS software
Ebook sources
Online publishing companies of the United States
Self-publishing companies
Short stories
2016 software |
https://en.wikipedia.org/wiki/The%20End%20of%20Average | The End of Average: How We Succeed in a World That Values Sameness is a book by Todd Rose. It was published by HarperCollins in 2016, and talks about the importance of individuality rather than the concept of average human beings.
In this book, the author argues that no individual can be accurately labeled as average. He presents an alternative to understanding individuals solely based on averages. He introduces three principles of individuality - the jaggedness principle (talent is always uneven), the context principle (traits are not fixed), and the pathways principle (we often take unconventional routes).
According to The New York Times, “Readers will be moved to examine their own averagerian prejudices, most so ingrained as to be almost invisible, all worthy of review.” As per Kirkus Reviews, the book is “an intriguing view into the evolution and imperfections of our current system but lacks a clear path toward implementing the proposed principles of individuality.”
References
Further reading
External links
2016 non-fiction books
American non-fiction books
HarperCollins books
Psychology books
Statistics books |
https://en.wikipedia.org/wiki/List%20of%20Chromista%20by%20conservation%20status | , the International Union for Conservation of Nature (IUCN) has evaluated the conservation status of 15 species within Chromista.
The IUCN has not evaluated any protist species other than those in Chromista. No Chromista subspecies or subpopulations have been evaluated. No evaluated Chromista species are confirmed to be extinct, but four are tagged as possibly extinct.
the New Zealand Threat Classification System has evaluated 38 species of macroalgae as Threatened and 23 as Data Deficient. Some of these species are only of concern nationally.
IUCN evaluations
Critically Endangered (possibly extinct)
Bifurcaria galapagensis - Galapagos stringweed
Desmarestia tropica - tropical acidweed
Dictyota galapagensis
Spatoglossum schmittii
Endangered
Sargassum setifolium - string sargassum
Vulnerable
Eisenia galapagensis - Galapagos kelp
Data Deficient
All are brown algae (Phaeophyceae):
Of the Dictyotaceae,
Dictyopteris diaphana
Dictyota major
Padina concrescens
Spatoglossum ecuadoreanum
Of the Chordariaceae,
Zosterocarpus abyssicola
Of the Sargassaceae,
Sargassum albemarlense
Sargassum galapagense
Sargassum templetonii
Of the Sporochnaceae,
Sporochnus rostratus
NZTCS evaluations
Nationally Critical
Dione arcuata - One Location
Range Restricted
Caulerpa racemosa - Secure Overseas
Caulerpa sertulariodes - Secure Overseas, One Location
Caulerpa webbiana - Secure Overseas, One Location
Chlidophyllon kaspar - One Location
Chordariopsis capensis - Data Poor, Secure Overseas
Chrysymenia? polydactyla - Data Poor, One Location
Codium geppiorum - Secure Overseas, One Location
Curdiea balthazar - One Location
Dichotomaria marginata - Secure Overseas, One Location
Durvillaea sp. Antipodes WELT A17080 - One Location
Durvillaea chathamensis - One Location
Galaxaura cohaerens - One Location
Galaxaura filamentosa - One Location
Galaxaura rugosa - One Location
Ganonema farinosa - One Location
Gelidium allanii - Threatened Overseas
“Gelidium” ceramoides
Gelid |
https://en.wikipedia.org/wiki/ESIM | An eSIM (embedded-SIM) is a form of SIM card that is embedded directly into a device. Instead of an integrated circuit located on a removable SIM card, typically made of PVC, an eSIM consists of software installed onto an eUICC chip permanently attached to a device. If the eSIM is eUICC-compatible, it can be re-programmed with new SIM information. Otherwise, the eSIM is programmed with its ICCID/IMSI and other information at the time it is manufactured, and cannot be changed.
Once an eSIM carrier profile has been installed on an eUICC, it operates the same as a physical SIM, complete with a unique ICCID and network authentication key generated by the carrier.
The eSIM standard was first released in 2016; since that point, eSIM has begun to replace physical SIM in domains including cellular telephony.
History
Since 2010, the GSMA had been discussing the possibility of a software-based SIM.
While Motorola noted that eUICC is geared at industrial devices, Apple "disagreed that there is any statement forbidding the use of an embedded UICC in a consumer product". Currently, the GSMA maintains two different versions of the standard: one for consumer devices and another for machine to machine (M2M) devices.
A first version of the standard was published in March 2016, followed by a second version in November 2016.
Implementation
In 2016, the Samsung Gear S2 Classic 3G smartwatch was the first device to implement an eSIM.
In 2017, during Mobile World Congress, Qualcomm introduced a technical solution, with a live demonstration, within its Snapdragon hardware chip associated with related software (secured Java applications).
Apple first introduced eSIM support in September 2017 with the Apple Watch Series 3. The first iPhone models to support it were the iPhone XS, iPhone XS Max, and iPhone XR, announced in September 2018. The first iPad model was the iPad Pro (3rd generation), announced in October 2018. In September 2022, Apple unveiled the iPhone 14, iPhone 14 P |
https://en.wikipedia.org/wiki/Phylogenetic%20inertia | Phylogenetic inertia or phylogenetic constraint refers to the limitations on the future evolutionary pathways that have been imposed by previous adaptations.
Charles Darwin first recognized this phenomenon, though the term was later coined by Huber in 1939. Darwin explained the idea of phylogenetic inertia based on his observations; he spoke about it when explaining the "Law of Conditions of Existence". Darwin also suggested that, after speciation, the organisms do not start over from scratch, but have characteristics that are built upon already existing ones that were inherited from their ancestors; and these characteristics likely limit the amount of evolution seen in that new taxa. This is the main concept of phylogenetic inertia.
Richard Dawkins also explained these constraints by likening natural selection to a river in his 1982 book The Extended Phenotype.
Examples of phylogenetic inertia
Body plan
Most terrestrial vertebrates have a body plan that consist of four limbs. The phylogenetic inertia hypothesis suggests that this body plan is observed, not because it happens to be optimal, but because tetrapods are derived from a clade of fishes (Sarcopterygii) which also have four limbs. Four limbs happened to be suitable means of locomotion, and so this body plan has not been selected against.
Humans do not have optimal structure for bipedalism, because much of the human body plan originally evolved under quadrupedal locomotion, and has since been constrained because of phylogenetic inertia.
Modes of reproduction
Birds are the only speciose group of vertebrates that are exclusively oviparous, or egg laying. It has been suggested that birds are phylogenetically constrained, as being derived from reptiles, and likely have not overcome this constraint or diverged far enough away to develop viviparity, or live birth.
Homologous structures
More specifically than the similarity in body plan, there are homologous bones across mammalian taxa. For example, t |
https://en.wikipedia.org/wiki/Whittle%20likelihood | In statistics, Whittle likelihood is an approximation to the likelihood function of a stationary Gaussian time series. It is named after the mathematician and statistician Peter Whittle, who introduced it in his PhD thesis in 1951.
It is commonly used in time series analysis and signal processing for parameter estimation and signal detection.
Context
In a stationary Gaussian time series model, the likelihood function is (as usual in Gaussian models) a function of the associated mean and covariance parameters. With a large number () of observations, the () covariance matrix may become very large, making computations very costly in practice. However, due to stationarity, the covariance matrix has a rather simple structure, and by using an approximation, computations may be simplified considerably (from to ). The idea effectively boils down to assuming a heteroscedastic zero-mean Gaussian model in Fourier domain; the model formulation is based on the time series' discrete Fourier transform and its power spectral density.
Definition
Let be a stationary Gaussian time series with (one-sided) power spectral density , where is even and samples are taken at constant sampling intervals .
Let be the (complex-valued) discrete Fourier transform (DFT) of the time series. Then for the Whittle likelihood one effectively assumes independent zero-mean Gaussian distributions for all with variances for the real and imaginary parts given by
where is the th Fourier frequency. This approximate model immediately leads to the (logarithmic) likelihood function
where denotes the absolute value with .
Special case of a known noise spectrum
In case the noise spectrum is assumed a-priori known, and noise properties are not to be inferred from the data, the likelihood function may be simplified further by ignoring constant terms, leading to the sum-of-squares expression
This expression also is the basis for the common matched filter.
Accuracy of approximation
The Whittle likelih |
https://en.wikipedia.org/wiki/Behavioral%20plasticity | Behavioral plasticity refers to a change in an organism's behavior that results from exposure to stimuli, such as changing environmental conditions. Behavior can change more rapidly in response to changes in internal or external stimuli than is the case for most morphological traits and many physiological traits. As a result, when organisms are confronted by new conditions, behavioral changes often occur in advance of physiological or morphological changes. For instance, larval amphibians changed their antipredator behavior within an hour after a change in cues from predators, but morphological changes in body and tail shape in response to the same cues required a week to complete.
Background
For many years, ethologists have studied the ways that behavior can change in response to changes in external stimuli or changes in the internal state of an organism. In a parallel literature, psychologists studying learning and cognition have spent years documenting the many ways that experiences in the past can affect the behavior an individual expresses at the current time. Interest in behavioral plasticity gained prominence more recently as an example of a type of phenotypic plasticity with major consequences for evolutionary biology.
Types
Behavioral plasticity can be broadly organized into two types: exogenous and endogenous. Exogenous plasticity refers to the changes in behavioral phenotype (i.e., observable behaviors) caused by an external stimulus, experience, or environment. Endogenous plasticity encompasses plastic responses that result from changes in internal cues, such as genotype, circadian rhythms, and menstruation.
These two broad categories can be further broken down into two other important classifications. When an external stimulus elicits or "activates" an immediate response (an immediate effect on behavior), then the organism is demonstrating contextual plasticity. This form of plasticity highlights the concept that external stimuli in a given context |
https://en.wikipedia.org/wiki/Duplex%20sequencing | Duplex sequencing is a library preparation and analysis method for next-generation sequencing (NGS) platforms that employs random tagging of double-stranded DNA to detect mutations with higher accuracy and lower error rates.
This method uses degenerate molecular tags in addition to sequencing adapters to recognize reads originating from each strand of DNA. The generated sequencing reads then will be analyzed using two methods: single-strand consensus sequences (SSCS) and duplex consensus sequences (DCS) assembly. Duplex sequencing theoretically can detect mutations with frequencies as low as 5 x 10−8 --that is more than 10,000 times higher in accuracy compared to the conventional next-generation sequencing methods.
The estimated error rate of standard next-generation sequencing platforms is 10−2 to 10−3 per base call. With this error rate, billions of base calls that are produced by NGS will result in millions of errors. The errors are introduced during sample preparation and sequencing such as polymerase chain reaction, sequencing, and image analysis errors. While the NGS platforms' error rate is acceptable in some applications such as detection of clonal variants, it is a major limitation for applications that require higher accuracy for detection of low-frequency variants such as detection of intra-organismal mosaicism, subclonal variants in genetically heterogeneous cancers, or circulating tumor DNA.
Several library preparation strategies have been developed that increase accuracy of NGS platforms such as molecular barcoding and circular consensus sequencing method. Like NGS platforms, the data generated by these methods originates from a single strand of DNA, and therefore the errors that are introduced during PCR amplification, tissue processing, DNA extraction, hybridization capture (where used) or DNA sequencing itself can still be distinguished as a true variant. The duplex sequencing method addresses this problem by taking advantage of the complementary |
https://en.wikipedia.org/wiki/Nexi | Nexi S.p.A. formerly known as Istituto Centrale delle Banche Popolari Italiane S.p.A. (ICBPI) is an Italian bank that specialises in payment systems such as Nexi Payments (formerly known as CartaSi). The bank was specialised as a central institution of Italian Popular Bank. ().
History
Istituto Centrale delle Banche Popolari Italiane (ICBPI) was founded by Banca Popolare di Cremona, Intra, Lecco, Lodi, Luino e Varese and Verona in 1939.
In 2006, it acquired Key Client Cards & Solutions, a spin-off of Deutsche Bank. Since January 28, 2008, the Automated Clearing House – ACH Sepa Compliant with ICCREA Banca has been activated. In April 2008, it incorporated the Centralized Services Joint Stock Company – Seceti. On 15 September, 2008, in Vienna ICBPI signed an agreement with Equens SE for the establishment of the Equens Italia joint venture, which will carry out the Automated Clearing House activity in Italy as insourcer of ICBPI and ICCREA.
At the end of 2008, the process of acquiring the majority of S.I. Holding (which controls 100% CartaSi S.p.A., Si Servizi S.p.A., Si Call S.p.A., Carta Facile S.p.A, CartaSi Capital S.p.A. and SiRe Business Services LTD), concluding the acquisition, following the obtaining of the required authorizations from the Italian Competition Authority, in particular the Bank of Italy and the Antitrust Authority, in June 2009.
In September 2010, the formalization of an agreement for the purchase of the custodian bank business from Banca Carige was announced for 19.5 million euro. In November 2010, the acquisition of the custodian bank activity from Banca Sella was announced, which at the same time became a shareholder of ICBPI with a 0.96% stake. In November 2010, the ICBPI was condemned by the Antitrust to pay a fine of 490,000 euros for agreements restricting competition in relation to the credit card sector, which was subsequently suspended by the Lazio TAR.
In 2015, the bank was acquired by a consortium of Bain Capital, Advent Intern |
https://en.wikipedia.org/wiki/Bare-metal%20server | In computer networking, a bare-metal server is a physical computer server that is used by one consumer, or tenant, only. Each server offered for rental is a distinct physical piece of hardware that is a functional server on its own. They are not virtual servers running in multiple pieces of shared hardware.
The term is used for distinguishing between servers that can host multiple tenants and which use virtualisation and cloud hosting. Unlike bare-metal servers, cloud servers are shared between multiple tenants. Each bare-metal server may run any amount of work for a user, or have multiple simultaneous users, but they are dedicated entirely to the entity who is renting them.
Bare-metal advocacy
Hypervisors provide some isolation between tenants but there can still be a noisy neighbour effect. If a physical server is multi-tenanted, peaks of load from one tenant may consume enough machine resources to temporarily affect other tenants. As the tenants are otherwise isolated, it is also hard to manage or load balance this. Bare-metal servers, and single tenancy, can avoid this. In addition, hypervisors provide weaker isolation and are much more risky from a security point-of-view compared to using separate machines. Attackers have always found vulnerabilities in the isolation software (such as hypervisors), covert channels are impractical to counter without physically separate machines, and shared hardware is vulnerable to defects in hardware protection mechanisms such as Rowhammer, Spectre, and Meltdown. As, once again, server costs are dropping as a proportion of total cost of ownership against their administration overhead, the classic solution of 'throwing hardware at the problem' becomes viable again.
Bare-metal cloud hosting
Infrastructure as a service, particularly through infrastructure as code, offers many advantages to make hosting conveniently manageable. Combining the features of both cloud hosting, and bare-metal servers, offers most of these, whil |
https://en.wikipedia.org/wiki/Predictive%20microbiology | Predictive Microbiology is the area of food microbiology where controlling factors in foods and responses of pathogenic and spoilage microorganisms are quantified and modelled by mathematical equations
It is based on the thesis that microorganisms' growth and environment are reproducible, and can be modeled. Temperature, pH and water activity impact bacterial behavior. These factors can be changed to control food spoilage.
Models can be used to predict pathogen growth in foods. Models are developed in several steps including design, development, validation, and production of an interface to display results. Models can be classified attending to their objective in primary models (describing bacterial growth), secondary models (describing factors affecting bacterial growth) or tertiary models (computer software programs)
References
Microbiology
Food safety |
https://en.wikipedia.org/wiki/Paktor | Paktor is a location-based dating and networking mobile application that connects mutually interested users and allows them to chat individually or as a group. Founded in Singapore, the app was launched in June 2013, and has over 5 billion swipes to date (October 2015). In late 2013, Paktor launched a subsidiary called GaiGai, an offline dating service that focuses on match-making and dating events. The word 'Paktor' originates from the Cantonese word for 'dating' (Chinese character 拍拖).
History
Founding
Paktor was founded by Joseph Phua and Ng Jing Shen.
Paktor was first launched in Singapore in June 2013. Despite the immature market, there was a lot of interest from consumers and according to a Paktor newsletter, there were 250000 matches in the first 2 weeks of 2014. The company also shared that it had received over 100 million profile ratings at that time.
The app is also available in Taiwan, Malaysia, Indonesia, Vietnam, Thailand, and South Korea.
In February 2017, Paktor acquired dating apps DOWN, Goodnight, and Groopify, folding them into Paktor Labs, a social app accelerator.
Change of Management
In November 2017, Paktor Group appointed co-founder and chief technology officer Ng Jing Shen as CEO of Paktor Group, taking over from current CEO Joseph Phua. Phua remained in the group CEO of M17 Entertainment which was formed after the merger of Paktor and 17 Media. The company also promoted current head of operations Shn Juay to Chief Operating Officer at Paktor Group.
In 2018, M17 Entertainment sold DOWN, the United States-based dating application, back to its founder Colin Hodge.
Company Overview
Financials
In March 2014, Paktor revealed its first round of funding with a US$500,000 pre-Series-A round that valued the brand at US$6.5 million. At this time, co-founder and CEO Joseph Phua was quoted as saying that the app was “easily 10-20 times ahead of the next player” referring to their main markets; Singapore, Malaysia, Vietnam, Thailand, Taiwan a |
https://en.wikipedia.org/wiki/Acorn%20TV | Acorn TV is a British-American subscription video streaming service offering television programming from the United Kingdom, Canada, Australia, New Zealand, other Commonwealth countries, Spain, and Mexico. In other countries, it is available on a variety of devices including Amazon Fire TV, Apple TV, Android TV, Chromecast, and Roku.
History
Acorn TV is owned by RLJ Entertainment, Inc. Acorn Media Group has distributed British television in the United States since 1994, originally selling VHS tapes before moving into DVD and Blu-ray media. Continuing the company's expansion into new formats, Acorn TV launched as a subsection of Acorn's direct-to-consumer e-commerce website in 2011. In 2013, Acorn TV was relaunched as a standalone service with expanded content offerings and monthly and annual subscription options. In 2013, the service began offering exclusive content, starting with the United States premiere of Doc Martin, Series 6. In 2015, Acorn TV was the only niche streaming service to have a program nominated for an Emmy when Curtain: Poirot's Last Case was nominated for Outstanding Television Movie. As of December 31, 2016, it had 430,000 paid subscribers.
Acorn TV launched in the United Kingdom as a service in its own right on April 29, 2020.
On 24 November 2022, Acorn TV announced without further elaboration that it would no longer be available in South Africa by the end of 2022, and requested that subscribers cancel their memberships. They previously launched in South Africa from December 2018.
Programming
Acorn TV offers a combination of new and classic mysteries, dramas, comedies, and documentaries. The service licenses content from producers and distributors including ITV, Channel 4, BBC Studios, All3Media, DRG, ZDF, and Content Media.
Original programming
Because its parent company, RLJ Entertainment, has a 64% stake in Agatha Christie Limited, the licensing arm of the Christie estate, Acorn TV was able to offer the United States premieres of the fi |
https://en.wikipedia.org/wiki/MinIO | MinIO is a High-Performance Object Storage released under GNU Affero General Public License v3.0. It is API compatible with the Amazon S3 cloud storage service. It can handle unstructured data such as photos, videos, log files, backups, and container images with a current maximum supported object size of 50TB.
History & development
MinIO's main developer is MinIO Inc, a Silicon Valleybased technology startup founded by Anand Babu Periasamy, Garima Kapoor, and Harshavardhana in November 2014.
MinIO has published a number of benchmarks to disclose both its own performance and the performance of an object storage in general. These benchmarks include comparisons to an Amazon S3 for Trino, Presto, and Spark, as well as throughput results for the S3Benchmark on HDD and NVMe drives.
Re-licensing
As of April 23, 2021, MinIO, Inc submitted a change that re-licensed the project from its previous Apache V2 to GNU Affero Public License Version 3 (AGPLv3)..
Architecture
MinIO storage stack has three major components: MinIO Server, MinIO Client (a.k.a. mc, which is a command-line client for the object and file management with any Amazon S3 compatible servers), and MinIO Client SDK, which can be used by application developers to interact with any Amazon S3 compatible server.
MinIO Server
MinIO cloud storage server is designed to be minimal and scalable. It is light enough to be bundled along with the application stack, similar to NodeJS and Redis.
MinIO is optimized for large enterprise deployments, including features like erasure coding, bitrot protection, encryption/WORM, identity management, continuous replication, global federation, and multi-cloud deployments via gateway mode.
MinIO server is hardware agnostic and thus can be installed both on physical and virtual machines or launched as Docker containers and deployed on container orchestration platforms like Kubernetes.
MinIO Client
MinIO Client provides an alternative to the standard UNIX commands (e.g. ls, cat |
https://en.wikipedia.org/wiki/Resource%20consumption | Resource consumption is about the consumption of non-renewable, or less often, renewable resources. Specifically, it may refer to:
water consumption
energy consumption
electric energy consumption
world energy consumption
natural gas consumption/gas depletion
oil consumption/oil depletion
logging/deforestation
fishing/overfishing
land use/land loss or
resource depletion and
general exploitation and associated environmental degradation
Measures of resource consumption are resource intensity and resource efficiency. Industrialization and globalized markets have increased the tendency for overconsumption of resources. The resource consumption rate of a nation does not usually correspond with the primary resource availability, this is called resource curse.
Unsustainable consumption by the steadily growing human population may lead to resource depletion and a shrinking of the earth's carrying capacity.
See also
Ecological footprint
Jevons paradox
Natural resource management
Scarcity
Uneconomic growth
References
External links
Ecology
Resource economics
Consumption |
https://en.wikipedia.org/wiki/Apple%E2%80%93FBI%20encryption%20dispute | The Apple–FBI encryption dispute concerns whether and to what extent courts in the United States can compel manufacturers to assist in unlocking cell phones whose data are cryptographically protected. There is much debate over public access to strong encryption.
In 2015 and 2016, Apple Inc. received and objected to or challenged at least 11 orders issued by United States district courts under the All Writs Act of 1789. Most of these seek to compel Apple "to use its existing capabilities to extract data like contacts, photos and calls from locked iPhones running on operating systems iOS 7 and older" in order to assist in criminal investigations and prosecutions. A few requests, however, involve phones with more extensive security protections, which Apple has no current ability to break. These orders would compel Apple to write new software that would let the government bypass these devices' security and unlock the phones.
The most well-known instance of the latter category was a February 2016 court case in the United States District Court for the Central District of California. The Federal Bureau of Investigation (FBI) wanted Apple to create and electronically sign new software that would enable the FBI to unlock a work-issued iPhone 5C it recovered from one of the shooters who, in a December 2015 terrorist attack in San Bernardino, California, killed 14 people and injured 22. The two attackers later died in a shootout with police, having first destroyed their personal phones. The work phone was recovered intact but was locked with a four-digit password and was set to eliminate all its data after ten failed password attempts (a common anti-theft measure on smartphones). Apple declined to create the software, and a hearing was scheduled for March 22. However, a day before the hearing was supposed to happen, the government obtained a delay, saying it had found a third party able to assist in unlocking the iPhone. On March 28, the government announced that the FBI had |
https://en.wikipedia.org/wiki/Mini-STX | Mini-STX (mSTX, Mini Socket Technology EXtended, originally "Intel 5x5") is a computer motherboard form factor that was released by Intel in 2015 (as "Intel 5x5").
These motherboards measure 147mm by 140mm (5.8" x 5.5"), making them larger than "4x4" NUC (102x102mm / 4.01" x 4.01" inches) and Nano-ITX (120x120mm / 4.7" x 4.7") boards, but notably smaller than the more common Mini-ITX (170x170mm / 6.7" x 6.7") boards. Unlike these standards, which use a square shape, the Mini-STX form factor is 7mm longer from front-to-rear, making it slightly rectangular.
Mini-STX design elements
The Mini-STX design suggests (but does not require) support for:
Socketed processors (e.g. LGA or PGA CPUs)
Onboard power regulation circuitry, enabling direct DC power input
IO ports embedded on the front and rear of the motherboard (akin to NUC, but unlike typical motherboards which often use headers instead to connect built-in ports on enclosures)
Adoption by manufacturers
This motherboard form factor is still not in particularly common use with consumer-PC manufacturers, although there are a few offerings:
ASRock offers both DeskMini kits (that use mini-STX boards) and standalone motherboards,
Asus offer VivoMini kits (that use mini-STX boards) and standalone motherboards,
Gigabyte offers a few motherboards, and
industrial PC suppliers (e.g. Kontron, Iesy, ASRock Industrial) also provide some options for mini-STX equipment.
Derivatives
ASRock developed a derivative of mini-STX, dubbed micro-STX, for their 'DeskMini GTX/RX' small form-factor PCs and industrial motherboards.
Micro-STX adds an MXM slot which allows the use of special PCI Express expansion cards, including graphics or machine learning accelerators, but increases the width of the board to be extended two inches, resulting in measurements of 147 x 188 mm (5.8" x 7.4")
References
http://www.pcworld.com/article/2975764/components/intel-unveils-ultra-tiny-fully-upgradable-5x5-mini-pc-motherboards.html
|
https://en.wikipedia.org/wiki/Synthetic%20immunology | Synthetic immunology is the rational design and construction of synthetic systems that perform complex immunological functions. Functions include using specific cell markers to target cells for destruction and or interfering with immune reactions. US Food and Drug Administration (FDA)-approved immune system modulators include anti-inflammatory and immunosuppressive agents, vaccines, therapeutic antibodies and Toll-like receptor (TLR) agonists.
History
The discipline emerged after 2010 following the development of genome editing technology including TALENS and CRISPR. In 2015, one project created T cells that became active only in the presence of a specific drug, allowing them to be turned on and off in situ. Another example is a T cell that targets only cells that display two separate markers.
In 2016, John Lin head of Pfizer's San Francisco biotech unit stated, “the immune system will be the most convenient vehicle for [engineered human cells], because they can move and migrate and play such important roles.”
Advances in systems biology support high-dimensional quantitative analysis of immune responses. Techniques include viral gene delivery, inducible gene expression, RNA-guided genome editing, and site-specific recombinases for applications related to biotechnology and cellular immunotherapy.
Types
Immunity-modulating organisms
Researchers are exploring the creation of 'smart' organisms such as bacteriophages and bacteria that can perform complex immunological tasks. Such strategies could produce organisms that perform multistep immune functions such as presenting antigen to and co-stimulating helper T cells in a specific manner, or providing integrated signals to B cells to induce affinity maturation and isotype switching during antibody production. Such engineered organisms have the potential be as safe and as inexpensive as probiotics but precise in carrying out targeted interventions.
Antibody-recruiting small molecules
Antibody therapeutics and othe |
https://en.wikipedia.org/wiki/Passive%20Wi-Fi | Passive Wi-Fi is a refinement of Wi-Fi technology that uses passive reflection to reduce energy consumption.
Wi-Fi energy use
Wi-Fi use can account for up to 60 percent of a smartphone’s energy consumption. When not connected to a network, Wi-Fi consumes energy because the device constantly searches for a signal.
Backscattering
The technique communicates via backscattering, reflecting incoming radio waves sent from a separate device. The technique is similar to contactless RFID chip cards although unlike such cards, the new technique does not require a special device to read the signal.
The project effectively decoupled the analog and the digital radio signals. Power-intensive functions – like producing a signal at a specific frequency are assigned to a single device in the network that is plugged into the grid. Smartphones modify and reflect this signal to communicate to the router. Prototype passive devices transferred data as far as 100 feet through walls at 11 megabits per second. The system used tens of microwatts of power, 10−4 less energy than conventional Wi-fi devices, and one thousandth the energy of Bluetooth LE and Zigbee communications standards.
Applications
Applications include smart home devices such as smoke detectors, temperature sensors and security cameras that will no longer require a power source.
References
Wi-Fi
Internet of things |
https://en.wikipedia.org/wiki/Librem | Librem is a line of computers manufactured by Purism, SPC featuring free (libre) software. The laptop line is designed to protect privacy and freedom by providing no non-free (proprietary) software in the operating system or kernel, avoiding the Intel Active Management Technology, and gradually freeing and securing firmware. Librem laptops feature hardware kill switches for the microphone, webcam, Bluetooth and Wi-Fi.
Models
Laptops
Librem 13, Librem 15 and Librem 14
In 2014, Purism launched a crowdfunding campaign on Crowd Supply to fund the creation and production of the Librem 15 laptop, conceived as a modern alternative to existing open-source hardware laptops, all of which used older hardware. The in the name refers to its 15-inch screen size. The campaign succeeded after extending the original campaign, and the laptops were shipped to backers. In a second revision of the laptop, hardware kill switches for the camera, microphone, Wi-Fi, and Bluetooth were added.
After the successful launch of the Librem 15, Purism created another campaign on Crowd Supply for a 13-inch laptop called the Librem 13, which also came with hardware kill switches similar to those on the Librem 15v2. The campaign was again successful and the laptops were shipped to customers.
Purism announced in December 2016 that it would start shipping from inventory rather than building to order with the new batches of Librem 15 and 13.
, Purism has one laptop model in production, the Librem 14 (version 1, US$1,370).
Comparison of laptops
Librem Mini
The Librem Mini is a small form factor desktop computer, which began shipping in June 2020.
Librem 5
On August 24, 2017, Purism started a crowdfunding campaign for the Librem 5, a smartphone aimed to run 100% free software, which would "[focus] on security by design and privacy protection by default". Purism claimed that the phone would become "the world's first ever IP-native mobile handset, using end-to-end encrypted decentralized communica |
https://en.wikipedia.org/wiki/Aeronca%20Museum | The Aeronca Museum is an online museum that hosts a collection of artifacts and aircraft produced by Aeronca (the Aeronautical Corporation of America) (a.k.a. Aeronca, Inc. or Aeronca Aircraft, now Magellan Aerospace Corp.), a prominent U.S. light aircraft manufacturer from 1928 to 1951. The museum is instituted to "preserve the knowledge and history of Aeronca aircraft in the public trust" and keep it "available for future generations," with the objective "to be the premier research library for all things Aeronca."
History
The museum began its development under President and Executive Director Todd Trainor, an aircraft restoration specialist, long involved with Aeronca planes, and son of late Aeronca restoration expert Tom Trainor.
A board was formed and the organization applied for 501(C)3 non-profit certification in early 2011; it holds that status now, and is currently operating as a public information service.
At present, the artifacts are at the museum's offices in Brighton, Michigan, and in hangars at the Brighton Airport, some undergoing restoration and preparation for presentation.
Currently, the museum largely serves as an archive of thousands of documents, drawings and blueprints, technical manuals and other research references related to Aeronca aircraft, and presents these online through its website, both commercially and as free-distribution information for Aeronca aircraft operators and enthusiasts.
Exhibits and artifacts
The museum is home to three complete Aeronca K projects (including one seaplane), one flyable Aeronca TAC, a C-3 basket case, and a KCA basket case, some in restoration—and over 10,000 other Aeronca artifacts and papers.
The museum has collected many documents, drawings, blueprints, manuals and other references from Aeronca, and from its successor, Magellan Aerospace, and from Aeronca enthusiasts, particularly those affiliated with the National Aeronca Association and the Fearless Aeronca Aviators, and makes them available thro |
https://en.wikipedia.org/wiki/Wax%20emulsion | Wax emulsions are stable mixtures of one or more waxes in water. Waxes and water are normally immiscible but can be brought together stably by the use of surfactants and a clever preparation process. Strictly speaking a wax emulsion should be called a wax dispersion since the wax is solid at room temperature. However, because the preparation takes place above the melting point of the wax, the actual process is called emulsification, hence the name wax emulsion. In praxis, wax dispersion is used for solvent based systems.
A wide range of emulsions based on different waxes and blends thereof are available, depending on the final application. Waxes that are found in wax emulsions can be of natural or synthetic origin. Common non-fossil natural waxes are carnaubawax, beeswax, candelilla wax or ricebran wax. Paraffin, microcrystalline and montanwax are the most used fossil natural waxes that are found in emulsions. Synthetic waxes that are used include (oxidised) LDPE and HDPE, maleic anhydride grafted polypropylene and Fischer-Tropsch waxes.
A range of different emulsifiers or surfactants are used to emulsify waxes. These can be anionic, cationic or non-ionic in nature. The most common however are fatty alcohol ethoxylates as non-ionic surfactants due to their superb stability against hard water, pH-shock and electrolytes. Some applications demand different emulsifier systems for example anionic surfactants for better hydrophobicity or cationic surfactants for better adhesion to certain materials like textile fibers.
Applications
Wax emulsions are widely used in a variety of technical applications like printing inks & lacquers, leather and textiles, paper, wood, metal, polishes, glass fiber sizing, glass bottle protection among other things. The most important properties that can be improved by the addition of wax emulsions are matting & gloss, hydrophobicity, soft touch, abrasion & rub resistance, scratch resistance, release, corrosion protection and anti-blocking. |
https://en.wikipedia.org/wiki/List%20of%20Passeriformes%20by%20population | This is a list of Passeriforme species by global population. While numbers are estimates, they have been made by the experts in their fields.
Passeriformes is the taxonomic order to which the perching birds belong.
Species by global population
See also
Lists of birds by population
Lists of organisms by population
References
Birds |
https://en.wikipedia.org/wiki/ClearPath%20Foundation | ClearPath Foundation is a nonprofit organization based in Charlotte, North Carolina, and Washington, D.C., that is focused on "conservative clean energy". According to a press release, the organization was founded to propose and support policies that accelerate clean energy without expanding the size of government.
History
ClearPath was founded by Republican entrepreneur and philanthropist Jay Faison in 2013 to change the conservative viewpoint on clean energy. After selling his majority stake in the audio-visual company SnapAV, Faison donated US$165 million to start ClearPath, with a mission of "accelerating conservative clean energy solutions."
Activities
ClearPath advances its mission through "strategic grant-making, advocacy, and digital platforms", and focuses on energy policy, polling, and analysis.
Principles
The ClearPath website lists five principles used for analyzing policy. ClearPath believes in "small government, free markets, and American innovation", as well as pollution risk management, cost-benefit analysis and energy security.
Policy Areas
Nuclear energy – ClearPath proposes expanded use of nuclear power in the United States, calling it "one of America's greatest success stories". In particular, they support existing nuclear plants, as well as advanced nuclear such as small modular reactors and generation IV reactors. In February 2016, ClearPath's CEO Jay Faison proposed developing "super-competitive, next-generation" nuclear reactors.
Hydropower – ClearPath supports hydropower development, calling it "one of America’s most overlooked energy resources". They propose expanding the electrification of non-powered dams, hydropower financing reform, and permitting reform.
Energy Innovation – ClearPath supports increased basic energy research funding for the Department of Energy (DOE) to accelerate development, and proposes paying for increased research through expanded oil and gas drilling. They also propose giving private companies greater ac |
https://en.wikipedia.org/wiki/43S%20preinitiation%20complex | The 43S preinitiation complex (43S PIC) is a ribonucleoprotein complex that exists during an early step of eukaryotic translation initiation. The 43S PIC contains the small ribosomal subunit (40S) bound by the initiation factors eIF1, eIF1A, eIF3, and the eIF2-Met-tRNAiMet-GTP ternary complex (eIF2-TC).
Function
The 43S is an important intermediate complex during cap-dependent translation initiation. In the canonical model of translation initiation, the 43S PIC is pre-formed as a stable complex and recruited to the 5' cap of eukaryotic messenger RNAs (mRNAs) by the eIF4F complex. The 43S PIC then "scans" in the 5' --> 3' direction along the mRNA in an ATP-dependent fashion (via eIF4A and/or other RNA helicases such as Ded1/DDX3 and DHX29) to locate the start codon. Start codon recognition occurs through base-pairing between the Met-tRNAiMet and AUG in the ribosomal P-site and a number of associated changes, and is followed by joining of the large 60S ribosomal subunit to form the 80S ribosome.
Structure
Due to its size and complexity, the 43S PIC has eluded high resolution structural characterization. However, combined approaches including cryo-EM, cross-linking, and the structural characterization of individual components, has led to models for the complex organization.
See also
Eukaryotic translation
40S ribosomal subunit
eIF3
References
Molecular biology
Gene expression
RNA-binding proteins
Protein biosynthesis
Protein complexes |
https://en.wikipedia.org/wiki/Bornhuetter%E2%80%93Ferguson%20method | The Bornhuetter–Ferguson method is a loss reserving technique in insurance.
Background
The Bornhuetter–Ferguson method was introduced in the 1972 paper "The Actuary and IBNR", co-authored by Ron Bornhuetter and Ron Ferguson.
Like other loss reserving techniques, the Bornhuetter–Ferguson method aims to estimate incurred but not reported insurance claim amounts. It is primarily used in the property and casualty and health insurance fields.
Generally considered a blend of the chain-ladder and expected claims loss reserving methods, the Bornhuetter–Ferguson method uses both reported or paid losses as well as an a priori expected loss ratio to arrive at an ultimate loss estimate. Simply, reported (or paid) losses are added to a priori expected losses multiplied by an estimated percent unreported. The estimated percent unreported (or unpaid) is established by observing historical claims experience.
The Bornhuetter–Ferguson method can be used with either reported or paid losses.
Methodology
There are two algebraically equivalent approaches to calculating the Bornhuetter–Ferguson ultimate loss.
In the first approach, undeveloped reported (or paid) losses are added directly to expected losses (based on an a priori loss ratio) multiplied by an estimated percent unreported.
In the second approach, reported (or paid) losses are first developed to ultimate using a chain-ladder approach and applying a loss development factor (LDF). Next, the chain-ladder ultimate is multiplied by an estimated percent reported. Finally, expected losses multiplied by an estimated percent unreported are added (as in the first approach).
The estimated percent reported is the reciprocal of the loss development factor.
Incurred but not reported claims can then be determined by subtracting reported losses from the Bornhuetter–Ferguson ultimate loss estimate.
References
Actuarial science |
https://en.wikipedia.org/wiki/Naren%20Gupta | Narendra K. Gupta (September 30, 1948December 25, 2021) was an Indian-American venture capital investor and technology entrepreneur. He founded the investment company Nexus Venture Partners, as well as various other investment and software firms. He was also CEO and president of Integrated Systems Inc. (ISI) for a decade and a half. In a career spanning over three decades he was recognized for his contributions to global technology and entrepreneurial ecosystems, and for venture capital investments in India.
Early life
Gupta was born on September 30, 1948. He received a Bachelor of Technology degree from the Indian Institute of Technology, Delhi, in 1969 majoring in mechanical engineering. He moved to the United States and received a Master of Science degree from the California Institute of Technology, in 1970 with a specialization in aeronautics and a Ph.D. from Stanford University in 1974 with a specialization in applied mechanics. He was a recipient of the president's gold medal for his performance at IIT Delhi and later received distinguished alumni awards from IIT in 1997 and from Caltech in 2004. He was also elected a fellow of the Institute of Electrical and Electronics Engineers in 1991.
Career
Gupta co-founded Integrated Systems Inc., an embedded software company, in 1980 and was the company's CEO for 15 years. In 1990, he took the company went public, and later merged with Wind River Systems. He remained on that board until it was acquired by Intel. He had more than two decades of experience in the field of early and early-growth stage investment in both America and India.
Nexus Venture Partners
In 2007, Gupta founded Nexus Venture Partners, a Silicon Valley investment firm, with offices in Menlo Park, California, and Mumbai and Bangalore in India. This is a "homegrown Indian operation [with] extensive reach and experience overseas [most notably in] Silicon Valley." It has made investments in a variety of technical and media startups around India and |
https://en.wikipedia.org/wiki/Kunstweg | Bürgi's Kunstweg is a set of algorithms invented by Jost Bürgi at the end of the 16th century. They can be used for the calculation of sines to an arbitrary precision. Bürgi used these algorithms to calculate a Canon Sinuum, a table of sines in steps of 2 arc seconds. It is thought that this table had 8 sexagesimal places. Some authors have speculated that this table only covered the range from 0 to 45 degrees, but nothing seems to support this claim. Such tables were extremely important for navigation at sea. Johannes Kepler called the Canon Sinuum the most precise known table of sines. Bürgi explained his algorithms in his work Fundamentum Astronomiae which he presented to Emperor Rudolf II. in 1592.
The principles of iterative sine table calculation through the Kunstweg are as follows: cells in a column sum up the values of the two previous cells in the same column. The final cell's value is divided by two, and the next iteration starts. Finally, the values of the last column get normalized. Rather accurate approximations of sines are obtained after few iterations.
As recently as 2015, Folkerts et al. showed that this simple process converges indeed towards the true sines. According to Folkerts, this was the first step towards difference calculus.
References
Algorithms
Trigonometry |
https://en.wikipedia.org/wiki/Cyber%20resilience | Cyber resilience refers to an entity's ability to continuously deliver the intended outcome, despite cyber attacks. Resilience to cyber attacks is essential to IT systems, critical infrastructure, business processes, organizations, societies, and nation-states.
Adverse cyber events are those that negatively impact the availability, integrity, or confidentiality of networked IT systems and associated information and services. These events may be intentional (e.g. cyber attack) or unintentional (e.g. failed software update) and caused by humans, nature, or a combination thereof.
Unlike cyber security, which is designed to protect systems, networks and data from cyber crimes, cyber resilience is designed to prevent systems and networks from being derailed in the event that security is compromised. Cyber security is effective without compromising the usability of systems and there is a robust continuity business plan to resume operations, if the cyber attack is successful.
Cyber resilience helps businesses to recognize that hackers have the advantage of innovative tools, element of surprise, target and can be successful in their attempt. This concept helps business to prepare, prevent, respond and successfully recover to the intended secure state. This is a cultural shift as the organization sees security as a full-time job and embedded security best practices in day-to-day operations. In comparison to cyber security, cyber resilience requires the business to think differently and be more agile on handling attacks.
The objective of cyber resilience is to maintain the entity's ability to deliver the intended outcome continuously at all times. This means doing so even when regular delivery mechanisms have failed, such as during a crisis or after a security breach. The concept also includes the ability to restore or recover regular delivery mechanisms after such events, as well as the ability to continuously change or modify these delivery mechanisms, if needed in the |
https://en.wikipedia.org/wiki/4D%20Inc. | 4D Inc is a US-based subsidiary of 4D SAS and publisher of 4D. It was established in 1984 and initially led by Guy Kawasaki. 4D is the developer and publisher of 4D (or 4th Dimension). 4D was founded in 1984 when development began for Silver Surfer (early code-name for 4D) and had its initial product release in 1987 with its own Programming Language.
After 16 years of operating as ACIUS from 1984 to 2000, the name was officially changed to 4D Inc.
4D is also the original developer of Wakanda.
References
Business software companies
Development software companies
Software companies established in 1984
1984 establishments in California
Software companies based in the San Francisco Bay Area
Technology companies based in the San Francisco Bay Area
Software companies of the United States |
https://en.wikipedia.org/wiki/4D%20SAS | 4D SAS is a French company owned by Laurent Ribardière. 4D has a US-based subsidiary 4D Inc. 4D was founded in 1984 when development began for Silver Surfer (early codename for 4D) and had its initial product release in 1987 with its own Programming Language. It is the developer and publisher of 4D (or 4th Dimension) and the original developer of Wakanda.
References
Business software companies
Development software companies
Software companies of France
French companies established in 1984
Software companies established in 1984
Computer companies established in 1984 |
https://en.wikipedia.org/wiki/Dmitri%20Leonidovich%20Romanowsky | Dmitri Leonidovich Romanowsky (sometimes spelled Dmitry and Romanowski, ; 1861–1921) was a Russian Emire physician who is best known for his invention of an eponymous histological stain called Romanowsky stain. It paved the way for the discovery and diagnosis of microscopic pathogens, such as malarial parasites.
Romanowsky was born in 1861 in Pskov Governorate, Russia. He attended the 6th Saint Petersburg Gymnasium. In 1880, he enrolled at the St. Petersburg University. He enrolled for two courses: natural science (physics and mathematics) and medicine. He concentrated on medicine in 1882 for a preparatory course to the Military Medical Academy. He graduated with honors in 1886. On 30 November 1886, he was appointed as a junior resident of the Ivangorod military hospital. After one month, he was transferred to the Revel local infirmary as an associate doctor. In 1889, he was attached to the Saint Petersburg Nikolaevsky Military hospital. He initially worked at the clinical department, and from May 1890, he was the head of the eye department. He obtained his medical degree in 1891. It was in that year that he published a paper in which he described the technique for staining malarial parasites from blood samples.
References
1861 births
1921 deaths
Saint Petersburg State University alumni
Ophthalmologists from the Russian Empire
19th-century physicians from the Russian Empire |
https://en.wikipedia.org/wiki/Computer%20bureau | A computer bureau is a service bureau providing computer services.
Computer bureaus developed during the early 1960s, following the development of time-sharing operating systems. These allowed the services of a single large and expensive mainframe computer to be divided up and sold as a fungible commodity. Development of telecommunications and the first modems encouraged the growth of computer bureau as they allowed immediate access to the computer facilities from a customer's own premises.
The computer bureau model shrank during the 1980s, as cheap commodity computers, particularly the PC clone but also the minicomputer allowed services to be hosted on-premises.
See also
Batch processing
Cloud computing
Grid computing
Service Bureau Corporation
Utility computing
References
Time-sharing
Computer systems
Business models |
https://en.wikipedia.org/wiki/Faddeev%E2%80%93LeVerrier%20algorithm | In mathematics (linear algebra), the Faddeev–LeVerrier algorithm is a recursive method to calculate the coefficients of the characteristic polynomial of a square matrix, , named after Dmitry Konstantinovich Faddeev and Urbain Le Verrier. Calculation of this polynomial yields the eigenvalues of as its roots; as a matrix polynomial in the matrix itself, it vanishes by the Cayley–Hamilton theorem. Computing the characteristic polynomial directly from the definition of the determinant is computationally cumbersome insofar as it introduces a new symbolic quantity ; by contrast, the Faddeev-Le Verrier algorithm works directly with coefficients of matrix .
The algorithm has been independently rediscovered several times in different forms. It was first published in 1840 by Urbain Le Verrier, subsequently redeveloped by P. Horst, Jean-Marie Souriau, in its present form here by Faddeev and Sominsky, and further by J. S. Frame, and others. (For historical points, see Householder. An elegant shortcut to the proof, bypassing Newton polynomials, was introduced by Hou. The bulk of the presentation here follows Gantmacher, p. 88.)
The Algorithm
The objective is to calculate the coefficients of the characteristic polynomial of the matrix ,
where, evidently, = 1 and 0 = (−1)n det .
The coefficients are determined by induction on , using an auxiliary sequence of matrices
Thus,
etc.,
...;
Observe terminates the recursion at . This could be used to obtain the inverse or the determinant of .
Derivation
The proof relies on the modes of the adjugate matrix, , the auxiliary matrices encountered.
This matrix is defined by
and is thus proportional to the resolvent
It is evidently a matrix polynomial in of degree . Thus,
where one may define the harmless ≡0.
Inserting the explicit polynomial forms into the defining equation for the adjugate, above,
Now, at the highest order, the first term vanishes by =0; whereas at the bottom order (constant in , fro |
https://en.wikipedia.org/wiki/Epoxydocosapentaenoic%20acid | Epoxide docosapentaenoic acids (epoxydocosapentaenoic acids, EDPs, or EpDPEs) are metabolites of the 22-carbon straight-chain omega-3 fatty acid, docosahexaenoic acid (DHA). Cell types that express certain cytochrome P450 (CYP) epoxygenases metabolize polyunsaturated fatty acid's (PUFAs) by converting one of their double bonds to an epoxide. In the best known of these metabolic pathways, cellular CYP epoxygenases metabolize the 20-carbon straight-chain omega-6 fatty acid, arachidonic acid, to epoxyeicosatrienoic acids (EETs); another CYP epoxygenase pathway metabolizes the 20-carbon omega-3 fatty acid, eicosapentaenoic acid (EPA), to epoxyeicosatetraenoic acids (EEQs). CYP epoxygenases similarly convert various other PUFAs to epoxides (see epoxygenase) These epoxide metabolites have a variety of activities. However, essentially all of them are rapidly converted to their corresponding, but in general far less active, Vicinal (chemistry) dihydroxy fatty acids by ubiquitous cellular Soluble epoxide hydrolase (sEH; also termed Epoxide hydrolase 2). Consequently, these epoxides, including EDPs, operate as short-lived signaling agents that regulate the function of their parent or nearby cells. The particular feature of EDPs (and EEQs) distinguishing them from EETs is that they derive from omega-3 fatty acids and are suggested to be responsible for some of the beneficial effects attributed to omega-3 fatty acids and omega-3-rich foods such as fish oil.
Structure
EDPs are epoxide eicosapentaenoic acid metabolites of DHA. DHA has 6 cis (see Cis–trans isomerism) Double bonds each one of which is located between carbons 4-5, 7-8, 10-11, 13-14, 16-17, or 19-20. Cytochrome P450 epoxygenases attack any one of these double bounds to form a respective docosapentaenoic acid (DPA) epoxide regioisomer (see Structural isomer, section on position isomerism (regioisomerism)). A given epoxygenase may therefore convert DHA to 4,5-EDP (i.e. 4,5-epoxy-7Z,10Z,13Z,16Z,19Z-DPA), 7,8-E |
https://en.wikipedia.org/wiki/Annual%20vs.%20perennial%20plant%20evolution | Annuality (living and reproducing in a single year) and perenniality (living more than two years) represent major life history strategies within plant lineages. These traits can shift from one to another over both macroevolutionary and microevolutionary timescales. While perenniality and annuality are often described as discrete either-or traits, they often occur in a continuous spectrum. The complex history of switches between annual and perennial habit involve both natural and artificial causes, and studies of this fluctuation have importance to sustainable agriculture. (Note that "perennial" here refers to both woody and herbaceous perennial species.)
According to some studies, either the trait of annuality or perenniality may be ancestral. This contradicts the commonly held belief that annuality is a derived trait from an ancestral perennial life form, as is suggested by a regarded plant population biology text.
Spatiotemporal scale
Above the species level, plant lineages clearly vary in their tendency for annuality or perenniality (e.g., wheat vs. oaks). On a microevolutionary timescale, a single plant species may show different annual or perennial ecotypes (e.g., adapted to dry or tropical range), as in the case of the wild progenitor of rice (Oryza rufipogon). Indeed, ability to perennate (live more than one year) may vary within a single population of a species.
Underlying mechanisms: Trade-offs
Annuality and perenniality are complex traits involving many underlying, often quantitative, genotypic and phenotypic factors. They are often determined by a trade-off between allocation to sexual (flower) structures and asexual (vegetative) structures. Switches between the annual and perennial habit are known to be common among herbaceous angiosperms.
Increased allocation to reproduction early in life generally leads to a decrease in survival later in life (senescence); this occurs in both annual and perennial semelparous plants. Exceptions to this pattern inc |
https://en.wikipedia.org/wiki/Archibald%20Burns%20%28photographer%29 | Archibald Burns (1831–1880) was a Scottish photographer based in Edinburgh, Scotland, and active from 1858 to 1880. He documented the city through various publications and recorded the historic buildings in a section of the city that was cleared for improvements in the 1860s.
Life and career
Burns became active in photography as an amateur in the 1850s. He became a member of the Photographic Society of Scotland in 1858 and was one of the first members of the Edinburgh Photographic Society in 1861. He pursued primarily landscape and architectural photography and capitalized on new tourist markets for illustrated books and views in the latter part of the nineteenth century.
Burns first professional photography studio was located at 22 Calton Stairs from 1861 to 1871, at which point he established his business at the Rock House until 1880. Through parts of 1870 and 1871, Burns shared the Rock House, which had earlier been the studio of important Scottish calotypists Hill & Adamson, with Glaswegian photographer Thomas Annan.
Burns died in 1880 and was buried at Warriston Cemetery. The contents of his studio – inventory, materials, and hardware – were put up for sale in May of that year.
Works
Archibald Burns promoted himself as a landscape photographer and sold individual prints, stereographs, cabinet cards, and magic lantern slides of views of Edinburgh and surrounding area.
Burns illustrated two books on Edinburgh published in 1868, three years before he took his series of photographs of closes and wynds for the Edinburgh Improvement Trust (January and February 1871). The text in Picturesque "Bits" from Old Edinburgh (1st ed. 1868) emphasizes the architectural history of Scotland and the importance of photography in preserving the knowledge of fading vernacular styles and ends with a questions regarding the future of Scottish architecture. The book is illustrated with 15 tipped-in albumen prints by Burns and 8 woodcuts by Charles Paton after drawings in Daniel |
https://en.wikipedia.org/wiki/Marie%20Pistilli%20Award | The Marie R. Pistilli Women in Engineering Achievement Award is issued annually since 2000 by the Design Automation Conference (DAC) to honor the outstanding achievements of women in Electronic Design Automation. It is named after the co-founder of DAC, Marie Pistilli. Originally named as the "Marie R. Pistilli Women in EDA Achievement Award", it is named the "Marie R. Pistilli Women in Engineering Achievement Award" since 2016.
Recipients
2000: Penny Herscher, then with Cadence Design Systems
2001: Deirdre Hanford, Synopsys, Inc.
2002: Ann Rincon, then senior technical staff member at IBM Microelectronics, Burlington, Vermont, "For her significant contributions in helping women advance in the field of DA technology"
2003: Karen Bartleson, Synopsis, Inc.
2004: Mary Jane Irwin, Pennsylvania State University
2005: Kathryn Kranen, Jasper Design Automation
2006: Ellen J. Yoffa, IBM
2007: Jan Willis, then with Cadence Design Systems - "Jan's leadership and dedication in serving as a mentor and role model to women in EDA, and young girls and boys as they begin to consider career options, exemplifies what the Marie R. Pistilli Women in Electronic Design Automation Achievement Award is all about."
2008: Louise Trevillyan IBM Research Center
2009: Telle Whitney
2010: Mar Hershenson, Vice President of Product Development in the Custom Design Business Unit at Magma Design Automation
2011: Limor Fix, senior principal engineer and director of academic programs and research at Intel
2012: Belle Wei, then Dean of Engineering, San Jose State University
2013: Nanette Collins, a public relations consultant from Boston, active in EDA industry
2014: Diana Marculescu, Professor of Electrical and Computer Engineering, Carnegie Mellon University
2015: Margaret Martonosi, Princeton University
2016: Soha Hassoun, Tufts University
2017: Janet Olson, then with Synopsys, Inc.
2018: Anne Cirkel, Senior Director for Technology Marketing at Mentor, a Siemens Business
2019: R. Iris Bahar, Brow |
https://en.wikipedia.org/wiki/National%20Computer%20Rank%20Examination | National Computer Rank Examination (NCRE) is a national exam held by China Education Department to test the computer proficiency and programming skill of non-computer specialized students and practitioners. The programming language can be chosen by examinees, including C, C++, Java, Visual Basic and Python 3. NCRE is widely recognized by enterprises and organizations in China.
Education in China
Computer programming |
https://en.wikipedia.org/wiki/Rise%20in%20core | The rise in core (RIC) method is an alternate reservoir wettability characterization method described by S. Ghedan and C. H. Canbaz in 2014. The method enables estimation of all wetting regions such as strongly water wet, intermediate water, oil wet and strongly oil wet regions in relatively quick and accurate measurements in terms of Contact angle rather than wettability index.
During the RIC experiments, core samples saturated with selected reservoir fluid were subjected to imbibition from a second reservoir fluid. RIC wettability measurements are compared with and modified – Amott test and USBM measurements using core plug pairs from different heights of a thick carbonate reservoir. Results show good coherence. The RIC method is an alternate method to Amott and USBM methods and that efficiently characterizes Reservoir Wettability.
Cut-off values vs wettability index
One study used the water advancing contact angle to estimate the wettability of fifty-five oil reservoirs. De-oxygenated synthetic formation brine and dead anaerobic crude was tested on quartz and calcite crystals at reservoir temperature. Contact angles from 0 to 75 degrees were deemed water wet, 75 to 105 degrees as intermediate and 105 to 180 degrees as oil wet. Although the range of wettabilities were divided into three regions, these were arbitrary divisions. The wettability of different reservoirs can vary within the broad spectrum from strongly water-wet to strongly oil-wet.
Another study described two initial conditions as reference and non-reference for calculating cut-off values by using advancing and receding contact angles and spontaneous imbibition data. Limiting value between water wet and intermediate zones was described as 62-degree. Similarly, cut-off values for advancing contact angle is described as 0 to 62 degrees for water wet region, 62 to 133 degrees for Intermediate-wet zone, and 133 to 180 degrees for Oil wet zone.
Chilingar and Yen examined extensive research work on 161 |
https://en.wikipedia.org/wiki/Giatec%20Scientific | Giatec Scientific Inc. is a Canadian-based company with headquarters located in Ottawa, Ontario. It is a developer and manufacturer of nondestructive testing quality control and condition assessment devices for the construction industry.
History
Giatec Scientific Inc. was co-founded by Pouria Ghods of Carleton University and Aali R. Alizadeh of the University of Ottawa in September 2010. The pair began working with advisers at Invest Ottawa, who arranged sources, funding and ideas to bring Giatec's products to the market.
The company's first product was a sensor to detect corrosion speed in the rebar/steel inside concrete. Unlike other non-destructive testing methods available at the time, Giatec used mobile-based applications software and smart technology to collect and analyze data.
In 2012, Giatec became independent of Invest Ottawa. That year, after the collapse of the Algo Centre Mall in Elliot Lake in 2012, Giatec's equipment was used in the forensic structural examination that was initiated as part of the public inquiry.
Giatec has developed a variety of testing devices and sensors for measurement of concrete permeability, electrical resistivity measurement of concrete, half-cell corrosion, corrosion rate, concrete temperature, and concrete maturity. In 2014, Giatec won the Rio Info 2013 Innovation Award, and in 2014, the company was included in the Ottawa Business Journal's annual list of "Startups to Watch". Giatec is also the recipient of Ottawa's Top 10 Fastest Growing Companies, and Canada's Top 500 Fastest Growing Companies in 2018.
Giatec also began to develop Internet of Things (IoT) applications for the construction industry through wireless concrete temperature and maturity sensors. In March 2015, the company released a new electrical resistivity monitoring device that sends data directly to a smartphone through a downloadable application. In October 2016, Giatec released Smart Concrete, a new IoT-based solution for ready-mix concrete producer |
https://en.wikipedia.org/wiki/DROWN%20attack | The DROWN (Decrypting RSA with Obsolete and Weakened eNcryption) attack is a cross-protocol security bug that attacks servers supporting modern SSLv3/TLS protocol suites by using their support for the obsolete, insecure, SSL v2 protocol to leverage an attack on connections using up-to-date protocols that would otherwise be secure. DROWN can affect all types of servers that offer services encrypted with SSLv3/TLS yet still support SSLv2, provided they share the same public key credentials between the two protocols. Additionally, if the same public key certificate is used on a different server that supports SSLv2, the TLS server is also vulnerable due to the SSLv2 server leaking key information that can be used against the TLS server.
Full details of DROWN were announced in March 2016, along with a patch that disables SSLv2 in OpenSSL; the vulnerability was assigned the ID . The patch alone will not be sufficient to mitigate the attack if the certificate can be found on another SSLv2 host. The only viable countermeasure is to disable SSLv2 on all servers.
The researchers estimated that 33% of all HTTPS sites were affected by this vulnerability as of March 1, 2016.
Details
DROWN is an acronym for "Decrypting RSA with Obsolete and Weakened eNcryption".
It exploits a vulnerability in the combination of protocols used and the configuration of the server, rather than any specific implementation error. According to the discoverers, the exploit cannot be fixed by making changes to client software such as web browsers.
The exploit includes a chosen-ciphertext attack with the use of a SSLv2 server as a Bleichenbacher oracle. SSLv2 worked by encrypting the master secret directly using RSA, and 40-bit export ciphersuites worked by encrypting only 40-bit of the master secret and revealing the other 88-bits as plaintext. The 48-byte SSLv3/TLS encrypted RSA ciphertext is "trimmed" to 40-bit parts and is then used in the SSLv2 ClientMasterKey message, which the server treats a |
https://en.wikipedia.org/wiki/Deep%20level%20underground | Deep level underground is construction that is or more below ground and not using the cut-and-cover method, especially train stations, air raid shelters and bunkers, and some tunnels and mines. Cut-and-cover is a simple method of construction for shallow tunnels where a trench is excavated and roofed over with an overhead support structure that is strong enough to carry the load of what is to be built above the tunnel.
History
Mining
Although some deep mining took place as early as the late Tudor period (in North-East England, and along the Firth of Forth coast in Scotland) deep shaft mining in Britain began to develop extensively in the late 18th century, with rapid expansion throughout the 19th century and early 20th century when the industry peaked. Before 1800, a great deal of coal was left in places as extraction was still primitive. As a result, in the deep Tyneside pits (300 to 1,000 ft deep) only about 40 percent of the coal could be extracted. The use of wooden pit props to support the roof was an innovation first introduced about 1800.
Transit systems
Before any plans were made for transit systems with tunnels and stations, several railway operators had used tunnels for freight and passenger trains, usually to reduce the grade of the railway line. Examples include Trevithick's Tunnel from 1804, built for the Penydarren locomotive, the 1829 Crown Street Tunnel at Liverpool and the long 1836 Lime Street Tunnel also at Liverpool, of which a part is still used today making it the world's oldest used tunnel.
The world's first urban underground railway was the Metropolitan Railway, which opened on January 10, 1863. It was built largely in shallow tunnels (see more at cut and cover) and is nowadays part of the London Underground. It was operated using steam trains, and despite the creation of numerous ventilation shafts, was unhealthy and uncomfortable for passengers and operating staff. Nevertheless, its trains were popular from the start and the Metropol |
https://en.wikipedia.org/wiki/Poly%28adp-ribose%29%20polymerase%20family%20member%2014 | Poly(ADP-ribose) polymerase family member 14 is a protein that, in humans, is encoded by the PARP14 gene.
Function
Poly(ADP-ribosyl)ation is an immediate DNA damage-dependent post-translational modification of histones and other nuclear proteins that contributes to the survival of injured proliferating cells. PARP14 belongs to the superfamily of enzymes that perform this modification (Ame et al., 2004 [PubMed 15273990]).
References
Further reading
Cellular processes
Molecular genetics
Mutation
Senescence |
https://en.wikipedia.org/wiki/Discrete%20Analysis | Discrete Analysis is a mathematics journal covering the applications of analysis to discrete structures. Discrete Analysis is an arXiv overlay journal, meaning the journal's content is hosted on the arXiv.
History
Discrete Analysis was created by Timothy Gowers to demonstrate that a high-quality mathematics journal could be inexpensively produced outside of the traditional academic publishing industry. The journal is open access, and submissions are free for authors.
The journal's 2018 MCQ is 1.21.
References
External links
Open access journals
Mathematics journals
Academic journals established in 2016
Continuous journals
Online-only journals
English-language journals |
https://en.wikipedia.org/wiki/Broker%20injection | Broker injection attack is a type of vulnerability that exploits misconfigured brokers, potentially allowing an attacker to read, write and inject information from/into their flow.
Description
There are many scenarios in which a broker is used to transport the information between tasks.
One of the most typical use cases is send e-mails in background. In this scenario we'll have two actors:
An information producer (a website, for example).
A worker or background process who actually sends the e-mail.
The producer needs an asynchronous and non-blocking way to send the email information to the worker.
This system is usually a broker. It takes the information from the web front-end and passes it to the worker, generating a new task in the worker. So, the worker has all the information to send the e-mail.
Taking the above scenario as an example, if we could access the broker, we would be able to make the worker generate new tasks with arbitrary data, unleashing a broker injection.
Attacks
With this in mind, we could make the following attacks:
Listing remote tasks.
Reading a remote task's contents.
Injection of tasks into remote processes.
Removing remote outstanding tasks.
Origin
The broker injection attack is not new, but it didn't have a name. This name was coined by Daniel García (cr0hn) at the RootedCON 2016 conference in Spain.
References
See also
Redis
RabbitMQ
ZeroMQ
Message broker
Celery (software)
External links
Official Redis security tips
Enteletaor: The broker injection tool
Broker injection in RootedCON 2016 (Spanish)
Hacking (computer security)
Machine code
Injection exploits
Computer network security |
https://en.wikipedia.org/wiki/Self-perpetuation | Self-perpetuation, the capability of something to cause itself to continue to exist, is one of the main characteristics of life. Organisms' capability of reproduction leads to self-perpetuation of the species, if not to the individual. Populations self-perpetuate and grow. Entire ecosystems show homeostasis, and thus perpetuate themselves. The slow modifying effect of succession and similar shifts in the composition of the system can, however, not be neglected in the long run. Overall, life's object's capabilities of self-perpetuation are always accompanied by evolution, a perfect steady state of the biological system is never reached. Sexual reproduction is also a form of imperfect self-replication and thus imperfect self-perpetuation because of recombination and mutation. Organisms are not like self-replicating machine but amass random modifications from generation to generation. The property of self-perpetuation in the strict sense thus only applies to life itself.
In a social context, self-perpetuation is tied to reflexivity and (usually) positive feedback loops:
Depending on the time scope or the context, self-perpetuation either depends on self-sustainability, or is equivalent to it. While we may talk about the self-sustainability of an ecosystem, this depends amongst other factor on the self-perpetuation of its constituting species.
In computer science, self-reproducing programs constitute an incomplete metaphor for self-perpetuation. A better analogue can be seen in computer viruses which are actually able to self-reproduce - given a suitable computing environment.
References
See also
self-stabilization, homeostasis
self-replication
self-reference
recursion
reproduction
feedback loop
cause and effect
von Neumann universal constructor
Homeostasis
Self-organization |
https://en.wikipedia.org/wiki/OpenWSN | OpenWSN is a project created at the University of California Berkeley and extended at the INRIA and at the Open University of Catalonia (UOC) which aims to build an open standard-based and open source implementation of a complete constrained network protocol stack for wireless sensor networks and Internet of Things. The root of OpenWSN is a deterministic MAC layer implementing the IEEE 802.15.4e TSCH based on the concept of Time Slotted Channel Hopping (TSCH). Above the MAC layer, the Low Power Lossy Network stack is based on IETF standards including the IETF 6TiSCH management and adaptation layer (a minimal configuration profile, 6top protocol and different scheduling functions). The stack is complemented by an implementation of 6LoWPAN, RPL in non-storing mode, UDP and CoAP, enabling access to devices running the stack from the native IPv6 through open standards.
OpenWSN is related to other projects including the following:
RIOT
OpenMote
OpenWSN is available for Linux, Windows and OS X platforms. Current release of OpenWSN is 1.14.0.
References
Wireless sensor network
Internet of things
Ambient intelligence
Emerging technologies |
https://en.wikipedia.org/wiki/Aksharapalli | Aksharapalli () is a certain type of alphasyllabic numeration scheme extensively used in the pagination of manuscripts produced in India in pre-modern times. The name Aksharapalli can be translated as the letter system. In this system the letters or the syllables of the script in which the manuscript is written are used to denote the numbers. In contrast to the Aksharapalli system, the ordinary decimal system is called the Ankapalli system.
Examples of syllables used to represent numerals
The following tables give examples of syllables used to represent numerals. The lists are not exhaustive.
Usage
When the Aksharapalli system is used, the various syllables that constitute a number are placed one below the other as in the Chinese language and they are written in the margins of the various leaves of the manuscript. This arrangement may be the consequence an attempt to save space for the contents of the manuscript. This method can be seen in the earliest available manuscript containing the Aksharapalli system which is a manuscript of sixth century CE.
History
Nothing definitely is known about the origin of the system. It is conjectured that the system might have evolved from the ciphered-additive numeral system of Brahmi. This system has been extensively used in Jain manuscripts up to the sixteenth century. The system has also survived for a long time in Nepal. The system was in use as late as the nineteenth century in those regions of India which now constitute the Kerala State.
References
Numeral systems
Indian mathematics |
https://en.wikipedia.org/wiki/Perfect%20%28server%20framework%29 | Perfect is an application server, web and server framework written in the programming language Swift. It provides tools for developing web and other representational state transfer (REST) services in Swift, and web server and datasource connectors, providing default structures for developers working with databases, web services, and web pages. Its main focus is to make mobile development easier by using the Swift language on both sides of the client–server model, the client mobile device, and the server, where it competes with other server-side development tools such as Node.js and Python. It is open-source software, released under an Apache License.
History
Perfect 1.0 was released on November 23, 2015, just before the Swift language became open-source, and draws much of its history from a programming language named Lasso, formerly part of Apple's product line.
Version 2.0 was rolled out September 6th 2016 and brought Swift 3 compatibility and a more modular approach to the framework. The revised approach enabled developers to choose the specific features to include in Perfect-oriented projects rather than having all features bundled in by default.
References
External links
Servers (computing)
See also
Vapor |
https://en.wikipedia.org/wiki/Surface%20activated%20bonding | Surface activated bonding (SAB) is a low-temperature wafer bonding technology with atomically clean and activated surfaces. Surface activation prior to bonding by using fast atom bombardment is typically employed to clean the surfaces. High-strength bonding of semiconductor, metal, and dielectric can be obtained even at room temperature.
Overview
In the standard SAB method, wafer surfaces are activated by argon fast atom bombardment in ultra-high vacuum (UHV) of 10−4–10−7 Pa. The bombardment removes adsorbed contaminants and native oxides on the surfaces. The activated surfaces are atomically clean and reactive for formation of direct bonds between wafers when they are brought into contact even at room temperature.
Researches on SAB
The SAB method has been studied for bonding of various materials, as shown in Table I.
The standard SAB, however, failed to bond some materials such as SiO2 and polymer films. The modified SAB was developed to solve this problem, by using a sputtering deposited Si intermediate layer to improve the bond strength.
The combined SAB has been developed for SiO2-SiO2 and Cu/SiO2 hybrid bonding, without use of any intermediate layer.
Technical specifications
References
Wafer bonding
Semiconductor technology
Electronics manufacturing
Packaging (microfabrication)
Semiconductor device fabrication
Microelectronic and microelectromechanical systems |
https://en.wikipedia.org/wiki/Non-Internet%20email%20address | A wide variety of non-Internet email address formats were used in early email systems before the ubiquity of the john.smith@example.com form used by Internet mail systems since the 1980s - and a few are still used in specialised contexts.
Single system
The earliest email addresses simply had to identify one user from another on one homogenous system, often a single host minicomputer or mainframe. They were therefore typically the user's login name on that system.
Examples of this style include:
ATS: 123
CompuServe: 432654,6564
MCI Mail: 373-1994
AOL: Steve Case
At a host
As computer systems became networked email addresses needed to be able to identify not only the user, but also which host or mail system they were on. Addresses of this type were used in a number of early email systems, including:
ARPANET: jim@washington
IBM Network Job Entry (NJE)
TSO/E TRANSMIT: user_id, node.user_id, node/user_id, nickname, distribution list name, (addressee,...)
VM SENDFILE: userid, userid AT YOURNODE, userid AT node, nickname
PROFS: userid@node
DECnet: host::user (e.g. DECWRL::WRL-TECHREPORTS)
cc:Mail: John Alexopoulos at MicroCircuits
MHS: Barry@MICROSOFT
Delivery path
Some email address schemes described the path through multiple hosts needed to deliver email. This worked well only if the first host given in the path was sufficiently well known for the sender's system to be able to contact it.
UUCP "bang path": reed!percival!bucket!lisag (example on a business card)
Hierarchical
Hierarchical addressing schemes are naturally able to expand. The modern Internet email address (e.g. john.smith@example.com), is of this type - but it was also used by a number of early systems, including:
Banyan VINES: Ed Hirsch@Faculty@Univ
Grey Book: USERID@UK.AC.CAM.ENG
FidoNET: lenz @ 2:331/113.1
Lotus Notes: Tyler Hamilton/Sales@Europe
Directory systems
In this type of system, there is no one unique address for a specific user, but instead a series of attributes, |
https://en.wikipedia.org/wiki/K-regular%20sequence | In mathematics and theoretical computer science, a k-regular sequence is a sequence satisfying linear recurrence equations that reflect the base-k representations of the integers. The class of k-regular sequences generalizes the class of k-automatic sequences to alphabets of infinite size.
Definition
There exist several characterizations of k-regular sequences, all of which are equivalent. Some common characterizations are as follows. For each, we take R′ to be a commutative Noetherian ring and we take R to be a ring containing R′.
k-kernel
Let k ≥ 2. The k-kernel of the sequence is the set of subsequences
The sequence is (R′, k)-regular (often shortened to just "k-regular") if the -module generated by Kk(s) is a finitely-generated R′-module.
In the special case when , the sequence is -regular if is contained in a finite-dimensional vector space over .
Linear combinations
A sequence s(n) is k-regular if there exists an integer E such that, for all ej > E and 0 ≤ rj ≤ kej − 1, every subsequence of s of the form s(kejn + rj) is expressible as an R′-linear combination , where cij is an integer, fij ≤ E, and 0 ≤ bij ≤ kfij − 1.
Alternatively, a sequence s(n) is k-regular if there exist an integer r and subsequences s1(n), ..., sr(n) such that, for all 1 ≤ i ≤ r and 0 ≤ a ≤ k − 1, every sequence si(kn + a) in the k-kernel Kk(s) is an R′-linear combination of the subsequences si(n).
Formal series
Let x0, ..., xk − 1 be a set of k non-commuting variables and let τ be a map sending some natural number n to the string xa0 ... xae − 1, where the base-k representation of x is the string ae − 1...a0. Then a sequence s(n) is k-regular if and only if the formal series is -rational.
Automata-theoretic
The formal series definition of a k-regular sequence leads to an automaton characterization similar to Schützenberger's matrix machine.
History
The notion of k-regular sequences was first investigated in a pair of papers by Allouche and Shallit. Prior to this, Berstel a |
https://en.wikipedia.org/wiki/Solutions%20for%20cavitation%20in%20marine%20propellers | Since the introduction of the marine propeller in the early 19th century, cavitation during operation has been a limiting factor in the efficiency of ships. Cavitation in marine propellers develops when the propeller operates at a high speed and reduces the efficiency of the propeller. Ever since the introduction of the propeller, solutions for cavitation have been developed and tested.
Nozzle System
A nozzle system uses a set of nozzles to help reduce and prevent the likelihood of cavitation in propellers. This system was developed by Samsung Shipping, based in South Korea. In order to reduce the possibility of cavitation happening in marine propellers, a set of nozzles are placed on the hull of the ship directly in front of the propeller. These nozzles spray out compressed air over the propeller that creates “a macro bubble”. This bubble completely encompasses the propeller that is in operation. With the differing characteristics of the seawater outside of the bubble and the air inside, a zone develops that has the ability to reduce the “resonance frequency”. Due to this reduction, cavitation is less likely to occur during the operation of a marine propeller.
To determine the effectiveness of this nozzle system, multiple tests have been carried out with the nozzles and without them. In these tests, it was discovered that the resonance frequencies and the likelihood of cavitation could be reduced by up to 75%. Those who conducted these tests also tried two different arrangements of the nozzles to find out which one was more effective. The first arrangement used only one nozzle, and though it used considerably less power than the other option, it was less effective at reducing cavitation. The multi-nozzle system, however, performed significantly better, but required more power to operate.
While this nozzle system has major drawbacks, particularly in its power requirements, the possibility of cavitation in the operation of marine propellers is reduced considerabl |
https://en.wikipedia.org/wiki/Simulation-based%20optimization | Simulation-based optimization (also known as simply simulation optimization) integrates optimization techniques into simulation modeling and analysis. Because of the complexity of the simulation, the objective function may become difficult and expensive to evaluate. Usually, the underlying simulation model is stochastic, so that the objective function must be estimated using statistical estimation techniques (called output analysis in simulation methodology).
Once a system is mathematically modeled, computer-based simulations provide information about its behavior. Parametric simulation methods can be used to improve the performance of a system. In this method, the input of each variable is varied with other parameters remaining constant and the effect on the design objective is observed. This is a time-consuming method and improves the performance partially. To obtain the optimal solution with minimum computation and time, the problem is solved iteratively where in each iteration the solution moves closer to the optimum solution. Such methods are known as ‘numerical optimization’, ‘simulation-based optimization’ or 'simulation-based multi-objective optimization' used when more than one objective is involved.
In simulation experiment, the goal is to evaluate the effect of different values of input variables on a system. However, the interest is sometimes in finding the optimal value for input variables in terms of the system outcomes. One way could be running simulation experiments for all possible input variables. However, this approach is not always practical due to several possible situations and it just makes it intractable to run experiments for each scenario. For example, there might be too many possible values for input variables, or the simulation model might be too complicated and expensive to run for a large set of input variable values. In these cases, the goal is to iterative find optimal values for the input variables rather than trying all possible v |
https://en.wikipedia.org/wiki/QRZ.com | QRZ.com is an amateur radio website listing almost every callsign in the world. Founded by Fred L. Lloyd in 1992, a considerable amount of effort was devoted to connecting with the FCC database to create a CD-ROM with all call signs issued in the United States. A copy of the CD-ROM is carried on board the International Space Station and one was also aboard the Russian Mir space station. QRZ.com is one of the most recognized websites for amateur radio enthusiasts. Information is pulled directly from the FCC database and from on-line databases of other nations, when available. Registered users can edit their data for accuracy and currency and many list additional information about their station, antennas and other life interests.
Etymology
QRZ, the name of the website, is the "Q" code for "Who is calling me?" and corresponds to the site's purpose of assisting amateur radio operators with the lookup of ham radio call signs from every country in the world.
Features
The website features a personal web page where registered amateur radio operators can post pictures of their ham radio shack, tell facts about themselves, and post their email and postal addresses for other radio amateurs to send their QSL cards and list any equipment they would like to sell. The website also features online discussion forums.
References
External links
Amateur radio
Online databases |
https://en.wikipedia.org/wiki/Subterminal%20object | In category theory, a branch of mathematics, a subterminal object is an object X of a category C with the property that every object of C has at most one morphism into X. If X is subterminal, then the pair of identity morphisms (1X, 1X) makes X into the product of X and X. If C has a terminal object 1, then an object X is subterminal if and only if it is a subobject of 1, hence the name. The category of categories with subterminal objects and functors preserving them is not accessible.
References
External links
Category theory |
https://en.wikipedia.org/wiki/Largest%20fungal%20fruit%20bodies | The largest mushrooms and conks are the largest known individual fruit bodies. These are known as sporocarps, or, more specifically, basidiocarps and ascocarps for the Basidiomycota and Ascomycota respectively. These fruit bodies have a wide variety of morphologies, ranging from the typical mushroom shape, to brackets (conks), puffballs, cup fungi, stinkhorns, crusts and corals. Many species of fungi, including yeasts, moulds and the fungal component of lichens, do not form fruit bodies in this sense, but can form visible presences such as cankers. Individual fruit bodies need not be individual biological organisms, and extremely large single organisms can be made up of a great many fruit bodies connected by networks of mycelia (including the "humongous fungus", a single specimen of Armillaria solidipes) can cover a very large area.
The largest identified fungal fruit body in the world is a specimen of Phellinus ellipsoideus (formerly Fomitiporia ellipsoidea). The species was discovered in 2008 by Bao-Kai Cui and Yu-Cheng Dai in Fujian Province, China. In 2011, the two of them published details of extremely large fruit body of the species that they had found on Hainan Island. The specimen, which was 20 years old, was estimated to weigh between . This was markedly larger than the previously largest recorded fungal fruit body, a specimen of Rigidoporus ulmarius found in the United Kingdom that had a circumference of .
List
Unidentified specimens
Two large specimens are excluded from the list above. The first, a polypore photographed in 1903 at Yeerongpilly, Brisbane, Queensland, measured about in width by top to bottom, emerging from a tree about two thick. It was sturdy enough to support the weight of two average women.
The second is more speculative. Somewhere in his world travels, writer/naturalist/explorer Ivan T. Sanderson encountered reports of a species of fungi which "weigh a ton, and upheave large trees". The earliest report appears to have been in the |
https://en.wikipedia.org/wiki/Analysis%20of%20similarities | Analysis of similarities (ANOSIM) is a non-parametric statistical test widely used in the field of ecology. The test was first suggested by K. R. Clarke as an ANOVA-like test, where instead of operating on raw data, operates on a ranked dissimilarity matrix.
Given a matrix of rank dissimilarities between a set of samples, each belonging to a single site (e.g. a single treatment group), the ANOSIM tests whether we can reject the null hypothesis that the similarity between sites is greater than or equal to the similarity within each site.
The test statistic R is calculated in the following way:
where B is the average of rank similarities of pairs of samples (or replicates) originating from different sites, W is the average of rank similarity of pairs among replicates within sites, and M = n(n − 1)/2 where n is the number of samples.
The test statistic R is constrained between the values −1 to 1, where positive numbers suggest more similarity within sites and values close to zero represent no difference between sites and within sites similarities. Negative R values suggest more similarity between sites than within sites and may raise the possibility of wrong assignment of samples to sites.
For the purpose of hypothesis testing, where the null hypothesis is that the similarities within sites are smaller or equal to the similarities between sites, the R statistic is usually compared to a set of R′ values that are achieved by means of randomly shuffling site labels between the samples and calculating the resulting R′, repeated many times. The percent of times that the actual R surpassed the permutations derived R′ values is the p-value for the actual R statistic.
Ranking of dissimilarity in ANOSIM and NMDS (non-metric multidimensional scaling) go hand in hand.
Combining both methods complement visualisation of group differences along with significance testing.
ANOSIM is implemented in several statistical software including PRIMER, the R Vegan package and PAST. |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.