source stringlengths 31 203 | text stringlengths 28 2k |
|---|---|
https://en.wikipedia.org/wiki/Versatile%20Video%20Coding | Versatile Video Coding (VVC), also known as H.266, ISO/IEC 23090-3, and MPEG-I Part 3, is a video compression standard finalized on 6 July 2020, by the Joint Video Experts Team (JVET), a joint video expert team of the VCEG working group of ITU-T Study Group 16 and the MPEG working group of ISO/IEC JTC 1/SC 29. It is the successor to High Efficiency Video Coding (HEVC, also known as ITU-T H.265 and MPEG-H Part 2). It was developed with two primary goalsimproved compression performance and support for a very broad range of applications.
Concept
In October 2015, the MPEG and VCEG formed the Joint Video Exploration Team (JVET) to evaluate available compression technologies and study the requirements for a next-generation video compression standard. The new standard has about 50% better compression rate for the same perceptual quality compared to HEVC, with support for lossless and subjectively lossless compression. It supports resolutions ranging from very low resolution up to 4K and 16K as well as 360° videos. VVC supports YCbCr 4:4:4, 4:2:2 and 4:2:0 with 8–10 bits per component, BT.2100 wide color gamut and high dynamic range (HDR) of more than 16 stops (with peak brightness of 1000, 4000 and 10000 nits), auxiliary channels (for depth, transparency, etc.), variable and fractional frame rates from 0 to 120 Hz and higher, scalable video coding for temporal (frame rate), spatial (resolution), SNR, color gamut and dynamic range differences, stereo/multiview coding, panoramic formats, and still-picture coding. Work on high bit depth support (12 to 16 bits per component) started in October 2020 and was included in the second edition published in 2022. Encoding complexity of several times (up to ten times) that of HEVC is expected, depending on the quality of the encoding algorithm (which is outside the scope of the standard). The decoding complexity is about twice that of HEVC.
VVC development has been made using the VVC Test Model (VTM), a reference software codebase t |
https://en.wikipedia.org/wiki/Multimodal%20sentiment%20analysis | Multimodal sentiment analysis is a technology for traditional text-based sentiment analysis, which includes modalities such as audio and visual data. It can be bimodal, which includes different combinations of two modalities, or trimodal, which incorporates three modalities. With the extensive amount of social media data available online in different forms such as videos and images, the conventional text-based sentiment analysis has evolved into more complex models of multimodal sentiment analysis, which can be applied in the development of virtual assistants, analysis of YouTube movie reviews, analysis of news videos, and emotion recognition (sometimes known as emotion detection) such as depression monitoring, among others.
Similar to the traditional sentiment analysis, one of the most basic task in multimodal sentiment analysis is sentiment classification, which classifies different sentiments into categories such as positive, negative, or neutral. The complexity of analyzing text, audio, and visual features to perform such a task requires the application of different fusion techniques, such as feature-level, decision-level, and hybrid fusion. The performance of these fusion techniques and the classification algorithms applied, are influenced by the type of textual, audio, and visual features employed in the analysis.
Features
Feature engineering, which involves the selection of features that are fed into machine learning algorithms, plays a key role in the sentiment classification performance. In multimodal sentiment analysis, a combination of different textual, audio, and visual features are employed.
Textual features
Similar to the conventional text-based sentiment analysis, some of the most commonly used textual features in multimodal sentiment analysis are unigrams and n-grams, which are basically a sequence of words in a given textual document. These features are applied using bag-of-words or bag-of-concepts feature representations, in which words or c |
https://en.wikipedia.org/wiki/CMOS%20amplifier | CMOS amplifiers (complementary metal–oxide–semiconductor amplifiers) are ubiquitous analog circuits used in computers, audio systems, smartphones, cameras, telecommunication systems, biomedical circuits, and many other systems. Their performance impacts the overall specifications of the systems. They take their name from the use of MOSFETs (metal–oxide–semiconductor field-effect transistors) as opposite to bipolar junction transistors (BJTs). MOSFETs are simpler to fabricate and therefore less expensive than BJT amplifiers, still providing a sufficiently high transconductance to allow the design of very high performance circuits. In high performance CMOS (complementary metal–oxide–semiconductor) amplifier circuits, transistors are not only used to amplify the signal but are also used as active loads to achieve higher gain and output swing in comparison with resistive loads.
CMOS technology was introduced primarily for digital circuit design. In the last few decades, to improve speed, power consumption, required area, and other aspects of digital integrated circuits (ICs), the feature size of MOSFET transistors has shrunk (minimum channel length of transistors reduces in newer CMOS technologies). This phenomenon predicted by Gordon Moore in 1975, which is called Moore’s law, and states that in about each 2 years, the number of transistors doubles for the same silicon area of ICs. Progress in memory circuits design is an interesting example to see how process advancement have affected the required size and their performance in the last decades. In 1956, a 5 MB Hard Disk Drive (HDD) weighed over a ton, while these days having 50000 times more capacity with a weight of several tens of grams is very common.
While digital ICs have benefited from the feature size shrinking, analog CMOS amplifiers have not gained corresponding advantages due to the intrinsic limitations of an analog design—such as the intrinsic gain reduction of short channel transistors, which affects th |
https://en.wikipedia.org/wiki/International%20Council%20of%20the%20Aeronautical%20Sciences | The International Council of the Aeronautical Sciences (ICAS) is a worldwide institution, established as an international forum for individual national aeronautical professional associations.
History
It was formed on 29 January 1957 at a conference in the US. The first ICAS Congress was held in Spain in 1958. Frank Wattendorf, of AGARD, was the first Director.
A second meeting was held in Paris, with Hugh Latimer Dryden of the National Advisory Committee for Aeronautics, and representatives from ONERA (Office National d'Etudes et de Recherches Aérospatiales), the Royal Aeronautical Society (RAeS), the WGL (now the Deutsche Gesellschaft für Luft- und Raumfahrt), the Association Française des Ingénieurs et Techniciens de l'Aéronautique (now the Association Aéronautique et Astronautique de France), and the Aeronautical Research Institute of Sweden.
Congress
It holds a biennial international congress in September. In 1986 it was held in London. In 2000 the congress was held in North Yorkshire. The 2018 Congress was held by Associação Brasileira de Engenharia e Ciências Mecânicas (ABCM) in Brazil. The 2020 congress will be held in Shanghai, China.
Presidents
Raymond Bisplinghoff 1978
Boris Laschka 1986
Paolo Santini 1990
Murray Scott 2013
Christian Mari 2015
Susan Ying 2017
Shinji Suzuki 2017
Dimitri Mavris 2023
Structure
The secretariat of ICAS is at Deutsche Gesellschaft für Luft- und Raumfahrt (DLR) in Bonn. It was first headquartered at the American Institute of Aeronautics and Astronautics (AIAA), the DLR from 1978, the RAeS in 1986, the Nederlandse Vereniging voor Luchtvaarttechniek from 1990, the AAAF (Association Aéronautique et Astronautique de France) from 1997, then Sweden from 2002, and Germany from January 2011.
See also
Council of European Aerospace Societies (CEAS)
Fédération Aéronautique Internationale (FAI)
International Astronautical Federation
International Energy Forum
References
External links
ICAS
ICAS 2020
1957 in aviation
Ae |
https://en.wikipedia.org/wiki/CD28%20family%20receptor | CD28 family receptors are a group of regulatory cell surface receptors expressed on immune cells. The CD28 family in turn is a subgroup of the immunoglobulin superfamily.
Two family members, CD28 and ICOS, act as positive regulators of T cell function while another three, BTLA, CTLA-4 and PD-1 act as inhibitors. Ligands for the CD28 receptor family include B7 family proteins.
CD28 receptors play a role in the development and proliferation of T cells. The CD28 receptors enhance signals from the T cell receptors (TCR) in order to stimulate an immune response and an anti-inflammatory response on regulatory T cells. Through the promotion of T cell function, CD28 receptors allow effector T cells to combat regulatory T cell-mediated suppression from adaptive immunity. CD28 receptors also elicit the prevention of spontaneous autoimmunity.
Function
CD28 receptors aid in other T cell processes such as cytoskeletal remodeling, production of cytokines and chemokines and intracellular biochemical reactions (i.e. phosphorylation, transcriptional signaling, and metabolism) that are key for T cell proliferation and differentiation. Ligation of CD28 receptors causes epigenetic, transcriptional and post-translational alterations in T cells. Specifically, CD28 costimulation controls many aspects within T cells, one being the expression of proinflammatory cytokine genes. A particular cytokine gene encodes for IL-2, which influences T cell proliferation, survival, and differentiation. The absence of CD28 costimulation results in the loss of IL-2 production causing the T cells to be anergic. Additionally, CD28 ligation causes arginine-methylation for many proteins. CD28 also drives transcription within T cells and produce signals that lead to IL-2 production and Bcl-xL regulation, an antiapoptotic protein, which are essential for T cell survival. CD28 receptors can be seen on 80% of human CD4+ and 50% of CD8+ T cells, in which this percentage decreases with age.
Clinical signifi |
https://en.wikipedia.org/wiki/Foldering | Foldering is the practice of communicating via messages saved to the "drafts" folder of an email or other electronic messaging account that is accessible by multiple people. The messages are never actually sent.
Foldering has been described as a digital equivalent of a dead drop.
History
Foldering was reportedly used by al-Qaeda at least as early as 2005 and it has also been used by drug cartels.
Notable cases
In 2012, David Petraeus was reported to have used foldering to communicate with Paula Broadwell.
In June 2018, Greg Andres cited Paul Manafort's use of foldering as evidence that Manafort engaged in deceptive behaviours.
Depictions in popular culture
Foldering was depicted in the movie Traitor (2008). The use of foldering can also be seen in Season 3 of Tom Clancy's Jack Ryan.
See also
Dead drop
Tradecraft
References
Data security
Espionage techniques |
https://en.wikipedia.org/wiki/Torservers.net | torservers.net is an independent network of non-profit organisations that provide nodes to the Tor anonymity network. The network started in June 2010 and currently transfers up to 7.4GB/s (~59.2Gb/s) of exit node traffic as of May 2022.
Torservers.net is known for operating servers with high network bandwidth and running them as exit nodes in the Tor network, which helps increase its speed and capacity. The group additionally helps provide lawyers for relay operators along with arranging operator meetups.
Funding
While Tor is free software that anyone can run, successful operation of Tor nodes may require technical expertise, access to high-bandwidth, and can involve legal complications in some jurisdictions. The Torservers.net network accepts financial donations as a way to sponsor additional nodes.
Bavarian Raid
On June 20, 2018, Bavarian police raided the home of the board members of the German non-profit Zwiebelfreunde, "Friends of the Onion," (part of torservers.net). Zwiebelfreunde helps collect donations from Europe for various non-commercial providers such as Riseup.net.
The police claim the raid was prompted by a blog post from an unrelated activist that promised violence against an upcoming Alternative for Germany convention in Augsburg. The blog post was published on a website that used a Riseup.net e-mail address. Riseup Collective is based in Seattle in the United States, and reported publicly that Zwiebelfreunde does not run its service.
On August 23 the German court at Landgericht München ruled that the raid and seizures was illegal. The hardware and documentation seized had been kept under seal, and purportedly were neither analyzed nor evaluated by the Bavarian police.
Members
Associated Whistleblowing Press (Belgium)
Access Now (USA)
Calyx Institute (USA)
Zwiebelfreunde e.V. (Germany)
Hart Voor Internetvrijheid (Netherlands)
Library Freedom Project (USA)
Noisebridge (USA)
External links
Official website
References
Tor (anon |
https://en.wikipedia.org/wiki/Galeas%20per%20montes | Galeas per montes (galleys across mountains) is the name given to a feat of military engineering made between December 1438 and April 1439 by the Republic of Venice, when several Venetian ships, including galleys and frigates were transported from the Adriatic Sea to Lake Garda. The operation required towing the ships upstream on the river Adige until Rovereto, then transporting the fleet by land to Torbole, on the Northern shores of the lake. The second leg of the journey was the most remarkable achievement, requiring a land journey 20 km through the Loppio Lake and the narrow .
Context
The Republic of Venice was at the time a power in the Mediterranean and, in the 15th century, it began an expansion phase towards the mainland of the current Lombardia and Veneto regions both through military conquest (e.g. Padua) or spontaneous "dedication", as in the case of Vicenza. The city of Brescia, located West of Lake Garda, allied with the Republic of Venice to escape the Duchy of Milan on November 20, 1426.
In 1438, the Duke of Milan Filippo Maria Visconti waged war against the Republic of Venice and, through a series of lucky victories, took control of Lombard lands up to the southern shores of Lake Garda. At the same time, the city of Brescia was under siege by the mercenary condottiero Niccolò Piccinino, on the Duke of Milan's payroll, and called on the Venetian Senate for assistance.
Piccinino took control of the entire Southern sector of the lake, so the Venetian warlord Gattamelata (Erasmo da Narni) could only access the lake its Northerns shore, namely Torbole or Riva. The Milanese army was also fortified in the castles of Peschiera del Garda and Desenzano, making a head-on clash too expensive. To avoid this problem, the Republic of Venice decided to prepare a military plan that would allow its troops (and navy) to surprise the Visconti army entering the lake from its Northern shore.
On December 1st 1438, after a very long session, the Republic's Minor Counci |
https://en.wikipedia.org/wiki/Ace%20Stream | Ace Stream is a peer-to-peer multimedia streaming protocol, built using BitTorrent technology. Ace Stream has been recognized by sources as a potential method for broadcasting and viewing bootlegged live video streams. The protocol functions as both a client and a server. When users stream a video feed using Ace Stream, they are simultaneously downloading from peers and uploading the same video to other peers.
History
Ace Stream began under the name TorrentStream as a pilot project to use BitTorrent technology to stream live video. In 2013 TorrentStream, was re-released under the name ACE Stream.
References
Computer networking
Applications of distributed computing
Cloud storage
Digital television
Distributed algorithms
Distributed data storage
Distributed data storage systems
File sharing networks
Film and video technology
Internet broadcasting
Streaming television
Multimedia
Peer-to-peer computing
Streaming media systems
Video hosting
Video on demand services |
https://en.wikipedia.org/wiki/Ultrawide%20formats | Ultrawide formats refers to photos, videos, and displays with aspect ratios greater than 2. There were multiple moves in history towards wider formats, including one by Disney, with some of them being more successful than others.
Cameras usually capture ultra-wide photos and videos using an anamorphic format lens, which shrinks the extended horizontal field-of-view (FOV) while saving on film or disk.
Historic displays
Before computer monitors became a separate product line, televisions were used as displays for computers such as Timex Sinclair 1000.
4:3
4:3 was the aspect ratio used by 35 mm silent films. By having televisions match this aspect ratio, movies originally filmed in 4:3 could be satisfactorily viewed on standard-definition television (SDTV).
Monitors around the turn of the century would often use resolutions like 640x480, 800x600, 1024x768 or 1200x900 in an 4:3 aspect ratio.
NTSC (480i)
National Television System Committee (NTSC) broadcasts were analogue and intended for analogue NTSC displays. The standard was developed and implemented by the NTSC in the United States in 1954. It also saw widespread international adoption by trade partners of the US. When converted to the Digital Video format, DV NTSC has a 3:2 aspect ratio, a resolution of 720x480i, and a refresh rate of 60 Hz.
PAL (576i)
Phase Alternating Line (PAL) broadcasts were analogue broadcasts meant for PAL analogue displays. The standard was developed in 1967 by the United Kingdom and Germany, and implemented in most countries. When converted to Digital Video format, DV PAL has a 5:4 resolution of 720×576i running at 50 Hz.
32:27
32:27 was originally developed for compressed video storage in cameras, meant to be displayed in anamorphic x1.5 as 16:9. 640×540i was such a 32:27 resolution running at 50 Hz and 100 Hz, meant for cameras.
Panasonic's DVCPRO HD with a resolution of 1280×1080i was latest in the line of 32:27 video formats for cameras. Hitachi's 42" and 50" 1280×1080i tel |
https://en.wikipedia.org/wiki/Self-adaptive%20mechanisms | Self-adaptive mechanisms, sometimes simply called adaptive mechanisms, in engineering, are underactuated mechanisms that can adapt to their environment. One of the most well-known example of this type of mechanisms are underactuated fingers, grippers, and robotic hands. Contrary to standard underactuated mechanisms where the motion is governed by the dynamics of the system, the motion of self-adaptive mechanisms is generally constrained by compliant elements cleverly located in the mechanisms.
Definition
Underactuated mechanisms have a lower number of actuators than the number of degrees of freedom (DOF). In a two-dimensional plane, a mechanism can have up to three DOF (two translations, one rotation), and in three-dimensional Euclidean space, up to six (three translations, three rotations). In the case of self-adaptive mechanisms, the lack of actuators is compensated by passive elements that constrain the motion of the system. Springs are a good example of such elements, but other can be used depending on the type of mechanisms.
One of the earliest example of self-adaptive mechanism is the flapping wing proposed by Leonardo da Vinci in the Codex Atlanticus.
Underactuated hands
The first commonly known underactuated finger was the Soft-Gripper designed by Shigeo Hirose in the late 1970s. The most common type of transmission mechanisms used in self-adaptive hands are linkages and tendons.
Kinetostatics
Underactuated fingers and hands are usually analyzed with respect to their kinetostatics (negligible kinetic energy, static analysis of a mechanism in motion) rather than the dynamics of the system, as the kinetic energy of these systems is generally negligible compared to the potential energy stored into the passive elements. The forces applied by each phalanx of an underactuated finger can be computed with the following expression:
where F is the vector made of the forces applied, J is the Jacobian matrix of the finger, T* is the transmission matrix, and t is |
https://en.wikipedia.org/wiki/Models%20of%20consciousness | Models of consciousness are used to illustrate and aid in understanding and explaining distinctive aspects of consciousness. Sometimes the models are labeled theories of consciousness. Anil Seth defines such models as those that relate brain phenomena such as fast irregular electrical activity and widespread brain activation to properties of consciousness such as qualia. Seth allows for different types of models including mathematical, logical, verbal and conceptual models.
Neuroscience
Neural correlates of consciousness
The Neural correlates of consciousness (NCC) formalism is used as a major step towards explaining consciousness. The NCC are defined to constitute the minimal set of neuronal events and mechanisms sufficient for a specific conscious percept, and consequently sufficient for consciousness. In this formalism, consciousness is viewed as a state-dependent property of some undefined complex, adaptive, and highly interconnected biological system.
Dehaene–Changeux model
The Dehaene–Changeux model (DCM), also known as the global neuronal workspace or the global cognitive workspace model is a computer model of the neural correlates of consciousness programmed as a neural network. Stanislas Dehaene and Jean-Pierre Changeux introduced this model in 1986. It is associated with Bernard Baars's Global workspace theory for consciousness.
Electromagnetic theories of consciousness
Electromagnetic theories of consciousness propose that consciousness can be understood as an electromagnetic phenomenon that occurs when a brain produces an electromagnetic field with specific characteristics. Some electromagnetic theories are also quantum mind theories of consciousness.
Orchestrated objective reduction
Orchestrated objective reduction (Orch-OR) model is based on the hypothesis that consciousness in the brain originates from quantum processes inside neurons, rather than from connections between neurons (the conventional view). The mechanism is held to be |
https://en.wikipedia.org/wiki/List%20of%20engineering%20physics%20schools | Engineering physics (or engineering science) can be studied at the bachelors, masters and Ph.D. levels at many universities, typically offered in a partnership between engineering faculties and the departments of physics.
Canada
In Canada, the Canadian Engineering Accreditation Board is responsible for accrediting undergraduate engineering physics programs, graduate study in aerospace engineering is also available at several Canadian post-secondary institutions, though Canadian post-graduate engineering programs do not require accreditation.
University of Alberta - Engineering Physics
University of British Columbia - Engineering Physics
Carleton University - Engineering Physics
Dalhousie University - Engineering Physics
McMaster University - Engineering Physics
Queen's University - Engineering Physics
Royal Military College of Canada - Engineering Physics
University of Saskatchewan - Engineering Physics
Simon Fraser University - Engineering Science
University of Toronto - Engineering Science
Only undergraduate engineering programs in Canada are accredited, and this is done by the Canadian Engineering Accreditation Board.
References
Engineering education
Lists of engineering schools |
https://en.wikipedia.org/wiki/Universal%20dielectric%20response | In physics and electrical engineering, the universal dielectric response, or UDR, refers to the observed emergent behaviour of the dielectric properties exhibited by diverse solid state systems. In particular this widely observed response involves power law scaling of dielectric properties with frequency under conditions of alternating current, AC. First defined in a landmark article by A. K. Jonscher in Nature published in 1977, the origins of the UDR were attributed to the dominance of many-body interactions in systems, and their analogous RC network equivalence.
The universal dielectric response manifests in the variation of AC Conductivity with frequency and is most often observed in complex systems consisting of multiple phases of similar or dissimilar materials. Such systems, which can be called heterogenous or composite materials, can be described from a dielectric perspective as a large network consisting of resistor and capacitor elements, known also as an RC network. At low and high frequencies, the dielectric response of heterogeneous materials is governed by percolation pathways. If a heterogeneous material is represented by a network in which more than 50% of the elements are capacitors, percolation through capacitor elements will occur. This percolation results in conductivity at high and low frequencies that is directly proportional to frequency. Conversely, if the fraction of capacitor elements in the representative RC network (Pc) is lower than 0.5, dielectric behavior at low and high frequency regimes is independent of frequency. At intermediate frequencies, a very broad range of heterogeneous materials show a well-defined emergent region, in which power law correlation of admittance to frequency is observed. The power law emergent region is the key feature of the UDR. In materials or systems exhibiting UDR, the overall dielectric response from high to low frequencies is symmetrical, being centered at the middle point of the emergent region, whic |
https://en.wikipedia.org/wiki/Comparison%20matrix | In linear algebra, let be a complex matrix. The comparison matrix of complex matrix A is defined as
See also
Hurwitz matrix
P-matrix
Perron–Frobenius theorem
Z-matrix
L-matrix
M-matrix
H-matrix (iterative method)
References
Matrices |
https://en.wikipedia.org/wiki/H-matrix%20%28iterative%20method%29 | In mathematics, an H-matrix is a matrix whose comparison matrix is an M-matrix. It is useful in iterative methods.
Definition: Let be a complex matrix. Then comparison matrix M(A) of complex matrix A is defined as where for all and for all . If M(A) is a M-matrix, A is a H-matrix.
Invertible H-matrix guarantees convergence of Gauss–Seidel iterative methods.
See also
Hurwitz matrix
P-matrix
Perron–Frobenius theorem
Z-matrix
L-matrix
M-matrix
Comparison matrix
References
Matrices |
https://en.wikipedia.org/wiki/Ecological%20unit | Ecological units, comprise concepts such as population, community, and ecosystem as the basic units, which are at the basis of ecological theory and research, as well as a focus point of many conservation strategies. The concept of ecological units continues to suffer from inconsistencies and confusion over its terminology. Analyses of the existing concepts used in describing ecological units have determined that they differ in respects to four major criteria:
The questions as to whether they are defined statistically or via a network of interactions,
If their boundaries are drawn by topographical or process-related criteria,
How high the required internal relationships are,
And if they are perceived as "real" entities or abstractions by an observer.
A population is considered to be the smallest ecological unit, consisting of a group of individuals that belong to the same species. A community would be the next classification, referring to all of the population present in an area at a specific time, followed by an ecosystem, referring to the community and it's interactions with its physical environment. An ecosystem is the most commonly used ecological unit and can be universally defined by two common traits:
The unit is often defined in terms of a natural border (maritime boundary, watersheds, etc.)
Abiotic components and organisms within the unit are considered to be interlinked.
See also
Biogeographic realm
Ecoregion
Ecotope
Holobiont
Functional ecology
Behavior settings
Regional geology
References
Ecology |
https://en.wikipedia.org/wiki/Force-sensing%20capacitor | A force-sensing capacitor is a material whose capacitance changes when a force, pressure or mechanical stress is applied. They are also known as "force-sensitive capacitors". They can provide improved sensitivity and repeatability compared to force-sensitive resistors but traditionally required more complicated electronics.
Operation principle
Typical force-sensitive capacitors are examples of parallel plate capacitors. For small deflections, there is a linear relationship between applied force and change in capacitance, which can be shown as follows:
The capacitance, , equals , where is permeability, is the area of the sensor and is the distance between parallel plates. If the material is linearly elastic (so follows Hooks Law), then the displacement, due to an applied force , is , where is the spring constant. Combining these equations gives the capacitance after an applied force as:
, where is the separation between parallel plates when no force is applied.
This can be rearranged to:
Assuming that , which is true for small deformations where , we can simplify this to:
C
It follows that:
C
C where , which is constant for a given sensor.
We can express the change in capacitance as:
Production
SingleTact makes force-sensitive capacitors using moulded silicon between two layers of polyimide to construct a 0.35mm thick sensor, with force ranges from 1N to 450N. The 8mm SingleTact has a nominal capacitance of 75pF, which increases by 2.2pF when the rated force is applied. It can be mounted on many surfaces for direct force measurement.
Uses
Force-sensing capacitors can be used to create low-profile force-sensitive buttons. They have been used in medical imaging to map pressures in the esophagus and to image breast and prostate cancer.
References
Capacitors
Sensors |
https://en.wikipedia.org/wiki/Belgian%20Scientific%20Expedition | Belgian Scientific Expedition was a scientific survey of the Great Barrier Reef, conducted in 1967–1968.
The Belgian Scientific Expedition to the Great Barrier Reef was a seven month expedition beginning in 1967, sponsored by the University of Liege, Belgium, the Belgium Ministry of Education and the National Foundation for Scientific Research. It indirectly honoured the Great Barrier Reef Expedition of 1928–1929, which was led by Maurice Yonge and a large group of researchers from Europe. This earlier expedition had studied the northern Great Barrier Reef primarily around Low Isles Reef.
The 1967 expedition, led by Professor Albert Distèche took place between Lady Musgrave Island and Lizard Island off the coast of Queensland on the Great Barrier Reef. Seventy-five ship's crew, many researchers and guests were involved in the expedition. Its primary objective was to make scientific marine biology films. Ron Taylor, who would become famous for his films and diving work with sharks was one of the cinematographers hired to undertake the underwater filming using a 35mm motion picture camera.
The former British warship, the De Moor was utilised for the study by the Belgian Navy. Captain Wally Muller was contracted to guide the De Moor through the Swain Reefs and remain with the expedition, on his charter vessel, the Careelah.
Coral reef scientists participated in the study as time permitted. The ship would return to shore every 10 days. Among these scientists were David Barnes from the Townsville area and Robert Endean from the University of Queensland. Sir Maurice Yonge would also visit during this expedition, in recognition of his earlier work in 1928.
Later studies of the Reef would be conducted and published as part of Project Stellaroid, which surveyed coral reefs in the North Pacific Ocean and their damage by the Crown of Thorns starfish.
References
Marine biology
Great Barrier Reef |
https://en.wikipedia.org/wiki/Janez%20Lawson | Janez Yvonne Lawson Bordeaux (February 22, 1930 – November 24, 1990) was an American chemical engineer who became one of NASA's computers. She was the first African-American hired into a technical position at Jet Propulsion Laboratory. She programmed the IBM 701.
Early life and education
Lawson was born on February 22, 1930, in Santa Monica, California. Her parents were Hilliard Lawson and Bernice Lawson. She attended Belmont High School and graduated in 1948. Lawson completed a bachelor's degree in chemical engineering at the University of California, Los Angeles in 1952. She was a straight-A student and President of the Delta Sigma Theta sorority.
Career
Despite her qualifications, Lawson could not get work as a chemical engineer because of her race and gender. She saw an advertisement for a job as a computer in Pasadena. There was discussion about whether or not she should get the job, but Macie Roberts stood up for her. Lawson got the job, and in 1953 was one of the first Jet Propulsion Laboratory employees to be sent to a training course at IBM. Lawson was the first African-American hired into a technical position at Jet Propulsion Laboratory. She was promoted to mathematician in 1954. She became skilled at programming during the course, using a keypunch and learning speedcoding. Lawson lived in Los Angeles and would commute for over an hour to the Jet Propulsion Laboratory every day. Lawson joined the Ramo-Wooldridge Corporation in the late 1950s.
References
1930 births
1990 deaths
20th-century African-American women
20th-century African-American scientists
20th-century American engineers
20th-century American mathematicians
20th-century women engineers
20th-century women mathematicians
African-American engineers
African-American women engineers
American chemical engineers
American computer programmers
American women engineers
Delta Sigma Theta members
Human computers
Jet Propulsion Laboratory faculty
Mathematicians from California
NASA people
Engineers |
https://en.wikipedia.org/wiki/Anomali | Anomali Inc. is an American cybersecurity company that develops and provides threat intelligence products. In 2023, the company moved into providing Security Analytics powered by AI.
History
Anomali was founded in 2013 under the name ThreatStream, by Greg Martin and Colby DeRodeff. At that time, the company's products provided filtering and customization options to give companies visibility into indicators of compromise (IOCs). In 2016, company rebranded as Anomali and introduced new products and a new approach to threat intelligence. This included providing SaaS and on-premise platforms that customers could use to upload their logs, and then have Anomali match data against threat intelligence for existing IOCs.
In 2013, the company launched the first version of ThreatStream, a threat intelligence platform (TIP), uses different sources to track known threats, monitor and detect security breaches. In 2016, when the company became known as Anomali, it launched its second product, Anomali, which later became Anomali Match, an enterprise threat detection service.
By 2018, Anomali had received $96.3 million in funding from 11 investors, including Paladin Capital Group, Institutional Venture Partners (IVP), GV (formerly Google Ventures), General Catalyst, Telstra Ventures, and Lumina Capital. The company works with government and business organizations such as the Bank of England, Citigroup, and Alaska Airlines.
In 2019, Anomali introduced Anomali Lens, a web-browser extension that highlights and collects relevant threat data from web pages. The data is added to ThreatStream and matched with internal network events using Anomali’s Match platform. Since being founded, Anomali has collaborated with partners spanning channel resellers, managed security services providers (MSSPs), systems integrators, and Commercial Threat Intelligence Feed providers to build out the Anomali Preferred Partner Store (Anomali APP Store). Anomali has established a collaborative relati |
https://en.wikipedia.org/wiki/PayMe | PayMe (officially known as PayMe from HSBC) is a mobile payment service from HSBC, currently available only for Hong Kong users with local phone numbers and banks. Users can pay businesses, transfer money to one another using a mobile app, linked to their credit card or (any local) bank account.
As of Nov 2022, the service had around 2.9 million active users, out of a Hong Kong population of 7.3 million, with a market share of 50%.
History
On its launch on 8 February 2017, heavy traffic meant that some users were unable to register, leading to media dubbing the service "Play Me". Topping up was only possible from a Visa or Mastercard credit card, with a limit of HK$10,000 per month.
In April 2018, the app won the FinTech Grand Award in the official Hong Kong ICT Awards.
From June 2018, users were able to link to a local bank account, with monthly top-up limits raised to HK$30,000, or HK$50,000 with a verified residential address. At the same time, PayMe launched an online shopping payment service, in collaboration with HKTVmall.
In July 2019, PayMe introduced the support of FPS (Faster Payment System) for transfer and top-up, along with a complete redesign of the app. At the same time, PayMe lowered the top-up limit for credit cards to HK$2,000 per month.
In May 2022, PayMe was selected to be one of the eligible electronic platforms for the consumption voucher scheme.
Features
PayMe was introduced as a standalone mobile app, offering P2P social payment. Users register via a Facebook account or Hong Kong mobile phone number and authorise access to a credit card or local bank account (not necessarily an HSBC account), from which the balance can be topped up, and a bank account to receive money.
There are no transaction fees.
When the PayMe account balance is too low to make a given transaction, the app automatically withdraws the necessary funds from the registered bank account or card.
The social networking interaction component allows users to send and r |
https://en.wikipedia.org/wiki/Network-based%20diffusion%20analysis | Network-based diffusion analysis (NBDA) is a statistical tool to detect and quantify social transmission of information or a behaviour in social networks (SNA, etc.). NBDA assumes that social transmission of a behavior follows the social network of associations or interactions among individuals, since individuals who spend a lot of time together, or who interact more have more opportunity to learn from each other. Therefore, NBDA infers social transmission if the spread of a novel behavior follows the social network of a population. NBDA thus allows the study of social learning to be linked to animal behavior research that uses social network analysis. NBDA was introduced by Franz & Nunn and further developed by Hoppitt, Boogert, & Laland.
Implementation
NBDA requires prior knowledge about the underlying social network of a population. In an observational study, the order (or timing) at which individuals in the population acquire a behaviour or information is recorded. NBDA then tests whether the spread of information or behaviour is explained by the previously determined network or not. Because more closely associated individuals are more likely to interact with each other, information is assumed to travel along social ties. If there is a good match between the diffusion of information and the underlying network social transmission is assumed. Otherwise, it is assumed that information was asocially acquired (e.g. trial and error, mistakes, etc.).
Application
NBDA does not only serve as a tool for the detection of social learning, but also allows the estimation of the strength of the social transmission effect. In addition, several individual-level variables can be included in the analysis, which have potential influence on an individual's learning rate (e.g. gender, rank or age), and can also be used to model the effect of, and statistically control for potential ecological and genetic influences. NBDA has been successfully used in a number of studies to identi |
https://en.wikipedia.org/wiki/Induction-induction | In intuitionistic type theory (ITT), some discipline within mathematical logic, induction-induction is for simultaneously declaring some inductive type and some inductive predicate over this type.
An inductive definition is given by rules for generating elements of some type. One can then define some predicate on that type by providing constructors for forming the elements of the predicate , such inductively on the way the elements of the type are generated. Induction-induction generalizes this situation since one can simultaneously define the type and the predicate, because the rules for generating elements of the type are allowed to refer to the predicate .
Induction-induction can be used to define larger types including various universe constructions in type theory. and limit constructions in category/topos theory.
Example 1
Present the type as having the following constructors , note the early reference to the predicate :
and-simultaneously present the predicate as having the following constructors :
if and then
if and and then .
Example 2
A simple common example is the Universe à la Tarski type former. It creates some inductive type and some inductive predicate . For every type in the type theory (except itself!), there will be some element of which may be seen as some code for this corresponding type ; The predicate inductively encodes each possible type to the corresponding element of ; and constructing new codes in will require referring to the decoding-as-type of earlier codes , via the predicate .
See also
Induction-recursion – for simultaneously declaring some inductive type and some recursive function on this type .
References
External links
A list of Peter Dybjer's publications on induction and induction-recursion
Type theory |
https://en.wikipedia.org/wiki/Broadcast%2C%20unknown-unicast%20and%20multicast%20traffic | Broadcast, unknown-unicast and multicast traffic (BUM traffic) is network traffic transmitted using one of three methods of sending data link layer network traffic to a destination of which the sender does not know the network address. This is achieved by sending the network traffic to multiple destinations on an Ethernet network. As a concept related to computer networking, it includes three types of Ethernet modes: broadcast, unicast and multicast Ethernet. BUM traffic refers to that kind of network traffic that will be forwarded to multiple destinations or that cannot be addressed to the intended destination only.
Overview
Broadcast traffic is used to transmit a message to any reachable destination in the network without the need to know any information about the receiving party. When broadcast traffic is received by a network switch it is replicated to all ports within the respective VLAN except the one from which the traffic comes from.
Unknown-unicast traffic happens when a switch receives unicast traffic intended to be delivered to a destination that is not in its forwarding information base. In this case the switch marks the frame for flooding and sends it to all forwarding ports within the respective VLAN. Forwarding this type of traffic can create unnecessary traffic that leads to poor network performance or even a complete loss of network service. This flooding of packets is known as a unicast flooding.
Multicast traffic allows a host to contact a subset of hosts or devices joined into a group. This causes the message to be broadcast when no group management mechanism is present. Flooding BUM frames is required in transparent bridging and in a data center context this does not scale well causing poor performance.
BUM traffic control
Throttling
One issue that may arise is that some network devices cannot handle high rates of broadcast, unknown-unicast or multicast traffic. In such cases, it is possible to limit the BUM traffic for specific ports in |
https://en.wikipedia.org/wiki/List%20of%20irreducible%20Tits%20indices | In the mathematical theory of linear algebraic groups, a Tits index (or index) is an object used to classify semisimple algebraic groups defined over a base field k, not assumed to be algebraically closed. The possible irreducible indices were classified by Jacques Tits, and this classification is reproduced below. (Because every index is a direct sum of irreducible indices, classifying all indices amounts to classifying irreducible indices.)
Organization of the list
An index can be represented as a Dynkin diagram with certain vertices drawn close to each other (the orbit of the vertices under the *-action of the Galois group of k) and with certain sets of vertices circled (the orbits of the non-distinguished vertices under the *-action). This representation captures the full information of the index except when the underlying Dynkin diagram is D4, in which case one must distinguish between an action by the cyclic group C3 or the permutation group S3.
Alternatively, an index can be represented using the name of the underlying Dykin diagram together with additional superscripts and subscripts, to be explained momentarily. This representation, together with the labeled Dynkin diagram described in the previous paragraph, captures the full information of the index.
The notation for an index is of the form gX, where
X is the letter of the underlying Dynkin diagram (A, B, C, D, E, F, or G),
n is the number of vertices of the Dynkin diagram,
r is the relative rank of the corresponding algebraic group,
g is the order of the quotient of the absolute Galois group that acts faithfully on the Dynkin diagram (so g = 1, 2, 3, or 6), and
t is either
the degree of a certain division algebra (that is, the square root of its dimension) arising in the construction of the algebraic group when the group is of classical type (A, B, C, or D), in which case t is written in parentheses, or
the dimension of the anisotropic kernel of the algebraic group when the group is of excepti |
https://en.wikipedia.org/wiki/Food%20powder | Food powder or powdery food is the most common format of dried solid food material that meets specific quality standards, such as moisture content, particle size, and particular morphology. Common powdery food products include milk powder, tea powder, cocoa powder, coffee powder, soybean flour, wheat flour, and chili powder. Powders are particulate discrete solid particles of size ranging from nanometres to millimetres that generally flow freely when shaken or tilted. The bulk powder properties are the combined effect of particle properties by the conversion of food products in solid state into powdery form for ease of use, processing and keeping quality. Various terms are used to indicate the particulate solids in bulk, such as powder, granules, flour and dust, though all these materials can be treated under powder category. These common terminologies are based on the size or the source of the materials.
The particle size, distribution, shape and surface characteristics and the density of the powders are highly variable and depend on both the characteristics of the raw materials and processing conditions during their formations. These parameters contribute to the functional properties of powders, including flowability, packaging density, ease of handling, dust forming, mixing, compressibility and surface activity.
Characteristics
Microstructure
Food powder may be amorphous or crystalline in their molecular level structure. Depending on the process applied, the powders can be produced in either of these forms. Powders in crystalline state possess defined molecular alignment in the long-range order, while amorphous state is disordered, more open and porous. Common powders found in crystalline states are salts, sugars and organic acids. Meanwhile, many food products such as dairy powders, fruit and vegetable powders, honey powders and hydrolysed protein powders are normally in amorphous state. The properties of food powders including their functionality and their |
https://en.wikipedia.org/wiki/Wireless%20data%20center | A Wireless Data center is a type of data center that uses wireless communication technology instead of cables to store, process and retrieve data for enterprises. The development of Wireless Data centers arose as a solution to growing cabling complexity and hotspots. The wireless technology was introduced by Shin et al., who replaced all cables with 60 GHz wireless connections at the Cayley data center.
Motivation
Most DCs deployed today can be classified as wired DCs because they use copper and optical fiber cables to handle intra- and inter-rack connections in the network. This approach has two problems, cable complexity and hotspots. Hotspots, also known as hot servers, are servers that generate high traffic compared to others in the network and they might become bottlenecks of the system. To address these problems several researchers propose the use of wireless communication into data center networks, to either augment existing wired data centers, or to realize a pure wireless data center
Although cable complexity at first seems like an esthetical problem, it can affect a DC in different ways. First, a significant manual effort is necessary to install and manage these cables. Apart from that, cables can additionally affect data center cooling. Finally, cables take up space, which could be used to add more servers. The use of wireless technologies could reduce the cable complexity and avoid the problems cited before, moreover, it would allow for automatic configurable link establishment between nodes with minimum effort.
Wireless links can be rearranged dynamically which makes it possible to perform adaptive topology adjustment. This means that the network can be rearranged to fulfil the real-time traffic demands of hotspots, thus solving the hot servers problem. Additionally, wireless connections do not rely on switches and therefore are free of problems such as single-point of failure and limited bisection bandwidth.
Requirements
The Data Center Network ( |
https://en.wikipedia.org/wiki/Three-Axis%20Acceleration%20Switch | The three-axis acceleration switch is a micromachined microelectromechanical systems (MEMS) sensor that detects whether an acceleration event has exceeded a predefined threshold. It is a small, compact device, only 5mm by 5mm, and measures acceleration in the x, y, and z axes. It was developed by the Army Research Laboratory for the purposes of traumatic brain injury (TBI) research and was first introduced in 2012 at the 25th International Conference on Micro Electro Mechanical Systems (MEMS).
The three-axis acceleration switch was designed to obtain acceleration data more effectively than a conventional accelerometer in order to more accurately characterize the forces and shocks responsible for TBI. While miniature accelerometers require a constant power draw, the three-axis acceleration switch only draws current when it senses an acceleration event, using up less energy and allowing the use of smaller batteries. The three-axis acceleration switch has shown to exhibit an expected battery lifetime that is about 100 times better than that of a digital accelerometer. In return, however, the acceleration switch has a lower resolution than that of a digital or analog accelerometer.
One potential application of the three-axis acceleration switch is in studying the head impacts of players in high-risk contact sports. Due to the size of conventional accelerometers, measuring the acceleration requires the device to be implemented inside the player's helmet, which is designed to mitigate the collision forces and thus may not accurately reflect the true level of injury potential. In contrast, the miniature nature of the acceleration switch makes it easier for the switch to be affixed directly onto the participant's head.
References
Accelerometers
Transducers
Microelectronic and microelectromechanical systems |
https://en.wikipedia.org/wiki/Exactis | Exactis LLC is a data broker established in 2015 and based in the U.S state of Florida. The firm reportedly handles business and consumer data in an effort to refine targeted advertising.
Data leak
Exactis became notable in June 2018, after a discovery by cybersecurity researcher Vinny Troia detailed how the organization made nearly 340 million detailed records about individual people available on a publicly accessible server, leaving those people at heightened risk of being impersonated, profiled or otherwise exploited through social engineering. The information has reportedly since been protected and removed from the public. However, the leak is notable due to the breach of data that was exposed, surpassing the Equifax breach which had exposed 145 million customers' personal data the year prior.
Exactis had reportedly accumulated some or all of its information without the knowledge or explicit consent of the data subjects.
References
Companies based in Florida
Data brokers
2015 establishments in Florida |
https://en.wikipedia.org/wiki/Data%20center%20security | Data center security is the set of policies, precautions and practices adopted at a data center to avoid unauthorized access and manipulation of its resources. The data center houses the enterprise applications and data, hence why providing a proper security system is critical. Denial of service (DoS), theft of confidential information, data alteration, and data loss are some of the common security problems afflicting data center environments.
Data security issues can be harmful to many companies sometimes, so it is very important to know what are the issues and find useful solutions for them. The purpose of data security is to protect digital information from unauthorized access. It is also important to note that data security is different from data privacy. There are many situations where data center security would be threatened on, especially for cloud-based data.
Overview
According to the Cost of a Data Breach Survey, in which 49 U.S. companies in 14 different industry sectors participated, they noticed that:
39% of companies say negligence was the primary cause of data breaches
Malicious or criminal attacks account for 37 percent of total breaches.
The average cost of a breach is $5.5 million.
Many big companies nowadays are using the cloud to store their and their customers' data, but the risks of saving data in the cloud can be enormous. Cyber attacks can be very harmful to many companies. There were 64% of companies worldwide that had troubles with cyber attacks in the year 2020 alone. Some cyber attacks targeted personal information such as identity theft can hurt someone's credits with life-changing influences.
The need for a secure data center
Physical security is needed to protect the value of the hardware therein.
Data protection
The cost of a breach of security can have severe consequences on both the company managing the data center and on the customers whose data are copied. The 2012 breach at Global Payments, a processing vendor for Visa, |
https://en.wikipedia.org/wiki/Outline%20VPN | Outline VPN is a free and open-source tool that deploys Shadowsocks servers on multiple cloud service providers. The software suite also includes client software for multiple platforms. Outline was developed by Jigsaw, a technology incubator created by Google.[3]
The Outline Server supports self-hosting, as well as cloud service providers including DigitalOcean, Rackspace, Google Cloud Platform, and Amazon EC2. Installation involves running a command on its command-line interface, or in the case of installing on DigitalOcean or Google Cloud, its graphical user interface.
Components
Outline has three main components:
The Outline Server acts as a proxy and relays connections between the client and the sites they want to access. It is based on Shadowsocks, and offers a REST API for management of the server by the Outline Manager application.
The Outline Manager is a graphical application used to deploy and manage access to Outline Servers. It supports Windows, macOS and Linux.
The Outline Client connects to the internet via the Outline Server. It supports Windows, macOS, Linux, ChromeOS, Android, and iOS.
Security and privacy
Outline uses the Shadowsocks protocol for communication between the client and server. Traffic is encrypted with the IETF ChaCha20 stream cipher (256-bit key) and authenticated with the IETF Poly1305 authenticator.
Outline is free and open-source, licensed under the Apache License 2.0, and was audited by Radically Open Security and claims not to log users' web traffic. The Outline Server supports unattended upgrades.
Outline is not a true VPN solution but rather a Shadowsocks-based proxy. The two technologies are similar in the way they can be used to redirect network traffic and make it appear as originating from another device (the server), and hide the traffic's final destination from observers and filters until it reaches the proxy server. However, a VPN has additional capabilities, such as encapsulating traffic within a virtual tun |
https://en.wikipedia.org/wiki/Dew%20computing | Dew computing is an information technology (IT) paradigm that combines the core concept of cloud computing with the capabilities of end devices (personal computers, mobile phones, etc.). It is used to enhance the experience for the end user in comparison to only using cloud computing. Dew computing attempts to solve major problems related to cloud computing technology, such as reliance on internet access. Dropbox is an example of the dew computing paradigm, as it provides access to the files and folders in the cloud in addition to keeping copies on local devices. This allows the user to access files during times without an internet connection; when a connection is established again, files and folders are synchronized back to the cloud server.
History
The term "dew computing", as used in information technology, first appeared in 2015 in IT literature and since then has become a field of its own. The cloud-dew architecture was proposed as a possible solution to the offline data accessibility problem. At first, its scope included only web applications; broader applications were later proposed.
Dew computing is a model which was derived from the original concept of cloud computing. Other models have also emerged from cloud computing, including fog computing, edge computing, dew computing, and others. Proponents claim that these novel models, such as dew computing, can provide better experiences for users.
Cloud computing provides universal access and scalability. However, having all the resources far from a user's control occasionally causes problems. In the classic cloud computing paradigm, when the internet connection to the servers is lost, the user is unable to access their data; dew computing aims to solve this problem.
Definition
As an information technology paradigm, dew computing seeks to use the capabilities of personal computers along with cloud services in a more reliable manner.
The key features of dew computing are independence and collaboration. Ind |
https://en.wikipedia.org/wiki/Mary%20Fenton | Mary Fenton alias Mehrbai (c. 1854 – c. 1896) was the first Gujarati, Parsi and Urdu theatre actress of European origin. Born to an Irish soldier in the British Indian Army, she fell in love and married Parsi actor-director Kavasji Palanji Khatau. He introduced her to acting and she had a successful stage career.
Early life
Mary Fenton was born in Landour near Mussoorie in India to Jannette and Mathew Fenton, an Irish retired soldier of the British Indian Army. She was baptized as Mary Jane Fenton, but there is no further information of her early life and education. Parsi theatre actor-director Kavasji Palanji Khatau was rehearsing for his play Inder Sabha, when Fenton had come to book the hall for her magic lantern show. She admired his acting, met him, fell in love and finally married him. Subsequently, she adopted a Parsi name Mehrbai. She already knew Hindi and Urdu, and in the 1870s Khatau gave her further training in singing and acting.
She created a sensation in the theatre due to her talent and relationship with Khatau. However, following a dispute between Khatau and the Empress Victoria Theatrical Company owner Jahangir Pestonjee Khambatta regarding Fenton's entry into theatre in 1878, Khatau left Bombay for Delhi and joined Alfred Theatre Company owned by Manek Master who also opposed Fenton. Consequently, Khatau started his own Alfred Company in 1881, where Fenton had a long and successful career.
Fenton and Khatau later separated. They had a son Jahangir Khatau.
Career
She was the first Anglo-Indian actress of the Parsi, Gujarati, and Urdu theatre. She became popular for her roles as the Parsi heroine. She acted in Nanabhai Ranina's Nazan Shirin (1881), Bamanji Kabra's Bholi Gul (Innocent Flower, 1882, based on Ellen Wood's English novel East Lynne), Agha Hasan Amanat's Urdu opera Inder Sabha, Khambatta's Khudadad (The Gift of God, 1898, based on Shakespeare's Pericles, Prince of Tyre), Gamde ni Gori (Village Nymph, 1890), Alauddin (1891), Tara Khurs |
https://en.wikipedia.org/wiki/Pseudomonas%20clemancea | Pseudomonas clemancea is a species of Pseudomonas bacteria which was first discovered in the North of England. The specific epithet clemancea was given by microbiologist Pattanathu Rahman at Teesside University to bestow the CLEMANCE (Clean Environment Management Centre). This bacterium has DNA coding distinct from existing species and has unique properties developed in response to the contaminated soil from which it comes.
Biosurfactant from Pseudomonas clemancea
Surfactants work by reducing surface tension between two liquids or a liquid and a solid. Biosurfactants, surface-active agents of biological origin, have environment-friendly properties; they are bio-degradable, non-toxic and can be made organically using local raw material and producers. Biosurfactants can be used in soaps, detergents, medical ointments, or as emulsifiers, i.e. within ice cream, facial cream, or sun lotion. P. clemancea produces rhamnolipids type of biosurfactants to detoxify oil and chemicals contaminants in the ground.
Further reading
A triangle study of human, instrument and bioelectronic nose for non-destructive sensing of seafood freshness
Production of rhamnolipid biosurfactants by Pseudomonas aeruginosa DS10-129 in a microfluidic bioreactor
Bacterial communities in systemic plant parts subjected to a fungal trunk disease
Advances in Pseudomonadaceae Research and Application: 2011 Edition
Environmental Sustainability: Role of Green Technologies
References
Pseudomonadales
Undescribed species |
https://en.wikipedia.org/wiki/ZMap%20%28software%29 | ZMap is a free and open-source security scanner that was developed as a faster alternative to Nmap. ZMap was designed for information security research and can be used for both white hat and black hat purposes. The tool is able to discover vulnerabilities and their impact, and detect affected IoT devices.
Using one gigabit per second of network bandwidth, ZMap can scan the entire IPv4 address space in 44 minutes on a single port. With a ten gigabit connection, ZMap scan can complete a scan in under five minutes.
Operation
ZMap iterates on techniques utilized by its predecessor, Nmap, by altering the scanning method in a few key areas. Nmap sends out individual signals to each IP address and waits for a reply. As replies return, Nmap compiles them into a database to keep track of responses, a process that slows down the scanning process. In contrast, ZMap uses cyclic multiplicative groups, which allows ZMap to scan the same space roughly 1,300 times faster than Nmap. The ZMap software takes every number from 1 to 232-1 and creates an iterative formula that ensures that each of the possible 32-bit numbers is visited once in a pseudorandom order. Building the initial list of numbers for every IP address takes upfront time, but it is a fraction of what is required to aggregate a list of every sent and received probe. This process ensures that once ZMap starts sending probes out to different IPs, an accidental denial of service could not occur because an abundance of transmissions would not converge on one subnet at the same time.
ZMap also speeds up the scanning process by sending a probe to every IP address only once by default, whereas Nmap resends a probe when it detects a connection delay or fails to get a reply. This results in about 2% of IP addresses being missed during a typical scan, but when processing billions of IP address, or potential IoT devices being targeted by cyberattackers, 2% is an acceptable tolerance.
Usage
ZMap can be used for both vulner |
https://en.wikipedia.org/wiki/Floc%20%28biofilm%29 | A floc is a type of microbial aggregate that may be contrasted with biofilms and granules, or else considered a specialized type of biofilm. Flocs appear as cloudy suspensions of cells floating in water, rather than attached to and growing on a surface like most biofilms. The floc typically is held together by a matrix of extracellular polymeric substance (EPS), which may contain variable amounts of polysaccharide, protein, and other biopolymers. The formation and the properties of flocs may affect the performance of industrial water treatment bioreactors such as activated sludge systems where the flocs form a sludge blanket.
Floc formation may benefit the constituent microorganisms in a number of ways, including protection from pH stress, resistance to predation, manipulation of microenvironments, and facilitation of mutualistic relationships in mixed microbial communities.
In general, the mechanisms by which flocculating microbial aggregates hold together are poorly understood. However, work on the activated sludge bacterium Zoogloea resiniphila has shown that PEP-CTERM proteins must be expressed for flocs to form; in their absence, growth is planktonic, even though exopolysaccharide is produced.
See also
Yeast flocculation#Process
References
Bacteriology
Biological matter
Environmental microbiology
Microbiology terms |
https://en.wikipedia.org/wiki/DIVA-GIS | DIVA-GIS is a free geographic information system software program used for the analysis of geographic data, in particular point data on biodiversity. The software was first designed for application to the study of wild potatoes in South America.
Development
DIVA-GIS was developed as a joint project by the International Potato Center in Peru, the International Plant Genetic Resources Institute, the Museum of Vertebrate Zoology at the University of California at Berkeley, the Secretariat of the Pacific Community, and the FAO. DIVA-GIS has a wide range of tools for evaluation of mapping and modeling of habitats. There is a command-line version of the program that has been developed, AVID-GIS.
Formats
DIVA-GIS can process data in all standard GIS formats, including data from ESRI's ArcGIS programs. The program runs on Windows and OS X. DIVA raster files generated may be imported and exported into R or the modeling program Maxent.
References
External links
Cross-platform software
Free GIS software |
https://en.wikipedia.org/wiki/CAINE%20Linux | CAINE Linux (Computer Aided INvestigative Environment) is an Italian Linux live distribution managed by Giovanni "Nanni" Bassetti. The project began in 2008 as an environment to foster digital forensics and incidence response (DFIR), with several related tools pre-installed.
Purpose
CAINE is a professional open source forensic platform that integrates software tools as modules along with powerful scripts in a graphical interface environment. Its operational environment was designed with the intent to provide the forensic professional all the tools required to perform the digital forensic investigate process (preservation, collection, examination and analysis). CAINE is a live Linux distribution so it can be booted from removable media (flash drive) or from an optical disk and run in memory. It can also be installed onto a physical or virtual system. In Live mode, CAINE can operate on data storage objects without having to boot up a supporting operating system. The latest version 11.0 can boot on UEFI/UEFI+Secure and Legacy BIOS allowing CAINE to be used on information systems that boot older operating systems (e.g. Windows NT) and newer platforms (Linux, Windows 10).
Requirements
CAINE is based on Ubuntu 18.04 64-bit, using Linux kernel 5.0.0-32. CAINE system requirements to run as a live disc are similar to Ubuntu 18.04. It can run on a physical system or in a virtual machine environment such as VMware Workstation.
Supported platforms
The CAINE Linux distribution has numerous software applications, scripts and libraries that can be used in a graphical or command line environment to perform forensic tasks. CAINE can perform data analysis of data objects created on Microsoft Windows, Linux and some Unix systems. One of the key forensic features since version 9.0 is that it sets all block devices by default to read-only mode. Write-blocking is a critical methodology to ensure that disks are not subject to writing operations by the operating system or forensic |
https://en.wikipedia.org/wiki/Hostinger | Hostinger International, Ltd is an employee-owned web hosting provider and an ICANN-accredited domain registrar. Established in 2004, the company is headquartered in Lithuania and employs more than 1,000 people. Hostinger is the parent company of 000webhost, Hosting24, Zyro, and Niagahoster.
Hostinger has been ranked among the fastest-growing companies in Europe for four consecutive years in the Financial Times' annual FT 1000 list.
History
Hostinger was founded in 2004 as Hosting Media.
In 2007, Hosting Media’s paid hosting offer was joined by a free web hosting service when the company founded 000webhost.
In 2008, the company launched Hosting24, a cPanel-based web hosting brand, in the United States. The data centers were located in Asheville, North Carolina, and the United Kingdom.
In 2011, Arnas Stuopelis joined the company as CEO. In the same year Hosting Media rebranded to Hostinger.
In 2013, Hostinger launched its subsidiary Niagahoster in Indonesia. It became one of the first hosting providers to offer tier-4 data centers.
In 2014, the company launched a Brazilian subsidiary, Weblink.
In 2019, Hostinger introduced a no-code, drag-and-drop website builder. Launched as a subsidiary under the name Zyro, the builder is supported by Hostinger’s hosting infrastructure and is aimed at small to medium-sized enterprises.
In 2023, Hostinger launched the AI Website Builder, a platform that uses artificial intelligence to generate a custom site from scratch. AI Builder automatically writes unique content, selects royalty-free stock images, and chooses a color palette and fonts that best suit the description.
In 2023, Daugirdas Jankus becomes Hostinger CEO. Arnas Stuopelis continues as Chairman of the Board.
Acquisitions
In 2019, the company partnered with LiteSpeed Web Server (LSWS). It also partnered with Google Cloud Platform in 2020.
In 2021, private equity company ConHostinger acquired an approximately 31% controlling stake in Hostinger with the goal o |
https://en.wikipedia.org/wiki/Data%20exfiltration | Data exfiltration occurs when malware and/or a malicious actor carries out an unauthorized data transfer from a computer. It is also commonly called data extrusion or data exportation. Data exfiltration is also considered a form of data theft. Since the year 2000, a number of data exfiltration efforts severely damaged the consumer confidence, corporate valuation, and intellectual property of businesses and national security of governments across the world.
Types of exfiltrated data
In some data exfiltration scenarios, a large amount of aggregated data may be exfiltrated. However, in these and other scenarios, it is likely that certain types of data may be targeted. Types of data that are targeted includes:
Usernames, associated passwords, and other system authentication related information
Information associated with strategic decisions
Cryptographic keys
Personal financial information
Social security numbers and other personally identifiable information (PII)
Mailing addresses
United States National Security Agency hacking tools
Techniques
Several techniques have been used by malicious actors to carry out data exfiltration. The technique chosen depends on a number of factors. If the attacker has or can easily gain physical or privileged remote access to the server containing the data they wish to exfiltrate, their chances of success are much better than otherwise. For example, it would be relatively easy for a system administrator to plant, and in turn, execute malware that transmits data to an external command and control server without getting caught. Similarly, if one can gain physical administrative access, they can potentially steal the server holding the target data, or more realistically, transfer data from the server to a DVD or USB flash drive. In many cases, malicious actors cannot gain physical access to the physical systems holding target data. In these situations, they may compromise user accounts on remote access applications using manufac |
https://en.wikipedia.org/wiki/Fortnite | Fortnite is an online video game developed by Epic Games and released in 2017. It is available in three distinct game mode versions that otherwise share the same general gameplay and game engine: Fortnite Battle Royale, a free-to-play battle royale game in which up to 100 players fight to be the last person standing; Fortnite: Save the World, a cooperative hybrid tower defense-shooter and survival game in which up to four players fight off zombie-like creatures and defend objects with traps and fortifications they can build; and Fortnite Creative, in which players are given complete freedom to create worlds and battle arenas.
Save the World and Battle Royale were released in 2017 as early access titles, while Creative was released on December 6, 2018.
While the Save the World and Creative versions have been successful for Epic Games, Fortnite Battle Royale in particular became an overwhelming success and a cultural phenomenon, drawing more than 125 million players in less than a year, earning hundreds of millions of dollars per month. Fortnite as a whole generated in gross revenue up until December 2019.
Save the World is available for macOS, PlayStation 4, Windows, and Xbox One, while Battle Royale and Creative were released for all those platforms, and also for Android and iOS devices and Nintendo Switch. The game also launched with the release of the ninth-generation PlayStation 5 and Xbox Series X/S consoles.
Game modes
Fortnite is distributed as three different game modes, using the same engine; each has similar graphics, art assets, and game mechanics.
Fortnite: Save the World is a player-versus-environment cooperative game, with four players collaborating towards a common objective on various missions. The game is set after a fluke storm appears across Earth, causing 98% of the population to disappear, and the survivors to be attacked by zombie-like "husks". The players take the role of commanders of home base shelters, collecting resources, saving s |
https://en.wikipedia.org/wiki/WDR%20paper%20computer | The WDR paper computer or Know-how Computer is an educational model of a computer consisting only of a pen, a sheet of paper, and individual matches in the most simple case. This allows anyone interested to learn how to program without having an electronic computer at their disposal.
The paper computer was created in the early 1980s when computer access was not yet widespread in Germany, to allow people to familiarize themselves with basic computer operation and assembly-like programming languages. It was distributed in over copies and at its time belonged to the computers with the widest circulation.
The Know-how Computer was developed by and Ulrich Rohde and was first presented in the television program WDR Computerclub in 1983. It was also published in German computer magazines mc and .
The original printed version of the paper computer has up to 21 lines of code on the left and eight registers on the right, which are represented as boxes that contain as many matches as the value in the corresponding register. A pen is used to indicate the line of code which is about to be executed. The user steps through the program, adding and subtracting matches from the appropriate registers and following program flow until the stop instruction is encountered.
The instruction set of five commands is small but Turing complete and therefore enough to represent all mathematical functions:
inc *register*: Add 1 to the register.
dec *register*: Subtract 1 from the register.
jmp *line*: Jumps to the specified line.
isz *register*: Checks if the register is zero. If so, skips a line. If not, continues normally.
stp: Stops the program.
In the original newspaper article about this computer, it was written slightly differently (translation):
+ = Add 1 to the contents of data register XX and increase (program step) by 1
- = Subtract 1 from the contents of data register XX and increase (program step) by 1
(J) = (Jump) to (line) XX
0 = Check if the content of the data re |
https://en.wikipedia.org/wiki/ADA%20%28smart%20antenna%29 | ADA (an abbreviation of adaptive antenna) is an advanced smart antenna, using multichannel controlled reception pattern antenna (CRPA), designed for various platforms, including UAVs, planes and ships, manufactured by the MLM factory of the Israeli Aerospace Industries (IAI).
The ADA enables vehicles, missiles and other platforms equipped with the smart antenna to navigate even when heavy jamming is conducted in order to block all reception of GPS and other GNSS signals. Using advanced techniques even very faint "good" signals are being extracted from the superposition of incoming waves that include noise and strong signals produced by the enemy (or friendly) jammers. These signals are then used (just like a regular GPS receiver) to triangulate the position and determine the GPS tune.
History
The theory of the ADA started to emerge as a theory of multichannel estimation. Its origins go back into methods developed in the 1920s that were used to determine direction of the arrival of radio signals by a set of two antennas based on the phase difference or amplitudes of their output voltages. Thus, the assessment of the directions of arrival of a single signal was conducted according to pointed type indicator readings or according to the Lissajous curves,
drawn by beam on the oscilloscope screen.
The existence of ADA was publicly unveiled in February 2017, along with a report that the Israeli Defence Forces has already decided to install the ADAs on "major platforms" in service. It was also displayed in the Aero India air show that was held from 14th to the 18th that month in Bangalore.
In July 2017 MLM has signed a contract with American Honeywell to jointly develop a GPS anti-jam navigation system, based on ADA and Honywell's devices.
References
External links
IAI ADA, by the IAI, on YouTube
IAI, Honeywell to jointly develop anti-jam system, by Globes, May 2018
IAI Unveils ADA-O to Help Land Platforms Deal with GNSS Anti-Jammers, by Israel Defense, Mar 2019 |
https://en.wikipedia.org/wiki/EASN%20Association | The European Aeronautics Science Network (EASN) is an international non-profit association dedicated to the advancement of the European aeronautics sciences and technologies. The establishment of EASN goes back to 2002 and it was based on two subsequent support actions funded by the European Commission while it was founded as an open, self-funded and self sustainable association in 2008. Specifically, the EASN Association is the Association of the European Academia active in supporting aviation research, facing the fragmentation of Academia and disseminating the new knowledge and breakthrough technologies incubated through aviation research. Headquartered in Brussels, Belgium, the EASN Association encounters members across 28 European countries with the number constantly increasing.
Overview
On 6 May 2008 the EASN Association was officially established by 22 distinguished personalities of the European Aeronautics research community, with the encouragement of the European Commission and the support of several university professors throughout Europe. The mission of the Association is to contribute to the advancement of the aeronautics sciences and technologies and support the education and research in Europe. The tools for accomplishing the above objectives include among others the organization of scientific events, research forums and collaboration with other networks and associations.
In 2017, the Association counted approximately 350 effective members Europe-wide. These effective members are individuals mainly from European universities active in aviation research. Furthermore, the Association has a number of about 50 associate members which are highly-ranked entities, such as Universities, University Departments, Research Establishments, SMEs, Industries and Laboratories. Among the Association’s associate members top universities are included such as KU Leuven, Polytechnic University of Milan, Warsaw University of Technology, University of Stuttgart, Politecnico |
https://en.wikipedia.org/wiki/Eliza%20Edwards | Eliza Edwards (1779–1846) was a human computer and daughter of Mary Edwards.
Early life and education
Edwards was born in Ludlow to Mary and John Edwards.
Career
Eliza was a human computer who took on the job of her mother Mary Edwards, working on the Nautical Almanac. She lost her job in 1829, during the formation of the Nautical Almanac Office. She was paid by the Board of Longitude.
References
1779 births
1846 deaths
Human computers
People from Ludlow
British women mathematicians
19th-century English mathematicians |
https://en.wikipedia.org/wiki/Sniper%20Fury | Sniper Fury is an online single player video game developed and published by Gameloft.
The action of Sniper Fury takes place in the near future. Technological development and geopolitical changes rendered former methods of resolving conflicts obsolete. Countries, corporations, and organizations employ services of highly trained professionals, who can eliminate specific targets with surgical precision.
Gameplay
The game revolves around mechanics common for most sniper games, where the player has to eliminate a great majority of his targets from long distance. To complete a task, the player can use a variety of futuristic gadgets, e.g. a detection device that will tell the location of every nearby human, or stimulants, which will boost reflexes to supernatural levels.
Release
Sniper Fury was released on 2 December 2015 for Android, iOS and Windows Phone. The Steam version was released on 13 June 2017.
Reception
Sniper Fury received "mixed or average" reviews, according to review aggregator Metacritic.
References
External links
Official website
Gameloft games
Sniper video games
Action-adventure games
IOS games
Online games
Multiplayer and single-player video games
Video games developed in France
Windows games
Windows Phone games
Android (operating system) games
2015 video games |
https://en.wikipedia.org/wiki/Mermaid%27s%20glove | Mermaid's glove is a common name referring to two different organisms:
Dictyota binghamiae, a seaweed
Haliclona oculata, a sponge |
https://en.wikipedia.org/wiki/Isosceles%20set | In discrete geometry, an isosceles set is a set of points with the property that every three of them form an isosceles triangle. More precisely, each three points should determine at most two distances; this also allows degenerate isosceles triangles formed by three equally-spaced points on a line.
History
The problem of finding the largest isosceles set in a Euclidean space of a given dimension was posed in 1946 by Paul Erdős. In his statement of the problem, Erdős observed that the largest such set in the Euclidean plane has six points. In his 1947 solution, Leroy Milton Kelly showed more strongly that the unique six-point planar isosceles set consists of the vertices and center of a regular pentagon. In three dimensions, Kelly found an eight-point isosceles set, six points of which are the same; the remaining two points lie on a line perpendicular to the pentagon through its center, at the same distance as the pentagon vertices from the center. This three-dimensional example was later proven to be optimal, and to be the unique optimal solution.
Decomposition into 2-distance sets
Kelly's eight-point three-dimensional isosceles set can be decomposed into two sets (the three points on a line perpendicular to the pentagon) and (the five vertices of the pentagon), with the property that each point in is equidistant from all points of . When such a decomposition is possible, in Euclidean spaces of any dimension, and must lie in perpendicular subspaces, must be an isosceles set within its subspace, and the set formed from by adding the point at the intersection of its two subspaces must also be an isosceles set within its subspace. In this way, an isosceles set in high dimensions can sometimes be decomposed into isosceles sets in lower dimensions. On the other hand, when an isosceles set has no decomposition of this type, then it must have a stronger property than being isosceles: it has only two distances, among all pairs of points.
Despite this decompositio |
https://en.wikipedia.org/wiki/Hybrid%20Access%20Networks | Hybrid Access Networks refer to a special architecture for broadband access networks where two different network technologies are combined to improve bandwidth. A frequent motivation for such Hybrid Access Networks to combine one xDSL network with a wireless network such as LTE. The technology is generic and can be applied to combine different types of access networks such as DOCSIS, WiMAX, 5G or satellite networks. The Broadband Forum has specified an architecture as a framework for the deployment of such converged networks.
Use cases
One of the main motivations for such Hybrid Access Networks is to provide faster Internet services in rural areas where it is not always cost-effective to deploy faster xDSL technologies such as G.Fast or VDSL2 that cannot cover long distances between the street cabinet and the home. Several governments, notably in Europe, required network operators to provide fast Internet services to all inhabitants with a minimum of 30 Mbps by 2020.
A second use case is to improve the reliability of the access link given that it is unlikely that both the xDSL network and the wireless network will fail at the same time.
A third motivation is the fast service turnup. The customer can immediately install the hybrid network access and use the wireless leg while the network operator is installing the wired part.
Technology
Several techniques are defined by the Broadband Forum to create Hybrid Access Networks. To illustrate them, we assume that the end user has an hybrid CPE (Customer-premises_equipment) router that is attached to both a wired access network such as xDSL and a wireless one such as LTE. Other deployments are possible, e.g., the end user might have two different access routers that are linked together by a cable instead of a single hybrid CPE router.
The first deployment scenario is where the network operator provides a hybrid CPE router to each subscriber but no specialised equipment in the operator's network. There are two possible |
https://en.wikipedia.org/wiki/Solid%20%28web%20decentralization%20project%29 | Solid (Social Linked Data) is a web decentralization project led by Sir Tim Berners-Lee, the inventor of the World Wide Web, originally developed collaboratively at the Massachusetts Institute of Technology (MIT). The project "aims to radically change the way Web applications work today, resulting in true data ownership as well as improved privacy" by developing a platform for linked-data applications that are completely decentralized and fully under users' control rather than controlled by other entities. The ultimate goal of Solid is to allow users to have full control of their own data, including access control and storage location. To that end, Tim Berners-Lee formed a company called Inrupt to help build a commercial ecosystem to fuel Solid.
History
Two decades after Berners-Lee invented the World Wide Web in 1989, he outlined the design issues of what later became the Solid project in drafts he wrote for the World Wide Web Consortium. Berners-Lee became increasingly dismayed at seeing his invention being abused, such as when Russian hackers interfered with the 2016 US elections, when the Facebook–Cambridge Analytica data scandal became public, when Facebook in 2012 conducted psychological experiments on nearly 700,000 users in secret, and when Google and Amazon applied for patents on devices that listen for emotional changes in human voices.
Berners-Lee felt that the Internet was in need of repair and conceived the Solid project as a first step to fix it, as a way to give individual users full control over the usage of their data. The Solid project is available to anyone to join and contribute, although Berners-Lee advises that people without coding skills should instead advocate publicly for changing the Internet.
In 2015, MIT received a gift from Mastercard to support the development of Solid. Berners-Lee's research team collaborated with the Qatar Computing Research Institute and Oxford University on Solid.
In 2018, Berners-Lee took a sabbatical from MIT |
https://en.wikipedia.org/wiki/Zero-divisor%20graph | In mathematics, and more specifically in combinatorial commutative algebra, a zero-divisor graph is an undirected graph representing the zero divisors of a commutative ring. It has elements of the ring as its vertices, and pairs of elements whose product is zero as its edges.
Definition
There are two variations of the zero-divisor graph commonly used.
In the original definition of , the vertices represent all elements of the ring. In a later variant studied by , the vertices represent only the zero divisors of the given ring.
Examples
If is a semiprime number (the product of two prime numbers)
then the zero-divisor graph of the ring of integers modulo (with only the zero divisors as its vertices) is either a complete graph or a complete bipartite graph.
It is a complete graph in the case that for some prime number . In this case the vertices are all the nonzero multiples of , and the product of any two of these numbers is zero modulo .
It is a complete bipartite graph in the case that for two distinct prime numbers and . The two sides of the bipartition are the nonzero multiples of and the nonzero multiples of , respectively. Two numbers (that are not themselves zero modulo ) multiply to zero modulo if and only if one is a multiple of and the other is a multiple of , so this graph has an edge between each pair of vertices on opposite sides of the bipartition, and no other edges. More generally, the zero-divisor graph is a complete bipartite graph for any ring that is a product of two integral domains.
The only cycle graphs that can be realized as zero-product graphs (with zero divisors as vertices) are the cycles of length 3 or 4.
The only trees that may be realized as zero-divisor graphs are the stars (complete bipartite graphs that are trees) and the five-vertex tree formed as the zero-divisor graph of .
Properties
In the version of the graph that includes all elements, 0 is a universal vertex, and the zero divisors can be identified as the vertic |
https://en.wikipedia.org/wiki/Albedometer | An albedometer is an instrument used to measure the albedo (reflecting radiation) of a surface. An albedometer is mostly used to measure the reflectance of earths surface. It is also useful to evaluate thermal effects in buildings and generation capacity with bifacial solar photovoltaic panels. Often it consists of two pyranometers: one facing up towards the sky and one facing down towards the surface. From the ratio of incoming and reflecting radiation the albedo can be calculated.
Measurement principle
The measurement of surface albedo of earths surface happens by using two pyranometers. The upfacing pyranometer measures the incoming global solar radiation. The downward facing pyranometer measures the reflected global solar radiation. The ratio of the reflected to the global radiation is the solar albedo and depends on the properties of the surface and the directional distribution of the incoming solar radiation. Typical values range from 4% for asphalt to 90% for fresh snow. Designs for a low-cost albedometer have been released with an open source hardware license which measures the reflection in 8 spectral bands in the visible light spectrum, additionally the system is equipped with a global navigation satellite system receiver, to georeference its position and an Inertial Measurement Unit to know its absolute orientation, make corrections in real time or detect errors.
Standards
ISO 9060
WMO No.8
ISO 9847
ASTM G207-11.
References
Meteorological instrumentation and equipment
Measuring instruments |
https://en.wikipedia.org/wiki/Rational%20thermodynamics | Rational thermodynamics is a school of thought in statistical thermodynamics developed in the 1960s.
Its introduction is attributed to Clifford Truesdell (1919–2000), Bernard Coleman (b. 1929) and Walter Noll (1925–2017).
The aim was to develop a mathematical model of thermodynamics that would go beyond the traditional "thermodynamics of irreversible processes" or TIP developed in the late 19th to early 20th centuries.
Truesdell's "flamboyant style" and "satirical verve" caused controversy between "rational thermodynamics" and proponents of traditional thermodynamics.
References
Clifford A. Truesdell, Rational Thermodynamics: A Course of Lectures on Selected Topics, Springer, (1969, 2nd ed. 1984).
Ingo Müller, Tommaso Ruggeri, Extended Rational Thermodynamics, Springer (1998), doi:10.1007/978-1-4612-2210-1.
See also
Archive for Rational Mechanics and Analysis
Thermodynamics |
https://en.wikipedia.org/wiki/Vincent%20Calvez | Vincent Calvez (born 24 August 1981) is a French mathematician. He is currently a directeur de recherche (senior researcher) at the Institute Camille Jordan at the Claude Bernard University Lyon 1. He is known for his work in mathematical modeling in biology, especially in the movement of bacteria.
Biography
Born in Saint-Malo, he attended lycee in Brest, he eventually gained admission to and attended École normale supérieure (Paris).
He obtained his agrégation in 2005 and did his thesis in mathematics under Bernoit Perthame at the University of Paris 6, finishing in 2007. In 2008, he took a position as a chargé de recherche at the École normale supérieure de Lyon in the Unité de mathématiques pures et appliquées (UMPA) in 2008. He defended his habilitation in 2015 and obtained the position of Directeur de Recherche at Claude Bernard University Lyon 1 in 2016.
Since 2014, he has been an associate editor of the Journal of Mathematical Biology, and, since 2018, he has been an associate editor of the SIAM Journal of Mathematical Analysis.
Aside from research, Calvez is also involved in the popularization of mathematics by coordinating the travelling exhibition “Mathàlyon,” intended for junior high and high school students.
Research
Calvez's early research involved the development of models for the collective movements of bacteria via chemotaxis using kinetic models. This model describes both the individual movements of bacteria and the transport of colonies as a whole, and satisfactorily explains the specific characters of this movement. Later, he worked in collaboration with biophysicists to find a new explanation for cell polarization and in collaboration with ecologists on models for invasion fronts.
Bibliography
Calvez, V.; Lenuzza, N.; Oelz, D.; Deslys, J.-P.; Laurent, P.; Mouthon, F.; Perthame, B. "Size distribution dependence of prion aggregates infectivity". Math. Biosci. 217 (2009), no. 1, 88–99.
Calvez, V.; Corrias, L. "The parabolic-paraboli |
https://en.wikipedia.org/wiki/Separation%20oracle | A separation oracle (also called a cutting-plane oracle) is a concept in the mathematical theory of convex optimization. It is a method to describe a convex set that is given as an input to an optimization algorithm. Separation oracles are used as input to ellipsoid methods.
Definition
Let K be a convex and compact set in Rn. A strong separation oracle for K is an oracle (black box) that, given a vector y in Rn, returns one of the following:
Assert that y is in K.
Find a hyperplane that separates y from K: a vector a in Rn, such that for all x in K.
A strong separation oracle is completely accurate, and thus may be hard to construct. For practical reasons, a weaker version is considered, which allows for small errors in the boundary of K and the inequalities. Given a small error tolerance d>0, we say that:
A vector y is d-near K if its Euclidean distance from K is at most d;
A vector y is d-deep in K if it is in K, and its Euclidean distance from any point in outside K is at least d.
The weak version also considers rational numbers, which have a representation of finite length, rather than arbitrary real numbers. A weak separation oracle for K is an oracle that, given a vector y in Qn and a rational number d>0, returns one of the following::
Assert that y is d-near K;
Find a vector a in Qn, normalized such that its maximum element is 1, such that for all x that are d-deep in K.
Implementation
A special case of a convex set is a set represented by linear inequalities: . Such a set is called a convex polytope. A strong separation oracle for a convex polytope can be implemented, but its run-time depends on the input format.
Representation by inequalities
If the matrix A and the vector b are given as input, so that , then a strong separation oracle can be implemented as follows. Given a point y, compute :
If the outcome is at most , then y is in K by definition;
Otherwise, there is at least one row of A, such that is larger than the correspondi |
https://en.wikipedia.org/wiki/BatiBUS | BatiBus was a network protocol for building automation that was introduced in 1989 and has since been succeeded by KNX. It was a relatively simple low-cost protocol that did not rely on dedicated chips.
The system was run by the BatiBus Club International (BCI), which was founded by the Swiss company Landis & Gyr and the French companies AIRELEC, Electricité de France and Merlin Gerin (who originated the concept). Predominately used in France and captured by French Electrical Standard NF C 46620, it provided layers 1, 2 and 7 of the OSI model. Approximately 500,000 BatiBus network units were installed, mainly in France.
BatiBus communicated over twisted pair lines and in topologies that could be divided into several segments. Each segment was powered with a 15 volt power supply rated at 150 milliamps. A device (node) could be reached at one of 240 possible addresses. In addition, 16 group addresses could be established under which all nodes in a group could be reached. The nodes avoided data collisions via CSMA/CA and had data flow controls. The maximum data transfer rate was 4800 bits/s.
References
Network protocols
Building automation |
https://en.wikipedia.org/wiki/RenderDoc | RenderDoc is a free and open source frame debugger that can be used to analyze single frames generated by other software programs such as games. RenderDoc can provide in-depth analysis of single frames from any application that uses Vulkan, D3D11, OpenGL & OpenGL ES, or D3D12. RenderDoc also allows the user to manipulate a captured frame to inspect different things such as pipeline stage, commands, texture maps, models, assets, and more. RenderDoc can also capture assets outside the view of the game's camera. RenderDoc supports analyzing frame rendering costs on the graphics processing unit.
Optimizing GPU programs is about finding bottlenecks. RenderDoc provides information of the calls on the GPU, the number of invocation of each shader, and the number of primitives and fragments generated.
Baldur Karlsson started RenderDoc development as a spare-time project out of need for a reliable debugger and was then expanded to support more platforms.
In 2014 Crytek announced the source code for RenderDoc is released for free.
At GDC 2018, AMD announced that Radeon GPU Profiler would include support for RenderDoc. RenderDoc also integrates with well known game engines such as Unity and Unreal Engine. Oculus maintains its own fork of RenderDoc.
References
External links
https://www.gamingonlinux.com/wiki/Using_Renderdoc_capture
Cross-platform software
Graphics standards
Video game development
Debuggers |
https://en.wikipedia.org/wiki/SuperCam | SuperCam is a suite of remote-sensing instruments for the Mars 2020 Perseverance rover mission that performs remote analyses of rocks and soils with a camera, two lasers and four spectrometers to seek organic compounds that could hold biosignatures of past microbial life on Mars, if it ever existed there.
SuperCam was developed in collaboration between the Research Institute in Astrophysics and Planetology () of the University of Toulouse in France, the French Space Agency (CNES), Los Alamos National Laboratory, the University of Hawaii and the University of Málaga in Spain. The Principal Investigator is Roger Wiens from Los Alamos National Laboratory. SuperCam is an improved version of the successful ChemCam instruments of the Curiosity rover that have been upgraded with two different lasers and detectors. SuperCam is used in conjunction with the AEGIS (Autonomous Exploration for Gathering Increased Science) targeting system, a program in which Vandi Verma, NASA roboticist and engineer, helped develop.
In April 2018, SuperCam entered the final stages of assembly and testing. The flight model was installed to the rover in June 2019. The rover mission was launched on 30 July 2020.
Instruments
For measurements of chemical composition, the instrument suite uses a version of the successful ChemCam instruments of the Curiosity rover that have been upgraded with two different lasers and detectors. SuperCam's instruments are able to identify the kinds of chemicals that could be evidence of past life on Mars. SuperCam is a suite of various instruments, and the collection of correlated measurements on a target can be used to determine directly the geochemistry and mineralogy of samples.
The suite has several integrated instruments: Raman spectroscopy, time-resolved fluorescence (TRF) spectroscopy, and Visible and InfraRed (VISIR) reflectance spectroscopy to provide preliminary information about the mineralogy and molecular structure of samples under consideration, as we |
https://en.wikipedia.org/wiki/Membranome%20database | Membranome database provides structural and functional information about more than 6000 single-pass (bitopic) transmembrane proteins from Homo sapiens, Arabidopsis thaliana, Dictyostelium discoideum, Saccharomyces cerevisiae, Escherichia coli and Methanocaldococcus jannaschii. Bitopic membrane proteins consist of a single transmembrane alpha-helix connecting water-soluble domains of the protein situated at the opposite sides of a biological membrane. These proteins are frequently involved in the signal transduction and communication between cells in multicellular organisms.
The database provides information about the individual proteins including computationally generated three-dimensional models of their transmembrane alpha-helices spatially arranged in the membrane, topology, intracellular localizations, amino acid sequences, domain architecture, functional annotation and available experimental structures from the Protein Data Bank. It also provides a classification of bitopic proteins into 15 functional classes, more than 700 structural superfamilies and 1400 families, along with 3D structures of bitopic protein complexes which are also classified to different families.
The second Membranome version provides 3D models of more than 2000 parallel homodimers formed by TM α-helices of bitopic proteins from different organisms which were generated using TMDOCK program. The models of the homodimers were verified through comparison with available experimental data for nearly 600 proteins. The database includes downloadable coordinate files of transmembrane helices and their homodimers with calculated membrane boundaries. Membranome 3.0 version incorporates models generated by AlphaFold 2.
The database website provides access to related webservers, FMAP and TMDOCK which have been developed for modeling individual alpha-helices and their dimeric complexes in membranes. The database and webservers were used in experimental and bioinformatics studies of bitopic membrane |
https://en.wikipedia.org/wiki/Cooperative%20Observer%20Program | The NOAA Cooperative Observer Program (COOP) is a citizen weather observer network run by the U.S. National Weather Service (NWS) and National Centers for Environmental Information (NCEI). Over 8,700 volunteers from the fifty states and all territories report at least daily a variety of weather conditions such as daily maximum and minimum temperatures, 24-hour precipitation totals, including snowfall, and significant weather occurrences throughout a day that are recorded via remarks in observer logs. Some stations also report stream stage or tidal levels.
Daily observations are reported electronically or over the phone, and monthly logs are submitted electronically or via the mail. Many stations are located in rural areas but the network also includes long-term stations in most urban centers. Observation locations include farms, in urban and suburban areas, National Parks, seashores, and mountaintops. Volunteers are trained by local NWS offices who provide rain gauges, snowsticks, thermometers, or other instruments. Data is initially received and analyzed by local NWS offices then ultimately stored and analyzed by NCEI, which also does final data quality checks. The program began with act of Congress in 1890 and grew out a network of observers developed by the Smithsonian Institution. It was a backbone of the U.S. climatological observation network and remains an important network in providing long-term observations of particular locations.
The Cooperative Weather Observer network consists of manual observations of only a few variables and consists of daily summaries rather than being continuous (i.e. real-time). Because of these limitations and other sensor limitations, as well as to attain a denser network of observations, there has been a move to supplement the coop program using automated weather stations since the 1990s. NWS sponsored programs include the Citizen Weather Observer Program (CWOP) and Community Collaborative Rain, Hail and Snow Network (CoCoRaH |
https://en.wikipedia.org/wiki/Journal%20of%20Mathematical%20Physics%2C%20Analysis%2C%20Geometry | The Journal of Mathematical Physics, Analysis, Geometry is a quarterly peer-reviewed scientific journal covering mathematics as applied to physics. It is published by the Verkin Institute for Low Temperature Physics and Engineering and was established in 1994 as Mathematical Physics, Analysis, Geometry. Papers are published in English, Ukrainian, and Russian. The journal is abstracted and indexed by Scopus. According to the Journal Citation Reports, the journal has a 2017 impact factor of 0.531.
Editors-in-chief
The following persons are or have been editors-in-chief:
Vladimir Marchenko: 1994—1999
Iossif Ostrovskii: 2000—2004
Leonid Pastur: 2005—present
History
The Kharkov Mathematical Society was founded in 1879 and, starting in 1880, the society published the journal named Communications of the Kharkov Mathematical Society (Russian Сообщения и протоколы заседаний математического общества при Императорском Харьковском университете). Publication was suspended in 1960, but in 1965 due to the efforts of Naum Akhiezer the journals Theory of functions, functional analysis and their applications, and Ukrainian Geometric Collection» were established. In 1994, these journals were merged by the Mathematical Division of the Verkin Institute to establish the current journal. The first editor was Vladimir Marchenko.
References
External links
Mathematics journals
Academic journals published in Ukraine
Quarterly journals
Academic journals established in 1994
Multilingual journals |
https://en.wikipedia.org/wiki/Seminal%20fluid%20protein | Seminal fluid proteins (SFPs) or accessory gland proteins (Acps) are one of the non-sperm components of semen. In many animals with internal fertilization, males transfer a complex cocktail of proteins in their semen to females during copulation. These seminal fluid proteins often have diverse, potent effects on female post-mating phenotypes. SFPs are produced by the male accessory glands.
Seminal fluid proteins frequently show evidence of elevated evolutionary rates and are often cited as an example of sexual conflict.
Proteomics
SFPs are best studied in mammals and insects, especially in the common fruit fly, Drosophila melanogaster. Most species produce a wide variety of proteins that are transferred to females. For example, approximately 150 SFPs have been identified in D. melanogaster, 46 in the mosquito Anopheles gambae, and around 160 in humans.
Elevated evolution
Even between closely related species, the seminal fluid proteome can vary greatly. SFPs show elevated rates of DNA sequence change compared to non-reproductive genes (measured by Ka/Ks ratio) in many orders, including Diptera (flies), Lepidoptera (butterflies and moths), Rodentia, and Primates.
Additionally, SFPs show high rates of gene turnover compared to non-reproductive genes.
Function
The function of SFPs is best understood in D. melanogaster. SFPs play a role in male–male sperm competition. One study that manipulated the amount of SFPs male D. melanogaster produced found that when males were in competition, males that produced more SFPs sired a larger proportion of offspring.
In many insect species, significant changes occur in female behavior and physiology following mating; the isolated receipt of SFPs has been shown to be responsible for many of these changes. In D. melanogaster females, over 160 genes show either up or down-regulation following isolated SFP receipt. These transcriptomic changes are not limited to the female's reproductive tract. SFPs lengthen the refractory peri |
https://en.wikipedia.org/wiki/Tapered%20floating%20point | In computing, tapered floating point (TFP) is a format similar to floating point, but with variable-sized entries for the significand and exponent instead of the fixed-length entries found in normal floating-point formats. In addition to this, tapered floating-point formats provide a fixed-size pointer entry indicating the number of digits in the exponent entry. The number of digits of the significand entry (including the sign) results from the difference of the fixed total length minus the length of the exponent and pointer entries.
Thus numbers with a small exponent, i.e. whose order of magnitude is close to the one of 1, have a higher relative precision than those with a large exponent.
History
The tapered floating-point scheme was first proposed by Robert Morris of Bell Laboratories in 1971, and refined with leveling by Masao Iri and Shouichi Matsui of University of Tokyo in 1981, and by Hozumi Hamada of Hitachi, Ltd.
Alan Feldstein of Arizona State University and Peter Turner of Clarkson University described a tapered scheme resembling a conventional floating-point system except for the overflow or underflow conditions.
In 2013, John Gustafson proposed the Unum number system, a variant of tapered floating-point arithmetic with an exact bit added to the representation and some interval interpretation to the non-exact values.
See also
Logarithmic number system (LNS)
Symmetric level-index arithmetic (SLI)
References
Further reading
. Previously published in:
Computer arithmetic |
https://en.wikipedia.org/wiki/NARCIS | NARCIS (National Academic Research and Collaboration Information System) of the Netherlands was an online portal for searching Dutch scientific research publications and data. As of July 2018, NARCIS indexed 268,989 data sets and 1,707,486 publications, including a significant proportion of open access works.
It started in 2004 as a project of the Koninklijke Nederlandse Akademie van Wetenschappen, Information Centre of the Radboud University of Nijmegen (METIS), Nederlandse Organisatie voor Wetenschappelijk Onderzoek, and Vereniging van Universiteiten. Since 2011 the (DANS) operated NARCIS from headquarters in The Hague. In 2015, it was decided to replace the Digital Author Identifier used until then with the International Standard Name Identifier or ORCID. As of 3 July 2023, the portal has been decommissioned.
See also
Open access in the Netherlands
References
Further reading
External links
Official site
(includes NARCIS)
Science and technology in the Netherlands
2004 establishments in the Netherlands
Dutch digital libraries
Open access (publishing)
Online databases |
https://en.wikipedia.org/wiki/Intelligent%20toilet | An intelligent toilet or smart toilet (sometimes referred to as a modern day bidet) is a bathroom plumbing fixture or type of electronic bidet toilet which incorporates traditional bidet cleansing (for genital, buttocks and anal hygiene), with the added enhancement of modern SMART home technology.
Modern intelligent toilets incorporate electronic bidet cleansing functions into a ceramic toilet bowl for a low-profile and modern aesthetic. Intelligent toilet bidet functions are managed through voice command, mobile app or remote control navigation, depending on the toilet’s manufacturer, make and model.
Lower-cost options include a detachable toilet seat referred to as an "electronic bidet (“e-bidet”)" seat, which are becoming increasingly popular as well.
Features
Electronic bidet shower for front and rear cleansing
Seat-warming
Adjustable water temperature
Deodorizer
Air dryer
Many models offer water-conservative 1.0-1.28GPF
Illuminating nightlight
Hygienic applications
Intelligent toilets and electronic bidets alleviate numerous health concerns, especially for users with sensitive or damaged skin. By eliminating fecal bacteria left behind by toilet paper, intelligent toilets also dramatically increase a user’s hygiene by preventing the spread of bacteria to each user’s hands. Intelligent toilets can aid:
Skin rashes
Hemorrhoid discomfort
Physical limitations (that may make traditional toilet paper wiping difficult)
Post-childbirth perineum cleansing and soothing
Environmental applications
Intelligent toilets can be environmentally beneficial as they reduce a user’s need for toilet paper, saving households money on paper products and allowing users to reduce their paper consumption over time.
Modern high-efficiency intelligent toilets also use 1.0-1.28 gallons of water per flush, reducing consumers’ water consumption and waste as well.
History
In the 1700s, Bidets introduced throughout numerous European countries as a standard for hygi |
https://en.wikipedia.org/wiki/TechGig.com | TechGig.com is a competitive programming and technology community website owned by Times Internet.
Overview
TechGig was founded in 2010 by Times Business Solutions(A Division of Times Internet) as technology community website on competitive programming, technology jobs, webinars and tech news. Programmers can take up skill tests and mock interviews in C, C++, C#, Java, .Net, MySQL, Linux, Unix, Ajax, Python, LAMP, JSON and other technologies. TechGig launched a recruitment platform where companies can hire candidates based on their test results. The usual contest formats are – coding, MCQs, skill tests, whitepapers, and business case studies. As of July 2018, it has an active community of 2.5 million developers.
Competitive programming
Code Gladiators is an annual coding competition started in 2014 to identify the best coders in India. Themes of the 2018 contest were Artificial Intelligence, Machine Learning, Alexa, Big Data, Blockchain, Cloud Computing, E-commerce, Mobility, and Internet of Things. The total bounty for 2018 contest was INR 7.5 million.
Geek Goddess is a coding contest for women programmers. The themes of 2018 contest were UI, IoT and Data Science with a total prize money of INR 0.75 Million.
Virtual Campus League is an Inter–college coding contest to identify the best programmers & tech enthusiast students across India.
Records
In 2017, TechGig Code Gladiators has been declared as the largest programming event by the Guinness World Records. The Guinness officials stated that TechGig had 81,641 unique submissions from programmers and the previous record was set in 2012 by Baidu which had 30,634 entrants.
The Code Gladiators competition had won the Limca Book National Record twice - for their 2015 and 2016 editions.
References
External links
TechGig Official Website
Programming contests |
https://en.wikipedia.org/wiki/Windows%2010%20Mobile%20version%20history | Windows 10 Mobile is a discontinued mobile operating system developed by Microsoft. It was released to manufacturing on November 20, 2015, and was made generally available on March 17, 2016. In October 2017, Microsoft announced that it would pause the active development of Windows 10 Mobile, and future development will be limited to maintenance releases and security patches. The last feature update is the Fall Creators Update. The last version of Windows 10 Mobile reached the end of life on January 14, 2020. Development for Windows 10 Mobile has completely ceased since then.
Version history
Version 1511 (November Update)
Windows 10 Mobile November Update, also known as version 1511 and codenamed "Threshold 2", is the first major update to Windows 10 Mobile and the only one in a series of updates under the "Threshold" codenames. It carries the build number 10.0.10586. It was released to the public on November 12, 2015.
The update reached end of service on January 9, 2018.
Version 1607 (Anniversary Update)
Windows 10 Mobile Anniversary Update, also known as version 1607 and codenamed "Redstone 1", is the second major update to Windows 10 Mobile and the first in a series of updates under the "Redstone" codenames. It carries the build number 10.0.14393. The first preview was released on February 19, 2016. It was released to the public on August 16, 2016.
The update reached end of service after the release of build 14393.2551 on October 9, 2018.
Version 1703 (Creators Update)
Windows 10 Mobile Creators Update also known as version 1703 and codenamed "Redstone 2", is the third major update to Windows 10 Mobile and the second in a series of updates under the "Redstone" codenames. It carries the build number 10.0.15063. The first preview was released to Insiders on August 17, 2016. It was released to the public on April 25, 2017.
The update reached end of service after the release of build 15063.1868 on June 11, 2019.
Version 1709 (Fall Creators Update)
Windows |
https://en.wikipedia.org/wiki/Binary%20angular%20measurement | The terms binary angular measurement (BAM) and binary angular measurement system (BAMS) refer to certain methodologies for representing and manipulating angles using binary (base 2) fixed-point arithmetic. The unit of angular measure used in those methods may be called binary radian (brad) or binary degree.
These representation of angles are often used in numerical control and digital signal processing applications, such as robotics, navigation, computer games, and digital sensors. On the other hand, this system is not adequate for situations where the number of full turns must be measured, e.g. to monitor the rotation of vehicle wheels or of leadscrews.
Representation
Unsigned fraction of turn
In this system, an angle is represented by an n-bit unsigned binary number in the sequence 0, ..., 2n−1 that is interpreted as an multiple of 1/2n of a full turn; that is, 360/2n degrees or 2π/2n radians. The number can also be interpreted as a fraction of a full turn between 0 (inclusive) and 1 (exclusive) represented in binary fixed-point format with a scaling factor of 1/2n. Multiplying that fraction by 360° or 2π gives the angle in degrees in the range 0 to 360, or in radians, in the range 0 to 2π, respectively.
For example, with n = 8, the binary integers (00000000)2 (fraction 0.00), (01000000)2 (0.25), (10000000)2 (0.50), and (11000000)2 (0.75) represent the angular measures 0°, 90°, 180°, and 270°, respectively.
The main advantage of this system is that the addition or subtraction of the integer numeric values with the n-bit arithmetic used in most computers produces results that are consistent with the geometry of angles. Namely, the integer result of the operation is automatically reduced modulo 2n, matching the fact that angles that differ by an integer number of full turns are equivalent. Thus one does not need to explicitly test or handle the wrap-around, as one must do when using other representations (such as number of degrees or radians in floating-po |
https://en.wikipedia.org/wiki/StackPath | StackPath is an American edge computing platform provider headquartered in Dallas, Texas. Its founding team was led by Lance Crosby, who also co-founded SoftLayer Technologies, acquired by IBM in 2013.
Acquisitions
MaxCDN (CDN), 2016
Staminus (DDoS mitigation), 2016
Fireblade (WAF), 2016
Cloak (VPN), 2016
Highwinds Network Group (CDN and VPN), 2017
Server Density (Monitoring)
Subsidiaries
MaxCDN
NetDNA, LLC was founded in 2009 as a content delivery network (CDN) with a focus on enterprise customers. The company was founded by David Henzel and Christopher Ueland in Los Angeles.
In 2010, NetDNA partnered with Wowza to launch the HDDN.com brand, a CDN for streaming video.
By 2010, the MaxCDN brand was created as a simpler CDN for both small and large businesses. MaxCDN, LLC operated as a division of NetDNA, LLC.
In 2011, Ben Neumann was CEO of NetDNA. In 2011, NetDNA completed a funding round with Chelsea Management in Los Angeles.
In 2013, NetDNA rebranded the company and its services as MaxCDN, conslidating other services such as HDDN.com under the same name, with the original NetDNA enterprise service rebranded as MaxCDN Enterprise.
MaxCDN was acquired by StackPath in 2016.
Divestitures
In 2019, StackPath sold its VPN lines of business, including IPVanish (acquired as part of the Highwinds Network Group) and Encrypt.me (the new brand of Cloak), to J2 Global.
In August 2023, Stackpath sold its CDN line of business (primarily approximately 100 select enterprise customer contracts as well as other assets) to Akamai following its decision to cease its content delivery network operations. The transaction did not include the acquisition of StackPath personnel or technology.
Investors
StackPath has received funding from investors including Abry Partners, Juniper Networks, and Cox Communications.
Founders
StackPath was founded May 5, 2015, by Lance Crosby, Greg Bock, Steven Canale, Ryan Carter, Paul Drew, Kenji Fukasawa, Jason Gulledge, Andrew Hi |
https://en.wikipedia.org/wiki/Dshell | Dshell is an open source, Python-based, forensic analysis framework developed by the U.S. Army Research Laboratory, MD. This tool provides users with the ability to develop custom analysis modules which helps them understand events of cyber intrusion. This framework handles stream reassembly of both IPv4 and IPv6 network traffic and also includes geolocation and IP-to-ASN mapping for each connection. Additionally, the framework plug-ins are designed to aid in the understanding of network traffic and present results to the user in a concise, useful manner. Since Dshell is written entirely in Python, the code base can be customized to particular problems by modifying an existing decoder to extract different information from existing protocols.
The U.S. Army Research Laboratory (ARL) released a version of Dshell to GitHub social coding website on December 17, 2014, with more than 100 downloads and 2,000 unique visitors in 18 countries. Before it was publicly released, Dshell had a small, select community of users in several government organizations. Users could use the tool to find the exact information they needed from network data including looking up names, reassembled website requests or decoded malware traffic. ARL chose to release Dshell to GitHub because sharing it with the world created more security teams gaining another specialized tool to keep their networks secure. Furthermore, increasing the security of the Internet as a whole by increasing the number of skilled eyes looking for bugs and potential improvements throughout the code.
In 2014, NASA released more than 1,000 open source projects. Other agencies, such as the National Security Agency, the National Guard and the Air Force Research Laboratory joined shortly after the following year.
GitHub was chosen for Dshell because it allows members to easily download software code, store edits, and provide a mechanism to offer feedback to the original designer. Additionally, rolling enhancements into the off |
https://en.wikipedia.org/wiki/Megaphorb | Megaphorbs are areas with a community of tall, exuberant, perennial herbs. The plant communities provide ground cover and often exist on the margins of wooded areas, terrain that has recently been cleared, or in disturbed areas adjacent to creeks or rivers. These plant communities are often found in mountainous areas.
Areas that a megaphorb layer frequents are margins of wooded areas or forests that have recently been deforested or where logging has occurred. Given the now newly open area it provides ample conditions for a megaphorb layer to produce. Other areas in which megaphorbs may appear are along the banks of streams and rivers where there is a lack of plant life and nothing obstructing the upwards growth of the plants. The root systems provide structure to riverbanks, reducing erosion in seasons of heavy rainfall. Once peak growing conditions are over, the megaphorb layer begins to die off, which provides decaying plant material to the surrounding area. This feeds detritivore species, and eventually boosts the nutrient ratio within the soil.
References
Ecology |
https://en.wikipedia.org/wiki/Organizing%20center | Organizing center may refer to:
Microtubule organizing center
Spemann's Organizer
Certain groups of cells in mesoderm formation, see FGF and mesoderm formation
Primitive streak in Amniotes responsible for gastrulation
a small cell group underneath the stem cells in Arabidopsis and other plants
animal cap cells treated with activin
Pattern formation
Developmental biology |
https://en.wikipedia.org/wiki/Aurora%20Pulsed%20Radiation%20Simulator | The Aurora Pulsed Radiation Simulator (also known as the Aurora flash x-ray simulator) is a 14 TW flash gamma-ray simulator that was designed to simulate the effects of a nuclear weapon's bremsstrahlung, or gamma radiation, pulses on military electronic systems. It was built in 1971 by the U.S. Defense Atomic Support Agency (DASA), which eventually became the Defense Threat Reduction Agency, and the U.S. Department of Energy (DOE).
More than long and weighing at 1,450 tons, the Aurora Simulator was the first gamma radiation simulator of its size in the world at the time. It was also one of only four large machines in the United States that were built specifically to test complete nuclear weapons packages, with the other three being the Hermes I to III simulators at Sandia Base, New Mexico. Situated at the Harry Diamond Laboratories (which later became a part of the Army Research Laboratory) in Adelphi, Maryland, it was used to test complete weapons electronics packages from the warheads of intercontinental ballistic missiles (ICBMs) to satellites. After more than 20 years of use during the Cold War, the Aurora Simulator was officially decommissioned and disassembled in 1996.
In 1986, the Aurora facility set the world record for the largest amount of high-power microwave power generated from a virtual cathode oscillator. As a result, HDL was recognized by the American Defense Preparedness Association (ADPA) in 1987.
History
Following the use of the atomic bomb in World War II, studies on flash radiography found that field emission flash x-rays, which were previously used to analyze explosions, could simulate radiation from a nuclear bomb. Given this realization, the U.S. military began to prioritize the development of flash x-ray machines to test parts of missile packages during the 1960s.
After the Soviet Union demonstrated the use of the world's first anti-ballistic missiles (ABM) in 1964, DASA launched a series of projects in response that aimed to hasten th |
https://en.wikipedia.org/wiki/Remote%20Activation%20Munition%20System | The Remote Activation Munition System (RAMS) is a radio frequency controlled system that is used to remotely detonate demolition charges. It can also be used to remotely operate electronic equipment such as beacons, laser markers, and radios.
RAMS was developed by a team of researchers led by James Chopak at the Army Research Laboratory from 1996 to 2000. The system consists of a transmitter and two different types of receivers, one to initiate blasting caps and one to initiate C4 directly.
RAMS was designed to serve as a more portable and convenient alternative to conventional remote activation systems like the model XM-122, which was considered too big, heavy, and fragile for efficient use. In addition, the XM-122 was limited in its range (about 1 km) and relied on very large high capacity batteries.
In contrast, the RAMS weighed only a couple pounds and its microprocessor-based transmitter was powered by at most seven standard 9-volt batteries. The device was capable of reaching a range up to 2 kilometers, and the combination of the crystal filter in the receivers and the FM detector circuit made it possible to maintain high signal sensitivity at a low power consumption rate.
In addition, the RAMS was operational in harsh environments with temperatures as low as and as high as . It was also capable of functioning when submerged in saltwater, up to depths of . However, testing performed by the Army Research Laboratory have found that due to the low power levels of the RAMS receiver’s electrical signals output, the system has demonstrated a noticeable level of unreliability in performance past a certain distance.
More modern versions of the RAMS can weigh as little as and can reach a range of more than 5 kilometers, allowing operators to stand farther away from the blast at a safer distance.
References
Radio waves
Military technology
Explosives engineering |
https://en.wikipedia.org/wiki/Magneto%20Inductive%20Remote%20Activation%20Munition%20System | The Magneto Inductive Remote Activation Munition System (MI-RAMS) is a variant of the Remote Munition System (RAMS) that uses electromagnetic induction to control electronic equipment, including demolition charges, munitions, and active barriers. The handheld MI-RAMS receiver consists of a box-shaped device with a fixed bulkhead-style receptacle connector on the top with a non-leaking metal shell threaded in the rear section of the connector and sealed with an O-ring.
With the use of quasi-static AC magnetic fields, MI-RAMS is capable of sending signals through ice, rock, soil, water, and concrete. As a result, MI-RAMS is often used to remotely control ordnance items and communication systems in areas in which radio frequency devices under-perform or fail. These areas include caves, bunkers, tunnels, dense jungle, ice fields, urban structures, and up to 66 feet underwater. The wireless channel created by MI-RAMS does not produce any far field (RF) emissions, which decreases likelihood of detection outside of the operating area. The MI-RAMS transmitters and receivers have also been designed to work with existing communication technology, allowing other types of handsets to link to the system and communicate with each other as long as one MI-RAMS unit is present.
MI-RAMS was designed and modified by researchers at the Army Research Laboratory for U.S. Army Combat Engineer Forces and Army and Navy Special Operations Forces (SEALs) to aid in establishing terrain dominance.
References
Military technology
Radio waves
Explosives engineering |
https://en.wikipedia.org/wiki/National%20Documentation%20Centre%20%28Greece%29 | The National Documentation Centre (EKT; ) is a Greek public organisation that promotes knowledge, research, innovation and digital transformation. It was established in 1980 with funding from the United Nations Development Programme with the aim to strengthen the collection and distribution of research-related material, and to ensure full accessibility to it. It has been designated as a National Scientific Infrastructure, a National Authority of the Hellenic Statistical System, and National Contact Point for European Research and Innovation Programmes. Since August 2019, it has been established as a discrete public-interest legal entity under private law, and is supervised by the Ministry of Digital Governance (Article 59 / Law 4623/2019). The management bodies of EKT are the Administrative Board and the Director who, since 2013, has been Dr. Evi Sachini.
Goals
EKT's institutional role is the collection, organisation, documentation, digital preservation and dissemination of scientific, research and cultural information, content and data produced in Greece. EKT’s specific objectives, as stated on its official website, focus, amongst others, on:
Ensuring the dissemination of the country's scientific output.
Meeting the needs of academia, policymakers and research and business communities for information and reliable data.
Increasing the digital scientific and cultural content that is available in a user-friendly form and with legitimate rights of use for different target groups
Promoting open access to publications and data in the academic and research communities.
Collaboration with academic libraries for the standardization in organising and distributing metadata and digital scientific content.
Collaboration and joint actions with libraries, archives, museums, scientific and cultural institutions which produce and manage content, focusing on the establishment of common interoperability standards and the availability of metadata and digital content.
Provi |
https://en.wikipedia.org/wiki/Measurable%20group | In mathematics, a measurable group is a special type of group in the intersection between group theory and measure theory. Measurable groups are used to study measures is an abstract setting and are often closely related to topological groups.
Definition
Let a group with group law
.
Let further be a σ-algebra of subsets of the set .
The group, or more formally the triple is called a measurable group if
the inversion is measurable from to .
the group law is measurable from to
Here, denotes the formation of the product σ-algebra of the σ-algebras and .
Topological groups as measurable groups
Every second-countable topological group can be taken as a measurable group. This is done by equipping the group with the Borel σ-algebra
,
which is the σ-algebra generated by the topology. Since by definition of a topological group, the group law and the formation of the inverse element is continuous, both operations are in this case also measurable from to and from to , respectively. Second countability ensures that , and therefore the group is also a measurable group.
Related concepts
Measurable groups can be seen as measurable acting groups that act on themselves.
References
Measure theory
Group theory |
https://en.wikipedia.org/wiki/Psammophory | Psammophory is a method by which certain plants armor themselves with sand on their body parts, lowering the chance of them being eaten by animals. Over 200 species of plants hailing from 88 genera in 34 families have been identified as psammorphorous.
The term was first proposed in 1989 by scientists studying the habits of the beetle Georissus which actively covers its elytra with sand or mud particles.
References
Botany |
https://en.wikipedia.org/wiki/Mixed%20binomial%20process | A mixed binomial process is a special point process in probability theory. They naturally arise from restrictions of (mixed) Poisson processes bounded intervals.
Definition
Let be a probability distribution and let be i.i.d. random variables with distribution . Let be a random variable taking a.s. (almost surely) values in . Assume that are independent and let denote the Dirac measure on the point .
Then a random measure is called a mixed binomial process iff it has a representation as
This is equivalent to conditionally on being a binomial process based on and .
Properties
Laplace transform
Conditional on , a mixed Binomial processe has the Laplace transform
for any positive, measurable function .
Restriction to bounded sets
For a point process and a bounded measurable set define the restriction of on as
.
Mixed binomial processes are stable under restrictions in the sense that if is a mixed binomial process based on and , then is a mixed binomial process based on
and some random variable .
Also if is a Poisson process or a mixed Poisson process, then is a mixed binomial process.
Examples
Poisson-type random measures are a family of three random counting measures which are closed under restriction to a subspace, i.e. closed under thinning, that are examples of mixed binomial processes. They are the only distributions in the canonical non-negative power series family of distributions to possess this property and include the Poisson distribution, negative binomial distribution, and binomial distribution. Poisson-type (PT) random measures include the Poisson random measure, negative binomial random measure, and binomial random measure.
References
Point processes |
https://en.wikipedia.org/wiki/Siemens%20Goole | Siemens Goole is a train factory located in Goole, East Riding of Yorkshire, England. Siemens Mobility assessed several sites in the United Kingdom before settling on Goole with an intent to build the plant if it were successful in gaining orders for new rolling stock. After a June 2018 announcement that Siemens had won the bid to build 94 London Underground New Tube for London trains for the Piccadilly line of the London Underground, Siemens confirmed it would go ahead with building the factory, though this was then subject to further delays as other rolling stock companies objected to Siemens being given the contract.
In November 2018, the High Court upheld the decision to award the contract to Siemens, but allowed the other train companies to pursue claims for damages. Groundwork on the site commenced in July 2020.
History
During a period of sustained new train building for the United Kingdom rail network, Siemens announced in March 2018 that it would build a train factory in the UK if it was successful in acquiring new contracts for trains. Siemens have already built for the UK market with their Desiro family design which includes the class 185 DMUs, and the class 350 and 450 EMUs among others. The company already employs over 4,400 people in the United Kingdom in rail and other transport related roles, with eight purpose built sites that provide rolling stock care.
After searching for a suitable site, Siemens acquired some land in Goole, in the East Riding of Yorkshire, stating that on winning orders, they would build a factory there. In June 2018, Siemens were awarded a £1.5 billion contract for 94 tube trains for the Piccadilly line. This triggered the plan to build the factory making Siemens the fourth train builder with an actual factory presence in the UK and the third new build factory within eight years. Previous to this, the Bombardier factory in Derby was the only UK based train builder.
The plant will cost £200 million and cover a area and employ |
https://en.wikipedia.org/wiki/PFCP | Packet Forwarding Control Protocol (PFCP) is a 3GPP protocol used on the Sx/N4 interface between the control plane and the user plane function, specified in TS 29.244. It is one of the main protocols introduced in the 5G Next Generation Mobile Core Network (aka 5GC), but also used in the 4G/LTE EPC to implement the Control and User Plane Separation (CUPS). PFCP and the associated interfaces seek to formalize the interactions between different types of functional elements used in the Mobile Core Networks as deployed by most operators providing 4G, as well as 5G, services to mobile subscribers. These 2 types of components are:
The Control Plane (CP) functional elements, handling mostly signaling procedures (e.g. network attachment procedures, management of User-data Plane paths and even delivery of some light-weight services as SMS)
The User-data Plane (UP) functional elements, handling mostly packet forwarding, based on rules set by the CP elements (e.g. packet forwarding for IPv4, IPv6 - or possibly even Ethernet with future 5G deployments - between the various supported wireless RANs and the PDN representing the Internet or an enterprise network).
PFCP's scope is similar to that of OpenFlow, however it was engineered to serve the particular use-case of Mobile Core Networks.
PFCP is also used on the interface between the control plane and user plane functions of a disaggregated BNG, as defined by the BroadBand Forum in TR-459.
Overview
Albeit similar to GTP in concepts and implementation, PFCP is complementary to it. It provides the control means for a signaling component of the Control-Plane to manage packet processing and forwarding performed by a User-Plane component. Typical EPC or 5G Packet Gateways are split by the protocol in 2 functional parts, allowing for a more natural evolution and scalability.
The PFCP protocol is used on the following 3GPP mobile core interfaces:
Sxa - between SGW-C and SGW-U
Sxb - between PGW-C and PGW-U
Sxc - between TDF |
https://en.wikipedia.org/wiki/List%20of%20transponder%20codes | The following list shows specific aeronautical transponder codes, and ranges of codes, that have been used for specific purposes in various countries. Traditionally each country has allocated transponder codes by their own scheme with little commonality across borders. The list is retained for historic interest.
Pilots are normally required to apply the code, allocated by air traffic control, to that specific flight. Occasionally countries may specify generic codes to be used in the absence of an allocated code. Such generic codes are specified in that country's Aeronautical Information Manual or Aeronautical Information Publication. There also are standard transponder codes for defined situations defined by the International Civil Aviation Organization (marked below as ICAO).
Transponder codes shown in this list in the color RED are for emergency use only such as an aircraft hijacking, radio communication failure or another type of emergency.
References
Encodings
Avionics |
https://en.wikipedia.org/wiki/DATAmatic%201000 | The DATAmatic 1000 is an obsolete computer system from Honeywell introduced in 1957. It uses vacuum tubes and crystal diodes for logic, and featured a unique magnetic tape format for storage.
The CPU uses a 48-bit word (plus four check bits). A word can hold 12 decimal digits (11 digits plus sign) or 8 six-bit alphanumeric characters. The system includes magnetic core storage of 2000 words or 24,000 digits in two banks, called "High-Speed Memory." Words in High-Speed Memory are also called "registers" in the documentation. The system also includes two input and two output tape buffer storage units of 62 words (744 digits) each.
The instructions are three address and all operations are storage-to-storage.
History
Datamatic Corporation was established in 1954 as a joint venture of Raytheon and Honeywell. In 1955 Honeywell bought out Raytheon's interest and the company became known as "Honeywell DATAmatic." Later Datamatic was renamed Honeywell Information Systems (HIS).
Tape format
The DATAmatic uses long reels of wide magnetic tape. The tape format uses 36 tracks ("channels")—31 data and 5 for checking. Data is recorded as blocks of 62 words. Within a block two words are recorded serially on each of the 31 tracks. A full reel of tape can contain 50,000 blocks (37,200,000 words). Data transfer rate is 60,000 digits per second. A DATAmatic 1000 system can attach up to 100 tape units.
When the tape is initially written data is recorded only in alternate blocks. When the end of the reel is reached the tape is recorded in reverse in previously unused blocks. This eliminates a rewind operation when reading or writing full reels. The format allows individual blocks to be rewritten in place.
Other peripherals
Standard card, paper tape, and line printer output was via offline input and output converters that wrote or read magnetic tape. Available peripheral devices included:
Punched card reader, standard 80-column punched cards, 900 cards per minute
Punched tape rea |
https://en.wikipedia.org/wiki/IndieAuth | IndieAuth is an open standard decentralized authentication protocol that uses OAuth 2.0 and enables services to verify the identity of a user represented by a URL, as well as to obtain an access token, that can be used to access resources under the control of the user.
IndieAuth is developed in the IndieWeb community and was published as a W3C Note. It was published as a W3C Note by the Social Web Working Group due to lacking the time needed to formally progress it to a W3C recommendation, despite having several interoperable implementations.
Implementations
WordPress IndieAuth Plugin
Known
Micro.blog
Grav (CMS) IndieAuth Plugin
Drupal IndieWeb Plugin
Cellar Door
See also
OpenID
WebID
References
External links
IndieAuth specification
World Wide Web Consortium standards
Password authentication
Federated identity
Identity management initiative
Computer access control protocols
WordPress |
https://en.wikipedia.org/wiki/Dynamic%20line%20rating%20for%20electric%20utilities | Dynamic line rating (DLR), also known as real-time thermal rating (RTTR), is an electric power transmission operation philosophy aiming at maximizing load, when environmental conditions allow it, without compromising safety. Research, prototyping and pilot projects were initiated in the 1990s, but the emergence of the "smart grid" stimulated electric utilities, scientists and vendors to develop comprehensive and sustainable solutions.
Principles and applications
The current-carrying capacity, or ampacity, of overhead lines starts with the type of conductor used. The conductor choice determines its electrical resistance and other physical parameters for dynamic line rating (DLR). Electric current passing through the conductor causes heating according to Joule's first law, resulting in the conductor expanding and the line sagging. Transmission ratings are set with a maximum allowable conductor temperature (annealing temperature) and minimum clearance rules to comply with legislation and regulation.
Ampacity has traditionally been limited by conductor thermal capacity defined in terms of stating rating (SR), based on a predetermined set of worst-case weather scenarios (typically a combination of cross-winds, high solar radiation and seasonal high ambient temperatures).
Real-time dynamic line rating
More often than not, there are unused margins between the limits defined by static ratings and the "true limits" measured by a DLR or real-time thermal rating (RTTR) system.
Several methods have been developed since the 1990s, most of which rely upon sensors deployed on overhead lines to measure parameters in real-time. Other systems utilize weather stations that monitor environmental conditions without contacting the line. Data received from any method is reported to a main computer for processing. Control center operators access usable data (line temperature, ratings, forecasts, historical values) in pseudo-real-time through a human-machine interface (HMI).
Wind genera |
https://en.wikipedia.org/wiki/Ribenboim%20Prize | The Ribenboim Prize, named in honour of Paulo Ribenboim, is awarded by the Canadian Number Theory Association for distinguished research in number theory by a mathematician who is Canadian or has close connections to Canadian mathematics. Normally the winner will have received their Ph.D. in the last 12 years. The winner is expected to give a plenary talk at the award ceremony.
Winners
See also
List of mathematics awards
References
Mathematics awards
Canadian awards
Awards established in 1999
.
1999 establishments in Canada |
https://en.wikipedia.org/wiki/Exapunks | Exapunks is a programming game developed by Zachtronics. It was released into early access on August 9, 2018, and fully released on October 22, 2018.
Gameplay
Exapunks takes place in an alternate timeline in the year 1997. The fictional world of Exapunks is heavily computerized, and a disease called "the phage" is ravaging the population, turning the bodies of those affected into computerized components. The player takes on the role of Moss, a hacker who breaks into computer systems in order to afford a $700/day drug to slow the progress of his phage affliction. His hacking missions are given to him by a mysterious artificial intelligence known as EMBER-2.
Each mission takes place inside a network of interconnected and specialized computer systems. Using programmable software agents called EXAs, the player must accomplish each given task by writing computer code to cleverly manipulate the data stored on the network's systems. The EXAs' instruction set features a few simple opcodes for movement, data processing, network messaging, and interfacing with files and registers. Due to their limited memory capacity, these tasks often require several agents working together in a highly coordinated fashion. EXA units also have the ability to replicate themselves inside the network. Typical missions include retrieving data from secured storage systems, hacking into company databases, and causing an automated teller machine to dispense free cash. Some puzzles also require the player to hack Moss's body to maintain his health. Some puzzles challenge the player to hacker battles, where they must pit their EXAs against an opponent's agents, for example altering a television station's program to broadcast Moss' content instead.
Players are generally free to write code for EXAs with as many EXAs as necessary, those are often limited by the number of opcodes that can be used. The player's solution must satisfy 100 different case scenarios iterating on the same problem. When the pl |
https://en.wikipedia.org/wiki/Genetics%20in%20fiction | Aspects of genetics including mutation, hybridisation, cloning, genetic engineering, and eugenics have appeared in fiction since the 19th century.
Genetics is a young science, having started in 1900 with the rediscovery of Gregor Mendel's study on the inheritance of traits in pea plants. During the 20th century it developed to create new sciences and technologies including molecular biology, DNA sequencing, cloning, and genetic engineering. The ethical implications were brought into focus with the eugenics movement.
Since then, many science fiction novels and films have used aspects of genetics as plot devices, often taking one of two routes: a genetic accident with disastrous consequences; or, the feasibility and desirability of a planned genetic alteration. The treatment of science in these stories has been uneven and often unrealistic. The film Gattaca did attempt to portray science accurately but was criticised by scientists.
Background
Modern genetics began with the work of the monk Gregor Mendel in the 19th century, on the inheritance of traits in pea plants. Mendel found that visible traits, such as whether peas were round or wrinkled, were inherited discretely, rather than by blending the attributes of the two parents. In 1900, Hugo de Vries and other scientists rediscovered Mendel's research; William Bateson coined the term "genetics" for the new science, which soon investigated a wide range of phenomena including mutation (inherited changes caused by damage to the genetic material), genetic linkage (when some traits are to some extent inherited together), and hybridisation (crosses of different species).
Eugenics, the production of better human beings by selective breeding, was named and advocated by Charles Darwin's cousin, the scientist Francis Galton, in 1883. It had both a positive aspect, the breeding of more children with high intelligence and good health; and a negative aspect, aiming to suppress "race degeneration" by preventing supposedly "d |
https://en.wikipedia.org/wiki/Dixon%20rings | Dixon rings are a form of random packing used in chemical processing. They consist of a stainless steel mesh formed into a ring with a central divider, and are intended to be packed randomly into a packed column. Dixon rings provide a large surface area and low pressure drop while maintaining a high mass transfer rate, making them useful for distillations and many other applications.
Background
Packed columns
Packed columns are used in a range of industries to allow intimate contact between two immiscible fluids which can be liquid/liquid or liquid/gas. The fluids are passed through in a countercurrent flow through a column.
Random column packing
Random column packing used to characterize the maximum volume fraction of a solid object obtained when they are packed randomly. This method of packing has been used since the early 1820s; the types of packing used were originally made out of glass spheres. However, in 1850 they were replaced by a more porous pumice stone and pieces of coke.
In the early 20th century Friedrich Raschig realized the importance of a high void fraction and having the internal surface of the packing media take part in the mass transfer. He designed the Raschig ring, which was more effective than previous forms of random packing and became very popular. Raschig rings are usually built from ceramic or metal and provided a large surface area within the column for interaction between liquid and gas vapors.
The development of the Dixon ring
In 1943 Dr Olaf George Dixon of ICI applied for a patent of a new product for column distillation. He used stainless steel mesh instead of sheet steel in the Lessing ring in order to improve the pressure drop of the packed column (in fact, they were called "wire gauze Lessing rings" in a 1949 publication).
High performance was achieved between radioactive and non-radioactive materials, and despite high cost they are still used in the nuclear industry e. g. for water detritiation.
Application
Dixon rin |
https://en.wikipedia.org/wiki/Manolis%20Kellis | Manolis Kellis (; born 1977) is a professor of Computer Science at the Massachusetts Institute of Technology (MIT) in the area of Computational Biology and a member of the Broad Institute of MIT and Harvard. He is the head of the Computational Biology Group at MIT and is a Principal Investigator in the Computer Science and Artificial Intelligence Lab (CSAIL) at MIT.
Kellis is known for his contributions to genomics, human genetics, epigenomics, gene regulation, genome evolution, disease mechanism, and single-cell genomics. He co-led the NIH Roadmap Epigenomics Project effort to create a comprehensive map of the human epigenome, the comparative analysis of 29 mammals to create a comprehensive map of conserved elements in the human genome, the ENCODE, GENCODE, and modENCODE projects to characterize the genes, non-coding elements, and circuits of the human genome and model organisms. A major focus of his work is understanding the effects of genetic variations on human disease, with contributions to obesity, diabetes, Alzheimer's disease, schizophrenia, and cancer.
Education and early career
Kellis was born in Greece, moved with his family to France when he was 12, and came to the U.S. in 1993. He obtained his PhD from MIT, where he worked with Eric Lander, founding director of the Broad Institute, and Bonnie Berger, professor at MIT and received the Sprowls award for the best doctorate thesis in Computer Science, and the first Paris Kanellakis graduate fellowship. Prior to computational biology, he worked on artificial intelligence, sketch and image recognition, robotics, and computational geometry, at MIT and at the Xerox Palo Alto Research Center.
Research and career
As of July 2018, Manolis Kellis has authored 187 journal publications that have been cited 68,380 times. He has helped direct several large-scale genomics projects, including the Roadmap Epigenomics project, the Encyclopedia of DNA Elements (ENCODE) project, the Genotype Tissue-Expression (GTEx) p |
https://en.wikipedia.org/wiki/L%C3%B6schian%20number | In number theory, the numbers of the form x2 + xy + y2 for integer x, y are called the Löschian numbers (or Loeschian numbers). These numbers are named after August Lösch. They are the norms of the Eisenstein integers. They are a set of whole numbers, including zero, and having prime factorization in which all primes congruent to 2 mod 3 have even powers (there is no restriction of primes congruent to 0 or 1 mod 3).
Properties
Every square number is a Löschian number (by setting x or y to 0).
Moreover, every number of the form for m and x integers is a Löschian number (by setting y=mx).
There are infinitely many Löschian numbers.
Given that odd and even integers are equally numerous, the probability that a Löschian number is odd is 0.75, and the probability that it is even is 0.25. This follows from the fact that is even only if x and y are both even.
The greatest common divisor and the least common multiple of any two or more Löschian numbers are also Löschian numbers.
The product of two Löschian numbers is always a Löschian number.
The product of a Löschian number and a non-Löschian number is never a Löschian number.
References
Integer sequences |
https://en.wikipedia.org/wiki/SVET%20plant%20growth%20system | SVET (, "light") was a plant cultivation unit on the Kristall module of the Mir space station. It was installed in 1990 and operated until 2001. Brassica rapa was successfully grown there in 1997.
The project was a joint Russian-Bulgarian one, developed at the Space Research and Technology Institute in Sofia.
References
Astrobiology
Mir
Space program of Bulgaria
Space-flown life
Bulgaria–Soviet Union relations |
https://en.wikipedia.org/wiki/Data%20Transfer%20Project | The Data Transfer Project (DTP) is an open-source initiative which features data portability between multiple online platforms. The project was launched and introduced by Google on July 20, 2018, and has currently partnered with Facebook, Microsoft, Twitter, and Apple.
Background
The project was formed by the Google Data Liberation Front in 2017, hoping to provide a platform that could allow individuals to move their online data between different platforms, without the need of downloading and re-uploading data. The ecosystem is achieved by extracting different files through various available APIs released by online platforms and translating such codes so that it could be compatible with other platforms. Similarly, the Data Transfer Project is currently being used as a part of Google Takeout and a similar program in Facebook (called "Access your information"), allowing the two personal data downloading services to be compatible with each other. This allows data to be easily transferred from the two platforms.
On July 20, 2018, the joint project was announced. The source code, which has been uploaded to GitHub, was mainly written by Google and Microsoft's engineers.
On July 30, 2019, Apple announced that it will be joining the project, allowing data portability in iCloud.
Implementations
On December 2, 2019, Facebook announced the ability for users to transfer photos and videos to Google Photos, originally available only in a select few countries. This expanded over the following months, and on June 4, 2020, Facebook announced full global availability of this feature.
See more
Data portability
Google Takeout
References
External links
Google
Computer-related introductions in 2018
Free network-related software
Interoperability |
https://en.wikipedia.org/wiki/Mars%20Multispectral%20Imager%20for%20Subsurface%20Studies | Mars Multispectral Imager for Subsurface Studies (MA-MISS) is a miniaturized imaging spectrometer designed to provide imaging and spectra by reflectance in the near-infrared (NIR) wavelength region and determine the mineral composition and stratigraphy. The instrument is part of the science payload on board the European Rosalind Franklin rover, tasked to search for biosignatures, and scheduled to land on Mars in spring 2023. MA-MISS is essentially inside a drill on the Rover, and will take measurements of the sub-surface directly.
MA-MISS will help on the search for biosignatures by studying minerals and ices in situ before the collection of samples. The instrument is integrated within the Italian core drill system called DEEDRI, and it will be dedicated to in situ studies of the mineralogy inside the excavated holes in terms of visible and infrared spectral reflectance.
The Principal Investigator is Maria Cristina De Sanctis, from the INAF (Istituto di Astrofisica Spaziale e Fisica Cosmica) in Italy.
Overview
The instrument is based on the design conceived by planetary scientist Angioletta Coradini in 2001. MA-MISS is integrated in the Rosalind Franklin rover 2-metre DEEDRI core drill and shares its structure and electronics. It will perform visible and near-infrared spectroscopy (NIR) in the 0.4 to 2.2 μm range of the excavated borehole wall.
A 5 watt lamp and an optical fiber array will provide the illumination of the target (about 100 μm spot) as well as collect the scattered light from the target. An optical fiber carries the light to the spectrometer.
It will be operated periodically during the pauses of the drilling activity and will also be able to provide images of the hole wall by a sapphire window connected to an array of optical fibres. MA-MISS will provide important information about mineralogy, petrology, and geological processes of sedimentary materials of the Martian subsurface. It will also give insights about materials that have not been |
https://en.wikipedia.org/wiki/Quartz%20Roasting%20Pits%20Complex | Quartz Roasting Pits Complex is a heritage-listed quartz roasting kiln located 10km north of Hill End, Mid-Western Regional Council, New South Wales, Australia. It was built from 1854 to 1855. It is also known as Cornish Roasting Pits. The property is owned by the New South Wales Office of Environment and Heritage. It was added to the New South Wales State Heritage Register on 2 April 1999.
History
The Hill End Quartz Roasting Pits Complex was established by the Colonial Gold Mining Company in 1855, on the traditional land of the Wiradjuri people, to provide gold extraction facilities for those working claims on the Tambaroora and Hill End goldfields. Although at this time alluvial mining was the primary method of gold recovery, the Tambaroora fields also supported the earliest attempts at reef mining in Australia, over fifteen years before the reef mining boom of the 1870s. The Quartz Roasting Pits Complex is one of the oldest goldrush sites in Australia and represents one of the first attempts to process gold bearing ore. It also represents an unusual technological solution to the problems initially experienced in extracting payable gold from the quartz reefs, in its development of the earliest form of quartz firing technology in Australia. With kilns for roasting gold bearing quartz, a sophisticated battery and dam system for crushing and washing the ore and houses for workers, the Complex provides tangible evidence of technological, social and domestic relationships during this very early stage of Australia's goldmining history. Gold bearing quartz was brought to the Complex from surrounding mines and roasted to make it easier to crush (a relatively unusual process) and then crushed in a steam powered battery. Gold was then extracted from the powdered ore through a process of sieving and washing.
Despite its impressive technological achievements, the operation, managed and probably designed by Alfred Spence, was short-lived, closing in 1856, only eighteen mon |
https://en.wikipedia.org/wiki/Trichoderma%20asperellum | Trichoderma asperellum Samuels, Lieckf & Nirenberg is a species of fungus in the family Hypocreaceae. It can be distinguished from T. viride by molecular and phenotypic characteristics. The most important molecular characteristics are divergent ITS-1 and 28S sequences and RFLP's of the endochitinase gene. Main phenotypic characters are conidial ornamentation and arrangement and branching of the conidiophores.
Important Isolates
This species has been used commercially and experimentally as a biopesticide for plant disease control: some commercial isolates were previously placed in T. harzianum.
References
External links
Trichoderma
Biopesticides
Biotechnology
Biological pest control
Fungi described in 1999 |
https://en.wikipedia.org/wiki/MicrOmega-IR | MicrOmega-IR is an infrared hyperspectral microscope that is part of the science payload on board the European Rosalind Franklin rover, tasked to search for biosignatures on Mars. The rover is planned to land on Mars in the mid- or late 2020s. MicrOmega-IR will analyse in situ the powder material derived from crushed samples collected by the rover's core drill.
Development
The MicrOmega mnemonic is derived from its French name Micro observatoire pour la mineralogie, l'eau, les glaces et l'activité; IR stands for infrared. It was developed by France's Institut d'Astrophysique Spatiale at the CNRS. France has also flown MicrOmega on other missions such as the 2011 Fobos-Grunt and the Hayabusa2 MASCOT mobile lander currently exploring asteroid Ryugu. France is also developing a variant called MacrOmega Near-IR Spectrometer for the Martian Moons Exploration (MMX) lander, a Japanese sample-return mission to Mars' moon Phobos.
The Principal Investigator of the MicrOmega-IR for the Rosalind Franklin rover is Jean-Pierre Bibring, a French astronomer and planetary scientist at the Institut d'Astrophysique Spatiale. Co-PIs are astrobiologists Frances Westall and Nicolas Thomas.
MicrOmega was developed by a consortium including:
IAS (Orsay, France)
LESIA (Meudon, France)
CBM (Orléans, France),
University Of Bern (Bern, Switzerland)
Russian Space Research Institute (IKI) (Moscow, Russia)
Overview
MicrOmega-IR is a visible and infrared hyperspectral microscope that is designed to characterize the texture and composition of crushed samples presented to the instrument. Its objective is to study mineral grain assemblages in detail to try to unravel their geological origin, structure and composition, including potential organics. These data will be vital for interpreting past and present geological processes and environments on Mars. Because MicrOmega-IR is an imaging instrument, it can also be used to identify grains that are particularly interesting, and assign them a |
https://en.wikipedia.org/wiki/Raman%20Laser%20Spectrometer | Raman Laser Spectrometer (RLS) is a miniature Raman spectrometer that is part of the science payload on board the European Space Agency Rosalind Franklin rover, tasked to search for biosignatures and biomarkers on Mars. The rover is planned to be launched in August–October 2022 and land on Mars in spring 2023.
Raman spectroscopy is a very useful technique employed to identify mineral phases produced by water-related processes. RLS will help to identify organic compounds and search for microbial life by identifying the mineral products and indicators of biologic activities. RLS will provide geological and mineralogical context information that will be scientifically cross-correlated with that obtained by other instruments.
Overview
Raman spectroscopy is sensitive to the composition and structure of any organic compound, making it a powerful tool for the definitive identification and characterisation of biomarkers, and providing direct information of potential biosignatures of past microbial life on Mars. This instrument will also provide general mineralogical information for igneous, metamorphous, and sedimentary processes.
RST will also correlate its spectral information with other spectroscopic and imaging instruments such as the Infrared Spectrometer and MicrOmega-IR. This will be the first Raman analyser to be deployed for a planetary exploration. The first version for the rover was presented by Fernando Rull-Perez and Sylvestre Maurice in 2003. The RLS is being developed by a European consortium integrated by Spanish, French, German and UK partners. The Principal Investigator is Fernando Rull-Perez, from Spanish Astrobiology Center. The co-investigator is from Observatoire Midi-Pyrénées (LAOMP), France.
The three major components are the Spectrometer Unit, the Control and Excitation Unit (includes the power converters), and Optical head.
Principle and operation
The RLS instrument provides a structural fingerprint by which molecules can be identified. It |
https://en.wikipedia.org/wiki/Infrared%20Spectrometer%20for%20ExoMars | Infrared Spectrometer for ExoMars (ISEM) is an infrared spectrometer for remote sensing that is part of the science payload on board the European Space Agency Rosalind Franklin rover, tasked to search for biosignatures and biomarkers on Mars. The rover is planned to be launched in August–October 2022 and land on Mars in spring 2023.
ISEM will provide context assessment of the surface mineralogy in the vicinity of the Rosalind Franklin rover for selection of potential astrobiological targets. The Principal Investigator is Oleg Korablev from the Russian Space Research Institute (IKI).
Overview
The Infrared Spectrometer for ExoMars (ISEM) is being developed by the Russian Space Research Institute (IKI). It will be the first instance of near-infrared spectroscopy (NIR) observations done from the Mars surface. The instrument will be installed on the Rosalind Franklin rover's mast to measure reflected solar radiation in the near infrared range for context assessment of the surface mineralogy in the vicinity of Rosalind Franklin for selection of potential astrobiological targets. As the number of samples obtained with the drill will be limited, the selection of high-value sites for drilling will be crucial. Working with PanCam (a high-resolution panoramic camera), ISEM will aid in the selection of potential targets, especially water-bearing minerals, for close-up investigations and drilling sites.
ISEM could detect, if present, organic compounds, including evolving trace gases such as hydrocarbons like methane in the Martian atmosphere.
Objectives
The stated science objectives of ISEM are:
Geological investigation and study a composition of Martian soils in the uppermost few millimeters of the surface.
Characterisation of the composition of surface materials, discriminating between various classes of silicates, oxides, hydrated minerals and carbonates.
Identification and mapping of the distribution of aqueous alteration products on Mars.
Real-time assessment of |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.