source
stringlengths
31
227
text
stringlengths
9
2k
https://en.wikipedia.org/wiki/Vibration%20isolation
Vibration isolation is the prevention of transmission of vibration from one component of a system to others parts of the same system, as in buildings or mechanical systems. Vibration is undesirable in many domains, primarily engineered systems and habitable spaces, and methods have been developed to prevent the transfer of vibration to such systems. Vibrations propagate via mechanical waves and certain mechanical linkages conduct vibrations more efficiently than others. Passive vibration isolation makes use of materials and mechanical linkages that absorb and damp these mechanical waves. Active vibration isolation involves sensors and actuators that produce disruptive interference that cancels-out incoming vibration. Passive isolation "Passive vibration isolation" refers to vibration isolation or mitigation of vibrations by passive techniques such as rubber pads or mechanical springs, as opposed to "active vibration isolation" or "electronic force cancellation" employing electric power, sensors, actuators, and control systems. Passive vibration isolation is a vast subject, since there are many types of passive vibration isolators used for many different applications. A few of these applications are for industrial equipment such as pumps, motors, HVAC systems, or washing machines; isolation of civil engineering structures from earthquakes (base isolation), sensitive laboratory equipment, valuable statuary, and high-end audio. A basic understanding of how passive isolation works, the more common types of passive isolators, and the main factors that influence the selection of passive isolators: Common passive isolation systems Pneumatic or air isolators These are bladders or canisters of compressed air. A source of compressed air is required to maintain them. Air springs are rubber bladders which provide damping as well as isolation and are used in large trucks. Some pneumatic isolators can attain low resonant frequencies and are used for isolating large industr
https://en.wikipedia.org/wiki/Institut%20oc%C3%A9anographique
The () is an ocean education organization based in Monaco. The institute manages two ocean museums (in Monaco and Paris) and lobbies globally for the preservation of the oceans' ecology. History The Institut océanographique was founded in 1906 by Albert I, Prince of Monaco (the International Hydrographic Organization was launched in Monaco in 1921). In 1957, Jacques Cousteau was appointed director of the Institut océanographique. In 1961, the Institut océanographique reached an agreement with the International Atomic Energy Agency to relocate the International Laboratory of Marine Radioactivity in Monaco. In 1998, the Institut océanographique organized the second International Aquariology Congress (the first edition was held in 1958). The European Union of Aquarium Curators was created during this event. The institute implemented the "Monaco Blue Initiative", a global initiative focused on deep sea biodiversity and large marine species, and backed by UNESCO. In 2016, hundreds of cubic meters of archives belonging to the Institut were found in the Schœlcher campus of the University of the French West Indies (Martinique). In 2019, the Institut océanographique invested 5 million euros in the opening of a care center for marine species, especially sick or injured turtles. Locations Oceanographic Museum of Monaco Institut océanographique de Paris
https://en.wikipedia.org/wiki/Deutsche%20Neurowissenschaften-Olympiade
The Deutsche Neurowissenschaften-Olympiade (DNO or German Neuroscience Olympiad) is a series of competitions for students in grades 8 to 13, aimed at promoting interest in neuroscience. The competition takes place at two levels: city and national. Each competition is in question and answer format and conducted in English. The winner of the national German Neuroscience Olympiad will represent Germany at the International Brain Bee competition. History The "Deutsche Neurowissenschaften-Olympiad" is part of the International Brain Bee program which oversees more than 150 competitions in over 45 countries each year. The program was founded in 1999 by Dr. Norbert Myslinski in the United States. The program has since expanded beyond the United States. Dr. Julianne McCall established the "German Brain Bee" competition in 2011, the first of its kind in Germany. The German Brain Bee takes place in Heidelberg and includes over 100 high schools across Germany. In 2015, "The German Brain Bee" was renamed the Deutsche Neurowissenschaften-Olympiad and is organized by the Deutsche Neurowissenschaften-Olympiad Association. In 2016, the board of the DNO began the process of expanding the competition throughout Germany. Mission The motto of DNO is "Driving Connectivity", the organization's mission is to motivate and inspire young people to learn about the human brain, foster interest in the various fields of neurosciences and pursuit of careers in basic neuroscience research. Through local and national competitions, summer research schools, practical courses, and other outreach programs DNO would like to encourage national and international exchange among students on their way towards a scientific career. The philosophy of DNO is that through encouraging young students to work together to face new and exciting challenges in neuroscience research, there can be a further understanding of brain functions. Competition The competition, which until 2016 has only taken place in Heid
https://en.wikipedia.org/wiki/K-independent%20hashing
In computer science, a family of hash functions is said to be k-independent, k-wise independent or k''-universal if selecting a function at random from the family guarantees that the hash codes of any designated k keys are independent random variables (see precise mathematical definitions below). Such families allow good average case performance in randomized algorithms or data structures, even if the input data is chosen by an adversary. The trade-offs between the degree of independence and the efficiency of evaluating the hash function are well studied, and many k-independent families have been proposed. Background The goal of hashing is usually to map keys from some large domain (universe) into a smaller range, such as bins (labelled ). In the analysis of randomized algorithms and data structures, it is often desirable for the hash codes of various keys to "behave randomly". For instance, if the hash code of each key were an independent random choice in , the number of keys per bin could be analyzed using the Chernoff bound. A deterministic hash function cannot offer any such guarantee in an adversarial setting, as the adversary may choose the keys to be the precisely the preimage of a bin. Furthermore, a deterministic hash function does not allow for rehashing: sometimes the input data turns out to be bad for the hash function (e.g. there are too many collisions), so one would like to change the hash function. The solution to these problems is to pick a function randomly from a large family of hash functions. The randomness in choosing the hash function can be used to guarantee some desired random behavior of the hash codes of any keys of interest. The first definition along these lines was universal hashing, which guarantees a low collision probability for any two designated keys. The concept of -independent hashing, introduced by Wegman and Carter in 1981, strengthens the guarantees of random behavior to families of designated keys, and adds a guarant
https://en.wikipedia.org/wiki/Kaynemaile%20Architectural%20Mesh
Kaynemaile is a chainmail fabric consisting of polycarbonate interlinked rings connected together by liquid-state assembly to form a flexible mesh sheet. It is made from polycarbonate and used in the architecture and design industry. It was invented by Kayne Horsham in 2004. History The first prototypes of Kaynemaile were created within Weta Workshop's Creatures, Armor and Weapons department for The Lord of The Rings movie trilogy. Real metal chainmail was identified as too heavy for the actors to wear or do stunts. Kaynemaile was created as a chainmail fabric that looked and moved like real chainmail, but without the weight, made of polycarbonate resin rings with no joints or seams in the links. The material was created by a team of New Zealand based chainmail technicians manually interconnect millions of polypropylene (PP) rings and used electroplating process to apply silver on the outside. The result produced the realistic looking chainmail armor used throughout the movies. The new chainmail rings were incorporated into the costume design of characters such as Aragorn and Boromir, the Gondorian race and Rohan armies, and many of the Orcs. The chainmail was nicknamed “Kayne's-mail” by Viggo Mortensen. Mass production A company, Kaynemaile Ltd, was co-founded in 2006 by Kayne Horsham and Robyn Downham to streamline the process and make it commercially available, as architectural mesh. Its creation is now automated through a liquid-state injection molding fabrication process. The mesh is used for to form large architectural building wraps, 3D shapes or seamless screens. It has been used as a lightweight, reflective, permeable membrane for buildings. Awards Architizer A+ Awards 2020, Product Winner in Facades-Building Envelopes & Cladding 2019 Architecture MasterPrize, Winner in Building Envelope & Construction Materials Wellington Gold Awards 2018, Finalist Global Gold NYCxDesign Award 2017, Best Architectural Product designEX Award 2014, Best New Inn
https://en.wikipedia.org/wiki/NUTS%20statistical%20regions%20of%20Serbia
As a candidate country of the European Union, Serbia (RS) is in the process of being included in the Nomenclature of Territorial Units for Statistics (NUTS). However, due to the ongoing dispute with Kosovo, it has not yet agreed with the European Commission and Eurostat. The proposed three NUTS levels are: Local administrative units Below the NUTS levels, the two LAU (Local Administrative Units) levels are: See also Administrative divisions of Serbia ISO 3166-2 codes of Serbia
https://en.wikipedia.org/wiki/Spoof%20surface%20plasmon
Spoof surface plasmons, also known as spoof surface plasmon polaritons and designer surface plasmons, are surface electromagnetic waves in microwave and terahertz regimes that propagate along planar interfaces with sign-changing permittivities. Spoof surface plasmons are a type of surface plasmon polariton, which ordinarily propagate along metal and dielectric interfaces in infrared and visible frequencies. Since surface plasmon polaritons cannot exist naturally in microwave and terahertz frequencies due to dispersion properties of metals, spoof surface plasmons necessitate the use of artificially-engineered metamaterials. Spoof surface plasmons share the natural properties of surface plasmon polaritons, such as dispersion characteristics and subwavelength field confinement. They were first theorized by John Pendry et al. Theory Surface plasmon polaritons (SPP) result from the coupling of delocalized electron oscillations ("surface plasmon") to electromagnetic waves ("polariton"). SPPs propagate along the interface between a positive- and a negative-permittivity material. These waves decay perpendicularly from the interface ("evanescent field"). For a plasmonic medium that is stratified along the z-direction in Cartesian coordinates, dispersion relation for SPPs can be obtained from solving Maxwell's equations: where is the wave vector that is parallel to the interface. It is in the direction of propagation. is the angular frequency. is the speed of light. and are the relative permittivies for metal and the dielectric. Per this relation, SPPs have shorter wavelengths than light in free space for a frequency band below surface plasmon frequency; this property, as well as subwavelength confinement, enables new applications in subwavelength optics and systems beyond the diffraction-limit. Nevertheless, for lower frequency bands such as microwave and terahertz, surface plasmon polariton modes are not supported; metals function approximately as perfect electric
https://en.wikipedia.org/wiki/Pan-African%20colours
Pan-African colours is a term that may refer to two different sets of colours: Green, yellow and red, the colours of the flag of Ethiopia, have come to represent the pan-Africanist ideology due to the country's history of having avoided being taken over by a colonial power. Numerous African countries have adopted the colours into their national flags, and they are similarly used as a symbol by many Pan-African organisations and the Rastafari movement. Red, black, and green, first introduced by Marcus Garvey in 1920, have also come to represent Pan-Africanism, and are shown on the pan-African flag. These colours have also been incorporated on national flags, and they have sometimes been used to represent black nationalism rather than Pan-Africanism. Green-yellow-red Green, yellow, and red are now found on the national flags of many African nations. The colour combination was borrowed from the flag of Ethiopia. The Ethiopian flag has influenced the flags of many Pan-African organizations and polities. Except for relatively brief periods of influence and occupation by the Kingdom of Italy, Ethiopia remained outside European control during the colonial era by defeating the Italian army at the battle of Adwa in 1896, ending the Italian protectorate. As a result, the country drew the admiration of many newly independent states in Africa. The adoption of the Ethiopian national colours by many Pan-African entities is a consequence of this. The first African state to adopt a gold, red and green flag upon independence was Ghana in 1957, designed by Theodosia Okoh. Red-black-green The Universal Negro Improvement Association and African Communities League (UNIA) founded by Marcus Garvey has a constitution that defines red, black, and green as the Pan-African colours: "red representing the noble blood that unites all people of African ancestry, the colour black for the people, green for the rich land of Africa." The UNIA flag was designated the official colours of Black A
https://en.wikipedia.org/wiki/History%20of%20software
Software is a set of programmed instructions stored in the memory of stored-program digital computers for execution by the processor. Software is a recent development in human history, and it is fundamental to the Information Age. Ada Lovelace's programs for Charles Babbage's Analytical Engine in the 19th century is often considered the founder of the discipline. However, the mathematician's efforts remained theoretical only, as the technology of Lovelace and Babbage's day proved insufficient to build his computer. Alan Turing is credited with being the first person to come up with a theory for software in 1935, which led to the two academic fields of computer science and software engineering. The first generation of software for early stored-program digital computers in the late 1940s had its instructions written directly in binary code, generally written for mainframe computers. Later, the development of modern programming languages alongside the advancement of the home computer would greatly widen the scope and breadth of available software, beginning with assembly language, and continuing through functional programming and object-oriented programming paradigms. Before stored-program digital computers Origins of computer science Computing as a concept goes back to ancient times, with devices such as the abacus, the Antikythera mechanism, Astrolabes, Mechanical Astronomical clocks and Mechanical Calculators. The Antikythera mechanism is an example for a highly complex ancient mechanical Astronomical device. However, these devices were pure hardware and had no software - their computing powers were directly tied to their specific form and engineering. Software requires the concept of a general-purpose processor - what is now described as a Turing machine - as well as computer memory in which reusable sets of routines and mathematical functions comprising programs can be stored, started, and stopped individually, and only appears recently in human history
https://en.wikipedia.org/wiki/Mivar-based%20approach
The Mivar-based approach is a mathematical tool for designing artificial intelligence (AI) systems. Mivar (Multidimensional Informational Variable Adaptive Reality) was developed by combining production and Petri nets. The Mivar-based approach was developed for semantic analysis and adequate representation of humanitarian epistemological and axiological principles in the process of developing artificial intelligence. The Mivar-based approach incorporates computer science, informatics and discrete mathematics, databases, expert systems, graph theory, matrices and inference systems. The Mivar-based approach involves two technologies: Information accumulation is a method of creating global evolutionary data-and-rules bases with variable structure. It works on the basis of adaptive, discrete, mivar-oriented information space, unified data and rules representation, based on three main concepts: “object, property, relation”. Information accumulation is designed to store any information with possible evolutionary structure and without limitations concerning the amount of information and forms of its presentation. Data processing is a method of creating a logical inference system or automated algorithm construction from modules, services or procedures on the basis of a trained mivar network of rules with linear computational complexity. Mivar data processing includes logical inference, computational procedures and services. Mivar networks allow us to develop cause-effect dependencies (“If-then”) and create an automated, trained, logical reasoning system. Representatives of Russian association for artificial intelligence (RAAI) – for example, V. I. Gorodecki, doctor of technical science, professor at SPIIRAS and V. N. Vagin, doctor of technical science, professor at MPEI declared that the term is incorrect and suggested that the author should use standard terminology. History While working in the Russian Ministry of Defense, O. O. Varlamov started developing the theory
https://en.wikipedia.org/wiki/Leapfrog%20filter
A leapfrog filter is a type of active circuit electronic filter that simulates a passive electronic ladder filter. Other names for this type of filter are active-ladder or multiple feedback filter. The arrangement of feedback loops in the signal flow-graph of the simulated ladder filter inspired the name leapfrog filter, which was coined by Girling and Good. The leapfrog filter maintains the low component sensitivity of the passive ladder filter that it simulates. Synthesis The definition and synthesis of leapfrog filters is described by Temes & LaPatra, Sedra & Brackett, Chen and Wait, Huelsman & Korn. Synthesis of leapfrog filters typically includes the following steps: Determine a prototype passive ladder filter that has the desired frequency response. Usually a doubly terminated prototype is used. Write the equations relating element current to voltage across the element in a form suitable for expression as a signal-flow graph. Draw the signal-flow graph. The nodes of the signal-flow graph will include both voltages and currents. The branch gains will include impedances and admittances. Convert all nodes of the signal-flow graph to voltages and all impedances to dimensionless transmittances. This is accomplished by dividing all impedance elements by , an arbitrary resistance and multiplying all admittance elements by . This scaling does not change the frequency response. Manipulate the signal-flow graph so that the gains feeding each summing node have the same signs. This is done as an implementation convenience. At the completion of this step, typically, all the feedback gains in the signal-flow graph will be +1 and the signs of the gain blocks in the forward path will alternate. As a result, some of the nodes, including the main output, may have a 180° phase inversion. This is usually of no consequence. The gain blocks are implemented with active filters and interconnected as indicated by the signal-flow graph. Often, state variable filter
https://en.wikipedia.org/wiki/Logic%20of%20graphs
In the mathematical fields of graph theory and finite model theory, the logic of graphs deals with formal specifications of graph properties using sentences of mathematical logic. There are several variations in the types of logical operation that can be used in these sentences. The first-order logic of graphs concerns sentences in which the variables and predicates concern individual vertices and edges of a graph, while monadic second-order graph logic allows quantification over sets of vertices or edges. Logics based on least fixed point operators allow more general predicates over tuples of vertices, but these predicates can only be constructed through fixed-point operators, restricting their power. A sentence may be true for some graphs, and false for others; a graph is said to model , written , if is true of the vertices and adjacency relation of . The algorithmic problem of model checking concerns testing whether a given graph models a given sentence. The algorithmic problem of satisfiability concerns testing whether there exists a graph that models a given sentence. Although both model checking and satisfiability are hard in general, several major algorithmic meta-theorems show that properties expressed in this way can be tested efficiently for important classes of graphs. Other topics of research in the logic of graphs include investigations of the probability that a random graph has a property specified within a particular type of logic, and methods for data compression based on finding logical sentences that are modeled by a unique graph. First order In the first-order logic of graphs, a graph property is expressed as a quantified logical sentence whose variables represent graph vertices, with predicates for equality and adjacency testing. Examples For instance, the condition that a graph does not have any isolated vertices may be expressed by the sentence where the symbol indicates the undirected adjacency relation between two vertices. This sen
https://en.wikipedia.org/wiki/Line%20number
In computing, a line number is a method used to specify a particular sequence of characters in a text file. The most common method of assigning numbers to lines is to assign every line a unique number, starting at 1 for the first line, and incrementing by 1 for each successive line. In the C programming language the line number of a source code line is one greater than the number of new-line characters read or introduced up to that point. Programmers could also assign line numbers to statements in older programming languages, such as Fortran, JOSS, and BASIC. In Fortran, not every statement needed a line number, and line numbers did not have to be in sequential order. The purpose of line numbers was for branching and for reference by formatting statements. Both JOSS and BASIC made line numbers a required element of syntax. The primary reason for this is that most operating systems at the time lacked interactive text editors; since the programmer's interface was usually limited to a line editor, line numbers provided a mechanism by which specific lines in the source code could be referenced for editing, and by which the programmer could insert a new line at a specific point. Line numbers also provided a convenient means of distinguishing between code to be entered into the program and direct mode commands to be executed immediately when entered by the user (which do not have line numbers). Largely due to the prevalence of interactive text editing in modern operating systems, line numbers are not a feature of most programming languages, even modern Fortran and Basic. History FORTRAN In Fortran, as first specified in 1956, line numbers were used to define input/output patterns, to specify statements to be repeated, and for conditional branching. For example: DIMENSION ALPHA(25), RHO(25) 1) FORMAT(5F12.4) 2) READ 1, ALPHA, RHO, ARG SUM = 0.0 DO 3 I=1, 25 IF (ARG-ALPHA(I)) 4,3,3 3) SUM = SUM + ALPHA(I) 4) VALUE = 3.14159*RHO(I-1) PRINT 1, ARG, SUM
https://en.wikipedia.org/wiki/Pequi%20oil
Pequi oil is a seed oil, extracted from the seeds of the Caryocar brasiliense, which is native to Brazil. The pale yellow mesocarp oil is extracted and used directly as a cooking oil. In Brazil, there are projects among indigenous people groups to develop pequi oil production as a means of economic development. Pequi oil contains 44.2% palmitic and 51.7% oleic acid, with linolenic and stearic acids in much smaller amounts.
https://en.wikipedia.org/wiki/NUTS%20statistical%20regions%20of%20Latvia
The three NUTS (Nomenclature of Territorial Units for Statistics) levels for Latvia (LV) are: NUTS codes LV0 Latvia LV00 Latvia LV003 Kurzeme LV005 Latgale LV006 Rīga LV007 Pierīga LV008 Vidzeme LV009 Zemgale Local administrative units NUTS 3 level is subdivided into LAU 1 units. LAU 1 units are subdivided into LAU 2 units. LAU (Local Administrative Units) in Latvia according to NUTS regulation (up to 31.12.2011): History of changes of LAU 2 codes till 1 July 2009 can be viewed here. LAU in Latvia according to NUTS regulation (starting 01.01.2012): * Due to Administrative Territorial Reform (1 July 2009) administrative districts no longer exist as administrative units in Latvia. Municipalities and parishes amalgamated and formed new LAU 2 units – municipalities. LAU 2 codes from 1 July 2009 updated version. Changes of LAU 2 codes on 1 July 2009 can be viewed here See also Subdivisions of Latvia ISO 3166-2 codes of Latvia FIPS region codes of Latvia Sources Hierarchical list of the Nomenclature of territorial units for statistics - NUTS and the Statistical regions of Europe Correspondence between the NUTS levels and the local administrative units of each EU country List of current NUTS codes Download current NUTS codes (ODS format) Municipalities of Latvia, Statoids.com Latvia Nuts Reform in Latvia
https://en.wikipedia.org/wiki/Project%20Muse
Project MUSE, a non-profit collaboration between libraries and publishers, is an online database of peer-reviewed academic journals and electronic books. Project MUSE contains digital humanities and social science content from over 250 university presses and scholarly societies around the world. It is an aggregator of digital versions of academic journals, all of which are free of digital rights management (DRM). It operates as a third-party acquisition service like EBSCO, JSTOR, OverDrive, and ProQuest. MUSE's online journal collections are available on a subscription basis to academic, public, special, and school libraries. Currently, more than 2,500 libraries worldwide subscribe. Electronic book collections became available for institutional purchase in January 2012. Thousands of scholarly books are available on the platform. History Project MUSE was founded in 1993 as a joint project between the Johns Hopkins University Press and the Milton S. Eisenhower Library at the Johns Hopkins University. With grants from the Andrew W. Mellon Foundation and the National Endowment for the Humanities, Project MUSE was launched online alongside the JHU Press Journals in 1995. Beginning in 2000, journals from other scholarly publishers were integrated into MUSE's online collections. Additional publishers have added journals each subsequent year. In January 2012, a new interface was launched which incorporated its current journal collection with electronic books published by members of the University Press Content Consortium (UPCC). The platform is powered by the WAIS searching utility called SWISH (Simple Web Indexing System for Humans), which allows Boolean searching in single issues, volumes, or across all 40+ titles. In cases where footnotes exist in articles, the footnote number is presented as a hyperlink to the article's bibliography or notes section. Journals Project MUSE offers tiered-pricing structures to meet budgetary and research needs of subscribing institu
https://en.wikipedia.org/wiki/Bousso%27s%20holographic%20bound
The Bousso bound captures a fundamental relation between quantum information and the geometry of space and time. It appears to be an imprint of a unified theory that combines quantum mechanics with Einstein's general relativity. The study of black hole thermodynamics and the information paradox led to the idea of the holographic principle: the entropy of matter and radiation in a spatial region cannot exceed the Bekenstein–Hawking entropy of the boundary of the region, which is proportional to the boundary area. However, this "spacelike" entropy bound fails in cosmology; for example, it does not hold true in our universe. Raphael Bousso showed that the spacelike entropy bound is violated more broadly in many dynamical settings. For example, the entropy of a collapsing star, once inside a black hole, will eventually exceed its surface area. Due to relativistic length contraction, even ordinary thermodynamic systems can be enclosed in an arbitrarily small area. To preserve the holographic principle, Bousso proposed a different law, which does not follow from black hole physics: the covariant entropy bound or Bousso bound. Its central geometric object is a lightsheet, defined as a region traced out by non-expanding light-rays emitted orthogonally from an arbitrary surface B. For example, if B is a sphere at a moment of time in Minkowski space, then there are two lightsheets, generated by the past or future directed light-rays emitted towards the interior of the sphere at that time. If B is a sphere surrounding a large region in an expanding universe (an anti-trapped sphere), then there are again two light-sheets that can be considered. Both are directed towards the past, to the interior or the exterior. If B is a trapped surface, such as the surface of a star in its final stages of gravitational collapse, then the lightsheets are directed to the future. The Bousso bound evades all known counterexamples to the spacelike bound. It was proven to hold when the entropy
https://en.wikipedia.org/wiki/CRAM-MD5
In cryptography, CRAM-MD5 is a challenge–response authentication mechanism (CRAM) based on the HMAC-MD5 algorithm. As one of the mechanisms supported by the Simple Authentication and Security Layer (SASL), it is often used in email software as part of SMTP Authentication and for the authentication of POP and IMAP users, as well as in applications implementing LDAP, XMPP, BEEP, and other protocols. When such software requires authentication over unencrypted connections, CRAM-MD5 is preferred over mechanisms that transmit passwords "in the clear," such as LOGIN and PLAIN. However, it can't prevent derivation of a password through a brute-force attack, so it is less effective than alternative mechanisms that avoid passwords or that use connections encrypted with Transport Layer Security (TLS). Protocol The CRAM-MD5 protocol involves a single challenge and response cycle, and is initiated by the server: Challenge: The server sends a base64-encoded string to the client. Before encoding, it could be any random string, but the standard that currently defines CRAM-MD5 says that it is in the format of a Message-ID email header value (including angle brackets) and includes an arbitrary string of random digits, a timestamp, and the server's fully qualified domain name. Response: The client responds with a string created as follows. The challenge is base64-decoded. The decoded challenge is hashed using HMAC-MD5, with a shared secret (typically, the user's password, or a hash thereof) as the secret key. The hashed challenge is converted to a string of lowercase hex digits. The username and a space character are prepended to the hex digits. The concatenation is then base64-encoded and sent to the server Comparison: The server uses the same method to compute the expected response. If the given response and the expected response match, then authentication was successful. Strengths The one-way hash and the fresh random challenge provide three types of security: Others
https://en.wikipedia.org/wiki/Alpha%20globulin
Alpha globulins are a group of globular proteins in plasma that are highly mobile in alkaline or electrically charged solutions. They inhibit certain blood proteases and show significant inhibitor activity. The alpha globulins typically have molecular weights of around 93 kDa. Examples Alpha globulins include certain hormones, proteins that transport hormones, and other compounds, including prothrombin and HDL. Alpha 1 globulins α1-antitrypsin Alpha 1-antichymotrypsin Orosomucoid (acid glycoprotein) Serum amyloid A Alpha 1-lipoprotein Protein HC Alpha 2 globulins Haptoglobin Alpha-2u globulin α2-macroglobulin Ceruloplasmin Thyroxine-binding globulin Alpha 2-antiplasmin Protein C Alpha 2-lipoprotein Angiotensinogen Cortisol binding globulin Vitamin D-binding protein
https://en.wikipedia.org/wiki/Tracks%20Ahead
Tracks Ahead is a television series about railroading, produced by Milwaukee PBS, originally solely for their station WMVS, then syndicated to public television stations, starting in 1990. In general, the series examines all aspects of railroading, both in the United States and in the rest of the world. Content covers a wide range of railroad-related materials. This includes scenic rail journeys, short-line railroads, layouts (in various gauges of model, tinplate, scale, garden), artists, photographers, and other railroad related material. Background At the dawn of cable television, Chuck Zehner, a Milwaukee train enthusiast, began producing and hosting the interview format show Just Trains on Milwaukee's local access channel on Viacom Cable. Eventually the show was picked up on the cable network around Milwaukee, After 72 shows Milwaukee's WMVS Channel 10 (PBS) agreed to air a new magazine format show On Track in the Milwaukee market. For the second season it was renamed Tracks Ahead and expanded to the PBS network. History The first season (released 1990) was hosted by Charles E. "Chuck" Zehner and the second season (released 1992) by Ward Kimball. Both were later repackaged and re-released with Spencer Christian as the host. All subsequent series have featured Christian. The primary audience for the series is women (ages 25–63) and children (ages 3–18) by 63.4%. The remaining audience is railroad interest groups. Season 5 was the first in 1080i high definition; season 6 was the first to incorporate 5.1 enhanced (surround sound) audio. Tracks Ahead 7 started airing in January 2009. As with the previous two seasons, it is in high definition, with digital 5.1 surround sound. The 14-part season includes segments from Japan, the Caribbean, Patagonia, and all around the United States. Tracks Ahead 8 starting airing in 2011 and marked the final season produced by series originator David K. Baule. Tracks Ahead 9 began production in 2012, with the final s
https://en.wikipedia.org/wiki/List%20of%20interactive%20geometry%20software
Interactive geometry software (IGS) or dynamic geometry environments (DGEs) are computer programs which allow one to create and then manipulate geometric constructions, primarily in plane geometry. In most IGS, one starts construction by putting a few points and using them to define new objects such as lines, circles or other points. After some construction is done, one can move the points one started with and see how the construction changes. History The earliest IGS was the Geometric Supposer, which was developed in the early 1980s. This was soon followed by Cabri in 1986 and The Geometer's Sketchpad. Comparison There are three main types of computer environments for studying school geometry: supposers, dynamic geometry environments (DGEs) and Logo-based programs. Most are DGEs: software that allows the user to manipulate ("drag") the geometric object into different shapes or positions. The main example of a supposer is the Geometric Supposer, which does not have draggable objects, but allows students to study pre-defined shapes. Nearly all of the following programs are DGEs. For a related, comparative physical example of these algorithms, see Lenart Sphere. License and platform The following table provides a first comparison of the different software according to their license and platform. 3D Software General features The following table provides a more detailed comparison : Macros Features related to macro constructions: (TODO) Loci Loci features related to IGS: (TODO) Proof We detail here the proof related features. (TODO) Measurements and calculation Measurement and calculation features related to IGS: (TODO) Graphics export formats Object attributes 2D programs C.a.R. C.a.R. is a free GPL analog of The Geometer's Sketchpad (GSP), written in Java. Cabri Cabri Cabri was developed by the French school of mathematics education in Grenoble (Laborde, 1993) CaRMetal CaRMetal is a free GPL software written in Java. Derived from C.
https://en.wikipedia.org/wiki/Potassium%20simplex%20optimized%20medium
Potassium Simplex Optimized Medium (KSOM) is a specialized medium primarily used for in vitro culture of mouse preimplantation embryos in research. Development The culture of preimplantation embryos is of paramount importance as it serves as a foundation for exploring embryo development, advancing assisted reproductive technology, and facilitating the generation of genetically modified animals. The primary motivation behind the establishment of KSOM was to address the phenomenon known as the two-cell block, wherein mouse embryos encounter developmental arrest at the two-cell stage. Researchers Lawitts and Biggers, in their groundbreaking work in the early 1990s, employed a sequential simplex optimization strategy to solve this issue. This method involves a systematic approach to optimize multiple variables simultaneously, allowing for the fine-tuning of media components. Their efforts culminated in the formulation of the Simplex Optimized Medium (SOM). Further refinements based on the intracellular K+/Na+ ratio in the two-cell stages led to the evolution of KSOM, potassium-supplemented SOM. Because KSOM uses a bicarbonate buffering mechanism, it is dependent on a CO2 incubator to maintain the right pH. The handling of embryos outside a CO2 incubator necessitates the use of HEPES buffered media, such as M2. Applications and advancements Following the foundational success of KSOM, researchers sought to further enhance the mediums' efficacy. This led to the introduction of amino acid supplementation, resulting in the variant known as KSOM/AA. The availability of amino acids in the medium further aided in the development of mouse embryos and augmented their growth potential. KSOM/AA has ever since become the most widely used medium for the culture of preimplantation mouse embryos in research. Ready-to-use KSOM/AA can be obtained from various commercial vendors but it can also be prepared in-house based on published protocols. See also Embryo culture Genetically
https://en.wikipedia.org/wiki/Critical%20point%20%28thermodynamics%29
In thermodynamics, a critical point (or critical state) is the end point of a phase equilibrium curve. One example is the liquid–vapor critical point, the end point of the pressure–temperature curve that designates conditions under which a liquid and its vapor can coexist. At higher temperatures, the gas cannot be liquefied by pressure alone. At the critical point, defined by a critical temperature Tc and a critical pressure pc, phase boundaries vanish. Other examples include the liquid–liquid critical points in mixtures, and the ferromagnet–paramagnet transition (Curie temperature) in the absence of an external magnetic field. Liquid–vapor critical point Overview For simplicity and clarity, the generic notion of critical point is best introduced by discussing a specific example, the vapor–liquid critical point. This was the first critical point to be discovered, and it is still the best known and most studied one. The figure to the right shows the schematic P-T diagram of a pure substance (as opposed to mixtures, which have additional state variables and richer phase diagrams, discussed below). The commonly known phases solid, liquid and vapor are separated by phase boundaries, i.e. pressure–temperature combinations where two phases can coexist. At the triple point, all three phases can coexist. However, the liquid–vapor boundary terminates in an endpoint at some critical temperature Tc and critical pressure pc. This is the critical point. The critical point of water occurs at and . In the vicinity of the critical point, the physical properties of the liquid and the vapor change dramatically, with both phases becoming even more similar. For instance, liquid water under normal conditions is nearly incompressible, has a low thermal expansion coefficient, has a high dielectric constant, and is an excellent solvent for electrolytes. Near the critical point, all these properties change into the exact opposite: water becomes compressible, expandable, a poor diele
https://en.wikipedia.org/wiki/Hua%20Luogeng
Hua Luogeng or Hua Loo-Keng (; 12 November 1910 – 12 June 1985) was a Chinese mathematician and politician famous for his important contributions to number theory and for his role as the leader of mathematics research and education in the People's Republic of China. He was largely responsible for identifying and nurturing the renowned mathematician Chen Jingrun who proved Chen's theorem, the best known result on the Goldbach conjecture. In addition, Hua's later work on mathematical optimization and operations research made an enormous impact on China's economy. He was elected a foreign associate of the US National Academy of Sciences in 1982. He was elected a member of the standing Committee of the first to sixth National people's Congress, Vice-Chairman of the sixth National Committee of the Chinese People's Political Consultative Conference (April 1985) and Vice-Chairman of the China Democratic League (1979). He joined the Communist Party of China in 1979. Hua did not receive a formal university education. Although awarded several honorary PhDs, he never got a formal degree from any university. In fact, his formal education only consisted of six years of primary school and three years of secondary school. For that reason, Xiong Qinglai, after reading one of Hua's early papers, was amazed by Hua's mathematical talent, and in 1931 Xiong invited him to study mathematics at Tsinghua University. Biography Early years (1910–1936) Hua Luogeng was born in Jintan, Jiangsu on 12 November, 1910. Hua's father was a small businessman. Hua met a capable math teacher in middle school who recognized his talent early and encouraged him to read advanced texts. After middle school, Hua enrolled in Chinese Vocational College in Shanghai, and there he distinguished himself by winning a national abacus competition. Although tuition fees at the college were low, living costs proved too high for his means, and Hua was forced to leave a term before graduating. After failing to find a
https://en.wikipedia.org/wiki/Pudendal%20venous%20plexus
The pudendal venous plexus (vesicoprostatic plexus) lies behind the arcuate pubic ligament and the lower part of the pubic symphysis, and in front of the bladder and prostate. Its chief tributary is the deep dorsal vein of the penis, but it also receives branches from the front of the bladder and prostate. It communicates with the vesical venous plexus and with the internal pudendal vein and drains into the vesical and hypogastric veins. See also Prostatic venous plexus
https://en.wikipedia.org/wiki/Genome%20mining
Genome mining describes the exploitation of genomic information for the discovery of biosynthetic pathways of natural products and their possible interactions. It depends on computational technology and bioinformatics tools. The mining process relies on a huge amount of data (represented by DNA sequences and annotations) accessible in genomic databases. By applying data mining algorithms, the data can be used to generate new knowledge in several areas of medicinal chemistry, such as discovering novel natural products. History In the mid- to late 1980s, researchers have increasingly focused on genetic studies with the advancing sequencing technologies. The GenBank database was established in 1982 for the collection, management, storage, and distribution of DNA sequence data due to the increasing availability of DNA sequences. With the increasing number of genetic data, biotechnological companies have been able to use human DNA sequence to develop protein and antibody drugs through genome mining since 1992. In the late 1990s, many companies, such as Amgen, Immunec, Genentech were able to develop drugs that progressed to the clinical stage by adopting genome mining. Since the Human Genome Project was completed in the early 2000, researchers have been sequencing the genomes of many microorganisms. Subsequently, many of these genomes have been carefully studied to identify new genes and biosynthetic pathways. Algorithms As large quantities of genomic sequence data began to accumulate in public databases, genetic algorithms became important to decipher the enormous collection of genomic data. They are commonly used to generate high-quality solutions to optimization and search problems by relying on bio-inspired operators such as mutation, crossover and selection. The followings are commonly used genetic algorithms: AntiSMASH (Antibiotics and Secondary Metabolite Analysis Shell) addresses secondary metabolite genome pipelines. PRISM (Prediction Informatics for Seco
https://en.wikipedia.org/wiki/Q-Bus
The Q-bus, also known as the LSI-11 Bus, is one of several bus technologies used with PDP and MicroVAX computer systems previously manufactured by the Digital Equipment Corporation of Maynard, Massachusetts. The Q-bus is a less expensive version of Unibus using multiplexing so that address and data signals share the same wires. This allows both a physically smaller and less-expensive implementation of essentially the same functionality. Over time, the physical address range of the Q-bus was expanded from 16 to 18 and then 22 bits. Block transfer modes were also added to the Q-bus. Main features of the Q-bus Like the Unibus before it, the Q-bus uses: Memory-mapped I/O Byte addressing A strict master-slave relationship between devices on the bus Asynchronous signaling Memory-mapped I/O means that data cycles between any two devices, whether CPU, memory, or I/O devices, use the same protocols. On the Unibus, a range of physical addresses are dedicated for I/O devices. The Q-bus simplifies this design by providing a specific signal (originally called BBS7, Bus Bank Select 7 but later generalized to be called BBSIO, Bus Bank Select I/O) that selects the range of addresses used by the I/O devices. Byte addressing means that the physical address passed on the Unibus is interpreted as the address of a byte-sized quantity of data. Because the bus actually contains a data path that is two bytes wide, address bit [0] is subject to special interpretation and data on the bus has to travel in the correct byte lanes. A strict Master-Slave relationship means that at any point in time, only one device can be the Master of the Q-bus. This master device can initiate data transactions which can then be responded to by a maximum of one selected slave device. (This had no effect on whether a given bus cycle is reading or writing data; the bus master can command either type of transaction.) At the end of the bus cycle, a bus arbitration protocol then selects the next device to
https://en.wikipedia.org/wiki/Vitamin%20A
Vitamin A is a fat-soluble vitamin and an essential nutrient for animals. The term "vitamin A" encompasses a group of chemically related organic compounds that includes retinol, retinal (also known as retinaldehyde), retinoic acid, and several provitamin (precursor) carotenoids, most notably beta-carotene. Vitamin A has multiple functions: it is essential for embryo development and growth, for maintenance of the immune system, and for vision, where it combines with the protein opsin to form rhodopsin the light-absorbing molecule necessary for both low-light (scotopic vision) and color vision. Vitamin A occurs as two principal forms in foods: A) retinol, found in animal-sourced foods, either as retinol or bound to a fatty acid to become a retinyl ester, and B) the carotenoids alpha-carotene, β-carotene, gamma-carotene, and the xanthophyll beta-cryptoxanthin (all of which contain β-ionone rings) that function as provitamin A in herbivore and omnivore animals which possess the enzymes that cleave and convert provitamin carotenoids to retinal and then to retinol. Some carnivore species lack this enzyme. The other carotenoids have no vitamin activity. Dietary retinol is absorbed from the digestive tract via passive diffusion. Unlike retinol, β-carotene is taken up by enterocytes by the membrane transporter protein scavenger receptor B1 (SCARB1), which is upregulated in times of vitamin A deficiency. Storage of retinol is in lipid droplets in the liver. A high capacity for long-term storage of retinol means that well-nourished humans can go months on a vitamin A- and β-carotene-deficient diet, while maintaining blood levels in the normal range. Only when the liver stores are nearly depleted will signs and symptoms of deficiency show. Retinol is reversibly converted to retinal, then irreversibly to retinoic acid, which activates hundreds of genes. Vitamin A deficiency is common in developing countries, especially in Sub-Saharan Africa and Southeast Asia. Deficiency can
https://en.wikipedia.org/wiki/Lagena%20%28anatomy%29
The lagena (from Greek λάγηνος : lágēnos for flask) is a structure found in humans and in animals. In human anatomy In the ear, the extremities of the ductus cochlearis are closed; the upper is termed the lagena and is attached to the cupula at the upper part of the helicotrema; the lower is lodged in the recessus cochlearis of the vestibule. In fish and amphibians The lagena is part of the vestibular system in fish and amphibians. It contains the otoliths asterisci. In fish, the lagena is implicated in hearing and the registration of vertical linear acceleration, in amphibians is the latter only.
https://en.wikipedia.org/wiki/Zoombak
Zoombak Inc was a U.S.-based company which developed GPS tracking devices for people and items. It used satellite-enabled GPS and a location network server for tracking. In 2011, Securus, Inc. acquired Zoombak, LLC from TruePosition, Inc., a subsidiary of Liberty Media. In March 2015 it was announced that BrickHouse Security acquired Securus, including Zoombak and related brands.
https://en.wikipedia.org/wiki/Hyperendemic
In epidemiology, the term hyperendemic disease is used to refer to a disease which is constantly and persistently present in a population at a high rate of incidence and/or prevalence (occurrence) and which equally affects (i.e. which is equally endemic in) all age groups of that population. It is one of the various degrees of endemicity (i.e. degrees of transmission of an infectious disease). Definitions According to a more precise definition given by the Robert Koch Institute in Germany, hyperendemicity is not necessarily associated with a high incidence rate. A hyperendemic disease is one which is ubiquitously present with ongoing circulation in an endemic region with a high prevalence rate. As a result, a hyperendemic region shows a relatively low incidence rate but at the same time it poses a high risk of infection to people coming into the region. According to another definition discussing malaria, a hyperendemic region is defined to be one with a seasonally high degree of endemicity where immunity does not succeed to prevent the effects of a disease for all age groups. Examples In the discussion of the dengue fever, a hyperendemic state is characterized by the continuous circulation of multiple viral serotypes in an area where a large pool of susceptible hosts and a competent vector (with or without seasonal variation) are constantly present. In another example, World Health Organization defines malaria to be hyperendemic if the percentage of persons with an enlarged spleen (spleen rate) is constantly greater than 50% for all age groups. Difference with similar epidemiological concepts Difference with holoendemic An endemic disease is one with a continuous occurrence at an expected frequency over a certain period of time and in a certain geographical location. Two terms are used when the degree of transmission or infection of an endemic disease is high: hyperendemic and holoendemic. One of the differences between hyperendemic and holoendemic diseases is
https://en.wikipedia.org/wiki/Scapy
Scapy is a packet manipulation tool for computer networks, originally written in Python by Philippe Biondi. It can forge or decode packets, send them on the wire, capture them, and match requests and replies. It can also handle tasks like scanning, tracerouting, probing, unit tests, attacks, and network discovery. Scapy provides a Python interface into libpcap or native raw sockets, in a similar way to that in which Wireshark provides a view and capture GUI. It differs by supporting packet injection, custom packet formats and scripting. While it is a command-line only tool, it can still interface with a number of other programs to provide visualisation including Wireshark, GnuPlot for providing graphs, graphviz or VPython for interactive displaying, etc. Starting with 2.4.0, Scapy supports Python 2.7 and 3.4+. It must not be confused with "scapy3k" (now renamed Kamene), which is an outdated independent and unmaintained fork.
https://en.wikipedia.org/wiki/EPIC%20204376071
EPIC 204376071 is an M-type star in the constellation of Scorpius. Parallax measurements by the Gaia space observatory put the star at a distance of about from Earth. It is likely a member of the Upper Sco association, and is young enough that it has not yet become a main-sequence star. Unusual light fluctuations of the star, including up to an 80% dimming in brightness (i.e., "single 80% deep occultation of 1-day duration"), were observed by astronomers. The unusual dimming was not only extremely deep, but also substantially asymmetric, with an egress about twice as long as the ingress. Nonetheless, such an unusual dimming for EPIC 204376071 is much greater than the 22% dimming observed for Tabby's star. Several explanations have been presented to explain the unusual dimming of the EPIC 204376071 star: one, orbiting dust or small particles; or two, a "transient accretion event of dusty material near the corotation radius of the star". The unusual lightcurve of the star is similar to the lightcurve of a candidate exoplanet, KIC 10403228 b, which may have been caused by a "tilted ring system" orbiting the planet. In the case of EPIC 204376071, an orbiting brown dwarf or large planet, with a ring system, could cause a similar lightcurve, according to the researchers. See also Disrupted planet List of stars that have unusual dimming periods
https://en.wikipedia.org/wiki/Huntington%27s%20Disease%20Association
The Huntington's Disease Association (HDA) is a charity that supports people in England and Wales affected by the genetic neurodegenerative brain condition Huntington's disease (HD). The HDA was founded in 1986 and is based in Liverpool. It supports a network of regional care advisors who offer care and support to people with and at risk of Huntington's disease and their families. The HDA also has a research programme that supports scientific and social research into HD. It has worked to establish and support a network of multidisciplinary clinics for HD patients in the UK. The charity was instrumental in the 2010 launch of an All-Party Parliamentary Group for Huntington's disease in the UK Parliament, chaired by Lord Walton of Detchant, and 2010 research revealing that the prevalence of HD is much higher than previously thought. The HDA won the 2005 UK Charity Award for healthcare and medical research and the 2005 National Health Service (NHS) Health and Social Care Award for long term conditions. The HDA's patrons are singer Tony Hadley and entertainer Shane Richie. The HDA is part of the UK HD Alliance, along with the Scottish Huntington's Association and HD Association Northern Ireland. It is also a member of the International Huntington Alliance.
https://en.wikipedia.org/wiki/Candied%20fruit
Candied fruit, also known as glacé fruit, is whole fruit, smaller pieces of fruit, or pieces of peel, placed in heated sugar syrup, which absorbs the moisture from within the fruit and eventually preserves it. Depending on the size and type of fruit, this process of preservation can take from several days to several months. This process allows the fruit to remain edible for up to a year. It has existed since the 14th century. The continual process of drenching the fruit in syrup causes the fruit to become saturated with sugar, preventing the growth of spoilage microorganisms due to the resulting unfavourable osmotic pressure. Fruits which are commonly candied include cherries, pineapple, greengages, pears, peaches and melon, as well as ginger root. The principal candied peels are orange and citron; these, together with candied lemon peel, are the usual ingredients of mixed chopped peel. Candied vegetables are also made, from vegetables such as pumpkin, turnip and carrot. Recipes vary from region to region, but the general principle is to boil the fruit, steep it in increasingly stronger sugar solutions for a number of weeks, and then dry off any remaining water. Uses As well as being eaten as snacks, candied fruits such as cherries and candied peels are commonly used in fruitcakes or pancakes. See also
https://en.wikipedia.org/wiki/BAR%20domain
In molecular biology, BAR domains are highly conserved protein dimerisation domains that occur in many proteins involved in membrane dynamics in a cell. The BAR domain is banana-shaped and binds to membrane via its concave face. It is capable of sensing membrane curvature by binding preferentially to curved membranes. BAR domains are named after three proteins that they are found in: Bin, Amphiphysin and Rvs. BAR domains occur in combinations with other domains Many BAR family proteins contain alternative lipid specificity domains that help target these protein to particular membrane compartments. Some also have SH3 domains that bind to dynamin and thus proteins like amphiphysin and endophilin are implicated in the orchestration of vesicle scission. N-BAR domain Some BAR domain containing proteins have an N-terminal amphipathic helix preceding the BAR domain. This helix inserts (like in the epsin ENTH domain) into the membrane and induces curvature, which is stabilised by the BAR dimer. Amphiphysin, endophilin, BRAP1/bin2 and nadrin are examples of such proteins containing an N-BAR. The Drosophila amphiphysin N-BAR (DA-N-BAR) is an example of a protein with a preference for negatively charged surfaces. Human proteins containing this domain AMPH; ARHGAP17; ARHGAP44; BIN1; BIN2; BIN3; SH3BP1; SH3GL1; SH3GL2; SH3GL3; SH3GLB1; SH3GLB2. F-BAR (EFC) domain F-BAR domains (for FCH-BAR, or EFC for Extended FCH Homology) are BAR domains that are extensions of the already established FCH domain. They are frequently found at the amino terminus of proteins. They can bind lipid membranes and can tubulate lipids in vitro and in vivo, but their exact physiological role still is under investigation. Examples of the F-BAR domain family are CIP4/FBP17/Toca-1, Syndapins (also called PACSINs) and muniscins. Gene knock-out of syndapin I in mice revealed that this brain-enriched isoform of the syndapin family is crucial for proper size control of synaptic vesicles and thereby inde
https://en.wikipedia.org/wiki/Coulure
Coulure (pronounced coo-LYUR) is a viticultural hazard that is the result of metabolic reactions to weather conditions that causes a failure of grapes to develop after flowering. In English the word shatter is sometimes used. Coulure is triggered by periods of cold, cloudy, rainy weather or very high out-of-season temperatures. The condition is most often manifested in the spring. It also occurs in vines that have little sugar content in their tissue. Flowers stay closed and are not fertilized. Thus the vines are not pollinated as the grape fails to develop and falls off. Coulure can also cause irregular bunches of grapes which are less compact than normal. These bunches are more sensitive to developing various grape diseases. The yield of a vine with coulure will decrease substantially. Grape varieties with high proclivity to coulure are Grenache, Malbec, Merlot, and Muscat Ottonel. Other causes of coulure may be vineyard conditions and practices, pruning too early or too severely, excessively fertile soils or overuse of fertilizers, and improper selection of rootstocks or clones. During the flowering part of the growing season (May–June in the Northern Hemisphere, November–December in the Southern Hemisphere), grapevines often need dry conditions with sufficient sunlight and ambient air temperature around for pollination to go smoothly. Less ideal conditions, particularly wet, rainy weather, increases the odds that a higher than normal numbers of flowers go unpollinated and coulure to occur. Coulure is a distinct phenomena unrelated to another viticultural hazard, millerandage, where the flowers are pollinated but the resulting berries develop without seeds and remain small. Like coulure, millerandage is often caused by inclement weather during the flowering and fruit set period and cause reduced yields. Cause and effect Coulure is caused by a carbohydrate deficiency in the plant tissues that causes the vine to conserve resources that would otherwise be funne
https://en.wikipedia.org/wiki/Intracellular%20digestion
Every organism requires energy to be active. However, to obtain energy from its outside environment, cells must not only retrieve molecules from their surroundings but also break them down. This process is known as intracellular digestion. In its broadest sense, intracellular digestion is the breakdown of substances within the cytoplasm of a cell. In detail, a phagocyte's duty is obtaining food particles and digesting it in a vacuole. For example, following phagocytosis, the ingested particle (or phagosome) fuses with a lysosome containing hydrolytic enzymes to form a phagolysosome; the pathogens or food particles within the phagosome are then digested by the lysosome's enzymes. Intracellular digestion can also refer to the process in which animals that lack a digestive tract bring food items into the cell for the purposes of digestion for nutritional needs. This kind of intracellular digestion occurs in many unicellular protozoans, in Pycnogonida, in some molluscs, Cnidaria and Porifera. There is another type of digestion, called extracellular digestion. In amphioxus, digestion is both extracellular and intracellular. Function Intracellular digestion is divided into heterophagic digestion and autophagic digestion. These two types take place in the lysosome and they both have very specific functions. Heterophagic intracellular digestion has an important job which is to break down all molecules that are brought into a cell by endocytosis. The degraded molecules need to be delivered to the cytoplasm; however, this will not be possible if the molecules are not hydrolyzed in the lysosome. Autophagic intracellular digestion is processed in the cell, which means it digests the internal molecules. Autophagy Generally, autophagy includes three small branches, which are macroautophagy, microautophagy, and chaperone-mediated autophagy. Occurrence Most organisms that use intracellular digestion belong to Kingdom Protista, such as amoeba and paramecium. Amoeba Amoeba u
https://en.wikipedia.org/wiki/Intake%20momentum%20drag
Intake momentum drag is an aerodynamic phenomenon which effects turbo prop and jet-powered aircraft. Causes Intake momentum drag is caused by the consequence of the speed of the air entering the engine increasing, but where the exit speed of the air from the engine remains constant. The outcome therefore is that the amount by which the engine increases air velocity, ostensibly by way of the compression process, is reduced. A repercussion of this causes a slight reduction in the thrust of a jet engine. Intake momentum drag yaw Intake momentum drag yaw is a further consequence of intake momentum drag which effects V/STOL (vertical and/or short take-off and landing) aircraft such as the Hawker Siddeley Harrier. Intake momentum drag yaw is an aspect in which the mass of air ingested by the intake of the engine, whilst the aircraft is in the hover during a crosswind, can result in a state of uncontrolled roll (a secondary aerodynamic effect of yaw). The phenomenon was identified during the test flying programme for the Harrier and which required precise investigation. This resulted in test pilot John Farley deliberately flying right into the edge of this condition repeatedly, so that a system to counteract the effect could be developed.
https://en.wikipedia.org/wiki/HB-F9P
The HB-F9P was a Sony MSX2-computer, launched in 1985. The abbreviation HB stands for Hit Bit. The HB-F9P was unusual in the sense that it did not come with a built in floppy disk drive, instead it had a built in software suite "Memovision" that would run automatically unless a program cartridge was inserted in the cartridge slot or the key was pressed during booting. Memovision could store data on the special HBI-55 (battery backed RAM) "data cartridge" that Sony originally had developed, for their HB-55 and HB-75 MSX computers. These systems contained a simple built in program called the "Personal Databank". Memovision was a continuation of the same idea, but much extended, as it contained stuff like a (birthday) calendar with alarm system, a "family databank" (a combination of a text editor and database) a built in calculator and a timer and time calculator, all rendered in pseudo 3D style. Variants Of the HB-F9 there were many different localized variants produced: HB-F9P, with QWERTY-keyboard layout was designed for the PAL-standard, and meant for European countries, except Germany, France, Spain and the former Soviet Union. HB-F9R, had a Russian keyboard. HB-F9D, had a QWERTZ-keyboard for the German market HB-F9F, had an AZERTY-keyboard for the French market, and was designed for the SECAM-standard. HB-F9S, had a modified QWERTY-keyboard for the Spanish market. Technical specifications Processor Sharp LH-0080 or NEC μPD780C memory ROM: 96 KB MSX BASIC version 2.0: 48 KB software suite: 48 KB RAM VRAM: 128 KB main memory: 128 KB Video VDP Yamaha V9938 text: 80×24, 40×24 en 32×24 (characters per line × lines) four colors, two foreground colors and two background colors graphical: maximal 512×212 pixels (16 colors of 512) and 256×212 (256 colors) colors: 512 maximal Engine MSX-Engine: S-1985 real-time clock with trickle-charged battery backup sound Yamaha YM2149 PSG 3 sound channels and one noise channel 8 octaves Interfaces power cord RF-output CV
https://en.wikipedia.org/wiki/Amanita%20chrysoblema
Amanita chrysoblema, with the common name American fly agaric, white variant, is a basidiomycete fungus of the genus Amanita. Although named chrysoblema, it is traditionally thought to be an Amanita muscaria variant, a group of fungi commonly known as fly agarics. A. chrysoblema is an uncommon fungus, distinguishable by an off-white to silvery-white cap with white warts. The cap has cuts on the side, but is otherwise similar to the usual fly agaric form. The fungi is poisonous due to high levels of ibotenic acid and muscimol. Taxonomy This white fly agaric was first described by science in 1880 by Peck, who classified it as an Amanita muscaria variant as A. muscaria var. alba. In 1918, Kauffmann named it Amanita chrysoblema during a study from Michigan, but the var. alba (or var. albus) name has been used in parallel to this since then. It is possible that this Amanita is not a muscaria, or fly agaric, but a species in its own right. This issue is currently under scientific scrutiny. The muscaria classification, may stem from it being wrongly treated as a white variant of the Amanita muscaria var. flavivolvata.
https://en.wikipedia.org/wiki/Theodor%20Fritsch
Theodor Fritsch (born Emil Theodor Fritsche; 28 October 1852 – 8 September 1933) was a German publisher and journalist. His antisemitic writings did much to influence popular German opinion against Jews in the late 19th and early 20th centuries. His writings also appeared under the pen names Thomas Frey, Fritz Thor, and Ferdinand Roderich-Stoltheim. He is not to be confused with his son, also Theodor Fritsch (1895–1946), likewise a bookseller and member of the SA. Life Fritsch was born Emil Theodor Fritsche, the sixth of seven children to Johann Friedrich Fritsche, a farmer in the village of Wiesenena (present-day Wiedemar) in the Prussian province of Saxony, and his wife August Wilhelmine, née Ohme. Four of his siblings died in childhood. He attended vocational school (Realschule) in Delitzsch where he learned casting and machine building. He then undertook study at the Royal Trade Academy (Königliche Gewerbeakademie) in Berlin, graduating as a technician in 1875. In the same year Fritsche found employment in a Berlin machine shop. He gained independence in 1879 through the founding of a technical bureau associated with a publishing firm. In 1880 he founded the Deutscher Müllerbund (Miller's League) which issued the publication Der Deutsche Müller (The German Miller). In 1905 he founded the "Saxon Small Business Association." He devoted himself to this organization and to the interests of crafts and small businesses (Mittelstand), as well as to the spread of antisemitic propaganda. When he changed his name to Fritsch is unclear. Publishing Fritsch created an early discussion forum, "Antisemitic Correspondence" in 1885 for antisemites of various political persuasions. In 1887 he sent several editions to Friedrich Nietzsche but was brusquely dismissed. Nietzsche sent Fritsch a letter in which he thanked him to be permitted "to cast a glance at the muddle of principles that lie at the heart of this strange movement", but requested not to be sent again such writi
https://en.wikipedia.org/wiki/John%20Lott%20%28mathematician%29
John William Lott (born January 12, 1959) is a professor of Mathematics at the University of California, Berkeley. He is known for contributions to differential geometry. Academic history Lott received his B.S. from the Massachusetts Institute of Technology in 1978 and M.A. degrees in mathematics and physics from University of California, Berkeley. In 1983, he received a Ph.D. in mathematics under the supervision of Isadore Singer. After postdoctoral positions at Harvard University and the Institut des Hautes Études Scientifiques, he joined the faculty at the University of Michigan. In 2009, he moved to University of California, Berkeley. Among his awards and honors: Sloan Research Fellowship (1989-1991) Alexander von Humboldt Fellowship (1991-1992) U.S. National Academy of Sciences Award for Scientific Reviewing (with Bruce Kleiner) Mathematical contributions A 1985 article of Dominique Bakry and Michel Émery introduced a generalized Ricci curvature, in which one adds to the usual Ricci curvature the hessian of a function. In 2003, Lott showed that much of the standard comparison geometry results for the Ricci tensor extend to the Bakry-Émery setting. For instance, if is a closed and connected Riemannian manifold with positive Bakry-Émery Ricci tensor, then the fundamental group of must be finite; if instead the Bakry-Émery Ricci tensor is negative, then the isometry group of the Riemannian manifold must be finite. The comparison geometry of the Bakry-Émery Ricci tensor was taken further in an influential article of Guofang Wei and William Wylie. Additionally, Lott showed that if a Riemannian manifold with smooth density arises as a collapsed limit of Riemannian manifolds with a uniform upper bound on diameter and sectional curvature and a uniform lower bound on Ricci curvature, then the lower bound on Ricci curvature is preserved in the limit as a lower bound on Bakry-Émery's Ricci curvature. In this sense, the Bakry-Émery Ricci tensor is shown to be natu
https://en.wikipedia.org/wiki/Biosafety%20Level%204%20Zoonotic%20Laboratory%20Network
The Biosafety Level 4 Zoonotic Laboratory Network is an international consortium of Biosafety Level 4 research laboratories. Its members are National Centre for Foreign Animal Disease (Canada) National Bio and Agro-Defense Facility (United States) Pirbright Institute (United Kingdom) Friedrich Loeffler Institute (Germany) CSIRO Australian Animal Health Laboratory (Australia)
https://en.wikipedia.org/wiki/Swedish%E2%80%93Russian%20Arc-of-Meridian%20Expedition
The Swedish–Russian Arc-of-Meridian expedition was a scientific expedition to Svalbard that took place from 1899 to 1902. The main purpose of the mission was to measure a meridian arc, in order to determine the earth flattening at the poles. The measurements were done by an establishing a chain of triangulation points from Keilhaufjellet in Sørkapp Land to Vesle Tavleøya north of Nordaustlandet. The Russians were responsible for the southern measurements, while the Swedes performed the northern measurements from a base at Crozierpynten on the eastern side of Sorgfjorden. A total arc of 4°10' was measured. The expedition also took the first photographs of the aurora.
https://en.wikipedia.org/wiki/Milk%20equivalent
Milk equivalent is a measure of the quantity of fluid milk used in a processed dairy product. Measured on a milkfat basis, it takes about 21.8 pounds of farm milk to make a pound of butter, and about 9.2 pounds to make a pound of American cheese. Measured on a skim solids basis, it takes about 11.6 pounds of farm milk to make a pound of nonfat dry milk. Farm milk weighs about 8.6 pounds per gallon.
https://en.wikipedia.org/wiki/SynOptics
SynOptics Communications was a Santa Clara, California-based early computer network equipment vendor from 1985 until 1994. SynOptics popularized the concept of the modular Ethernet hub and high-speed Ethernet networking over copper twisted-pair and fiber optic cables. History SynOptics Communications was founded in 1985 by Andrew K. Ludwick and Ronald V. Schmidt, both of whom worked at Xerox's Palo Alto Research Center (PARC). The most significant product that Synoptics produced was LattisNet (originally named AstraNet) in 1987. This meant that unshielded twisted-pair cabling already installed in office buildings could be re-utilized for computer networking instead of special coaxial cables. The star network topology made the network much easier to manage and maintain. Together these two innovations directly led to the ubiquity of Ethernet networks. Before the final standard version of what is known today as the 10BASE-T protocol, there were several different methods and standards for running Ethernet over twisted-pair cabling at various speeds, such as StarLAN. LattisNet was similar to the final 10BASE-T protocol except that it had slightly different voltage and signal characteristics. Synoptics updated their product line to the 10BASE-T specification once it was published. Through the late 1980s and into the early 1990s, SynOptics produced a series of innovative products including early 10BASE-2 hubs, pre-standard (LattisNet), and 100BASE-TX products. The company was the market leader in Ethernet LAN hubs over rivals 3Com and Cabletron. Despite intense competition that drove down prices, Synoptics' annual revenue grew to a high of $700 million in 1993. To move away from the rapidly commoditizing Layer 1/2 Ethernet equipment market and grow their market share in the increasingly lucrative and more profitable Layer 3 networking arena, SynOptics merged with Billerica, Massachusetts based Wellfleet Communications on July 6, 1994, in a US$ 2.7 Billion dollar de
https://en.wikipedia.org/wiki/Metastasis%20suppressor
A metastasis suppressor is a protein that acts to slow or prevent metastases (secondary tumors) from spreading in the body of an organism with cancer. Metastasis is one of the most lethal cancer processes. This process is responsible for about ninety percent of human cancer deaths. Proteins that act to slow or prevent metastases are different from those that act to suppress tumor growth. Genes for about a dozen such proteins are known in humans and other animals. Background The treatment of cancer usually aims to destroy and/or stop the growth of the primary tumor. Major improvements in the methods of surgery, radiation and chemotherapy have taken place, but corresponding improvements in patient survival have not always followed. Treatments that focus on the primary cancer typically do not address metastasis. Metastasis suppressors act by different mechanisms than tumor suppressors and do not affect primary tumors. Tumor suppressors, however, also inhibit metastasis, since metastasis is dependent upon tumorigenicity. Metastasis suppressors were first identified using microcell-mediated chromosome transfer (MMCT), which introduces chromosomes into intact recipient cells. Chromosomes 1, 6, 7, 8, 10, 11, 12, 16 and 17 harbor metastasis suppressor genes. MicroRNAs (miRNAs) are a class of gene regulators that bind the 3′ untranslated regions of target messenger RNAs, leading to either suppression of their translation or acceleration of their degradation. In cell MDA-MB-231 and its metastatic variant, six miRNAs displayed lower expression in metastatic cells. Among them, miR-335 and miR-126 suppress metastasis without affecting primary tumor growth. miR-335 targets multiple pathways, including SOX4, MERTK, PTPRN2 and TNC, which contribute to metastasis-suppression. miR-335 expression is correlated with metastasis-free survival in clinical breast cancer. Clinical applications Metastasis suppressors can potentially serve as prognostic markers, therapeutic targets an
https://en.wikipedia.org/wiki/Annual%20Review%20of%20Ecology%2C%20Evolution%2C%20and%20Systematics
{{Infobox journal | title = Annual Review of Ecology, Evolution, and Systematics | cover = Annual Review of Ecology, Evolution, and Systematics cover.png | discipline = Ecology, evolutionary biology, and systematics | formernames = Annual Review of Ecology and Systematics (1970–2002)' | editor = Douglas J. Futuyma | abbreviation = Annu. Rev. Ecol. Evol. Syst. | publisher = Annual Reviews | country = US | frequency = Annually | history = 1970–present, years old | impact = 11.8 | impact-year = 2022 | website = http://www.annualreviews.org/journal/ecolsys | ISSN = 1545-2069 | eISSN = 1545-2069 | LCCN = 2003213237 | OCLC = 51651460 | CODEN = ARECBC }} The Annual Review of Ecology, Evolution, and Systematics is an annual scientific journal published by Annual Reviews. The journal was established in 1970 as the Annual Review of Ecology and Systematics and changed its name beginning in 2003. It publishes invited review articles on topics considered to be timely and important in the fields of ecology, evolutionary biology, and systematics. As of 2023, Journal Citation Reports gave the journal a 2022 impact factor of 11.8, ranking it third of 169 journals in the "Ecology" category and third of 52 journals in "Evolutionary Biology". History The Annual Review of Ecology and Systematics was first published in 1970, with Richard F. Johnston as its first editor. In 1975 it began publishing biographies of notable ecologists in the prefatory chapter. In 2003, its name was changed to its current form, the Annual Review of Ecology, Evolution, and Systematics. As of 2023, Journal Citation Reports gave the journal a 2022 impact factor of 11.8, ranking it third of 169 journals in the "Ecology" category and third of 52 journals in "Evolutionary Biology". It defines its scope as covering significant developments in the field of ecology, evolution, and systematics of all life on earth. This includes reviews about molecular evolution, phylogeny, speciation, population dynamics, conse
https://en.wikipedia.org/wiki/Vitamin%20D%20and%20Omega-3%20Trial
Vitamin D and Omega-3 Trial (VITAL) was a clinical trial designed to investigate the use of daily dietary supplements of vitamin D and fish oil. The sponsor of the study was Brigham and Women's Hospital, collaborating with The National Cancer Institute, National Heart, Lung, and Blood Institute, Office of Dietary Supplements, National Institute of Neurological Disorders and Stroke, National Center for Complementary and Integrative Health, Pharmavite LLC, Pronova BioPharma and BASF. The studied aimed to enroll 20,000 participants (women 55 or over, men 50 or over) who were randomized into one of four groups: daily vitamin D (2000 IU) and fish oil (1 g); daily vitamin D and fish-oil placebo; daily vitamin-D placebo and fish oil; daily vitamin-D placebo and fish-oil placebo. Participants answered annual questionnaires to determine effects the risks of developing cancer, heart disease, stroke, osteoporosis, diabetes, memory loss and depression. The outcome of this study was: "The results of this trial indicate that supplementation with either n–3 fatty acid at a dose of 1 g/day or vitamin D3 at a dose of 2000 IU/day was not effective for primary prevention of CV or cancer events among healthy middle-aged men and women over 5 years of follow-up. There was also no difference in progression/development of CKD among patients with type 2 diabetes. This is one of the largest trials on this topic. The finding of a lower MI risk with n–3 fatty acid is hypothesis generating and deserves further study. The authors also noted some interaction with baseline fish consumption, with greater CV benefit observed among participants who had low fish intake at baseline."
https://en.wikipedia.org/wiki/Drawdown%20chart
Drawdown charts are rectangular pieces of non-fluorescent paper which are used to test a variety of coating properties. These properties include opacity, spreading rate, penetration, and flow & leveling behavior. This non-fluorescent material has to be especially rugged in order to maintain its structure and give reliable readings, as the coatings tested are often corrosive or abrasive. These charts are necessary in the testing of any coating as they give reliable and accurate readings for any type of coating before the coating is applied to the intended material. Drawdown bars Drawdown bars are used in collaboration with drawdown charts. The bars are generally made of stainless steel or aluminum and touch the chart at each end, while the center is slightly raised to a desired height in order to obtain the desired film thickness. As with the charts, the bars have to be especially resistant to corrosion because of the corrosive nature of some coatings. There are a multiplicity of more complicated designs including some in which the paint is already in the chart apparatus, but the bar method described above is the most commonly used. Procedure Initially the desired coating is applied in a puddle on the drawdown chart. Then the bar is placed at the top of the chart, and drawn down over the paint in order to spread it evenly along the length of the chart. Once this evenly spread paint dries, it can be tested for a variety of properties.
https://en.wikipedia.org/wiki/Power%20diagram
In computational geometry, a power diagram, also called a Laguerre–Voronoi diagram, Dirichlet cell complex, radical Voronoi tesselation or a sectional Dirichlet tesselation, is a partition of the Euclidean plane into polygonal cells defined from a set of circles. The cell for a given circle C consists of all the points for which the power distance to C is smaller than the power distance to the other circles. The power diagram is a form of generalized Voronoi diagram, and coincides with the Voronoi diagram of the circle centers in the case that all the circles have equal radii. Definition If C is a circle and P is a point outside C, then the power of P with respect to C is the square of the length of a line segment from P to a point T of tangency with C. Equivalently, if P has distance d from the center of the circle, and the circle has radius r, then (by the Pythagorean theorem) the power is d2 − r2. The same formula d2 − r2 may be extended to all points in the plane, regardless of whether they are inside or outside of C: points on C have zero power, and points inside C have negative power. The power diagram of a set of n circles Ci is a partition of the plane into n regions Ri (called cells), such that a point P belongs to Ri whenever circle Ci is the circle minimizing the power of P. In the case n = 2, the power diagram consists of two halfplanes, separated by a line called the radical axis or chordale of the two circles. Along the radical axis, both circles have equal power. More generally, in any power diagram, each cell Ri is a convex polygon, the intersection of the halfspaces bounded by the radical axes of circle Ci with each other circle. Triples of cells meet at vertices of the diagram, which are the radical centers of the three circles whose cells meet at the vertex. Related constructions The power diagram may be seen as a weighted form of the Voronoi diagram of a set of point sites, a partition of the plane into cells within which one of the sites is
https://en.wikipedia.org/wiki/Aestivation%20%28botany%29
Aestivation or estivation is the positional arrangement of the parts of a flower within a flower bud before it has opened. Aestivation is also sometimes referred to as praefoliation or prefoliation, but these terms may also mean vernation: the arrangement of leaves within a vegetative bud. Aestivation can be an important taxonomic diagnostic; for example Malvaceae flower buds have valvate sepals, with the exception of the genera Fremontodendron and Chiranthodendron, which have sometimes been misplaced as a result. Terminology The terms used to describe aestivation are the same as those used to describe leaf vernation. Classes of aestivation include: crumpled decussate imbricate – overlapping contorted or twisted – every petal or sepal is outside its neighbour on one margin, and inside its neighbour on the other margin. cochleate – spirally twisted. contortiplicate – contorted and also plicate. quincuncial – with five parts, where two petals or sepals are outside all others, two are inside all others, and the fifth is outside on one margin and inside on the other. induplicate – folded inwards. open – petals or sepals do not overlap or even touch each other . reduplicate – folded outwards. valvate – margins of adjacent petals or sepals touch each other without overlapping. vexillary – a special type of aestivation occurring in plants like pea; in this type of aestivation a large petal called standard encloses two smaller petals.
https://en.wikipedia.org/wiki/Minor%20allele%20frequency
Minor allele frequency (MAF) is the frequency at which the second most common allele occurs in a given population. They play a surprising role in heritability since MAF variants which occur only once, known as "singletons", drive an enormous amount of selection. Single nucleotide polymorphisms (SNPs) with a minor allele frequency of 0.05 (5%) or greater were targeted by the HapMap project. MAF is widely used in population genetics studies because it provides information to differentiate between common and rare variants in the population. As an example, a 2015 study sequenced the whole genomes of Sardinian individuals. The authors classified the variants found in the study in three classes according to their MAF. It was observed that rare variants (MAF < 0.05) appeared more frequently in coding regions than common variants (MAF > 0.05) in this population. Interpreting MAF data 1. Introduce the reference of a SNP of interest, as an example: rs429358, in a database (dbSNP or other). 2. Find MAF/MinorAlleleCount link. MAF/MinorAlleleCount: C=0.1506/754 (1000 Genomes, where number of genomes sampled = N = 2504); where C is the minor allele for that particular locus; 0.1506 is the frequency of the C allele (MAF), i.e. 15% within the 1000 Genomes database; and 754 is the number of times this SNP has been observed in the population of the study. See also Allele frequency
https://en.wikipedia.org/wiki/FASEB%20Excellence%20in%20Science%20Award
The Excellence in Science Award was established by the Federation of American Societies for Experimental Biology (FASEB) in 1989 to recognize outstanding achievement by women in biological science. All women who are members of one or more of the societies of FASEB are eligible for nomination. Nominations recognize a woman whose career achievements have contributed significantly to further our understanding of a particular discipline by excellence in research. The award includes a $10,000 unrestricted research grant, funded by Eli Lilly and Company. Award recipients Source: FASEB 1989 Marian Koshland 1990 Elizabeth Hay 1991 Ellen Vitetta 1992 Bettie Sue Masters 1993 Susan Leeman 1994 Lucille Shapiro 1995 Philippa Marrack 1996 Zena Werb 1997 Claude Klee 1998 Eva Neer 1999 Helen Blau 2000 Peng Loh 2001 Laurie Glimcher 2002 Phyllis Wise 2003 Joan A. Steitz 2004 Janet Rossant 2005 Anita Roberts 2006 Marilyn Farquhar and Elaine Fuchs 2007 Frances Arnold 2008 Mina J. Bissell 2009 Susan L. Lindquist 2010 Susan S. Taylor 2011 Gail R. Martin 2012 Susan R. Wessler 2013 Terry Orr-Weaver 2014 Kathryn V. Anderson 2015 Diane Griffin 2016 Bonnie Bassler 2017 Diane Mathis 2018 Lynne E. Maquat 2019 Barbara B. Kahn 2020 : Lifetime Achievement : Brigid Hogan Mid-Career Investigator : Aviv Regev Early-Career Investigator : Karen Schindler 2021: Lifetime Achievement : M. Celeste Simon Mid-Career Investigator : Valentina Greco Early-Career Investigator : Cigall Kadoch 2022: Lifetime Achievement : Arlene H. Sharpe Mid-Career Investigator : Sallie R. Permar Early-Career Investigator : Smita Krishnaswamy 2023: Lifetime Achievement : Elaine S. Jaffe Mid-Career Investigator : Paola Arlotta Early-Career Investigator : Diana Libuda See also List of biology awards
https://en.wikipedia.org/wiki/Jackson%20integral
In q-analog theory, the Jackson integral series in the theory of special functions that expresses the operation inverse to q-differentiation. The Jackson integral was introduced by Frank Hilton Jackson. For methods of numerical evaluation, see and . Definition Let f(x) be a function of a real variable x. For a a real variable, the Jackson integral of f is defined by the following series expansion: Consistent with this is the definition for More generally, if g(x) is another function and Dqg denotes its q-derivative, we can formally write or giving a q-analogue of the Riemann–Stieltjes integral. Jackson integral as q-antiderivative Just as the ordinary antiderivative of a continuous function can be represented by its Riemann integral, it is possible to show that the Jackson integral gives a unique q-antiderivative within a certain class of functions (see ). Theorem Suppose that If is bounded on the interval for some then the Jackson integral converges to a function on which is a q-antiderivative of Moreover, is continuous at with and is a unique antiderivative of in this class of functions. Notes
https://en.wikipedia.org/wiki/Society%20for%20Indecency%20to%20Naked%20Animals
The Society for Indecency to Naked Animals, or SINA [pronounced "sinna"], was a satirical hoax concocted by serial prankster Alan Abel. The group used the language and rhetoric of conservative moralists for the ostensible aim of clothing "indecent" naked animals, including domestic pets, barnyard animals, and large wildlife. The society and its aims were then largely presented in the public arena on news and talk shows by comedian/improviser Buck Henry (in the guise of SINA president "G. Clifford Prout") from 1959 through 1963. History In 1959, Alan Abel wrote a satirical story about an imaginary organization for The Saturday Evening Post but the editors rejected it. Abel then transformed his story into a series of press releases from the organization that garnered media attention. Looking to expand the range and impact of the hoax, Abel persuaded the actor Buck Henry to assume the role of the group president, G. Clifford Prout, Jr., in U.S. television news and talk show appearances. Henry, who had improv training, was able to play Prout with an intense deadpan sincerity, as well as to stay in character through unscripted interviews. During the interviews, Prout was often presented as an amusing eccentric, but was otherwise taken seriously by the broadcasters who interviewed him. Abel (seen far less frequently) played the group's executive vice-president Bruce Spencer. Over time, the history of SINA and some of its specific aims were codified into a coherent (if unlikely) backstory. The group had been founded some years before by G. Clifford Prout, Sr., and was being carried on by his son. An alleged debate within SINA was how large an animal had to be to require clothing; the official position quoted by Prout was "any dog, cat, horse or cow that stands higher than 4 inches or longer than 6 inches." Slogans such as "Decency today means morality tomorrow" and "A nude horse is a rude horse" were offered. The group received widespread American media cove
https://en.wikipedia.org/wiki/Geobacter%20chapellei
Geobacter chapellei is a Gram-negative, strictly anaerobic, mesophilic and non-motile bacterium from the genus of Geobacter which has been isolated fromaq uifer sediments from the Atlantic Coastal Plain in the United States. See also List of bacterial orders List of bacteria genera
https://en.wikipedia.org/wiki/44th%20meridian%20west
The meridian 44° west of Greenwich is a line of longitude that extends from the North Pole across the Arctic Ocean, Greenland, the Atlantic Ocean, South America, the Southern Ocean, and Antarctica to the South Pole. The 44th meridian west forms a great circle with the 136th meridian east. From Pole to Pole Starting at the North Pole and heading south to the South Pole, the 44th meridian west passes through: {| class="wikitable plainrowheaders" ! scope="col" width="120" | Co-ordinates ! scope="col" | Country, territory or sea ! scope="col" | Notes |- | style="background:#b0e0e6;" | ! scope="row" style="background:#b0e0e6;" | Arctic Ocean | style="background:#b0e0e6;" | |- | style="background:#b0e0e6;" | ! scope="row" style="background:#b0e0e6;" | Lincoln Sea | style="background:#b0e0e6;" | |- | ! scope="row" | |Nansen Land |- | style="background:#b0e0e6;" | ! scope="row" style="background:#b0e0e6;" | J.P. Koch Fjord | style="background:#b0e0e6;" | |- | ! scope="row" | |Freuchen Land |- | style="background:#b0e0e6;" | ! scope="row" style="background:#b0e0e6;" | Nordenskiöld Fjord | style="background:#b0e0e6;" | |- | ! scope="row" | | Mainland |- | style="background:#b0e0e6;" | ! scope="row" style="background:#b0e0e6;" | Prince Christian Sound | style="background:#b0e0e6;" | |- | ! scope="row" | | Sammisoq and Egger Island |- | style="background:#b0e0e6;" | ! scope="row" style="background:#b0e0e6;" | Atlantic Ocean | style="background:#b0e0e6;" | |-valign="top" | ! scope="row" | | Maranhão Piauí — Bahia — Minas Gerais — , passing just west of Belo Horizonte (at ) Rio de Janeiro — |- | style="background:#b0e0e6;" | ! scope="row" style="background:#b0e0e6;" | Sepetiba Bay | style="background:#b0e0e6;" | |- | ! scope="row" | | Rio de Janeiro — Restinga da Marambaia (sandbank) |- | style="background:#b0e0e6;" | ! scope="row" style="background:#b0e0e6;" | Atlantic Ocean | style="background:#b0e0e6;" | |- | style="background:#b0e0e6;" | ! scope="r
https://en.wikipedia.org/wiki/Gravitational%20shielding
The term gravitational shielding refers to a hypothetical process of shielding an object from the influence of a gravitational field. Such processes, if they existed, would have the effect of reducing the weight of an object. The shape of the shielded region would be similar to a shadow from the gravitational shield. For example, the shape of the shielded region above a disk would be conical. The height of the cone's apex above the disk would vary directly with the height of the shielding disk above the Earth. Experimental evidence to date indicates that no such effect exists. Gravitational shielding is considered to be a violation of the equivalence principle and therefore inconsistent with both Newtonian theory and general relativity. The concept of gravity shielding is a common concept in science fiction literature, especially for space travel. One of the first and best known examples is the fictional gravity shielding substance "Cavorite" that appears in H. G. Wells' classic 1901 novel The First Men in the Moon. Wells was promptly criticized for using it by Jules Verne. Tests of the equivalence principle , no experiment was successful in detecting positive shielding results. To quantify the amount of shielding, at the beginning of 20th century Quirino Majorana suggested an extinction coefficient h that modifies Newton's gravitational force law as follows: The best laboratory measurements have established an upper bound limit for shielding of 4.3×10−15 m²/kg. The best estimate based on the most accurate gravity anomaly data during the 1997 solar eclipse has provided a new constraint on the shielding parameter 6×10−19 m²/kg. However, astronomical observations impose much more stringent limits. Based on lunar observations available in 1908, Poincaré established that h can be no greater than 10−18 m²/kg. Subsequently, this bound has been greatly improved. Eckhardt showed that lunar ranging data implies an upper bound of 10−22 m²/kg, and Williams, et al., ha
https://en.wikipedia.org/wiki/Diagram%20%28mathematical%20logic%29
In model theory, a branch of mathematical logic, the diagram of a structure is a simple but powerful concept for proving useful properties of a theory, for example the amalgamation property and the joint embedding property, among others. Definition Let be a first-order language and be a theory over For a model of one expands to a new language by adding a new constant symbol for each element in where is a subset of the domain of Now one may expand to the model The positive diagram of , sometimes denoted , is the set of all those atomic sentences which hold in while the negative diagram, denoted thereof is the set of all those atomic sentences which do not hold in . The diagram of is the set of all atomic sentences and negations of atomic sentences of that hold in Symbolically, . See also Elementary diagram
https://en.wikipedia.org/wiki/Sialidase-4
Sialidase-4 is an enzyme that in humans is encoded by the NEU4 gene. Function This gene belongs to a family of glycohydrolytic enzymes which remove sialic acid residues from glycoproteins and glycolipids. Interactions Sialidase-4 has been shown to interact with phospholipid scramblase 1.
https://en.wikipedia.org/wiki/Proizvolov%27s%20identity
In mathematics, Proizvolov's identity is an identity concerning sums of differences of positive integers. The identity was posed by Vyacheslav Proizvolov as a problem in the 1985 All-Union Soviet Student Olympiads. To state the identity, take the first 2N positive integers, 1, 2, 3, ..., 2N − 1, 2N, and partition them into two subsets of N numbers each. Arrange one subset in increasing order: Arrange the other subset in decreasing order: Then the sum is always equal to N2. Example Take for example N = 3. The set of numbers is then {1, 2, 3, 4, 5, 6}. Select three numbers of this set, say 2, 3 and 5. Then the sequences A and B are: A1 = 2, A2 = 3, and A3 = 5; B1 = 6, B2 = 4, and B3 = 1. The sum is which indeed equals 32. Proof A slick proof of the identity is as follows. Note that for any , we have that :. For this reason, it suffices to establish that the sets and : coincide. Since the numbers are all distinct, it therefore suffices to show that for any , . Assume the contrary that this is false for some , and consider positive integers . Clearly, these numbers are all distinct (due to the construction), but they are at most : a contradiction is reached. Notes
https://en.wikipedia.org/wiki/Cell%20culture%20assay
A cell culture assay is any method used to assess the cytotoxicity of a material. This refers to the in vitro assessment of a material to determine whether it releases toxic chemicals in the cell. It also determines if the quantity is sufficient to kill cells, either directly or indirectly, through the inhibition of cell metabolic pathways. Cell culture evaluations are the precursor to whole animal studies and are a way to determine if significant cytotoxicity exists for the given material. Cell culture assays are standardized by ASTM, ISO, and BSI (British Standards Institution.) See also Microphysiometry
https://en.wikipedia.org/wiki/COVID-19%20naming
During the early days of the COVID-19 pandemic, the disease and virus were sometimes called "coronavirus", "Wuhan coronavirus", or "Wuhan pneumonia". In January 2020, the World Health Organization (WHO) tentatively named it "2019-nCoV", short for "2019 Novel Coronavirus", or "2019 Novel Coronavirus Acute Respiratory Disease". This naming was based on the organization's 2015 guidelines for naming novel viruses and diseases, avoiding the use of geographic locations (such as Wuhan), in part to prevent social stigma. A similar structure has also been used by the AP when referring to virus variants, for example, referring to it as the "Delta variant" rather than the "South African variant". On 11 February 2020, the WHO named the disease COVID-19 (short for coronavirus disease 2019). That same day, the International Committee on Taxonomy of Viruses (ICTV) formally announced it had named the causative virus as SARS-CoV-2 (severe acute respiratory syndrome coronavirus 2) based upon its genetic similarity to the 2003 SARS-CoV. WHO Director General Tedros Adhanom Ghebreyesus explained that CO stands for coronavirus, VI for virus, and D stands for disease, while 19 stands for the year that the outbreak was first detected. The separation between the disease and the causative virus is based on the same nomenclature policies that separate AIDS and the virus which causes it, HIV. Chinese virus From January to March 2020, US President Donald Trump repeatedly described the virus as the "Chinese virus". In March 2020, the president claimed to have abandoned the term, telling Fox News "we shouldn't make any more of a big deal out of it". On March 18 and 19, 2020, Trump twice defended using the term "Chinese virus" amid instances of bigotry against Asians in the United States. Trump referred to it as "the China Virus" at least as late as January 2021. This description was also used by members of the Spanish far-right political party Vox, especially by its leader Santiago Abascal i
https://en.wikipedia.org/wiki/English%20Qaballa
English Qaballa (EQ) is a system of Hermetic Qabalah, supported by a system of arithmancy that interprets the letters of the English alphabet via an assigned set of values, discovered by James Lees in 1976. It is the result of an intent to understand, interpret, and elaborate on the mysteries of Aleister Crowley's received text, Liber AL vel Legis, the Book of the Law. According to Jake Stratton-Kent, "the English Qaballa is a qabalah and not a system of numerology. A qabalah is specifically related to three factors: one, a language; two, a 'holy' text or texts; three, mathematical laws at work in these two." This system has also been referred to as the ALW cipher, New Aeon English Qabalah or NAEQ by other writers. Background The practice of using alphabetic letters to represent numbers developed in the Greek city of Miletus, and is thus known as the Milesian system. Early examples include vase graffiti dating to the 6th century BCE. Aristotle wrote that the Pythgoraean tradition, founded in the 6th century by Pythagoras of Samos, practiced isopsephy, the Greek predecessor of Hebrew gematria. Pythagoras was a contemporary of the philosophers Anaximander, Anaximenes, and the historian Hecataeus, all of whom lived in Miletus, across the sea from Samos. The Milesian system was in common use by the reign of Alexander the Great (336–323 BCE) and was adopted by other cultures during the subsequent Hellenistic period. It was officially adopted in Egypt during the reign of Ptolemy II Philadelphus (284–246 BCE). The first system of English gematria was used by the poet John Skelton in 1523 in his poem "The Garland of Laurel". In 1532, Heinrich Cornelius Agrippa defined an analogue of the Greek system of isopsephy in his work De Occulta Philosopha. Agrippa based his system on the order of the Classical Latin alphabet, appending the four additional letters in use at the time after Z, including the still-in-use letters J (600), U (700), and W (900). Agrippa was the mentor o
https://en.wikipedia.org/wiki/HTML5%20Boilerplate
HTML5 Boilerplate is an HTML, CSS and JavaScript template (or boilerplate) for creating HTML5 websites with cross-browser compatibility.
https://en.wikipedia.org/wiki/Steiner%20conic
The Steiner conic or more precisely Steiner's generation of a conic, named after the Swiss mathematician Jakob Steiner, is an alternative method to define a non-degenerate projective conic section in a projective plane over a field. The usual definition of a conic uses a quadratic form (see Quadric (projective geometry)). Another alternative definition of a conic uses a hyperbolic polarity. It is due to K. G. C. von Staudt and sometimes called a von Staudt conic. The disadvantage of von Staudt's definition is that it only works when the underlying field has odd characteristic (i.e., ). Definition of a Steiner conic Given two pencils of lines at two points (all lines containing and resp.) and a projective but not perspective mapping of onto . Then the intersection points of corresponding lines form a non-degenerate projective conic section (figure 1) A perspective mapping of a pencil onto a pencil is a bijection (1-1 correspondence) such that corresponding lines intersect on a fixed line , which is called the axis of the perspectivity (figure 2). A projective mapping is a finite product of perspective mappings. Simple example: If one shifts in the first diagram point and its pencil of lines onto and rotates the shifted pencil around by a fixed angle then the shift (translation) and the rotation generate a projective mapping of the pencil at point onto the pencil at . From the inscribed angle theorem one gets: The intersection points of corresponding lines form a circle. Examples of commonly used fields are the real numbers , the rational numbers or the complex numbers . The construction also works over finite fields, providing examples in finite projective planes. Remark: The fundamental theorem for projective planes states, that a projective mapping in a projective plane over a field (pappian plane) is uniquely determined by prescribing the images of three lines. That means that, for the Steiner generation of a conic section, besides two p
https://en.wikipedia.org/wiki/IntervalZero
IntervalZero, Inc. develops hard real-time software and its symmetric multiprocessing (SMP) enabled RTX and RTX64 software transform the Microsoft Windows general-purpose operating system (GPOS) into a real-time operating system (RTOS). IntervalZero and its engineering group regularly release new software (cf its history). Its most recent product, RTX64, focuses on 64-bit and symmetric multiprocessing (SMP) to replace dedicated hardware based systems such as digital signal processors (DSPs) or field-programmable gate arrays (FPGAs) with multicore PCs. For instance, an audio mixing surface manufacturer which largely deployed DSP based systems, switched to personal computer (PC) based systems, dedicating multi-core processors for the real time audio processing. Founded in July 2008 by a group of former Ardence executives, IntervalZero is headed by CEO Jeffrey D. Hibbard. The firm has offices in Waltham, MA; Nice, France; Munich, Germany, and Taiwan, ROC. This global presence is important because these solutions are deployed worldwide, primarily in industrial automation, military, aerospace, medical devices, digital media, and test and simulation software. The corporate name, IntervalZero, comes from the technical definition of the optimal experience between a system command and execution. History IntervalZero's lineage traces back to 1980, when a group of Massachusetts Institute of Technology engineers started VenturCom and began to develop expertise in embedded technology. It was during this time that Venix was developed and marketed. Their first innovation was to focus on Windows NT 4.0 as a possible real-time solution for the Industry in 1995 by releasing RTX. Since then, a lot of controllers are PC and Windows based. Their second innovation came as a second product, Component Integrator, which makes Windows NT 4.0 an embedded OS. It was licensed by Microsoft a few years later and became the origin of Windows NT Embedded. In 2004, VenturCom, was renamed A
https://en.wikipedia.org/wiki/Ethane-1%2C1-dithiol
Ethane-1,1-dithiol is an organosulfur compound with formula CH3CH(SH)2. It is a colourless smelly liquid that is added to or found in some foods. The compound is an example of a geminal dithiol. Use Flavouring uses of ethane-1,1-dithiol may include drinks, oil, gravy, soup, meat, fruit, seasonings, and snacks. Maximum concentrations in use that are generally recognised as safe (GRAS) is five parts per million (ppm) but typical uses are around 0.2 ppm. It is supplied as a 1% solution in ethanol, due to its strong offensive smell. In the diluted form with 4% ethyl acetate and ethanol the CAS number is 69382-62-3. Toxicity may be due to metabolism products hydrogen sulfide and acetaldehyde, however as used it has a margin of safety of over 10,000,000. Other ways that it is modified in the body apart from hydrolysis is methylation to 1-methylsulfanyl-ethanethiol, oxidation of the sulfur to an ethyl sulfonate, glucuronidation of the sulfur, or combination with cysteine by way of a disulfide bridge. Identifiers Since it is used in the food industry there are codes issued by various authorities. It is identified as JECFA number 1660. The Flavor and Extract Manufacturers Association (FEMA) id is 4111. The European designation for the flavouring is Fl 12.293. Natural occurrence It can be produced during fermentation of grapes. It is used as a food flavouring. It is found naturally in the scent of durian. Properties The nuclear magnetic resonance spectrum shows three environments for protons, in the ratio of 3:2:1 corresponding to CH3 SH and CH. Reactions Ethane-1,1-dithiol has several reactions known that are important in white wine flavours. In the presence of oxygen, ethane-1,1-dithiol is converted to cis/trans-3,6-dimethyl-1,2,4,5-tetrathiane which has a rubbery aroma. This molecule has a ring with four sulfur atoms and two carbons, two ethane-1,1-dithiol molecules become linked at their sulfur atoms with the loss of hydrogen. This can further oxidise to cis/trans
https://en.wikipedia.org/wiki/Eagle%20effect
The Eagle effect, Eagle phenomenon, or paradoxical zone phenomenon, named after Harry Eagle who first described it, originally referred to the paradoxically reduced antibacterial effect of penicillin at high doses, though recent usage generally refers to the relative lack of efficacy of beta lactam antibacterial drugs on infections having large numbers of bacteria. The former effect is paradoxical because the effectiveness of an antibiotic generally rises with increasing drug concentration. Mechanism Proposed mechanisms: Reduced expression of penicillin binding proteins during stationary growth phase Induction of microbial resistance mechanisms (such as beta lactamases with short half-lives) by high drug concentrations Precipitation of antimicrobial drug in vitro, possibly also leading to the crystallized drug being mis-detected as colonies of the microbe. Self-antagonising the receptor with which it binds (penicillin binding proteins, for example, in the case of a penicillin). Penicillin is a bactericidal antibiotic that works by inhibiting cell wall synthesis but this synthesis only occurs when bacteria are actively replicating (or in the log phase of growth). In cases of extremely high bacterial burden (such as with Group A Strep), bacteria may be in the stationary phase of growth. In this instance since no bacteria are actively replicating (presumably due to nutrient restriction) penicillin has no activity. This is why adding an antibiotic like clindamycin, which acts ribosomally, kills some of the bacteria and returns them to the log phase of growth.
https://en.wikipedia.org/wiki/Mahler%20measure
In mathematics, the Mahler measure of a polynomial with complex coefficients is defined as where factorizes over the complex numbers as The Mahler measure can be viewed as a kind of height function. Using Jensen's formula, it can be proved that this measure is also equal to the geometric mean of for on the unit circle (i.e., ): By extension, the Mahler measure of an algebraic number is defined as the Mahler measure of the minimal polynomial of over . In particular, if is a Pisot number or a Salem number, then its Mahler measure is simply . The Mahler measure is named after the German-born Australian mathematician Kurt Mahler. Properties The Mahler measure is multiplicative: where is the norm of . Kronecker's Theorem: If is an irreducible monic integer polynomial with , then either or is a cyclotomic polynomial. (Lehmer's conjecture) There is a constant such that if is an irreducible integer polynomial, then either or . The Mahler measure of a monic integer polynomial is a Perron number. Higher-dimensional Mahler measure The Mahler measure of a multi-variable polynomial is defined similarly by the formula It inherits the above three properties of the Mahler measure for a one-variable polynomial. The multi-variable Mahler measure has been shown, in some cases, to be related to special values of zeta-functions and -functions. For example, in 1981, Smyth proved the formulas where is the Dirichlet L-function, and where is the Riemann zeta function. Here is called the logarithmic Mahler measure. Some results by Lawton and Boyd From the definition, the Mahler measure is viewed as the integrated values of polynomials over the torus (also see Lehmer's conjecture). If vanishes on the torus , then the convergence of the integral defining is not obvious, but it is known that does converge and is equal to a limit of one-variable Mahler measures, which had been conjectured by Boyd. This is formulated as follows: Let denote the int
https://en.wikipedia.org/wiki/Battery%20terminal
Battery terminals are the electrical contacts used to connect a load or charger to a single cell or multiple-cell battery. These terminals have a wide variety of designs, sizes, and features that are often not well documented. Automotive battery terminals Automotive batteries typically have one of three types of terminals. In recent years, the most common design was the SAE Post, consisting of two lead posts in the shape of truncated cones, positioned on the top of the battery, with slightly different diameters to ensure correct electrical polarity. The "JIS" type is similar to the SAE but smaller, once again positive is larger than negative but both are smaller than their SAE counterparts. Most older Japanese cars were fitted with JIS terminals. General Motors, and other automobile manufacturers, have also begun using side-post battery terminals, which consist of two recessed female 3/8" threads (SAE 3/8-16) into which bolts or various battery terminal adapters are to be attached. These side posts are of the same size and do not prevent incorrect polarity connections. L terminals consist of an L-shaped post with a bolt hole through the vertical side. These are used on some European cars, motorcycles, lawn and garden devices, snowmobiles, and other light-duty vehicles. Some batteries sizes are available with terminals in many different configurations, but two main configurations are: positive on left and negative on the right corner negative on the left and positive on the right corner. Terminals can also be both on the long or short side of the battery, or diagonally opposed, or in the middle. Purchasing the wrong configuration may prevent battery cables from reaching the battery terminals. Marine battery terminals Marine batteries typically have two posts, a 3/8"-16 threaded post for the positive terminal, and a 5/16"-18 threaded post for the negative terminal. Zinc battery terminals Zinc battery terminals are an environmentally friendly alternative to
https://en.wikipedia.org/wiki/Justniffer
Justniffer is a TCP packet sniffer. It can log network traffic in a 'standard' (web server-like) or in a customized way. It can also log response times, useful for tracking network services performances (e.g. web server, application server, etc.). The output format of the traffic can be easily customized. An example written in Python (delivered with the official package) stores the transferred contents in an output directory separated by domains. This means that the transferred files like html, css, javascript, images, sounds, etc. can be saved to a directory. Overview justniffer was born to help in troubleshooting performance in network TCP-based services: HTTP, RTSP, SIP, SMTP, IMAP, POP, LDAP, Telnet etc. It can collect low and high level protocol and performance information, reconstructing the TCP flow in a reliable way using portions of the Linux kernel code. It uses a slightly modified version of the libnids libraries that already include a modified version of Linux kernel code in a more reusable way. It can be extended with external scripts (bash, Python, or any executable) and generate logs in a customizable way The man page for justniffer explains all the options. External links Official website Justniffer - tcp packet sniffer Lars Michelsen Examples Examples Grab http traffic and observe Grab Http traffic See also Comparison of packet analyzers tcpdump, a packet analyzer pcap, an application programming interface (API) for capturing network traffic snoop, a command line packet analyzer included with Solaris wireshark, a network packet analyzer dsniff, a packet sniffer and set of traffic analysis tools netsniff-ng, a free Linux networking toolkit ngrep, a tool that can match regular expressions within the network packet payloads etherape, a network mapping tool that relies on sniffing traffic tcptrace, a tool for analyzing the logs produced by tcpdump Microsoft Network Monitor, a packet analyzer Network analyzers Free network management so
https://en.wikipedia.org/wiki/Foundational%20Questions%20Institute
The Foundational Questions Institute, styled FQxI (formerly FQXi), is an organization that provides grants to "catalyze, support, and disseminate research on questions at the foundations of physics and cosmology." It was founded in 2005 by cosmologists Max Tegmark and Anthony Aguirre,. It has run multiple worldwide grant competitions (in 2006, 2008, 2010, 2013, 2015, 2017 and 2019), the first of which provided US$2M to 30 projects. It also runs frequent essay contests open to the general public with $40,000 in prizes awarded by a jury panel and the best texts published in book format. FQxI is an independent, philanthropically funded non-profit organization, run by scientists for scientists. The $6.2 million seed funding was donated by the John Templeton Foundation, whose goal is to reconcile science and religion. Tegmark has stated that the money came with "no strings attached"; The Boston Globe stated FQxI is run by "two well-respected researchers who say they are not religious. The institute's scientific advisory board is also filled with top scientists." Critics of the John Templeton Foundation such as Sean Carroll have also stated they were satisfied that the FQxI is independent. Notable members FQXi members include Scott Aaronson Anthony Aguirre Yakir Aharonov John Carlos Baez Julian Barbour John D. Barrow Jacob Biamonte Raphael Bousso Sean Carroll David Chalmers Paul Davies David Deutsch George F. R. Ellis Nicolas Gisin Brian Greene Sabine Hossenfelder Robert Lawrence Kuhn Seth Lloyd Roger Penrose Lisa Randall Martin Rees Carlo Rovelli Lee Smolin Leonard Susskind Gerard 't Hooft Max Tegmark Vlatko Vedral Steven Weinberg Frank Wilczek Stephen Wolfram Anton Zeilinger Wojciech Zurek Among others
https://en.wikipedia.org/wiki/Electroneurogram
An electroneurogram is a method used to visualize directly recorded electrical activity of neurons in the central nervous system (brain, spinal cord) or the peripheral nervous system (nerves, ganglions). The acronym ENG is often used. An electroneurogram is similar to an electromyogram (EMG), but the latter is used to visualize muscular activity. An electroencephalogram (EEG) is a particular type of electroneurogram in which several electrodes are placed around the head and the general activity of the brain is recorded, without having very high resolution to distinguish between the activity of different groups of neurons. An electroneurogram is usually obtained by placing an electrode in the neural tissue. The electrical activity generated by neurons is recorded by the electrode and transmitted to an acquisition system, which usually allows to visualize the activity of the neuron. Each vertical line in an electroneurogram represents one neuronal action potential. Depending on the precision of the electrode used to record neural activity, an electroneurogram can contain the activity of a single neuron to thousands of neurons. Researchers adapt the precision of their electrode to either focus on the activity of a single neuron or the general activity of a group of neurons, both strategies having their advantages. External links An example of a neural recording Electrophysiology
https://en.wikipedia.org/wiki/Spawning%20networks
Spawning networks are a new class of programmable networks that automate the life cycle process for the creation, deployment, and management of network architecture. These networks are capable of spawning distinct child virtual networks with their own transport, control, and management systems. This definition is created because the deployment of new network architectures, services, and protocols is often manual, ad hoc, and time-consuming. The definition was introduced in a paper titled Spawning Networks, published in IEEE Networks by a group of researchers from Columbia University, University of Hamburg, Intel Corporation, Hitachi Limited, and Nortel Networks. The authors are Andrew T. Campbell, Michael E. Kounavis, Daniel A. Villela, of Columbia University, John B. Vicente, of Intel Corporation, Hermann G. De Meer, of University of Hamburg, Kazuho Miki, of Hitachi Limited, and Kalai S. Kalaichelvan, of Nortel Networks. There was also a paper titled "The Genesis Kernel: A Programming System for Spawning Network Architectures", Michael E. Kounavis, Andrew T. Campbell, Stephen Chou, Fabien Modoux, John Vicente and Hao Zhuang. A first implementation of Spawning Networks was realized at Columbia University as part of the Ph.D thesis work of Michael Kounavis. This implementation is based on the design of the Genesis Kernel, a programming system consisting of three layers: A transport environment which is a collection of programmable virtual routers, a programming environment which offers open access to the programmable data path and a life cycle environment which is responsible for spawning and managing network architectures. One of the concepts used in design of the Genesis Kernel is the creation of a network architecture based on a profiling script specifying the architecture components and their interaction.
https://en.wikipedia.org/wiki/Social%20behavior
Social behavior is behavior among two or more organisms within the same species, and encompasses any behavior in which one member affects the other. This is due to an interaction among those members. Social behavior can be seen as similar to an exchange of goods, with the expectation that when you give, you will receive the same. This behavior can be affected by both the qualities of the individual and the environmental (situational) factors. Therefore, social behavior arises as a result of an interaction between the two—the organism and its environment. This means that, in regards to humans, social behavior can be determined by both the individual characteristics of the person, and the situation they are in. A major aspect of social behavior is Personal information, which is the basis for survival and reproduction. Social behavior is said to be determined by two different processes, that can either work together or oppose one another. The dual-systems model of reflective and impulsive determinants of social behavior came out of the realization that behavior cannot just be determined by one single factor. Instead, behavior can arise by those consciously behaving (where there is an awareness and intent), or by pure impulse. These factors that determine behavior can work in different situations and moments, and can even oppose one another. While at times one can behave with a specific goal in mind, other times they can behave without rational control, and driven by impulse instead. There are also distinctions between different types of social behavior, such as mundane versus defensive social behavior. Mundane social behavior is a result of interactions in day-to-day life, and are behaviors learned as one is exposed to those different situations. On the other hand, defensive behavior arises out of impulse, when one is faced with conflicting desires. The development of social behavior Social behavior constantly changes as one continues to grow and develop, reaching
https://en.wikipedia.org/wiki/Shepherd%20moon
A shepherd moon, also called a herder moon or watcher moon, is a small natural satellite that clears a gap in planetary-ring material or keeps particles within a ring contained. The name is a result of their limiting the "herd" of the ring particles as a shepherd. Due to their gravitational effect, they pick up particles and deflect them from their original orbits through orbital resonance. This causes gaps in the ring system, such as the particularly striking Cassini Division, as well as other characteristic bands, or strange "twisted" deformation of rings. Discovery The existence of shepherd moons was theorized in early 1979. Observations of the rings of Uranus show that they are very thin and well defined, with sharp gaps between rings. To explain this, Goldreich and Tremaine suggested that two small satellites that were undetected at the time might be confining each ring. The first images of shepherd satellites were taken later that year by Voyager 1. Examples Jupiter Several of Jupiter's small innermost moons, namely Metis and Adrastea, are within Jupiter's ring system and are also within Jupiter's Roche limit. It is possible that these rings are composed of material that is being pulled off these two bodies by Jupiter's tidal forces, possibly facilitated by impacts of ring material on their surfaces. Saturn The complex ring system of Saturn has several such satellites. These include Prometheus (F ring), Daphnis (Keeler Gap), Pan (Encke Gap), Janus, and Epimetheus (both A ring). Uranus Uranus also has shepherd moons on its ε ring, Cordelia and Ophelia. They are interior and exterior shepherds, respectively. Both moons are well within Uranus's synchronous orbit radius, and their orbits are therefore slowly decaying due to tidal deceleration. Neptune Neptune's rings are very unusual in that they first appeared to be composed of incomplete arcs in Earth-based observations, but Voyager 2's images showed them to be complete rings with bright clumps. It is
https://en.wikipedia.org/wiki/Dragon%27s%20Dogma%20Online
Dragon's Dogma Online was a massively multiplayer online role-playing game (MMORPG) developed and published by Capcom for PlayStation 3, PlayStation 4 and Microsoft Windows in 2015; As of 2019, the servers have been shut down. A follow-up to the 2012 action role-playing game Dragon's Dogma, the game had the player take the role of an "Arisen" in service to the White Dragon, charged with defending the land of Lestania from monster attacks. Gameplay carried over the basic battle and Pawn companion systems from Dragon's Dogma while incorporating features of the MMORPG genre. The game was conceived in 2012 alongside Dark Arisen, an expanded version of Dragon's Dogma. Development was delayed into 2013 due to the development of Dark Arisen. Returning staff included director Kento Kinoshita, producer Minae Matsukawa and composer Tadayoshi Makino. The scenario was written by Kazushige Nojima, noted for his work on the Final Fantasy series. The game reached one million downloads within a month of its release, and was positively reviewed. Gameplay Dragon's Dogma Online was a massively multiplayer online role-playing game (MMORPG) where players take on the role of the "Arisen", a warrior of the White Dragon who defends the land of Lestania from monsters. The Arisen is a customizable avatar created by the player at the beginning of the game, with their gender, physique and other features such as their voice being customizable. The game's hub is the White Dragon Shrine, with the rest of Lestania being available for exploration following the opening quest. Players progressed through the game by completing quests, which were accepted from the game's non-playable characters and divided into three categories; main quests connected to the central narrative, World Quests which were multiplayer-exclusive quests where up to eight players would cooperate during the quest; and Solo Quests which one player can complete. Quests gave experience and monetary rewards, and were gated based o
https://en.wikipedia.org/wiki/Gas%20networks%20simulation
Gas networks simulation or gas pipeline simulation is a process of defining the mathematical model of gas transmission and gas distribution systems, which are usually composed of highly integrated pipe networks operating over a wide range of pressures. Simulation allows to predict the behaviour of gas network systems under different conditions. Such predictions can be effectively used to guide decisions regarding the design and operation of the real system. Simulation types Depending on the gas flow characteristics in the system there are two states that can be matter of simulation: Steady state – the simulation does not take into account the gas flow characteristics' variations over time and described by the system of algebraic equations, in general nonlinear ones. Unsteady state (transient flow analysis) – described either by a partial differential equation or a system of such equations. Gas flow characteristics are mainly functions of time. Network topology In the gas networks simulation and analysis, matrices turned out to be the natural way of expressing the problem. Any network can be described by set of matrices based on the network topology. The network consists of one source node (reference node) L1, four load nodes (2, 3, 4 and 5) and seven pipes or branches. For network analysis it is necessary to select at least one reference node. Mathematically, the reference node is referred to as the independent node and all nodal and branch quantities are dependent on it. The pressure at source node is usually known, and this node is often used as the reference node. However, any node in the network may have its pressure defined and can be used as the reference node. A network may contain several sources or other pressure-defined nodes and these form a set of reference nodes for the network.The load nodes are points in the network where load values are known. These loads may be positive, negative or zero. A negative load represents a demand for gas from the
https://en.wikipedia.org/wiki/Calix%20Limited
Calix Limited is an Australian technology company whose core technology is a "kiln" built in Bacchus Marsh that produces "mineral honeycomb". Calix's technology includes work on CO2 capture to address global sustainability challenges across several industries including wastewater treatment, aquaculture, advanced energy storage and cement / lime production. Products Calix's products include ACTI-Mag, BOOSTER-Mag and AQUA-Cal+. ACTI-Mag - Wastewater management AQUA-Cal+ - Sustainable aquaculture farming, lake & pond remediation BOOSTER-Mag - Crop protection History The origins of Calix began in 2005, with work on flash calcining. In 2007 their first pilot calciner facility was built to test what was possible with a fully established CFC 850 calciner plant at Bacchus Marsh in 2011. In 2012 Calix acquired mining tenements at Myrtle Springs, South Australia. Between 2013 and 2017 Calix established a plant in South East Asia and won several awards and funding for a variety of sustainable technologies including Advanced Battery research, and Agricultural applications. Calix listed publicly in the Australian Securities Exchange (ASX) in 2018 and in coordination with the LEILAC (Low Emissions Intensity Lime & Cement) project in Europe, began construction of a pilot plant in Lixhe, Belgium. In 2019 Calix acquired the US-based company Inland Empire Resources. Project LEILAC Calix is a key partner in, and provides the core technology towards, Project LEILAC (Low Emissions Intensity Lime & Cement), as part of Horizon 2020, a programme established by the European Commission. Financial information Calix Limited is a listed Australian public company, having listed on the Australian Securities Exchange (ASX) in 2018. Prior to its listing, Calix was backed by investors including Och-Ziff Capital Management Group and Washington H Soul Pattinson. Since its foundation in 2005, Calix Limited has to date committed more than $40 million to commercialising its technologies and proc
https://en.wikipedia.org/wiki/Gross%20anatomy
Gross anatomy is the study of anatomy at the visible or macroscopic level. The counterpart to gross anatomy is the field of histology, which studies microscopic anatomy. Gross anatomy of the human body or other animals seeks to understand the relationship between components of an organism in order to gain a greater appreciation of the roles of those components and their relationships in maintaining the functions of life. The study of gross anatomy can be performed on deceased organisms using dissection or on living organisms using medical imaging. Education in the gross anatomy of humans is included training for most health professionals. Techniques of study Gross anatomy is studied using both invasive and noninvasive methods with the goal of obtaining information about the macroscopic structure and organisation of organs and organ systems. Among the most common methods of study is dissection, in which the corpse of an animal or a human cadaver is surgically opened and its organs studied. Endoscopy, in which a video camera-equipped instrument is inserted through a small incision in the subject, may be used to explore the internal organs and other structures of living animals. The anatomy of the circulatory system in a living animal may be studied noninvasively via angiography, a technique in which blood vessels are visualised after being injected with an opaque dye. Other means of study include radiological techniques of imaging, such as X-ray and MRI. In medical and healthcare professional education Most health profession schools, such as medical, physician assistant, and dental schools, require that students complete a practical (dissection) course in gross human anatomy. Such courses aim to educate students in basic human anatomy and seek to establish anatomical landmarks that may later be used to aid medical diagnosis. Many schools provide students with cadavers for investigation by dissection, aided by dissection manuals, as well as cadaveric atlases (e.g. Ne
https://en.wikipedia.org/wiki/%C3%81lvaro%20R%C3%ADos%20Poveda
Álvaro Ríos Poveda (born 3 February 1974, Cali, Colombia) is a Colombian electronic engineer, university professor, and researcher who specializes in biomedical engineering and mechatronics. He has performed research on myoelectric prostheses, sensory feedback, and bionic vision technologies. Early life and education He began his studies at San Juan Berchmans school in Cali, Colombia. Ríos earned an undergraduate degree in Electronic Engineering at Pontifical Xavierian University and completed his master's at Simon Bolivar University and doctorate studies at USF in Biomedical Engineering. Along with that, he obtained an MBA from ISEAD. His professional career began with research in neural prostheses and bionics systems. Ríos researched biomedical engineering, artificial intelligence, and robotics in control systems. Career Ríos is a researcher and university professor in undergraduate and postgraduate studies in Europe and Mexico. From a young age, his interest had grown into motor limitations. In 1996, he developed prosthetic systems that allow patients to have greater limb functionality while ensuring accessibility for these systems in developing countries. In 1997, a sensory feedback system for prostheses was presented at France's World Congress on Biomedical Engineering. Ríos's public work includes myoelectric prosthesis with sensory feedback, presented at MEC'02: The Next Generation. His work mainly aims of his work is to control prosthetics more naturally, utilizing artificial intelligence, neural control, machine learning, and gesture control. Research He has remained a member of the Publications Committee of the International Federation of Medical and Biological Engineering and a founding member of the Colombian Association of Biomedical Engineering. Since its inception in 2017, Ríos has been participating at every CBS IEEE International Conference on Bionic Systems and Cyborg and giving guest participation at IEEE's CBS 2017. Later in 2018, he represe
https://en.wikipedia.org/wiki/Glycomimetic
Glycomimetic is a term used to refer to molecules that have structures similar to carbohydrates, but with some variation. This will normally result in modified biological properties. Introduction Often, modification of the structure will take place around the glycosidic linkage. Replacement of one or other of the glycosidic oxygen atoms by carbon, sulfur, nitrogen etc. will alter the properties of the glycosidic bond. The molecules produced in this way would be called carbasugars or C-glycosides, thiosugars or thioglycosides, or iminosugars or glycosylamines. When nitrogen is introduced, the glycomimetic may become positively charged at physiological pH, meaning that it may act as an enzyme inhibitor, either by Coulombic interaction with carboxylate amino acid side-chains in the enzyme active site, or by mimicking positive-charge build-up at the transition state of the reaction, or both. Iminosugars (sometimes referred to erroneously as azasugars) are classic examples of molecules with this behaviour. Glycosylamines typically have a lower stability, being easily hydrolysed, which means that to exploit an exocyclic nitrogen substituent at C-1, further modification is necessary. An example of this would be the additional substitution of the ring-oxygen for carbon as is seen in valienamine. Altering the structure of a carbohydrate will normally result in several changes to the properties of the molecule. As well as changing the stability of the glycosidic bond, the ring-conformation may be affected. Also the conformation of the glycosidic bond may be affected. As well as obvious changes in the immediate vicinity of the substitution, e.g. that replacement of an acetal oxygen by methylene (CH2) would result in loss of a hydrogen-bond participatory atom, such a substitution is expected to have more subtle effects resulting from a change in the dipole of the molecule, such as slight changes in hydrogen bonding or pKa values of the unchanged hydroxyl groups. Substitution
https://en.wikipedia.org/wiki/City%20network
City networks can either refer to a membership organization city leaders join to connect their city to other municipalities, or to a geographical concept used to describe inter-connectivity of cities on different levels (trade, railways, culture etc.). City networks in international cooperation In the perspective of international cooperation, the term "city network" refers to a membership organization that cities join either to take part to specific projects, to be represented by geographical specificity, or to assert political commitments. One of the main reasons why cities join a city network is to learn good practices from peer cities that have similar characteristics and face comparable challenges (infrastructure, urban transport, water and sanitation, smart city, etc.). City networks as a geographical concept City networks are a geographical concept studying connections between cities by placing the cities as nodes on a network. In modern conceptions of cities, these networks play an important role in understanding the nature of cities. City networks can identify physical connections to other places, such as railways, canals, scheduled flights, or telecommunication networks, typically done using graph theory. City networks also exist in immaterial form, such as trade, global finance, markets, migration, cultural links, shared social spaces or shared histories. There are also networks of religious nature, in particular through pilgrimage. The city itself is then regarded as the node where different networks run together. Some urban thinkers have argued that cities can only be understood if the context of the city's connections is understood. It has been argued that city networks are a key ingredient of what defines a city, along with the number of people (density) and the particular way of life in cities.
https://en.wikipedia.org/wiki/Quantum%20satis
Quantum satis (abbreviation q.s. or Q.S.) is a Latin term meaning the amount which is enough. It has its origins as a quantity specification in medicine and pharmacology, where a similar term quantum sufficit ("as much as is sufficient") has been used (abbreviated Q.S.). Quantum satis is also used in the same function in food regulations and food safety laws in the European Community (EC/EU). The specification of quantum satis for an ingredient essentially means "Add as much of this ingredient as is needed to achieve the desired result, but not more." In food safety regulations in the EU it is a catch-all restriction for artificial food ingredients (especially food additives) which are harmless enough to have no specific quantity restriction. It serves to protect consumers from the addition of excessive and unnecessary amounts of such artificial food additives in their foodstuffs and compels producers to: Introduce minimal additives to food for human consumption Observe Good Manufacturing Practice Refrain from wilful consumer deception For example, European Union directive 94/36/EC (which regulates the use of food colors) explains in Article 2 (7): "In the Annexes to this Directive 'quantum satis' means that no maximum level is specified. However, coloring matters shall be used according to good manufacturing practice at a level not higher than is necessary to achieve the intended purpose and provided that they do not mislead the customer". The words quantum satis are used with reference to a number of substances in the Annexes III and IV to the EU directive 94/36/EC.
https://en.wikipedia.org/wiki/Variational%20perturbation%20theory
In mathematics, variational perturbation theory (VPT) is a mathematical method to convert divergent power series in a small expansion parameter, say , into a convergent series in powers , where is a critical exponent (the so-called index of "approach to scaling" introduced by Franz Wegner). This is possible with the help of variational parameters, which are determined by optimization order by order in . The partial sums are converted to convergent partial sums by a method developed in 1992. Most perturbation expansions in quantum mechanics are divergent for any small coupling strength . They can be made convergent by VPT (for details see the first textbook cited below). The convergence is exponentially fast. After its success in quantum mechanics, VPT has been developed further to become an important mathematical tool in quantum field theory with its anomalous dimensions. Applications focus on the theory of critical phenomena. It has led to the most accurate predictions of critical exponents. More details can be read here.
https://en.wikipedia.org/wiki/Clearance%20rate
In criminal justice, clearance rate is calculated by dividing the number of crimes that are "cleared" (a charge being laid) by the total number of crimes recorded. Clearance rates are used by various groups as a measure of crimes solved by the police. Clearance rates can be problematic for measuring the performance of police services and for comparing various police services. This is because a police force may employ a different way of measuring clearance rates. For example, each police force may have a different method of recording when a "crime" has occurred and different criteria for determining when a crime has been "cleared." A given police force may appear to have a much better clearance rate because of its calculation methodology. Some U.S. police forces have been criticized for overuse of "exceptional clearance", which is intended to classify as "cleared" cases where probable cause to arrest a suspect exists, but police are unable to do so for reasons outside their control (such as death or incarceration in a foreign country). In system conflict theory, it is argued that clearance rates cause the police to focus on appearing to solve crimes (generating high clearance rate scores) rather than actually solving crimes. Further focus on clearance rates may result in effort being expended to attribute crimes (correctly or incorrectly) to a criminal, which may not result in retribution, compensation, rehabilitation or deterrence.
https://en.wikipedia.org/wiki/Papiine%20betaherpesvirus%203
Papiine betaherpesvirus 3 (PaHV-3) is a species of virus in the genus Cytomegalovirus, subfamily Betaherpesvirinae, family Herpesviridae, and order Herpesvirales.
https://en.wikipedia.org/wiki/Unicity%20distance
In cryptography, unicity distance is the length of an original ciphertext needed to break the cipher by reducing the number of possible spurious keys to zero in a brute force attack. That is, after trying every possible key, there should be just one decipherment that makes sense, i.e. expected amount of ciphertext needed to determine the key completely, assuming the underlying message has redundancy. Claude Shannon defined the unicity distance in his 1949 paper "Communication Theory of Secrecy Systems". Consider an attack on the ciphertext string "WNAIW" encrypted using a Vigenère cipher with a five letter key. Conceivably, this string could be deciphered into any other string—RIVER and WATER are both possibilities for certain keys. This is a general rule of cryptanalysis: with no additional information it is impossible to decode this message. Of course, even in this case, only a certain number of five letter keys will result in English words. Trying all possible keys we will not only get RIVER and WATER, but SXOOS and KHDOP as well. The number of "working" keys will likely be very much smaller than the set of all possible keys. The problem is knowing which of these "working" keys is the right one; the rest are spurious. Relation with key size and possible plaintexts In general, given particular assumptions about the size of the key and the number of possible messages, there is an average ciphertext length where there is only one key (on average) that will generate a readable message. In the example above we see only upper case English characters, so if we assume that the plaintext has this form, then there are 26 possible letters for each position in the string. Likewise if we assume five-character upper case keys, there are K=265 possible keys, of which the majority will not "work". A tremendous number of possible messages, N, can be generated using even this limited set of characters: N = 26L, where L is the length of the message. However, only a smaller set
https://en.wikipedia.org/wiki/Atomic%20energy
Atomic energy or energy of atoms is energy carried by atoms. The term originated in 1903 when Ernest Rutherford began to speak of the possibility of atomic energy. H. G. Wells popularized the phrase "splitting the atom", before discovery of the atomic nucleus. Atomic energy includes: Nuclear binding energy, the energy required to split a nucleus of an atom. Nuclear potential energy, the potential energy of the particles inside an atomic nucleus. Nuclear reaction, a process in which nuclei or nuclear particles interact, resulting in products different from the initial ones; see also nuclear fission and nuclear fusion. Radioactive decay, the set of various processes by which unstable atomic nuclei (nuclides) emit subatomic particles. The energy of inter-atomic or chemical bonds, which holds atoms together in compounds. Atomic energy is the source of nuclear power, which uses sustained nuclear fission to generate heat and electricity. It is also the source of the explosive force of an atomic bomb.
https://en.wikipedia.org/wiki/Plantarum%20seu%20stirpium%20historia
'Plantarum seu stirpium historia' is an illustrated botanical text by the Flemish physician Lobelius (Matthias de l'Obel (1538–1616) and published in Antwerp in 1576. Later, he translated it into Flemish in 1581 with the title Kruydtboeck. This was l'Obel's second work, following publication of the Stirpium adversaria nova in London in 1570. This work was intended as a companion publication to his Stirpium adversaria nova, and incorporates a revised version of the latter, as Nova stirpium adversaria. History L'Obel's first work was the Stirpium aduersaria noua, published in London in 1571. The Plantarum seu stirpium historia of 1576 was intended as a companion piece, and was published in conjunction with a re-issue of Stirpium aduersaria noua, as Nova stirpium adversaria In 1571 l'Obel published his Flemish translation of Plantarum, as Kruydtboeck (Herb book), and in 1605 he reissued the Stirpium. The Plantarum deals with several subjects, including descriptions and illustrations (more than 2,000) of plants known to l'Obel, with plant names given in Latin, German, English, French, Flemish, Italian, and Spanish. It also includes a history of botany and the use of plants in treating diseases. It was published by Christophe Plantin, whose workshop handled most of the major botanical works of the period and accumulated a large stock of woodcuts which were recycled between all these publications. The work is notable as one of the earliest attempts to classify plants by their natural characteristics rather than medicinal properties. Once Kruydtboeck was published, Plantin issued the engravings as a separate album, arranged according to L'Obel's classification. Structure The initial section is the new Stirpium observations (671 pages), which is liberally illustrated, followed by the revised Nova stirpium adversarial. The former contains 1,473 woodcuts. While many of these were reused from other botanical works, principally L'Écluse, about half were original to this
https://en.wikipedia.org/wiki/Atmospheric%20reanalysis
An atmospheric reanalysis (also: meteorological reanalysis and climate reanalysis) is a meteorological and climate data assimilation project which aims to assimilate historical atmospheric observational data spanning an extended period, using a single consistent assimilation (or "analysis") scheme throughout. Operational data analysis In operational numerical weather prediction, forecast models are used to predict future states of the atmosphere, based on how the climate system evolves with time from an initial state. The initial state provided as input to the forecast must consist of data values for a range of "prognostic" meteorological fields – that is, those fields which determine the future evolution of the model. Spatially varying fields are required in the form used by the model, for example at each intersection point on a regular grid of longitude and latitude circles, and initial data must be valid at a single time that corresponds to the present or the recent past. By contrast, the available observational data usually do not include all of the model's prognostic fields, and may include other additional fields; these data also have different spatial distribution from the forecast model grid, are valid over a range of times rather than a single time, and are also subject to observational error. The technique of data assimilation is therefore used to produce an analysis of the initial state, which is a best fit of the numerical model to the available data, taking into account the errors in the model and the data. Uses In addition to initializing operational forecasts, the analyses themselves are a valuable tool for subsequent meteorological and climatological studies. However, an operational analysis dataset, i.e. the analysis data which were used for the real-time forecasts, will typically suffer from inconsistency if it spans any extended period of time, because operational analysis systems are frequently being improved. A reanalysis project involves
https://en.wikipedia.org/wiki/Tree%20of%20life%20%28Kabbalah%29
The Tree of Life (Hebrew: עֵץ חַיִּים ʿĒṣ Ḥayyīm) is a diagram used in Kabbalah and various other mystical traditions. It is usually referred to as the Kabbalistic tree of life in order to distinguish it from the biblical tree of life and the archetypal tree of life found in many cultures. Scholars have asserted that the concept of a tree of life with different spheres encompassing aspects of reality traces its origins back to Assyria in the 9th century BC. The Assyrians assigned values and specific numbers to their deities similar to those used by the later Jewish Kabbalah. The beginnings of the Jewish Kabbalah are traced back by scholars to the Medieval Age, originating in the Bahir and the Zohar. Although the earliest extant Hebrew kabbalistic manuscripts dating to the late 13th century contain diagrams, including one labelled "Tree of Wisdom," the now iconic "Tree of life" emerged over the course of the fourteenth century. Scholars now believe that it should be regarded as primarily indebted to the Porphyrian tree rather than to any speculative ancient sources, Assyrian or otherwise. The iconic representation first appeared in print on the cover of the Latin translation of Gates of Light in the year 1516. Scholars have traced the origin of the art in the Porta Lucis cover to Johann Reuchlin. Description The tree of life usually consists of 10 or 11 nodes symbolizing different archetypes and 22 paths connecting the nodes. The nodes are often arranged into three columns to represent that they belong to a common category. In the Jewish Kabbalah, the nodes are called sephiroth. They are usually represented as spheres and the paths are usually represented as lines. The nodes usually represent encompassing aspects of existence, God, or the human psyche. The paths usually represent the relationship between the concepts ascribed to the spheres or a symbolic description of the requirements to go from one sphere to another. The columns are usually symbolized as pil
https://en.wikipedia.org/wiki/Chenopodium%20nuttalliae
Chenopodium nuttalliae is a species of edible plant native to Mexico. It is known by the common names huauzontle (literally "hairy amaranth", from the Nahuatl huauhtli 'amaranth' and tzontli 'hair') and Aztec broccoli. Other variations of the name include huauhzontle, huazontle, huanzontle, and guausoncle. It is related to other commonly-consumed plants such as quinoa, amaranth, and epazote, as well as the common American weeds goosefoot and lambsquarters. The plant grows upright branches with red tinted green leafy stems. Huauzontle stems superficially resemble baby broccoli, although the stems are much thinner, and support fewer of the leaves. Like amaranth, huauzontle is very resistant to cold and dry climates, and grows even in poor soil. In addition, it has a high nutritional level, which makes it an excellent grain alternative in regions with difficulties sowing other types of cereals. During the rule of Moctezuma, huauzontle and amaranth were the fourth most important crops, after maize, beans, and chia. Many towns paid tribute to the Aztec empire in huauzontle. Culinary use As with other members of the goosefoot family, the leaves, branches, flowers (inflorescence), and seeds of huauzontle are all edible. The plant is used both as a herb and as a vegetable in Mexican cuisine. While it is eaten throughout Mexico, it is most commonly consumed in the center of the country, especially in the states of Tlaxcala, Mexico, Guerrero, Morelos, and in the south of Mexico City. The most popular dish is huauzontle pancakes stuffed with cheese and topped with tomato sauce. Alternatively, huauzontles can be encased in an egg batter and deep-fried with a stick of salty Mexican cheese. Huauzontle is used to season salads, ahuautles in pasilla sauce, and beef fillets. The mature seeds can also be ground into flour to make tortillas. Like quinoa, another plant in the genus Chenopodium, huauzontle contains saponins, albeit in lesser quantities. While saponins are toxic to
https://en.wikipedia.org/wiki/Experimental%20testing%20of%20time%20dilation
Time dilation as predicted by special relativity is often verified by means of particle lifetime experiments. According to special relativity, the rate of a clock C traveling between two synchronized laboratory clocks A and B, as seen by a laboratory observer, is slowed relative to the laboratory clock rates. Since any periodic process can be considered a clock, the lifetimes of unstable particles such as muons must also be affected, so that moving muons should have a longer lifetime than resting ones. A variety of experiments confirming this effect have been performed both in the atmosphere and in particle accelerators. Another type of time dilation experiments is the group of Ives–Stilwell experiments measuring the relativistic Doppler effect. Atmospheric tests Theory The emergence of the muons is caused by the collision of cosmic rays with the upper atmosphere, after which the muons reach Earth. The probability that muons can reach the Earth depends on their half-life, which itself is modified by the relativistic corrections of two quantities: a) the mean lifetime of muons and b) the length between the upper and lower atmosphere (at Earth's surface). This allows for a direct application of length contraction upon the atmosphere at rest in inertial frame S, and time dilation upon the muons at rest in S′. Time dilation and length contraction Length of the atmosphere: The contraction formula is given by , where L0 is the proper length of the atmosphere and L its contracted length. As the atmosphere is at rest in S, we have γ=1 and its proper Length L0 is measured. As it is in motion in S′, we have γ>1 and its contracted length L′ is measured. Decay time of muons: The time dilation formula is , where T0 is the proper time of a clock comoving with the muon, corresponding with the mean decay time of the muon in its proper frame. As the muon is at rest in S′, we have γ=1 and its proper time T′0 is measured. As it is moving in S, we have γ>1, therefore its proper ti
https://en.wikipedia.org/wiki/Biaxial%20nematic
A biaxial nematic is a spatially homogeneous liquid crystal with three distinct optical axes. This is to be contrasted to a simple nematic, which has a single preferred axis, around which the system is rotationally symmetric. The symmetry group of a biaxial nematic is i.e. that of a rectangular right parallelepiped, having 3 orthogonal axes and three orthogonal mirror planes. In a frame co-aligned with optical axes the second rank order parameter tensor of a biaxial nematic has the form where is the standard nematic scalar order parameter and is a measure of the biaxiality. The first report of a thermotropic biaxial nematic appeared in 2004 based on a boomerang shaped oxadiazole bent-core mesogen. The biaxial nematic phase for this particular compound only occurs at temperatures around 200 °C and is preceded by as yet unidentified smectic phases. It is also found that this material can segregate into chiral domains of opposite handedness. For this to happen the boomerang shaped molecules adopt a helical superstructure. In one azo bent-core mesogen a thermal transition is found from a uniaxial Nu to a biaxial nematic Nb mesophase, as predicted by theory and simulation. This transition is observed on heating from the Nu phase with Polarizing optical microscopy as a change in Schlieren texture and increased light transmittance and from x-ray diffraction as the splitting of the nematic reflection. The transition is a second order transition with low energy content and therefore not observed in differential scanning calorimetry. The positional order parameter for the uniaxial nematic phase is 0.75 to 1.5 times the mesogen length and for the biaxial nematic phase 2 to 3.3 times the mesogen length. Another strategy towards biaxial nematics is the use of mixtures of classical rodlike mesogens and disklike discotic mesogens. The biaxial nematic phase is expected to be located below the minimum in the rod-disk phase diagram. In one study a miscible system of ro
https://en.wikipedia.org/wiki/GMS%20%28software%29
GMS (Groundwater Modeling System) is water modeling application for building and simulating groundwater models from Aquaveo. It features 2D and 3D geostatistics, stratigraphic modeling and a unique conceptual model approach. Currently supported models include MODFLOW, MODPATH, MT3DMS, RT3D, FEMWATER, SEEP2D, and UTEXAS. Version 6 introduced the use of XMDF (eXtensible Model Data Format), which is a compatible extension of HDF5. The purpose of this is to allow internal storage and management of data in a single HDF file, rather than using many flat files. History GMS was initially developed in the late 1980s and early 1990s on Unix workstations by the Engineering Computer Graphics Laboratory at Brigham Young University. The development of GMS was funded primarily by The United States Army Corps of Engineers and was known—until version 4.0, released in late 1999—as the Department of Defense Groundwater Modeling System, or DoD GMS. It was ported to Microsoft Windows in the mid 1990s. Version 3.1 was the last version that supported HP-UX, IRIX, OSF/1, and Solaris platforms. Development of GMS—along with WMS and SMS—was transferred to Aquaveo when it formed in April 2007. A study published in the Journal of Agricultural and Applied Economics in August 2000 stated that "GMS provides an interface to the groundwater flow model, MODFLOW, and the contaminant transport model, MT3D. MODFLOW is a three-dimensional, cell-centered, finite-difference, saturated-flow model capable of both steady-state and transient analyses...These two models, when put together, provide a comprehensive tool for examining groundwater flow and nitrate transport and accumulation". The study was designed to help develop a "permit scheme to effectively manage nitrate pollution of groundwater supplies for communities in rural areas without hindering agricultural production in watersheds". Version history Reception A 2001 report prepared for the Iowa Comprehensive Petroleum Underground Storage Tank Fu
https://en.wikipedia.org/wiki/Mononuclear%20phagocyte%20system
In immunology, the mononuclear phagocyte system or mononuclear phagocytic system (MPS) also known as the reticuloendothelial system or macrophage system is a part of the immune system that consists of the phagocytic cells located in reticular connective tissue. The cells are primarily monocytes and macrophages, and they accumulate in lymph nodes and the spleen. The Kupffer cells of the liver and tissue histiocytes are also part of the MPS. The mononuclear phagocyte system and the monocyte macrophage system refer to two different entities, often mistakenly understood as one. "Reticuloendothelial system" is an older term for the mononuclear phagocyte system, but it is used less commonly now, as it is understood that most endothelial cells are not macrophages. The mononuclear phagocyte system is also a somewhat dated concept trying to combine a broad range of cells, and should be used with caution. Cell types and locations The spleen is the second largest unit of the mononuclear phagocyte system. The monocyte is formed in the bone marrow and transported by the blood; it migrates into the tissues, where it transforms into a histiocyte or a macrophage. Macrophages are diffusely scattered in the connective tissue and in liver (Kupffer cells), spleen and lymph nodes (sinus histiocytes), lungs (alveolar macrophages), and central nervous system (microglia). The half-life of blood monocytes is about 1 day, whereas the life span of tissue macrophages is several months or years. The mononuclear phagocyte system is part of both humoral and cell-mediated immunity. The mononuclear phagocyte system has an important role in defense against microorganisms, including mycobacteria, fungi, bacteria, protozoa, and viruses. Macrophages remove senescent erythrocytes, leukocytes, and megakaryocytes by phagocytosis and digestion. Functions Formation of new red blood cells (RBCs) and white blood cells (WBCs). Destruction of senescent RBCs. Formation of plasma proteins. Formation of