source
stringlengths
31
227
text
stringlengths
9
2k
https://en.wikipedia.org/wiki/Debugfs
debugfs is a special file system available in the Linux kernel since version 2.6.10-rc3. It was written by Greg Kroah-Hartman. debugfs is a simple-to-use RAM-based file system specially designed for debugging purposes. It exists as a simple way for kernel developers to make information available to user space. Unlike , which is only meant for information about a process, or sysfs, which has strict one-value-per-file rules, debugfs has no rules at all. Developers can put any information they want there. Use To compile a Linux kernel with the debugfs facility, the option must be set to yes. It is typically mounted at with a command such as: mount -t debugfs none /sys/kernel/debug It can be manipulated using several calls from the C header file , which include: for creating a file in the debug filesystem. for creating a directory inside the debug filesystem. for creating a symbolic link inside the debug filesystem. for removing a debugfs entry from the debug filesystem.
https://en.wikipedia.org/wiki/Futile%20cycle
A futile cycle, also known as a substrate cycle, occurs when two metabolic pathways run simultaneously in opposite directions and have no overall effect other than to dissipate energy in the form of heat. The reason this cycle was called "futile" cycle was because it appeared that this cycle operated with no net utility for the organism. As such, it was thought of being a quirk of the metabolism and thus named a futile cycle. After further investigation it was seen that futile cycles are very important for regulating the concentrations of metabolites. For example, if glycolysis and gluconeogenesis were to be active at the same time, glucose would be converted to pyruvate by glycolysis and then converted back to glucose by gluconeogenesis, with an overall consumption of ATP. Futile cycles may have a role in metabolic regulation, where a futile cycle would be a system oscillating between two states and very sensitive to small changes in the activity of any of the enzymes involved. The cycle does generate heat, and may be used to maintain thermal homeostasis, for example in the brown adipose tissue of young mammals, or to generate heat rapidly, for example in insect flight muscles and in hibernating animals during periodical arousal from torpor. It has been reported that the glucose metabolism substrate cycle is not a futile cycle but a regulatory process. For example, when energy is suddenly needed, ATP is replaced by AMP, a much more reactive adenine. Example The simultaneous carrying out of glycolysis and gluconeogenesis is an example of a futile cycle, represented by the following equation: For example, during glycolysis, fructose-6-phosphate is converted to fructose-1,6-bisphosphate in a reaction catalysed by the enzyme phosphofructokinase 1 (PFK-1). But during gluconeogenesis (i.e. synthesis of glucose from pyruvate and other compounds) the reverse reaction takes place, being catalyzed by fructose-1,6-bisphosphatase (FBPase-1). Giving an overall reaction of:
https://en.wikipedia.org/wiki/GPS%20signals
GPS signals are broadcast by Global Positioning System satellites to enable satellite navigation. Receivers on or near the Earth's surface can determine location, time, and velocity using this information. The GPS satellite constellation is operated by the 2nd Space Operations Squadron (2SOPS) of Space Delta 8, United States Space Force. GPS signals include ranging signals, used to measure the distance to the satellite, and navigation messages. The navigation messages include ephemeris data, used in trilateration to calculate the position of each satellite in orbit, and information about the time and status of the entire satellite constellation, called the almanac. There are four GPS signal specifications designed for civilian use. In order of date of introduction, these are: L1 C/A, L2C, L5 and L1C. L1 C/A is also called the legacy signal and is broadcast by all currently operational satellites. L2C, L5 and L1C are modernized signals, and only broadcast by newer satellites (or not yet at all), and , none are yet considered to be fully operational for civilian use. In addition, there are restricted signals with published frequencies and chip rates but encrypted coding intended to be used only by authorized parties. Some limited use of restricted signals can still be made by civilians without decryption; this is called codeless and semi-codeless access, and is officially supported. The interface to the User Segment (GPS receivers) is described in the Interface Control Documents (ICD). The format of civilian signals is described in the Interface Specification (IS) which is a subset of the ICD. Common characteristics The GPS satellites (called space vehicles in the GPS interface specification documents) transmit simultaneously several ranging codes and navigation data using binary phase-shift keying (BPSK). Only a limited number of central frequencies are used; satellites using the same frequency are distinguished by using different ranging codes; in other words,
https://en.wikipedia.org/wiki/TV%20Network%20Protocol
The TV Network Protocol or TVNP as it is more commonly referred to is an open network protocol developed to enable CCTV systems from any manufacturer to be integrated into an existing CCTV network. It provides high levels of support for audio routing, video routing and camera control. The protocol was developed by Philips Projects (now Tyco Integrated Systems) on behalf of the Traffic Control Systems Unit (TCSU), now a part of Transport for London (TfL). Tyco acts as the standards and approvals house for companies who want to implement the protocol. The protocol's roots can be traced back to the Highways Agency HDLC standard. It is the property of TfL and is independent of any supplier. As of late of 2011 there are at least eight manufacturers who have a partial or full TVNP interface, including: BAE Systems (previously Petards) Chubb (previously Initial Fire and Security) Honeywell Infinitronix Meyertech Costain (previously Simulation Systems Limited) Synectics Tyco (previously Philips Projects). TVNP layers are broadly based on the OSI model. TVNP Layer 2 and 3 correspond to OSI Layers 2 and 3. When used over RS-232 only, TVNP Layer 1 corresponds to OSI Layer 1. TVNP Layer 4 is equivalent to OSI Layer 7. Structuring the TVNP in such a way means that as future needs and provisions change, aspects of one layer can be enhanced or modified without the need for change to the other layers. Layer 1 (L1) For serial RS-232 L1 is the Physical Protocol Layer that defines the electrical signals and interconnect requirements at the communication interface port(s) of the CCTV system. V3.0 of the specification allows UDP/IP, typically over Ethernet, to be used for L1. This option is not a physical protocol layer in the OSI sense. Layer 2 (L2) is the Frame Protocol Layer, sometimes referred to as the Link Layer. Its purpose is to detect and correct errors in the stream of data passing between any two adjacent CCTV systems, so that CCTV network messages are not recei
https://en.wikipedia.org/wiki/Event%20chain%20methodology
Event chain methodology is a network analysis technique that is focused on identifying and managing events and relationship between them (event chains) that affect project schedules. It is an uncertainty modeling schedule technique. Event chain methodology is an extension of quantitative project risk analysis with Monte Carlo simulations. It is the next advance beyond critical path method and critical chain project management. Event chain methodology tries to mitigate the effect of motivational and cognitive biases in estimating and scheduling. It improves accuracy of risk assessment and helps to generate more realistic risk adjusted project schedules. History Event chain methodology is an extension of traditional Monte Carlo simulation of project schedules where uncertainties in task duration and costs are defined by statistical distribution. For example, task duration can be defined by three point estimates: low, base, and high. The results of analysis is a risk adjusted project schedule, crucial tasks, and probabilities that project will be completed on time and on budget. Defining uncertainties using statistical distribution provide accurate results if there is a reliable historical data about duration and cost of similar tasks in previous projects. Another approach is to define uncertainties using risk events or risk drivers, which can be assigned to different tasks or resources. Information about probabilities and impact of such events is easier to elicit, which improves accuracy of analysis. Risks can be recorded in the Risk register. Event chain methodology was first proposed in the period of 2002–2004. It is fully or partially implemented in a number of software application. Event Chain Methodology is based on six principles and has a number of outcomes. Principles Moment of risk and state of activity Activities (tasks) are not a continuous uniform procedure. Tasks are affected by external events, which transform an activity from one state to another.
https://en.wikipedia.org/wiki/Perpendicular%20axis%20theorem
The perpendicular axis theorem (or plane figure theorem) states that the moment of inertia of a planar lamina (i.e. 2-D body) about an axis perpendicular to the plane of the lamina is equal to the sum of the moments of inertia of the lamina about the two axes at right angles to each other, in its own plane intersecting each other at the point where the perpendicular axis passes through it. Define perpendicular axes , , and (which meet at origin ) so that the body lies in the plane, and the axis is perpendicular to the plane of the body. Let Ix, Iy and Iz be moments of inertia about axis x, y, z respectively. Then the perpendicular axis theorem states that This rule can be applied with the parallel axis theorem and the stretch rule to find polar moments of inertia for a variety of shapes. If a planar object has rotational symmetry such that and are equal, then the perpendicular axes theorem provides the useful relationship: Derivation Working in Cartesian coordinates, the moment of inertia of the planar body about the axis is given by: On the plane, , so these two terms are the moments of inertia about the and axes respectively, giving the perpendicular axis theorem. The converse of this theorem is also derived similarly. Note that because in , measures the distance from the axis of rotation, so for a y-axis rotation, deviation distance from the axis of rotation of a point is equal to its x coordinate.
https://en.wikipedia.org/wiki/Overdrive%20voltage
Overdrive voltage, usually abbreviated as VOV, is typically referred to in the context of MOSFET transistors. The overdrive voltage is defined as the voltage between transistor gate and source (VGS) in excess of the threshold voltage (VTH) where VTH is defined as the minimum voltage required between gate and source to turn the transistor on (allow it to conduct electricity). Due to this definition, overdrive voltage is also known as "excess gate voltage" or "effective voltage." Overdrive voltage can be found using the simple equation: VOV = VGS − VTH. Technology VOV is important as it directly affects the output drain terminal current (ID) of the transistor, an important property of amplifier circuits. By increasing VOV, ID can be increased until saturation is reached. Overdrive voltage is also important because of its relationship to VDS, the drain voltage relative to the source, which can be used to determine the region of operation of the MOSFET. The table below shows how to use overdrive voltage to understand what region of operation the MOSFET is in: A more physics-related explanation follows: In an NMOS transistor, the channel region under zero bias has an abundance of holes (i.e., it is p-type silicon). By applying a negative gate bias (VGS < 0) we attract more holes, and this is called accumulation. A positive gate voltage (VGS > 0) will attract electrons and repel holes, and this is called depletion because we are depleting the number of holes. At a critical voltage called the threshold voltage (VTH) the channel will actually be so depleted of holes and rich in electrons that it will INVERT to being n-type silicon, and this is called the inversion region. As we increase this voltage, VGS, beyond VTH, we are said to be then overdriving the gate by creating a stronger channel, hence the overdrive (often called Vov, Vod, or Von) is defined as (VGS − VTH). See also MOSFET Threshold voltage Electronic amplifier Short-channel effect Biasing
https://en.wikipedia.org/wiki/Constant-Q%20transform
In mathematics and signal processing, the constant-Q transform and variable-Q transform, simply known as CQT and VQT, transforms a data series to the frequency domain. It is related to the Fourier transform and very closely related to the complex Morlet wavelet transform. Its design is suited for musical representation. The transform can be thought of as a series of filters fk, logarithmically spaced in frequency, with the k-th filter having a spectral width δfk equal to a multiple of the previous filter's width: where δfk is the bandwidth of the k-th filter, fmin is the central frequency of the lowest filter, and n is the number of filters per octave. Calculation The short-time Fourier transform of x[n] for a frame shifted to sample m is calculated as follows: Given a data series at sampling frequency fs = 1/T, T being the sampling period of our data, for each frequency bin we can define the following: Filter width, δfk. Q, the "quality factor": This is shown below to be the integer number of cycles processed at a center frequency fk. As such, this somewhat defines the time complexity of the transform. Window length for the k-th bin: Since fs/fk is the number of samples processed per cycle at frequency fk, Q is the number of integer cycles processed at this central frequency. The equivalent transform kernel can be found by using the following substitutions: The window length of each bin is now a function of the bin number: The relative power of each bin will decrease at higher frequencies, as these sum over fewer terms. To compensate for this, we normalize by N[k]. Any windowing function will be a function of window length, and likewise a function of window number. For example, the equivalent Hamming window would be Our digital frequency, , becomes . After these modifications, we are left with Variable-Q bandwidth calculation The variable-Q transform is the same as constant-Q transform, but the only difference is the filter Q is variable,
https://en.wikipedia.org/wiki/Dual%20cone%20and%20polar%20cone
Dual cone and polar cone are closely related concepts in convex analysis, a branch of mathematics. Dual cone In a vector space The dual cone C of a subset C in a linear space X over the reals, e.g. Euclidean space Rn, with dual space X is the set where is the duality pairing between X and X, i.e. . C is always a convex cone, even if C is neither convex nor a cone. In a topological vector space If X is a topological vector space over the real or complex numbers, then the dual cone of a subset C ⊆ X is the following set of continuous linear functionals on X: , which is the polar of the set -C. No matter what C is, will be a convex cone. If C ⊆ {0} then . In a Hilbert space (internal dual cone) Alternatively, many authors define the dual cone in the context of a real Hilbert space (such as Rn equipped with the Euclidean inner product) to be what is sometimes called the internal dual cone. Using this latter definition for C, we have that when C is a cone, the following properties hold: A non-zero vector y is in C if and only if both of the following conditions hold: y is a normal at the origin of a hyperplane that supports C. y and C lie on the same side of that supporting hyperplane. C is closed and convex. implies . If C has nonempty interior, then C is pointed, i.e. C* contains no line in its entirety. If C is a cone and the closure of C is pointed, then C has nonempty interior. C is the closure of the smallest convex cone containing C (a consequence of the hyperplane separation theorem) Self-dual cones A cone C in a vector space X is said to be self-dual if X can be equipped with an inner product ⟨⋅,⋅⟩ such that the internal dual cone relative to this inner product is equal to C. Those authors who define the dual cone as the internal dual cone in a real Hilbert space usually say that a cone is self-dual if it is equal to its internal dual. This is slightly different from the above definition, which permits a change of inner product. For
https://en.wikipedia.org/wiki/Separation%20principle
In control theory, a separation principle, more formally known as a principle of separation of estimation and control, states that under some assumptions the problem of designing an optimal feedback controller for a stochastic system can be solved by designing an optimal observer for the state of the system, which feeds into an optimal deterministic controller for the system. Thus the problem can be broken into two separate parts, which facilitates the design. The first instance of such a principle is in the setting of deterministic linear systems, namely that if a stable observer and a stable state feedback are designed for a linear time-invariant system (LTI system hereafter), then the combined observer and feedback is stable. The separation principle does not hold in general for nonlinear systems. Another instance of the separation principle arises in the setting of linear stochastic systems, namely that state estimation (possibly nonlinear) together with an optimal state feedback controller designed to minimize a quadratic cost, is optimal for the stochastic control problem with output measurements. When process and observation noise are Gaussian, the optimal solution separates into a Kalman filter and a linear-quadratic regulator. This is known as linear-quadratic-Gaussian control. More generally, under suitable conditions and when the noise is a martingale (with possible jumps), again a separation principle applies and is known as the separation principle in stochastic control. The separation principle also holds for high gain observers used for state estimation of a class of nonlinear systems and control of quantum systems. Proof of separation principle for deterministic LTI systems Consider a deterministic LTI system: where represents the input signal, represents the output signal, and represents the internal state of the system. We can design an observer of the form and state feedback Define the error e: Then Now we can write the closed-loo
https://en.wikipedia.org/wiki/Continental%20rise
The continental rise is a low-relief zone of accumulated sediments that lies between the continental slope and the abyssal plain. It is a major part of the continental margin, covering around 10% of the ocean floor. Formation This geologic structure results from deposition of sediments, mainly due to mass wasting, the gravity-driven downhill motion of sand and other sediments. Mass wasting can occur gradually, with sediments accumulating discontinuously, or in large, sudden events. Large mass wasting occurrences are often triggered by sudden events such as earthquakes or oversteepening of the continental slope. More gradual accumulation of sediments occurs when hemipelagic sediments suspended in the ocean slowly settle to the ocean basin. Slope Because the continental rise lies below the continental slope and is formed from sediment deposition, it has a very gentle slope, usually ranging from 1:50 to 1:500. As the continental rise extends seaward, the layers of sediment thin, and the rise merges with the abyssal plain, typically forming a slope of around 1:1000. Accompanying Structures Alluvial Fans Deposition of sediments at the mouth of submarine canyons may form enormous fan-shaped accumulations called submarine fans on both the continental slope and continental rise. Alluvial or sedimentary fans are shallow cone-shaped reliefs at the base of the continental slope that merge together, forming the continental rise. Erosional submarine canyons slope downward and lead to alluvial fan valleys with increasing depth. It is in this zone that sediment is deposited, forming the continental rise. Alluvial fans such as the Bengal Fan, which stretches , make up one of the largest sedimentary structures in the world. Many alluvial fans also contain critical oil and natural gas reservoirs, making them key points for the collection of seismic data. Abyssal Plain Beyond the continental rise stretches the abyssal plain, which lies on top of basaltic oceanic crust and spa
https://en.wikipedia.org/wiki/Hot%20zone%20%28environment%29
Hot zone, also written as hot-zone or hotzone, refers to an area or region that is significantly affected by environmental hazards or risks. It may refer to a location where there is high pollution, contamination, or a concentration of hazardous substances or activities. Etymology The term hot zone was likely coined during the Cold War where it described locations rendered hazardous due to nuclear contamination. The term was later extended to areas or locations considered to be hazardous such as Level-4 biosafety labs, places in which there is active conflict, and so forth. The term hot zone was popularized by the 1995 book The Hot Zone by Richard Preston, and its film adaptation Outbreak, released the same year. Types of hot zones Biological The biological zone describes an area or location where there is a risk of exposure to biological agents or contaminants that can cause harm to human health or the environment. It is often associated with situations involving the release, presence, or spread of infectious diseases, pathogens, or biological hazards. Precautions are taken in a gradient level of protection. In 2009, the outbreak of swine influenza happened in most places of the whole world. The swine influenza originated from a Mexican woman, and it was transmitted from person to person by air with a rapid speed. North America and Mexico were the first places to be affected by the virus. According to the World Health Organization (WHO), there were 14,142 deaths of swine influenza, and 1,311,522 people had this virus in their bodies in 2009. Spain, China, the United States, and South Korea are considered to be hot zones of swine influenza. There were 155,051 people who had this virus in Spain, 120,498 in China, 107,939 in United States and 101,182 in South Korea. The people in those areas can very easily get this virus. The virus is transmittable through the air, such as through coughing, sneezing or touching something containing the virus. Epidemics of a
https://en.wikipedia.org/wiki/Magnetic%20mirror%20point
In astrophysics, a magnetic mirror point is a point where the motion of a charged particle trapped in a magnetic field (such as the (approximately) dipole field of the Earth) reverses its direction. More precisely, it is the point where the projection of the particle's velocity vector in the direction of the field vector is equal to zero. Whenever charged particles from the sun hit Earth's magnetosphere, it is observed that the magnetic field of Earth reverses direction. Since the forces that generate our magnetic field are constantly changing, the field itself is also in continual flux, its strength waxing and waning over time. This causes the location of Earth's magnetic north and south poles to gradually shift, and to even completely flip locations every 300,000 years or so. See also Magnetic mirror L-shell Dipole model of the Earth's magnetic field List of artificial radiation belts
https://en.wikipedia.org/wiki/International%20Service%20for%20the%20Acquisition%20of%20Agri-biotech%20Applications
The International Service for the Acquisition of Agri-biotech Applications (ISAAA) is a non-profit international organization that shares agricultural biotechnology, focusing on genetic engineering. Structure ISAAA operates three regional centers; ISAAA SEAsiaCenter, ISAAA AfriCenter and ISAAA AmeriCenter. ISAAA SEAsiaCenter is hosted by the International Rice Research Institute (IRRI) in Los Baños, Laguna, Philippines. This center also serves as the Global Coordination Office as well as the home of the Global Knowledge Center on Crop Biotechnology. ISAAA AfriCenter is hosted by the International Livestock Research Institute (ILRI) located in Nairobi, Kenya. ISAAA AmeriCenter is located in Cornell University, Ithaca, New York. It serves as the administrative and financial headquarters of the organization. Donor organisations The ISAAA receives funding from both public and private donors. Some of the ISAAA's funding agencies and companies include the USDA, US Grains Council, Monsanto, Bayer, two banks – Fondazione Bussolera in Italy and Ibercaja in Spain, USAID and the Agricultural Biotechnology Support Project II. Annual Report on the Global Status of Commercialized Biotech/GM Crops The organization releases an annual publication on the global status of commercially approved genetically engineered crops. The publication is authored by Clive James, the founder and chair emeritus of ISAAA. The annual brief provides research on global trends in the adoption of major biotech crops since they were first planted commercially. Various environmental groups have accused the ISAAA of inflating the size and impact of genetically modified crops in their report. James says that the report is based on a multiple public and private sources and that he considers it conservative. The 2015 report says that "18 million farmers planted 179.7 million hectares of biotech crops in 28 countries, a marginal decrease of 1% (1.8 million hectares) from 2014." As per International Servic
https://en.wikipedia.org/wiki/Flash%20cut
A flash cut, also called a flash cutover, is an immediate change in a complex system, with no phase-in period. In the United States, some telephone area codes were split or overlaid immediately, rather than being phased in with a permissive dialing period. An example is telephone area code 213, which serves downtown Los Angeles and its immediate environs, split in January 1951 into 213 and 714 all at once. Another example is an immediate switch from an analog television channel to a digital television channel on the same frequency, where the two cannot operate in parallel without interference. A flash cut can also define a procedure in which multiple components of computer infrastructure are upgraded in multiple ways, all at once, with no phase-in period. In film, an extremely brief shot, sometimes as short as one frame, which is nearly subliminal in effect. Also a series of short staccato shots that create a rhythmic effect. See also Big bang adoption Flag day (software) Dagen H, when Sweden switched from driving on the left-hand side of the road to the right Smash cut, an abrupt change of scene in a motion picture
https://en.wikipedia.org/wiki/Burgers%20vector
In materials science, the Burgers vector, named after Dutch physicist Jan Burgers, is a vector, often denoted as , that represents the magnitude and direction of the lattice distortion resulting from a dislocation in a crystal lattice. The vector's magnitude and direction is best understood when the dislocation-bearing crystal structure is first visualized without the dislocation, that is, the perfect crystal structure. In this perfect crystal structure, a rectangle whose lengths and widths are integer multiples of (the unit cell edge length) is drawn encompassing the site of the original dislocation's origin. Once this encompassing rectangle is drawn, the dislocation can be introduced. This dislocation will have the effect of deforming, not only the perfect crystal structure, but the rectangle as well. The said rectangle could have one of its sides disjoined from the perpendicular side, severing the connection of the length and width line segments of the rectangle at one of the rectangle's corners, and displacing each line segment from each other. What was once a rectangle before the dislocation was introduced is now an open geometric figure, whose opening defines the direction and magnitude of the Burgers vector. Specifically, the breadth of the opening defines the magnitude of the Burgers vector, and, when a set of fixed coordinates is introduced, an angle between the termini of the dislocated rectangle's length line segment and width line segment may be specified. When calculating the Burgers vector practically, one may draw a rectangular counterclockwise circuit (Burgers circuit) from a starting point to enclose the dislocation (see the picture above). The Burgers vector will be the vector to complete the circuit, i.e., from the end to the start of the circuit. The direction of the vector depends on the plane of dislocation, which is usually on one of the closest-packed crystallographic planes. The magnitude is usually represented by the equation (For BCC
https://en.wikipedia.org/wiki/Thailand%20Center%20of%20Excellence%20for%20Life%20Sciences
The Thailand Center of Excellence for Life Sciences (TCELS) was founded in 2004 by the government of Thailand. TCELS is a public organization under the auspices of the Ministry of Higher Education, Science, Research and Innovation. TCELS has the responsibility of providing a link between innovation in life sciences and investment, and spurring domestic and international partnership in the life science business in Thailand. History TCELS was founded in 2004. Initially, TCELS was established as an organization under the umbrella of the Office of Knowledge Management and Development (OKMD), which houses a group of public organizations under the supervision of the Office of the Prime Minister. On 27 May 2011, TCELS was elevated to a public organization under the Ministry of Science and Technology (MST). Mission Support and develop the life sciences business and industry Promote and support innovations, research, and knowledge related to the commercialization of life sciences products and services Develop and support the necessary infrastructure and human capacity for life sciences business and industry Create a strategic plan for developing life sciences business and industry Serve as the coordination center for facilitating cooperation among domestic and international organizations for life sciences business and industry Serve as Thailand’s life sciences business information and knowledge center. Focus areas Pharmaceuticals and biotechnology A key project is pharmacogenomics. TCELS supports the Medical Genomic Center to promote awareness of this diagnostic tool. Natural products TCELS has supported a project with the aim of developing new products from Hevea brasiliensis. The research is conducted by a team from Prince of Songkhla University. Biomedical engineering TCELS conducts various projects in medical robotics, medical devices, and operates the Advanced Dental Technology Center (ADTEC). Medical services Advanced Cell and Gene Therapies Program Automated Ce
https://en.wikipedia.org/wiki/Causal%20sets
The causal sets program is an approach to quantum gravity. Its founding principles are that spacetime is fundamentally discrete (a collection of discrete spacetime points, called the elements of the causal set) and that spacetime events are related by a partial order. This partial order has the physical meaning of the causality relations between spacetime events. The program is based on a theorem by David Malament that states that if there is a bijective map between two past and future distinguishing space times that preserves their causal structure then the map is a conformal isomorphism. The conformal factor that is left undetermined is related to the volume of regions in the spacetime. This volume factor can be recovered by specifying a volume element for each space time point. The volume of a space time region could then be found by counting the number of points in that region. Causal sets was initiated by Rafael Sorkin who continues to be the main proponent of the program. He has coined the slogan "Order + Number = Geometry" to characterize the above argument. The program provides a theory in which space time is fundamentally discrete while retaining local Lorentz invariance. Definition A causal set (or causet) is a set with a partial order relation that is Reflexive: For all , we have . Antisymmetric: For all , we have and implies . Transitive: For all , we have and implies . Locally finite: For all , we have is a finite set. We'll write if and . The set represents the set of spacetime events and the order relation represents the causal relationship between events (see causal structure for the analogous idea in a Lorentzian manifold). Although this definition uses the reflexive convention we could have chosen the irreflexive convention in which the order relation is irreflexive and asymmetric. The causal relation of a Lorentzian manifold (without closed causal curves) satisfies the first three conditions. It is the local finiteness cond
https://en.wikipedia.org/wiki/Quality%20engineering
Quality engineering is the discipline of engineering concerned with the principles and practice of product and service quality assurance and control. In software development, it is the management, development, operation and maintenance of IT systems and enterprise architectures with a high quality standard. Description Quality engineering is the discipline of engineering that creates and implements strategies for quality assurance in product development and production as well as software development. Quality Engineers focus on optimizing product quality which W. Edwards Deming defined as: Quality engineering body of knowledge includes: Management and leadership The quality system Elements of a quality system Product and process design Classification of quality characteristics Design inputs and review Design verification Reliability and maintainability Product and process control Continuous improvement Quality control tools Quality management and planning tools Continuous improvement techniques Corrective action Preventive action Statistical process control (SPC) Risk management Roles Auditor: Quality engineers may be responsible for auditing their own companies or their suppliers for compliance to international quality standards such as ISO9000 and AS9100. They may also be independent auditors under an auditing body. Process quality: Quality engineers may be tasked with value stream mapping and statistical process control to determine if a process is likely to produce defective product. They may create inspection plans and criteria to ensure defective parts are detected prior to completion. Supplier quality: Quality engineers may be responsible for auditing suppliers or performing root cause and corrective action at their facility or overseeing such activity to prevent the delivery of defective product. Software IT services are increasingly interlinked in workflows across platform boundaries, device and organisational boundaries, for exam
https://en.wikipedia.org/wiki/Compiler%20Description%20Language
Compiler Description Language (CDL) is a programming language based on affix grammars. It is very similar to Backus–Naur form (BNF) notation. It was designed for the development of compilers. It is very limited in its capabilities and control flow, and intentionally so. The benefits of these limitations are twofold. On the one hand, they make possible the sophisticated data and control flow analysis used by the CDL2 optimizers resulting in extremely efficient code. The other benefit is that they foster a highly verbose naming convention. This, in turn, leads to programs that are, to a great extent, self-documenting. The language looks a bit like Prolog (this is not surprising since both languages arose at about the same time out of work on affix grammars). However, as opposed to Prolog, control flow in CDL is deterministically based on success/failure, i.e., no other alternatives are tried when the current one succeeds. This idea is also used in parsing expression grammars. CDL3 is the third version of the CDL language, significantly different from the previous two versions. Design The original version, designed by Cornelis H. A. Koster at the University of Nijmegen, which emerged in 1971, had a rather unusual concept: it had no core. A typical programming language source is translated to machine instructions or canned sequences of those instructions. Those represent the core, the most basic abstractions that the given language supports. Such primitives can be the additions of numbers, copying variables to each other, and so on. CDL1 lacks such a core. It is the responsibility of the programmer to provide the primitive operations in a form that can then be turned into machine instructions by means of an assembler or a compiler for a traditional language. The CDL1 language itself has no concept of primitives, no concept of data types apart from the machine word (an abstract unit of storage - not necessarily a real machine word as such). The evaluation rules a
https://en.wikipedia.org/wiki/Hot%20and%20cold%20cognition
Hot cognition is a hypothesis on motivated reasoning in which a person's thinking is influenced by their emotional state. Put simply, hot cognition is cognition coloured by emotion. Hot cognition contrasts with cold cognition, which implies cognitive processing of information that is independent of emotional involvement. Hot cognition is proposed to be associated with cognitive and physiological arousal, in which a person is more responsive to environmental factors. As it is automatic, rapid and led by emotion, hot cognition may consequently cause biased decision making. Hot cognition may arise, with varying degrees of strength, in politics, religion, and other sociopolitical contexts because of moral issues, which are inevitably tied to emotion. Hot cognition was initially proposed in 1963 by Robert P. Abelson. The idea became popular in the 1960s and the 1970s. An example of a biased decision caused by hot cognition would be a juror disregarding evidence because of an attraction to the defendant. Decision making with cold cognition is more likely to involve logic and critical analysis. Therefore, when an individual engages in a task while using cold cognition, the stimulus is likely to be emotionally neutral and the "outcome of the test is not motivationally relevant" to the individual. An example of a critical decision using cold cognition would be concentrating on the evidence before drawing a conclusion. Hot and cold cognition form a dichotomy within executive functioning. Executive functioning has long been considered as a domain general cognitive function, but there has been support for separation into "hot" affective aspects and "cold" cognitive aspects. It is recognized that executive functioning spans across a number of cognitive tasks, including working memory, cognitive flexibility and reasoning in active goal pursuit. The distinction between hot and cool cognition implies that executive function may operate differently in different contexts. The dist
https://en.wikipedia.org/wiki/Wyoming%20statistical%20areas
The U.S. currently has ten statistical areas that have been delineated by the Office of Management and Budget (OMB). In July, 2023, the OMB delineated two metropolitan statistical areas and eight micropolitan statistical areas in Wyoming. The most populous of these statistical areas is the Cheyenne, WY Metropolitan Statistical Area with a 2020 Census population of 100,512. Statistical areas The Office of Management and Budget (OMB) has designated more than 1,000 statistical areas for the United States and Puerto Rico. These statistical areas are important geographic delineations of population clusters used by the OMB, the United States Census Bureau, planning organizations, and federal, state, and local government entities. The OMB defines a core-based statistical area (commonly referred to as a CBSA) as "a statistical geographic entity consisting of the county or counties (or county-equivalents) associated with at least one core of at least 10,000 population, plus adjacent counties having a high degree of social and economic integration with the core as measured through commuting ties with the counties containing the core." The OMB further divides core-based statistical areas into metropolitan statistical areas (MSAs) that have "a population of at least 50,000" and micropolitan statistical areas (μSAs) that have "a population of at least 10,000, but less than 50,000." The OMB defines a combined statistical area (CSA) as "a geographic entity consisting of two or more adjacent core-based statistical areas with employment interchange measures of at least 15%." The primary statistical areas (PSAs) include all combined statistical areas and any core-based statistical area that is not a constituent of a combined statistical area. Table The table below describes the 10 United States statistical areas and 23 counties of the State of Wyoming with the following information: The core based statistical area (CBSA) as designated by the OMB. The CBSA population according t
https://en.wikipedia.org/wiki/Field%20Programmable%20Nanowire%20Interconnect
Field Programmable Nanowire Interconnect (often abbreviated FPNI) is a new computer architecture developed by Hewlett-Packard. This is a defect-tolerant architecture, using the results of the Teramac experiment. Technology The design combines a nanoscale crossbar switch structure with conventional CMOS to create a hybrid chip that is simpler to fabricate and offers greater flexibility in the choice of nanoscale devices. The FPNI improves on a field-programmable gate array (FPGA) architecture by lifting the configuration bit and associated components out of the semiconductor plane and replacing them in the interconnect with nonvolatile switches, which decreases both the area and power consumption of the circuit -- while providing up to eight times the density at less cost. This is an example of a more comprehensive strategy for improving the efficiency of existing semiconductor technology: placing a level of intelligence and configurability in the interconnect can have a profound effect on integrated circuit performance, and can be used to significantly extend Moore's Law without having to shrink the transistors. External links http://www.iop.org/EJ/abstract/0957-4484/18/3/035204 Nanotechnology journal, Issue 3 (24 January 2007)Nano/CMOS architectures using a field-programmable nanowire interconnect Gate arrays
https://en.wikipedia.org/wiki/Phosphorylation%20cascade
A phosphorylation cascade is a sequence of signaling pathway events where one enzyme phosphorylates another, causing a chain reaction leading to the phosphorylation of thousands of proteins. This can be seen in signal transduction of hormone messages. A signaling pathway begins at the cell surface where a hormone or protein binds to a receptor at the extracellular matrix. The interactions between the molecule and receptor cause a conformational change at the receptor, which activates multiple enzymes or proteins. These enzymes activate secondary messengers, which leads to the phosphorylation of thousands of proteins. The end product of a phosphorylation cascade is the changes occurring inside the cell. One best example that explains this phenomenon is mitogen-activated protein (MAP) kinase or ERK kinase. MAP kinase not only plays an important function during growth of cell in the M phase phosphorylation cascade but also plays an important role during the sequence of signaling pathway. In order to regulate its functions so it does not cause chaos, it can only be active when both tyrosine and threonine/serine residues are phosphorylated.
https://en.wikipedia.org/wiki/1/2%20%2B%201/4%20%2B%201/8%20%2B%201/16%20%2B%20%E2%8B%AF
In mathematics, the infinite series is an elementary example of a geometric series that converges absolutely. The sum of the series is 1. In summation notation, this may be expressed as The series is related to philosophical questions considered in antiquity, particularly to Zeno's paradoxes. Proof As with any infinite series, the sum is defined to mean the limit of the partial sum of the first terms as approaches infinity. By various arguments, one can show that this finite sum is equal to As approaches infinity, the term approaches 0 and so tends to 1. History Zeno's paradox This series was used as a representation of many of Zeno's paradoxes. For example, in the paradox of Achilles and the Tortoise, the warrior Achilles was to race against a tortoise. The track is 100 meters long. Achilles could run at 10 m/s, while the tortoise only 5. The tortoise, with a 10-meter advantage, Zeno argued, would win. Achilles would have to move 10 meters to catch up to the tortoise, but the tortoise would already have moved another five meters by then. Achilles would then have to move 5 meters, where the tortoise would move 2.5 meters, and so on. Zeno argued that the tortoise would always remain ahead of Achilles. The Dichotomy paradox also states that to move a certain distance, you have to move half of it, then half of the remaining distance, and so on, therefore having infinitely many time intervals. This can be easily resolved by noting that each time interval is a term of the infinite geometric series, and will sum to a finite number. The Eye of Horus The parts of the Eye of Horus were once thought to represent the first six summands of the series. In a myriad ages it will not be exhausted A version of the series appears in the ancient Taoist book Zhuangzi. The miscellaneous chapters "All Under Heaven" include the following sentence: "Take a chi long stick and remove half every day, in a myriad ages it will not be exhausted." See also 0.999... 1/2 − 1/4
https://en.wikipedia.org/wiki/1/4%20%2B%201/16%20%2B%201/64%20%2B%201/256%20%2B%20%E2%8B%AF
In mathematics, the infinite series is an example of one of the first infinite series to be summed in the history of mathematics; it was used by Archimedes circa 250–200 BC. As it is a geometric series with first term and common ratio , its sum is Visual demonstrations The series lends itself to some particularly simple visual demonstrations because a square and a triangle both divide into four similar pieces, each of which contains the area of the original. In the figure on the left, if the large square is taken to have area 1, then the largest black square has area  ×  = . Likewise, the second largest black square has area , and the third largest black square has area . The area taken up by all of the black squares together is therefore , and this is also the area taken up by the gray squares and the white squares. Since these three areas cover the unit square, the figure demonstrates that Archimedes' own illustration, adapted at top, was slightly different, being closer to the equation See below for details on Archimedes' interpretation. The same geometric strategy also works for triangles, as in the figure on the right: if the large triangle has area 1, then the largest black triangle has area , and so on. The figure as a whole has a self-similarity between the large triangle and its upper sub-triangle. A related construction making the figure similar to all three of its corner pieces produces the Sierpiński triangle. Proof by Archimedes Archimedes encounters the series in his work Quadrature of the Parabola. He finds the area inside a parabola by the method of exhaustion, and he gets a series of triangles; each stage of the construction adds an area times the area of the previous stage. His desired result is that the total area is times the area of the first stage. To get there, he takes a break from parabolas to introduce an algebraic lemma: Proposition 23. Given a series of areas , of which A is the greatest, and each is equal to four times th
https://en.wikipedia.org/wiki/Shope%20papilloma%20virus
The Shope papilloma virus (SPV), also known as cottontail rabbit papilloma virus (CRPV) or Kappapapillomavirus 2, is a papillomavirus which infects certain leporids, causing keratinous carcinomas resembling horns, typically on or near the animal's head. The carcinomas can metastasize or become large enough to interfere with the host's ability to eat, causing starvation. Richard E. Shope investigated the horns and discovered the virus in 1933, an important breakthrough in the study of oncoviruses. The virus was originally discovered in cottontail rabbits in the Midwestern U.S. but can also infect brush rabbits, black-tailed jackrabbits, snowshoe hares, European rabbits, and domestic rabbits. History In the 1930s, hunters in northwestern Iowa reported that the rabbits they shot had several "horn" protrusions on many parts of their bodies including their faces and necks. The virus is also a possible source of myths about the jackalope, a rabbit with the horns of an antelope, and related cryptids such as the wolpertinger. Stories and illustrations of horned rabbits appear in scientific treatises dating back many years, such as the Tableau encyclopédique et méthodique, from 1655. The Iowa reports led cancer researcher Richard E. Shope to investigate, and he discovered the virus in 1933. He separated the virus from horny warts on cottontail rabbits and made one of the first mammalian tumor virus discoveries. Shope determined the protrusions were keratinous carcinomas due to the infection of CRPV. Shope's research led to the development of the first mammalian model of cancer caused by a virus. He was able to isolate virus particles from tumors on captured animals and use these to inoculate domestic rabbits, which then developed similar tumors. This has contributed to our understanding of fundamental mechanisms in neoplasia, or the formation of a new, abnormal growth of tissue. The virus was sequenced in 1984, showing substantial sequence similarities to HPV1a. It has be
https://en.wikipedia.org/wiki/Insect%20biodiversity
Insect biodiversity accounts for a large proportion of all biodiversity on the planet—over half of the estimated 1.5 million organism species described are classified as insects. Species diversity Estimates of the total number of insect species or those within specific orders are often highly variable. Globally, averages of these predictions estimate there are around 1.5 million beetle species and 5.5 million insect species with around 1 million insect species currently found and described. Between 950,000–1,000,000 of all described species are insects, so over 50% of all described eukaryotes (1.8 million) are insects (see illustration). With only 950,000 known non-insects, if the actual number of insects is 5.5 million, they may represent over 80% of the total, and with only about 20,000 new species of all organisms being described each year, most insect species likely will remain undescribed, unless species descriptions greatly increase in rate. Of the 24 orders of insects, four dominate in terms of numbers of described species, with at least 670,000 species included in Coleoptera, Diptera, Hymenoptera and Lepidoptera. The fossil record concerning insects stretches back for hundreds of millions of years. It suggests there are ongoing background levels of both new species appearing and extinctions. Very occasionally, the record also appears to show mass extinctions of insects. The Permian–Triassic extinction event saw the greatest level of insect extinction, with the Cretaceous–Paleogene being the second highest. Insect diversity has recovered after past mass extinctions, due to periods where new species originate with increased frequency, though the recovery can take millions of years. In the Holocene Several studies seemed to indicate that some insect populations are in decline in the late 20th and early 21st centuries, and has also been popularized as the windshield phenomenon. For many studies, factors such as abundance, biomass, and species richness are o
https://en.wikipedia.org/wiki/Comparison%20of%20bootloaders
The following tables compare general and technical information for a number of available bootloaders. General information Technical information Note: The column MBR (Master Boot Record) refers to whether or not the boot loader can be stored in the first sector of a mass storage device. The column VBR (Volume Boot Record) refers to the ability of the boot loader to be stored in the first sector of any partition on a mass storage device. Storage medium support Operating system support File-system support Non-journaled Journaled Read-only Other features Notes BOOT Loaders
https://en.wikipedia.org/wiki/Eventually%20%28mathematics%29
In the mathematical areas of number theory and analysis, an infinite sequence or a function is said to eventually have a certain property, if it doesn't have the said property across all its ordered instances, but will after some instances have passed. The use of the term "eventually" can be often rephrased as "for sufficiently large numbers", and can be also extended to the class of properties that apply to elements of any ordered set (such as sequences and subsets of ). Notation The general form where the phrase eventually (or sufficiently large) is found appears as follows: is eventually true for ( is true for sufficiently large ), where and are the universal and existential quantifiers, which is actually a shorthand for: such that is true or somewhat more formally: This does not necessarily mean that any particular value for is known, but only that such an exists. The phrase "sufficiently large" should not be confused with the phrases "arbitrarily large" or "infinitely large". For more, see Arbitrarily large#Arbitrarily large vs. sufficiently large vs. infinitely large. Motivation and definition For an infinite sequence, one is often more interested in the long-term behaviors of the sequence than the behaviors it exhibits early on. In which case, one way to formally capture this concept is to say that the sequence possesses a certain property eventually, or equivalently, that the property is satisfied by one of its subsequences , for some . For example, the definition of a sequence of real numbers converging to some limit is: For each positive number , there exists a natural number such that for all , . When the term "eventually" is used as a shorthand for "there exists a natural number such that for all ", the convergence definition can be restated more simply as: For each positive number , eventually . Here, notice that the set of natural numbers that do not satisfy this property is a finite set; that is, the set is empty or has
https://en.wikipedia.org/wiki/Bandelet%20%28computer%20science%29
Bandelets are an orthonormal basis that is adapted to geometric boundaries. Bandelets can be interpreted as a warped wavelet basis. The motivation behind bandelets is to perform a transform on functions defined as smooth functions on smoothly bounded domains. As bandelet construction utilizes wavelets, many of the results follow. Similar approaches to take account of geometric structure were taken for contourlets and curvelets. See also Wavelet Multiresolution analysis Scale space
https://en.wikipedia.org/wiki/GPS%20animal%20tracking
GPS animal tracking is a process whereby biologists, scientific researchers, or conservation agencies can remotely observe relatively fine-scale movement or migratory patterns in a free-ranging wild animal using the Global Positioning System (GPS) and optional environmental sensors or automated data-retrieval technologies such as Argos satellite uplink, mobile data telephony or GPRS and a range of analytical software tools. A GPS tracking device will generally record and store location data at a pre-determined interval or on interrupt by an environmental sensor. These data may be held pending recovery of the device or relayed to a central data store or internet-connected computer using an embedded cellular (GPRS), radio, or satellite modem. The animal's location can then be plotted against a map or chart in near real-time or, when analysing the track later, using a GIS package or custom software. GPS tracking devices may also be attached to domestic animals, such as pets, pedigree livestock and working dogs. Some owners use these collars for geofencing of their pets. GPS wildlife tracking can place additional constraints on size and weight and may not allow for post-deployment recharging or replacement of batteries or correction of attachment. As well as allowing in-depth study of animal behaviour and migration, the high-resolution tracks available from a GPS-enabled system can potentially allow for tighter control of animal-borne communicable diseases such as the H5N1 strain of avian influenza. Attachment Collar attachment Collar attachment is the primary technique where the subject has a suitable body type and behaviour. Tracking collars are typically used on the animal's neck (assuming the head has a larger circumference than the neck) but also on a limb, perhaps around an ankle. Suitable animals for neck attachment include primates, large cats, some bears, etc. Limb attachment works well in animals such as kiwi, where the foot is much larger than the ankle
https://en.wikipedia.org/wiki/Sun%20Web%20Developer%20Pack
The Sun Web Developer Pack (SWDP) is a collection of open source software released by Sun Microsystems for developing web applications that run on Java EE application servers. The SWDP is targeted at software developers interested in writing web applications that use Web 2.0 technologies such as Ajax, REST, Atom, and JavaScript. Software Included in the SWDP The SWDP consists of the following software: Scripting language support Project Phobos, a project that allows you to write web applications in JavaScript or other scripting languages Ajax technologies Project jMaki, a framework for creating Ajax-enabled web applications in Java, PHP, or Phobos Project Dynamic Faces, a framework for creating Ajax-enabled JavaServer Faces applications REST RESTful web services, an API for creating REST web services in Java WADL ROME, a Java API for parsing and generating RSS and Atom web feeds Atom Server (The ROME Propono subproject), a prototype Java API and framework for creating a web feed server for Atom feeds Release history Release 1 of the SWDP was made public on March 12, 2007. External links Sun Web Developer Pack home page Documentation Online SWDP Tutorial SWDP Tutorial bundle download Getting Started with the SWDP java.net project pages Project jMaki Project Dynamic Faces Project Phobos WADL ROME API Scripting languages Web services Java (programming language) Sun Microsystems Sun Microsystems software Web frameworks Java enterprise platform Software architecture
https://en.wikipedia.org/wiki/Web%20Application%20Description%20Language
The Web Application Description Language (WADL) is a machine-readable XML description of HTTP-based web services. WADL models the resources provided by a service and the relationships between them. WADL is intended to simplify the reuse of web services that are based on the existing HTTP architecture of the Web. It is platform and language independent and aims to promote reuse of applications beyond the basic use in a web browser. WADL was submitted to the World Wide Web Consortium by Sun Microsystems on 31 August 2009, but the consortium has no current plans to standardize it. WADL is the REST equivalent of SOAP's Web Services Description Language (WSDL), which can also be used to describe REST web services. Format The service is described using a set of resource elements. Each resource contains param elements to describe the inputs, and method elements which describe the request and response of a resource. The request element specifies how to represent the input, what types are required and any specific HTTP headers that are required. The response describes the representation of the service's response, as well as any fault information, to deal with errors. Example The following listing shows an example of a WADL description for the Yahoo News Search application. <application xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" xsi:schemaLocation="http://wadl.dev.java.net/2009/02 wadl.xsd" xmlns:tns="urn:yahoo:yn" xmlns:yn="urn:yahoo:yn" xmlns:ya="urn:yahoo:api" xmlns:xsd="http://www.w3.org/2001/XMLSchema" xmlns="http://wadl.dev.java.net/2009/02"> <grammars> <include href="NewsSearchResponse.xsd"/> <include href="Error.xsd"/> </grammars> <resources base="http://api.search.yahoo.com/NewsSearchService/V1/"> <resource path="newsSearch"> <method name="GET" id="search"> <request> <param name="appid" type="xsd:string" style="query" required="true"/> <param name="query" type="xsd:stri
https://en.wikipedia.org/wiki/Middlings%20purifier
A middlings purifier is a device used in the production of flour to remove the husks from the kernels of wheat. It was developed in Minnesota by Edmund LaCroix, a French inventor hired by Cadwallader C. Washburn and George Christian of the Washburn "A" Mill. It was developed to complement the emerging roller mill technique of the late 19th century, which used corrugated metal rollers instead of abrasive grindstones to grind wheat into flour. The middlings purifier was used to separate the bran from the usable part of the flour. The machine developed by LaCroix passed the partially ground middlings over a screen, and a stream of air blew away the particles of bran. This process was used because winter wheat, sown in the fall and harvested early the next summer was not feasible to grow in Minnesota. Spring wheat was sown in the spring and harvested in late summer. This could be grown by Minnesota farmers, but the conventional techniques of grinding grain between millstones ended up producing a darker flour than consumers desired. It was also difficult to mix the gluten and the starch completely. After Washburn's company developed the roller-milling technique with the use of a middlings purifier, they tried to monopolize the method, but the Pillsbury Company and other competitors were able to duplicate the process thanks to employees who left Washburn and passed along trade secrets. The development of the middling purifier may have been based on the invention of the purifier by Ignaz Paul, an Austrian Miller and inventor (1778 - 1842) early in the 19th century. Washburn later teamed up with John Crosby to form the Washburn-Crosby Company, which eventually became General Mills.
https://en.wikipedia.org/wiki/Vehicular%20metrics
There are a broad range of metrics that denote the relative capabilities of various vehicles. Most of them apply to all vehicles while others are type-specific. See also
https://en.wikipedia.org/wiki/IntruShield
The McAfee IntruShield is a network-based intrusion prevention sensor appliance that is used in prevention of zero-day, DoS (Denial of Service) attacks, spyware, malware, botnets and VoIP threats. It is now called McAfee Network Security Platform.
https://en.wikipedia.org/wiki/Boundaries%20in%20landscape%20history
Boundaries—particularly field boundaries—are among the oldest features in an English rural landscape. Although a boundary itself is an abstract concept, the boundary can often be seen by differences in land use on either side. Longevity of boundaries Boundaries - a real or imagined line that marks the limit of something. Many field boundaries in the central region of England originated with the enclosure of the previous open fields in the 18th or 19th century. In a few instances, current field boundaries (particularly in the West Country) have been shown to have originated in the Bronze Age or Iron Age. With a few exceptions, however, the attempt to establish pre-Saxon boundaries has been "largely fruitless". Areas that were never formally enclosed might yet prove a more fruitful area of research. Hedgerow dating The presence of bluebells in a hedge is often an indicator of an early hedge. It has been proposed that boundary hedges can be dated by hedgerow dating. This involves counting the number of species in a 27-metre section of hedge. In its simplest form each separate species suggests an age of 100 years. A variety of additional complexities have been suggested, but results have been mixed and the technique remains controversial. Parish boundaries Parish boundaries are of particular interest to landscape historians, since they are often inherited from land holdings that date back to the middle Saxon period or earlier. The coincidence of another landscape feature with a parish boundary can be used to date that feature—for example in the Time Team episode screened on 11 March 2007, a mill leat was determined to pre-date the Norman conquest because it coincided with a parish boundary. The boundaries of a few Anglo-Saxon estates were described in the boundary clauses of Anglo-Saxon Charters. These boundary clauses can sometimes be used to characterise the landscape at the time. In some cases, it has been possible to show that the boundaries of these Anglo-Saxon
https://en.wikipedia.org/wiki/Static%20synchronous%20compensator
A static synchronous compensator (STATCOM), is a shunt-connected, reactive compensation device used on transmission networks. It uses power electronics to form a voltage-source converter that can act as either a source or sink of reactive AC power to an electricity network. It is a member of the FACTS family of devices. STATCOMS are alternatives to other passive reactive power devices, such as capacitors and inductors (reactors). They have a variable reactive power output, can change their output in terms of milliseconds, and able to supply and consume both capacitive and inductive vars. While they can be used for voltage support and power factor correction, their speed and capability are better suited for dynamic situations like supporting the grid under fault conditions or contingency events. The use of voltage-source based FACTs device had been desirable for some time, as it helps mitigate the limitations of current-source based devices whose reactive output decreases with system voltage. However, limitations in technology have historically prevented wide adoption of STATCOMs. When gate turn-off thyristors (GTO) became more widely available in the 1990s and had the ability to switch both on and off at higher power levels, the first STATCOMs began to be commercially available. These devices typically used 3-level topologies and Pulse-Width Modulation (PWM) to simulate voltage waveforms. Modern STATCOMs now make use of Insulated-Gate Bipolar Transistors (IGBTs), which allow for faster switching at high-power levels. 3-level topologies have begun to give way to Multi-Modular Converter (MMC) Topologies, which allow for more levels in the voltage waveform, reducing harmonics and improving performance. History When AC won the War of Currents in the late 19th century, and electric grids began expanding and connecting cities and states, the need for reactive compensation became apparent. While AC offered benefits with transformation and reduced current, the alternati
https://en.wikipedia.org/wiki/List%20of%20plant%20orders
This article lists the living orders of the Viridiplantae, based primarily on the work of Ruggiero et al. 2015. Living order of Lycophytes and ferns are taken from Christenhusz et al. 2011b and Pteridophyte Phylogeny Group. Living orders of Gymnosperms are added from Christenhusz et al. 2011a while extinct orders are from Anderson, Anderson & Cleal 2007. Division Prasinodermophyta Class Prasinodermophyceae Order Prasinodermatales Class Palmophyllophyceae Order Prasinococcales Order Palmophyllales Division Chlorophyta Subdivision Prasinophytina Class Mamiellophyceae Order Monomastigales Order Dolichomastigales Order Mamiellales Class Pyramimonadophyceae Order Pyramimonadales Subdivision Chlorophytina Class Nephroselmidophyceae Order Nephroselmidales Class Picocystophyceae Order Picocystales Order ?Pseudoscourfieldiales Class Chloropicophyceae Order Chloropicales Class Pedinophyceae Order ?Scourfieldiales Order Marsupiomonadales Order Pedinomonadales Class Chlorodendrophyceae Order Chlorodendrales Class Trebouxiophyceae Order ?Phyllosiphonales Order Chlorellales Order Prasiolales Order Microthamniales Order Trebouxiales Class Ulvophyceae Order Ignatiales Order Oltmannsiellopsidales Order Scotinosphaerales Order Ulotrichales Order Ulvales Order Trentepohliales Order Cladophorales Order Dasycladales Order Bryopsidales Class Chlorophyceae Order Chaetopeltidales Order Chaetophorales Order Chlamydomonadales Order Chlorococcales Order Microsporales Order Oedogoniales Order Sphaeropleales Order Tetrasporales Division Streptophyta Subdivision Chlorokybophytina Class Mesostigmatophyceae Order Mesostigmatales Class Chlorokybophyceae Order Chlorokybales Subdivision Klebsormidiophyinta Class Klebsormidiophyceae Order Klebsormidiales Subdivision Charophytina Class Charophyceae Order †Sycidiales Order †Chovanellales Order †Moellerinales Order Charales (Stoenworts & musk grasses) Subdivision Coleochaetophytina Cla
https://en.wikipedia.org/wiki/Test%20%28biology%29
In biology, a test is the hard shell of some spherical marine animals and protists, notably sea urchins and microorganisms such as testate foraminiferans, radiolarians, and testate amoebae. The term is also applied to the covering of scale insects. The related Latin term testa is used for the hard seed coat of plant seeds. Etymology The anatomical term "test" derives from the Latin testa (which means a rounded bowl, amphora or bottle). Structure The test is a skeletal structure, made of hard material such as calcium carbonate, silica, chitin or composite materials. As such, it allows the protection of the internal organs and the attachment of soft flesh. In sea urchins The test of sea urchins is made of calcium carbonate, strengthened by a framework of calcite monocrystals, in a characteristic "stereomic" structure. These two ingredients provide sea urchins with a great solidity and a moderate weight, as well as the capacity to regenerate the mesh from the cuticle. According to a 2012 study, the skeletal structures of sea urchins consist of 92% of "bricks" of calcite monocrystals (conferring solidity and hardness) and 8% of a "mortar" of amorphous lime (allowing flexibility and lightness). This lime is constituted itself of 99.9% of calcium carbonate, with 0.1% structural proteins, which make sea urchins animals with an extremely mineralized skeleton (which also explains their excellent conservation as fossils). In foraminiferans The test of foraminifera, a group of single-celled organisms, is extremely evolutionarily diverse. Many different methods of constructing the test are present, from lacking a test in Reticulomyxa, proteinaceous tests in the "allogromiids", agglomerated tests made from foreign particles in many groups including textulariids, silica tests in silicoloculinids, and aragonite or calcite tests in many forms including miliolids and rotaliids. It can be of many types, including proteinaceous, agglutinated (exogenous agglomerate), porcelai
https://en.wikipedia.org/wiki/Outline%20of%20zoology
The following outline is provided as an overview of and topical guide to zoology: Zoology – study of animals. Zoology, or "animal biology", is the branch of biology that relates to the animal kingdom, including the identification, structure, embryology, evolution, classification, habits, and distribution of all animals, both living and extinct, and how they interact with their ecosystems. The term is derived from Ancient Greek word ζῷον (zōon), i.e. "animal" and λόγος, (logos), i.e. "knowledge, study". To study the variety of animals that exist (or have existed), see list of animals by common name and lists of animals. Essence of zoology Animal Fauna Branches of zoology Branches by group studied Arthropodology - study of arthropods as a whole Carcinology - the study of crustaceans Myriapodology - study of milli- and centipedes Arachnology - study of spiders and related animals such as scorpions, pseudoscorpions, and harvestmen, collectively called arachnids Acarology - study of mites and ticks Entomology - study of insects Coleopterology - study of beetles Lepidopterology - study of butterflies Melittology - study of bees Myrmecology - study of ants Orthopterology - study of grasshoppers Herpetology - study of amphibians and reptiles Batrachology - study of amphibians including frogs and toads, salamanders, newts, and caecilians Cheloniology - study of turtles and tortoises Saurology - study of lizards Serpentology - study of snakes Ichthyology - study of fish Malacology - study of mollusks Conchology - study of shells Teuthology - study of cephalopods Mammalogy - study of mammals Cetology - study of cetaceans Primatology - study of primates Ornithology - study of birds Parasitology - study of parasites, their hosts, and the relationship between them Helminthology - study of parasitic worms (helminths) Planktology - study of plankton, various small drifting plants, animals and microorganisms that inhabit bodies of water Protozoology
https://en.wikipedia.org/wiki/Complete%20theory
In mathematical logic, a theory is complete if it is consistent and for every closed formula in the theory's language, either that formula or its negation is provable. That is, for every sentence the theory contains the sentence or its negation but not both (that is, either or ). Recursively axiomatizable first-order theories that are consistent and rich enough to allow general mathematical reasoning to be formulated cannot be complete, as demonstrated by Gödel's first incompleteness theorem. This sense of complete is distinct from the notion of a complete logic, which asserts that for every theory that can be formulated in the logic, all semantically valid statements are provable theorems (for an appropriate sense of "semantically valid"). Gödel's completeness theorem is about this latter kind of completeness. Complete theories are closed under a number of conditions internally modelling the T-schema: For a set of formulas : if and only if and , For a set of formulas : if and only if or . Maximal consistent sets are a fundamental tool in the model theory of classical logic and modal logic. Their existence in a given case is usually a straightforward consequence of Zorn's lemma, based on the idea that a contradiction involves use of only finitely many premises. In the case of modal logics, the collection of maximal consistent sets extending a theory T (closed under the necessitation rule) can be given the structure of a model of T, called the canonical model. Examples Some examples of complete theories are: Presburger arithmetic Tarski's axioms for Euclidean geometry The theory of dense linear orders without endpoints The theory of algebraically closed fields of a given characteristic The theory of real closed fields Every uncountably categorical countable theory Every countably categorical countable theory A group of three elements See also Lindenbaum's lemma Łoś–Vaught test
https://en.wikipedia.org/wiki/Impedance%20parameters
Impedance parameters or Z-parameters (the elements of an impedance matrix or Z-matrix) are properties used in electrical engineering, electronic engineering, and communication systems engineering to describe the electrical behavior of linear electrical networks. They are also used to describe the small-signal (linearized) response of non-linear networks. They are members of a family of similar parameters used in electronic engineering, other examples being: S-parameters, Y-parameters, H-parameters, T-parameters or ABCD-parameters. Z-parameters are also known as open-circuit impedance parameters as they are calculated under open circuit conditions. i.e., Ix=0, where x=1,2 refer to input and output currents flowing through the ports (of a two-port network in this case) respectively. The Z-parameter matrix A Z-parameter matrix describes the behaviour of any linear electrical network that can be regarded as a black box with a number of ports. A port in this context is a pair of electrical terminals carrying equal and opposite currents into and out-of the network, and having a particular voltage between them. The Z-matrix gives no information about the behaviour of the network when the currents at any port are not balanced in this way (should this be possible), nor does it give any information about the voltage between terminals not belonging to the same port. Typically, it is intended that each external connection to the network is between the terminals of just one port, so that these limitations are appropriate. For a generic multi-port network definition, it is assumed that each of the ports is allocated an integer n ranging from 1 to N, where N is the total number of ports. For port n, the associated Z-parameter definition is in terms of the port current and port voltage, and respectively. For all ports the voltages may be defined in terms of the Z-parameter matrix and the currents by the following matrix equation: where Z is an N × N matrix the elements of wh
https://en.wikipedia.org/wiki/Symmetry%20operation
In group theory, geometry, representation theory and molecular geometry, a symmetry operation is a geometric transformation of an object that leaves the object looking the same after it has been carried out. For example, as transformations of an object in space, rotations, reflections and inversions are all symmetry operations. Such symmetry operations are performed with respect to symmetry elements (for example, a point, line or plane). In the context of molecular symmetry, a symmetry operation is a permutation of atoms such that the molecule or crystal is transformed into a state indistinguishable from the starting state. Two basic facts follow from this definition, which emphasizes its usefulness. Physical properties must be invariant with respect to symmetry operations. Symmetry operations can be collected together in groups which are isomorphic to permutation groups. In the context of molecular symmetry, quantum wavefunctions need not be invariant, because the operation can multiply them by a phase or mix states within a degenerate representation, without affecting any physical property. Molecules Identity Operation The identity operation corresponds to doing nothing to the object. Because every molecule is indistinguishable from itself if nothing is done to it, every object possesses at least the identity operation. The identity operation is denoted by or . In the identity operation, no change can be observed for the molecule. Even the most asymmetric molecule possesses the identity operation. The need for such an identity operation arises from the mathematical requirements of group theory. Reflection through mirror planes The reflection operation is carried out with respect to symmetry elements known as planes of symmetry or mirror planes. Each such plane is denoted as (sigma). Its orientation relative to the principal axis of the molecule is indicated by a subscript. The plane must pass through the molecule and cannot be completely outside it. I
https://en.wikipedia.org/wiki/Half-time%20%28music%29
In popular music, half-time is a type of meter and tempo that alters the rhythmic feel by essentially doubling the tempo resolution or metric division/level in comparison to common-time. Thus, two measures of approximate a single measure of , while a single measure of 4/4 emulates 2/2. Half-time is not to be confused with alla breve or odd time. Though notes usually get the same value relative to the tempo, the way the beats are divided is altered. While much music typically has a backbeat on quarter note (crotchet) beats two and four, half time would increase the interval between backbeats to double, thus making it hit on beats three and seven, or the third beat of each measure (count out of an 8 beat measure (bar), common practice in half time): 1 2 3 4 1 2 3 4 1 2 3 4 5 6 7 8 1 2 3 4 Essentially, a half time 'groove' is one that expands one measure over the course of two. The length of each note is doubled while its frequency is halved. Common-time Time signatures are defined by how they divide the measure (in , complex triple time, each measure is divided in three, each of which is divided into three eighth notes: 3×3=9). In "common" time, often considered , each level is divided in two (simple duple time: 2×2=4). In a common-time rock drum pattern each measure (a whole note) is divided in two by the bass drum (half note), each half is divided in two by the snare drum (quarter note, collectively the bass and snare divide the measure into four), and each quarter note is divided in two by a ride pattern (eighth note). "Half"-time refers to halving this division (divide each measure into quarter notes with the ride pattern), while "double"-time refers to doubling this division (divide each measure into sixteenth notes with the ride pattern). Half-time A classic example is the half-time shuffle, a variation of a shuffle rhythm, which is used extensively in hip-hop and some blues music. Some of the variations of
https://en.wikipedia.org/wiki/Phylloclade
Phylloclades and cladodes are flattened, photosynthetic shoots, which are usually considered to be modified branches. The two terms are used either differently or interchangeably by different authors. Phyllocladus, a genus of conifer, is named after these structures. Phylloclades/cladodes have been identified in fossils dating from as early as the Permian. Definition and morphology The term "phylloclade" is from the Neo-Latin phyllocladium, itself derived from Greek phyllo, leaf, and klados, branch. Definitions of the terms "phylloclade" and "cladode" vary. All agree that they are flattened structures that are photosynthetic and resemble leaf-like branches. In one definition, phylloclades are a subset of cladodes, namely those that greatly resemble or perform the function of leaves, as in Butcher's broom (Ruscus aculeatus) as well as Phyllanthus and some Asparagus species. By an alternative definition, cladodes are distinguished by their limited growth and that they involve only one or two internodes. By this definition, some of the most leaf-like structures are cladodes, rather than phylloclades. By that definition, Phyllanthus has phylloclades, but Ruscus and Asparagus have cladodes. Another definition uses "phylloclade" to refer a portion of a leaf-like stem or branch with multiple nodes and internodes, and "cladode" for a single internode of a phylloclade. Although phylloclades are usually interpreted as modified branches, developmental studies have shown that they are intermediate between leaves and branches as their name indicates. Molecular genetic investigations have confirmed these findings. For example, Hirayama et al. (2007) showed that the phylloclade of Ruscus aculeatus "is not homologous to either the shoot or the leaf, but that it has a double organ identity," which means that it combines shoot and leaf processes. Similar structures Aristate leaves end in a stiff point that may continue the primary leaf vein; this can resemble the stem end o
https://en.wikipedia.org/wiki/Magnetocapacitance
Magnetocapacitance is a property of some dielectric, insulating materials, and metal–insulator–metal heterostructures that exhibit a change in the value of their capacitance when an external magnetic field is applied to them. Magnetocapacitance can be an intrinsic property of some dielectric materials, such as multiferroic compounds like BiMnO3, or can be a manifest of properties extrinsic to the dielectric but present in capacitance structures like Pd, Al2O3, and Al.
https://en.wikipedia.org/wiki/Google%20Guice
Google Guice (pronounced like "juice") is an open-source software framework for the Java platform developed by Bob Lee and Kevin Bourrillion at Google and released under the Apache License. It provides support for dependency injection using annotations to configure Java objects. Dependency injection is a design pattern whose core principle is to separate behavior from dependency resolution. Guice allows implementation classes to be bound programmatically to an interface, then injected into constructors, methods or fields using an @Inject annotation. When more than one implementation of the same interface is needed, the user can create custom annotations that identify an implementation, then use that annotation when injecting it. Being the first generic framework for dependency injection using Java annotations in 2008, Guice won the 18th Jolt Award for best Library, Framework, or Component. See also Spring Framework
https://en.wikipedia.org/wiki/Remote%20graphics%20unit
A remote graphics unit (RGU) is a device that allows a computer to be separated from some input/output devices such as keyboard, mouse, speakers, and display monitors. The key part being remoted is the graphics sub-system of the computer. History RGUs may have their origin with experiments with graphics controllers on mainframe computers in the 1970s. RGUs have been mostly associated with high end workstations running Unix-like operating systems or Windows since the late 1990s. Generally RGUs are used for special applications like remote sensing, financial services commodity trading desks, computer-aided design, etc. Depending on how one chooses to define RGUs, dedicated X terminals may also be included. Application Usually the reasons that might lie behind the desire to separate the user interface of a computer from the actual computer itself would be: securing computers away from users for corporate or government security, to reduce heat and noise in rooms with many computer operators, or to facilitate computer maintenance by placing all computers in very close proximity to one another. KVM interoperability Unlike other technologies used to achieve this, such as KVM Extension (or Remote KVM) and DVI Extension for example, a remote graphics unit will effectively split a computer's PCI or PCI-Express bus and transmit only bus commands over to the user side. With KVM Extension and DVI Extension, the graphics processing is done by a traditional graphics processing unit (GPU) on the computer side. Bus data is much smaller than rendered graphics data so the theory behind the remote graphics unit is that it is possible to achieve higher resolutions and better graphics performance when there is a large separation in distance between the user-side input/output devices and the computer side. Examples An example of a product line that was commercialized using RGU as the description of the technology is the Matrox Extio series. Extio is a brand that is marketing sho
https://en.wikipedia.org/wiki/Comparison%20of%20DVR%20software%20packages
This is a comparison of digital video recorder (DVR), also known as personal video recorder (PVR), software packages. Note: this is may be considered a comparison of DVB software, not all listed packages have recording capabilities. General information Basic general information for popular DVR software packages - not all actually record. Features Information about what common and prominent DVR features are implemented natively (without third-party add-ons unless stated otherwise): Video format support Information about what video codecs are implemented natively (without third-party add-ons) in the PVRs. Information about what video codecs are implemented natively (without third-party add-ons) in the PVRs. Network support Each features is in context of computer-to-computer interaction. All features must be available after the default install otherwise the feature needs a footnote. 1 Yes with registry change 2 Yes with retail third-party plugin 3 Yes with free supported third-party plugin 4 Yes with free unsupported third-party plugin 5 Yes with free third-party software Web Guide 4 6 Yes with add-on software called DVBLink Server 7 Yes with using symlinks, or just adding folders in settings TV tuner hardware TV gateway network tuner TV servers DVRs require TV tuner cards to receive signals. Many DVRs, as seen above, can use multiple tuners. HdHomerun has CableCARD Models (HDHomeRun Prime) and OTA Models (HDHomeRun Connect) that are networked TV Tuners See also List of free television software Comparison of video player software Home cinema Home theater PC (HTPC) Digital video recorder Hard disk recorder DVD recorder Quiet PC Media server Notes External links FLOSS Media Centers Comparison Chart PVR software packages Television technology Television time shifting technology
https://en.wikipedia.org/wiki/T-MPLS
T-MPLS or Transport MPLS is a transport network layer technology that uses extensions to a subset of the existing MPLS standards and is designed specifically for application in transport networks. Work to define T-MPLS was started by the ITU-T in February 2006. It was intended specifically as a connection-oriented packet-switched (co-ps) application offering a simpler implementation by removing MPLS features that are not relevant to co-ps applications and adding mechanisms that provide support of critical transport functionality. ITU-T ceased work on T-MPLS in December 2008, in favour of MPLS-TP standardization. T-MPLS uses the same architectural principles of layered networking that are used in other technologies like SDH and OTN. Service providers have already developed management processes and work procedures based on these principles for use in networks that use those other technologies. In this way T-MPLS was intended to provide a reliable packet-based technology that is familiar and also aligned with circuit-based transport networking; it supports current organizational processes and large-scale work procedures. T-MPLS is a low cost Layer 2 technology that provides QoS, end-to-end OA&M and protection switching. The following ITU-T Recommendations exist for T-MPLS. These Recommendations will be superseded (i.e., replaced) by new revisions that apply to MPLS-TP and reference the joint work being undertaken by the ITU-T and IETF. A further set of ITU-T Recommendations were at a draft stage when work on T-MPLS was halted. These may be resumed at a later stage to reference the material developed as part of the MPLS-TP effort. After IETF raised concerns over T-MPLS technology, mainly about incompatibility with the already established IP/MPLS, the ITU-T and the IETF started a joint activity to solve potential issues. The decision was to transfer control to IETF to develop a new MPLS profile specialized for transport (MPLS-TP) with input from ITU recommendation
https://en.wikipedia.org/wiki/PC%20Tools%20%28company%29
PC Tools (founded in 2003), formerly known as WinGuides.com, was a software company acquired by Symantec in 2008; the new owner eventually discontinued the PC Tools name. Company headquarters were in Australia, with offices in Luxembourg, the United States, United Kingdom, Ireland and Ukraine. The company had previously developed and distributed security and optimization software for the Mac OS X and Microsoft Windows platforms. Products By 29 November 2006 software owned by PC Tools had been downloaded over 125 million times. PC Tools Browser Defender PC Tools Browser Defender, also called Browser Defender for short, is a browser toolbar for Internet Explorer and Mozilla Firefox browsers on Windows based computers. Browser Defender allows for safe web surfing. PC Tools iAntiVirus iAntivirus was updated in 2012 and rebranded under Symantec's Norton brand. PC Tools iAntiVirus is free antivirus software for Intel based Apple Macintosh computers running Mac OS 10.5 (Leopard) and Mac OS 10.6 (Snow Leopard) initially released in June 2008, used to detect and remove malware, spyware and malicious exploits, using both signature-based and heuristic detection. AntiVirus was criticized because it only scans for Macintosh viruses, ignoring Windows and Linux viruses. It was praised for its speed and low usage of system resources. PC Tools Internet Security PC Tools Internet Security, was the combination of the Spyware Doctor product the Firewall product and the Anti Spam product. It provided the functionality of all three stand alone products into a single seamless product. Symantec is no longer offering this product as of 18 May 2013. It was payware designed for Windows 8 (32-/64-bit), Windows 7 (32-/64-bit), Windows Vista (32-/64-bit) and Windows XP (32-bit). PC Tools Registry Mechanic PC Tools Registry Mechanic, the first software PC Tools released, scanned the Windows registry to find errors. Version 11, released on 31 October 2011, is the last one. PC Tools Spywar
https://en.wikipedia.org/wiki/Peano%20existence%20theorem
In mathematics, specifically in the study of ordinary differential equations, the Peano existence theorem, Peano theorem or Cauchy–Peano theorem, named after Giuseppe Peano and Augustin-Louis Cauchy, is a fundamental theorem which guarantees the existence of solutions to certain initial value problems. History Peano first published the theorem in 1886 with an incorrect proof. In 1890 he published a new correct proof using successive approximations. Theorem Let be an open subset of with a continuous function and a continuous, explicit first-order differential equation defined on D, then every initial value problem for f with has a local solution where is a neighbourhood of in , such that for all . The solution need not be unique: one and the same initial value may give rise to many different solutions . Proof By replacing with , with , we may assume . As is open there is a rectangle . Because is compact and is continuous, we have and by the Stone–Weierstrass theorem there exists a sequence of Lipschitz functions converging uniformly to in . Without loss of generality, we assume for all . We define Picard iterations as follows, where . , and . They are well-defined by induction: as is within the domain of . We have where is the Lipschitz constant of . Thus for maximal difference , we have a bound , and By induction, this implies the bound which tends to zero as for all . The functions are equicontinuous as for we have so by the Arzelà–Ascoli theorem they are relatively compact. In particular, for each there is a subsequence converging uniformly to a continuous function . Taking limit in we conclude that . The functions are in the closure of a relatively compact set, so they are themselves relatively compact. Thus there is a subsequence converging uniformly to a continuous function . Taking limit in we conclude that , using the fact that are equicontinuous by the Arzelà–Ascoli theorem. By the fundamental theorem of calcul
https://en.wikipedia.org/wiki/Escherichia%20coli%20O121
Escherichia coli O121 is a pathogenic serotype of Escherichia coli, associated with Shiga toxin, intestinal bleeding, and hemolytic-uremic syndrome (HUS). HUS, if left untreated, can lead to kidney failure. Most serotypes of E. coli—a widespread species of bacteria residing in the lower intestines of mammals—are beneficial or do not cause disease. Unlike other pathogenic serotypes, such as E. coli O157:H7 (also an enterohemorrhagic E. coli), little is known in detail about the public health significance of O121. Therefore, O121 is sometimes roughly classified as a type of “non-O157 Shiga toxin–producing E. coli” (non-O157 STEC). A U.S. outbreak of E. coli O121 in 2013 sickened 24 people in 15 states according to a statement released by the CDC. New York officials found the bacterium strain in an open package of Farm Rich brand chicken quesadillas from an ill person’s home; parent company Rich Products Corp. of Buffalo, New York is now recalling these and several other items and the CDC, USDA, and FDA are now investigating to find the precise source of the outbreak. In 2016, General Mills recalled 10 million pounds of wheat flour tied to an E. coli O121 outbreak. See also Escherichia coli O157:H7
https://en.wikipedia.org/wiki/Nagata%20ring
In commutative algebra, an N-1 ring is an integral domain whose integral closure in its quotient field is a finitely generated -module. It is called a Japanese ring (or an N-2 ring) if for every finite extension of its quotient field , the integral closure of in is a finitely generated -module (or equivalently a finite -algebra). A ring is called universally Japanese if every finitely generated integral domain over it is Japanese, and is called a Nagata ring, named for Masayoshi Nagata, or a pseudo-geometric ring if it is Noetherian and universally Japanese (or, which turns out to be the same, if it is Noetherian and all of its quotients by a prime ideal are N-2 rings). A ring is called geometric if it is the local ring of an algebraic variety or a completion of such a local ring , but this concept is not used much. Examples Fields and rings of polynomials or power series in finitely many indeterminates over fields are examples of Japanese rings. Another important example is a Noetherian integrally closed domain (e.g. a Dedekind domain) having a perfect field of fractions. On the other hand, a principal ideal domain or even a discrete valuation ring is not necessarily Japanese. Any quasi-excellent ring is a Nagata ring, so in particular almost all Noetherian rings that occur in algebraic geometry are Nagata rings. The first example of a Noetherian domain that is not a Nagata ring was given by . Here is an example of a discrete valuation ring that is not a Japanese ring. Choose a prime and an infinite degree field extension of a characteristic field , such that . Let the discrete valuation ring be the ring of formal power series over whose coefficients generate a finite extension of . If is any formal power series not in then the ring is not an N-1 ring (its integral closure is not a finitely generated module) so is not a Japanese ring. If is the subring of the polynomial ring in infinitely many generators generated by the squares and cubes of all
https://en.wikipedia.org/wiki/Body%20force
In physics, a body force is a force that acts throughout the volume of a body. Forces due to gravity, electric fields and magnetic fields are examples of body forces. Body forces contrast with contact forces or surface forces which are exerted to the surface of an object. Fictitious forces such as the centrifugal force, Euler force, and the Coriolis effect are other examples of body forces. Definition Qualitative A body force is simply a type of force, and so it has the same dimensions as force, [M][L][T]−2. However, it is often convenient to talk about a body force in terms of either the force per unit volume or the force per unit mass. If the force per unit volume is of interest, it is referred to as the force density throughout the system. A body force is distinct from a contact force in that the force does not require contact for transmission. Thus, common forces associated with pressure gradients and conductive and convective heat transmission are not body forces as they require contact between systems to exist. Radiation heat transfer, on the other hand, is a perfect example of a body force. More examples of common body forces include; Gravity, Electric forces acting on an object charged throughout its volume, Magnetic forces acting on currents within an object, such as the braking force that results from eddy currents, Fictitious forces (or inertial forces) can be viewed as body forces. Common inertial forces are, Centrifugal force, Coriolis force, Euler force (or transverse force), which occurs in a rotating reference frame when the rate of rotation of the frame is changing However, fictitious forces are not actually forces. Rather they are corrections to Newton's second law when it is formulated in an accelerating reference frame. (Gravity can also be considered a fictitious force in the context of General Relativity.) Quantitative The body force density is defined so that the volume integral (throughout a volume of interest) of it gives the total
https://en.wikipedia.org/wiki/Nador%20transmitter
Nador transmitter is the main transmission facility for longwave and shortwave of Medi 1 Radio, a privately owned broadcasting company of Morocco. It is situated approximately 18 kilometres south of the city of Nador and a few kilometres south of Selouane at 35°2'29"N and 2°55'7"W. The longwave transmitter of the Nador facility, which works at 171 kHz, had originally a transmission power of 2000 kilowatts. After 2009 the original equipment had been replaced by a new set of Thomson Broadcast S7HP solid-state transmitters with an output of 1600 kW. This modernization project incorporated a refurbishment of the three guyed masts, each 380 metres tall, thus the tallest structures in Africa after the demolition of the OMEGA Navigation System in Paynesville, Liberia in 2011. In addition the Nador facility is also equipped with two Thomson 250 kW shortwave transmitters. One of these transmitters carried the main program of Morocco's state broadcaster SNRT on varying frequencies around 15345 kHz until it was switched off in September 2012. The other transmitter was used for Medi 1 on 9575 kHz, unheard since May 2017 as well. Sources and references External links Entries at skyscraperpage Radio masts and towers Towers in Morocco Transmitter Buildings and structures in Oriental (Morocco) Shortwave radio stations
https://en.wikipedia.org/wiki/P-TEFb
The positive transcription elongation factor, P-TEFb, is a multiprotein complex that plays an essential role in the regulation of transcription by RNA polymerase II (Pol II) in eukaryotes. Immediately following initiation Pol II becomes trapped in promoter proximal paused positions on the majority of human genes (Figure 1). P-TEFb is a cyclin dependent kinase that can phosphorylate the DRB sensitivity inducing factor (DSIF) and negative elongation factor (NELF), as well as the carboxyl terminal domain of the large subunit of Pol II and this causes the transition into productive elongation leading to the synthesis of mRNAs. P-TEFb is regulated in part by a reversible association with the 7SK snRNP. Treatment of cells with the P-TEFb inhibitors DRB or flavopidirol leads to loss of mRNA production and ultimately cell death. Discovery, Composition and Structure P-TEFb was identified and purified as a factor needed for the generation of long run-off transcripts using an in vitro transcription system derived from Drosophila cells. It is a cyclin dependent kinase containing the catalytic subunit, Cdk9, and a regulatory subunit, cyclin T in Drosophila. In humans there are multiple forms of P-TEFb which contain Cdk9 and one of several cyclin subunits, cyclin T1, T2, and K. P-TEFb associates with other factors including the bromodomain protein BRD4, and is found associated with a large complex of proteins called the super elongation complex. Importantly, for the AIDS virus, HIV, P-TEFb is targeted by the HIV Tat protein which bypasses normal cellular P-TEFb control and directly brings P-TEFb to the promoter proximal paused polymerase in the HIV genome. The structures of human P-TEFb containing Cdk9 and cyclin T1 and the HIV Tat•P-TEFb complex have been solved using X-ray crystallography. The first structure solved demonstrated that the two subunits were arranged as has been found in other cyclin dependent kinases. Three amino acid substitutions were inadvertently introduc
https://en.wikipedia.org/wiki/Hardy%E2%80%93Littlewood%20maximal%20function
In mathematics, the Hardy–Littlewood maximal operator M is a significant non-linear operator used in real analysis and harmonic analysis. Definition The operator takes a locally integrable function f : Rd → C and returns another function Mf. For any point x ∈ Rd, the function Mf returns the maximum of a set of reals, namely the set of average values of f for all the balls B(x, r) of any radius r at x. Formally, where |E| denotes the d-dimensional Lebesgue measure of a subset E ⊂ Rd. The averages are jointly continuous in x and r, therefore the maximal function Mf, being the supremum over r > 0, is measurable. It is not obvious that Mf is finite almost everywhere. This is a corollary of the Hardy–Littlewood maximal inequality. Hardy–Littlewood maximal inequality This theorem of G. H. Hardy and J. E. Littlewood states that M is bounded as a sublinear operator from Lp(Rd) to itself for p > 1. That is, if f ∈ Lp(Rd) then the maximal function Mf is weak L1-bounded and Mf ∈ Lp(Rd). Before stating the theorem more precisely, for simplicity, let {f > t} denote the set {x | f(x) > t}. Now we have: Theorem (Weak Type Estimate). For d ≥ 1, there is a constant Cd > 0 such that for all λ > 0 and f ∈ L1(Rd), we have: With the Hardy–Littlewood maximal inequality in hand, the following strong-type estimate is an immediate consequence of the Marcinkiewicz interpolation theorem: Theorem (Strong Type Estimate). For d ≥ 1, 1 < p ≤ ∞, and f ∈ Lp(Rd), there is a constant Cp,d > 0 such that In the strong type estimate the best bounds for Cp,d are unknown. However subsequently Elias M. Stein used the Calderón-Zygmund method of rotations to prove the following: Theorem (Dimension Independence). For 1 < p ≤ ∞ one can pick Cp,d = Cp independent of d. Proof While there are several proofs of this theorem, a common one is given below: For p = ∞, the inequality is trivial (since the average of a function is no larger than its essential supremum). For 1 < p < ∞, first we shall use the f
https://en.wikipedia.org/wiki/Photoreceptor%20protein
Photoreceptor proteins are light-sensitive proteins involved in the sensing and response to light in a variety of organisms. Some examples are rhodopsin in the photoreceptor cells of the vertebrate retina, phytochrome in plants, and bacteriorhodopsin and bacteriophytochromes in some bacteria. They mediate light responses as varied as visual perception, phototropism and phototaxis, as well as responses to light-dark cycles such as circadian rhythm and other photoperiodisms including control of flowering times in plants and mating seasons in animals. Structure Photoreceptor proteins typically consist of a protein attached to a non-protein chromophore (sometimes referred as photopigment, even so photopigment may also refer to the photoreceptor as a whole). The chromophore reacts to light via photoisomerization or photoreduction, thus initiating a change of the receptor protein which triggers a signal transduction cascade. Chromophores found in photoreceptors include retinal (retinylidene proteins, for example rhodopsin in animals), flavin (flavoproteins, for example cryptochrome in plants and animals) and bilin (biliproteins, for example phytochrome in plants). The plant protein UVR8 is exceptional amongst photoreceptors in that it contains no external chromophore. Instead, UVR8 absorbs light through tryptophan residues within its protein coding sequence. Photoreceptors in animals Melanopsin: in vertebrate retina, mediates pupillary reflex, involved in regulation of circadian rhythms Photopsin: reception of various colors of light in the cone cells of vertebrate retina Rhodopsin: green-blue light reception in the rod cells of vertebrate retina Protein Kinase C: mediates photoreceptor deactivation, and retinal degeneration OPN5: sensitive to UV-light Photoreceptors in plants UVR8: UV-B light reception Cryptochrome: blue and UV-A light reception Phototropin: blue and UV-A light perception (to mediate phototropism and chloroplast movement) Zeitlupe: blue light en
https://en.wikipedia.org/wiki/Robert%20Dunnell
Robert Chester Dunnell (December 4, 1942 – December 13, 2010) was an archaeologist known for his contribution in archaeological systematics, measurement and explanation of the archaeological record, evolutionary archaeology, and the archaeology of eastern North America. Dunnell received his PhD from Yale University in 1967. He was a professor of anthropology at the University of Washington until his retirement in 1996 after which he was emeritus at the University of Washington as well as Mississippi State University. Among Dunnell's contribution to archaeology was the recognition of the role the theory of biological evolution as a means of explaining cultural phenomena. In addition, he argued that "cultural evolution" which has its roots in 19th Century social scientists such as Lewis Henry Morgan and Herbert Spencer is distinct from "scientific evolution" which Darwinian in character. Cultural evolution is vitalistic and assumes a direction to the nature of change (i.e., progress). Darwinian evolution, Dunnell argues, holds that evolution is a two-step process in which variability generation is separate from mechanisms that sort that variability. While advocating "scientific evolution" as the basis for anthropological theory, Dunnell argued that the use of a strictly biological model was insufficient to explain cultural variability. He argued that a more comprehensive version of evolutionary theory is needed that considered cultural inheritance as an additional means of the transmission of variability between individuals. Overall, Dunnell advocated the use of a Darwinian model. Dunnell's approach advocates the evolutionary model to explain (cultural) variation, while exposing the pitfalls of using analogy to explain historical events. Dunnell's geographical interests included the U.S. Southeast. Selected bibliography Dunnell, Robert C., 1978. Style and Function: A Fundamental Dichotomy. American Antiquity, 43(2), pp. 192–202. Dunnell, Robert C., 1980.
https://en.wikipedia.org/wiki/Zoidogamy
Zooidogamy is a type of plant reproduction in which male gametes (antherozoids) swim in a path of water to the female gametes (archegonium). Zoidogamy is found in algae, bryophytes, pteridophytes, and some gymnosperms (others use siphonogamy). Zoidogamy relates to evolution, as it provides a pathway from wind-borne abiotic pollination and similar mechanisms to fluid-based mechanisms used in most animals.
https://en.wikipedia.org/wiki/Site-specific%20recombination
Site-specific recombination, also known as conservative site-specific recombination, is a type of genetic recombination in which DNA strand exchange takes place between segments possessing at least a certain degree of sequence homology. Enzymes known as site-specific recombinases (SSRs) perform rearrangements of DNA segments by recognizing and binding to short, specific DNA sequences (sites), at which they cleave the DNA backbone, exchange the two DNA helices involved, and rejoin the DNA strands. In some cases the presence of a recombinase enzyme and the recombination sites is sufficient for the reaction to proceed; in other systems a number of accessory proteins and/or accessory sites are required. Many different genome modification strategies, among these recombinase-mediated cassette exchange (RMCE), an advanced approach for the targeted introduction of transcription units into predetermined genomic loci, rely on SSRs. Site-specific recombination systems are highly specific, fast, and efficient, even when faced with complex eukaryotic genomes. They are employed naturally in a variety of cellular processes, including bacterial genome replication, differentiation and pathogenesis, and movement of mobile genetic elements. For the same reasons, they present a potential basis for the development of genetic engineering tools. Recombination sites are typically between 30 and 200 nucleotides in length and consist of two motifs with a partial inverted-repeat symmetry, to which the recombinase binds, and which flank a central crossover sequence at which the recombination takes place. The pairs of sites between which the recombination occurs are usually identical, but there are exceptions (e.g. attP and attB of λ integrase). Classification: tyrosine- vs. serine- recombinases [[Image:SUrot.png|thumb|right|450px|Fig. 2. Ser-Recombinases: The (essentially irreversible) subunit-rotation pathway.Contrary to Tyr-recombinases, the four participating DNA strands are cut in sync
https://en.wikipedia.org/wiki/RIKEN%20MDGRAPE-3
MDGRAPE-3 is an ultra-high performance petascale supercomputer system developed by the Riken research institute in Japan. It is a special purpose system built for molecular dynamics simulations, especially protein structure prediction. MDGRAPE-3 consists of 201 units of 24 custom MDGRAPE-3 chips (4,824 total), plus additional dual-core Intel Xeon processors (codename "Dempsey") which serve as host machines. In June 2006 Riken announced its completion, achieving the petaFLOPS level of floating point arithmetic performance. This was more than three times faster than the 2006 version of the IBM Blue Gene/L system, which then led the TOP500 list of supercomputers at 0.28 petaFLOPS. Because it's not a general-purpose machine capable of running the LINPACK benchmarks, MDGRAPE-3 does not qualify for the TOP500 list. See also Supercomputing in Japan MDGRAPE-4
https://en.wikipedia.org/wiki/Eyespot%20apparatus
The eyespot apparatus (or stigma) is a photoreceptive organelle found in the flagellate or (motile) cells of green algae and other unicellular photosynthetic organisms such as euglenids. It allows the cells to sense light direction and intensity and respond to it, prompting the organism to either swim towards the light (positive phototaxis), or away from it (negative phototaxis). A related response ("photoshock" or photophobic response) occurs when cells are briefly exposed to high light intensity, causing the cell to stop, briefly swim backwards, then change swimming direction. Eyespot-mediated light perception helps the cells in finding an environment with optimal light conditions for photosynthesis. Eyespots are the simplest and most common "eyes" found in nature, composed of photoreceptors and areas of bright orange-red red pigment granules. Signals relayed from the eyespot photoreceptors result in alteration of the beating pattern of the flagella, generating a phototactic response. Microscopic structure Under the light microscope, eyespots appear as dark, orange-reddish spots or stigmata. They get their color from carotenoid pigments contained in bodies called pigment granules. The photoreceptors are found in the plasma membrane overlaying the pigmented bodies. The eyespot apparatus of Euglena comprises the paraflagellar body connecting the eyespot to the flagellum. In electron microscopy, the eyespot apparatus appears as a highly ordered lamellar structure formed by membranous rods in a helical arrangement. In Chlamydomonas, the eyespot is part of the chloroplast and takes on the appearance of a membranous sandwich structure. It is assembled from chloroplast membranes (outer, inner, and thylakoid membranes) and carotenoid-filled granules overlaid by plasma membrane. The stacks of granules act as a quarter-wave plate, reflecting incoming photons back to the overlying photoreceptors, while shielding the photoreceptors from light coming from other directio
https://en.wikipedia.org/wiki/Allele-specific%20oligonucleotide
An allele-specific oligonucleotide (ASO) is a short piece of synthetic DNA complementary to the sequence of a variable target DNA. It acts as a probe for the presence of the target in a Southern blot assay or, more commonly, in the simpler dot blot assay. It is a common tool used in genetic testing, forensics, and molecular biology research. An ASO is typically an oligonucleotide of 15–21 nucleotide bases in length. It is designed (and used) in a way that makes it specific for only one version, or allele, of the DNA being tested. The length of the ASO, which strand it is chosen from, and the conditions by which it is bound to (and washed from) the target DNA all play a role in its specificity. These probes can usually be designed to detect a difference of as little as 1 base in the target's genetic sequence, a basic ability in the assay of single-nucleotide polymorphisms (SNPs), important in genotype analysis and the Human Genome Project. To be detected after it has bound to its target, the ASO must be labeled with a radioactive, enzymatic, or fluorescent tag. The Illumina Methylation Assay technology takes advantage of ASO to detect one base pair difference (cytosine versus thymine) to measure methylation at a specific CpG site. Example The human disease sickle cell anemia is caused by a genetic mutation in the codon for the sixth amino acid of the blood protein beta-hemoglobin. The normal DNA sequence G-A-G codes for the amino acid glutamate, while the mutation changes the middle adenine to a thymine, leading to the sequence G-T-G (G-U-G in the mRNA). This altered sequence substitutes a valine into the final protein, distorting its structure. To test for the presence of the mutation in a DNA sample, an ASO probe would be synthesized to be complementary to the altered sequence, here labeled as "S". As a control, another ASO would be synthesized for the normal sequence "A". Each ASO is fully complementary to its target sequence (and will bind strongly), but has
https://en.wikipedia.org/wiki/Zakai%20equation
In filtering theory the Zakai equation is a linear stochastic partial differential equation for the un-normalized density of a hidden state. In contrast, the Kushner equation gives a non-linear stochastic partial differential equation for the normalized density of the hidden state. In principle either approach allows one to estimate a quantity function (the state of a dynamical system) from noisy measurements, even when the system is non-linear (thus generalizing the earlier results of Wiener and Kalman for linear systems and solving a central problem in estimation theory). The application of this approach to a specific engineering situation may be problematic however, as these equations are quite complex. The Zakai equation is a bilinear stochastic partial differential equation. It was named after Moshe Zakai. Overview Assume the state of the system evolves according to and a noisy measurement of the system state is available: where are independent Wiener processes. Then the unnormalized conditional probability density of the state at time t is given by the Zakai equation: where the operator As previously mentioned, is an unnormalized density and thus does not necessarily integrate to 1. After solving for , integration and normalization can be done if desired (an extra step not required in the Kushner approach). Note that if the last term on the right hand side is omitted (by choosing h identically zero), the result is a nonstochastic PDE: the familiar Fokker–Planck equation, which describes the evolution of the state when no measurement information is available. See also Kushner equation Kalman filter Wiener filter
https://en.wikipedia.org/wiki/The%20Billion-Dollar%20Molecule
The Billion-Dollar Molecule is a book by journalist Barry Werth about the founding and early research efforts of the American biotechnology company Vertex Pharmaceuticals, which was founded in 1989 by Joshua Boger and was among the first biotechnology companies to adopt an explicit strategy of rational drug design as opposed to techniques based on combinatorial chemistry. This book is notable as an inside look at a biotechnology company, and the stresses and marketing pressures on funding research into drug design. This book is a mixture of finance and technology. In February, 2014, Barry Werth published a follow-on book, The Antidote, that looks at Vertex 20 years later after his original effort.
https://en.wikipedia.org/wiki/Invitation%20system
An invitation system is a method of encouraging people to join an organization, such as a club or a website. In regular society, it refers to any system whereby new members are chosen; they cannot simply apply. In relation to websites and other technology-related organisations, the term refers to a more specific situation whereby invitations are sent, but there is never any approval needed from other members. Popular alternatives to this specific version are open registration and closed registration. Open registration is where any user can freely join. Closed registration involves an existing member recommending a new member and approval is sought amongst existing members. The basis of the invitation system is that a member can grant approval to a new user without having to consult any other members. Existing members may receive a set number of invitations (sometimes in the form of tokens) to allow others to join the service. Those invited to a website are typically sent either a specialized URL or a single-use pass code. Applications Invitation systems for websites are usually temporary. They are typically used for services in private beta testing, in order to control the number of users on the service. In other cases, they can be used due to limited availability of server resources. There are a growing number of sites which use invitation systems on a permanent basis to create exclusivity and to control quality of user-generated content. Rarely, they may be used on a permanent basis in order to aggregate social network statistics (all users will ultimately have a traceable connection to all others). They are sometimes used to avoid abusive types or spammers, by relying on mutual trust between all members. Variations Sometimes, a tiered invitation system may be in place, wherein those higher up the hierarchy will have the power to grant more invitations, whereas low-ranking members may receive few or even no invitation rights. Examples Some prominent se
https://en.wikipedia.org/wiki/Newton%27s%20inequalities
In mathematics, the Newton inequalities are named after Isaac Newton. Suppose a1, a2, ..., an are real numbers and let denote the kth elementary symmetric polynomial in a1, a2, ..., an. Then the elementary symmetric means, given by satisfy the inequality If all the numbers ai are non-zero, then equality holds if and only if all the numbers ai are equal. It can be seen that S1 is the arithmetic mean, and Sn is the n-th power of the geometric mean. See also Maclaurin's inequality
https://en.wikipedia.org/wiki/Vitali%20covering%20lemma
In mathematics, the Vitali covering lemma is a combinatorial and geometric result commonly used in measure theory of Euclidean spaces. This lemma is an intermediate step, of independent interest, in the proof of the Vitali covering theorem. The covering theorem is credited to the Italian mathematician Giuseppe Vitali. The theorem states that it is possible to cover, up to a Lebesgue-negligible set, a given subset E of Rd by a disjoint family extracted from a Vitali covering of E. Vitali covering lemma There are two basic version of the lemma, a finite version and an infinite version. Both lemmas can be proved in the general setting of a metric space, typically these results are applied to the special case of the Euclidean space . In both theorems we will use the following notation: if is a ball and , we will write for the ball . Finite version Theorem (Finite Covering Lemma). Let be any finite collection of balls contained in an arbitrary metric space. Then there exists a subcollection of these balls which are disjoint and satisfy Proof: Without loss of generality, we assume that the collection of balls is not empty; that is, n > 0. Let be the ball of largest radius. Inductively, assume that have been chosen. If there is some ball in that is disjoint from , let be such ball with maximal radius (breaking ties arbitrarily), otherwise, we set m := k and terminate the inductive definition. Now set . It remains to show that for every . This is clear if . Otherwise, there necessarily is some such that intersects and the radius of is at least as large as that of . The triangle inequality then easily implies that , as needed. This completes the proof of the finite version. Infinite version Theorem (Infinite Covering Lemma). Let be an arbitrary collection of balls in a separable metric space such that where denotes the radius of the ball B. Then there exists a countable sub-collection such that the balls of are pairwise disjoint, and satisfyAnd mor
https://en.wikipedia.org/wiki/Duress%20code
A duress code is a covert distress signal used by an individual who is being coerced by one or more hostile persons. It is used to warn others that they are being forced to do something against their will. Typically, the warning is given via some innocuous signal embedded in normal communication, such as a code-word or phrase spoken during conversation to alert other personnel. Alternatively, the signal may be incorporated into the authentication process itself, typically in the form of a panic password, distress password, or duress PIN that is distinct from the user's normal password or PIN. These concepts are related to a panic alarm and often achieve the same outcome. Civilian usage Some home and property alarm systems have duress PINs, where the last two digits of the reset code are switched around. Entering the code when under duress from an assailant can trigger a silent alarm, alerting police or security personnel in a covert manner. The implementation of this feature has not been without controversy, as it has been claimed to lead to false alarms. A similar mechanism, SafetyPIN, has been proposed for use in ATMs. In 2010, the Federal Trade Commission issued a report studying the viability of such mechanisms for ATMs. They noted duress PINs have never been actually implemented in any ATM, and conclude that the costs of deployment outweighs the likelihood they will actually deter criminal activity. When a duress PIN is used to trigger a silent alarm, an adversary can always request the PIN in advance and ensure the appropriately modified PIN is entered instead. If the adversary does not know which PIN is correct, they may choose randomly between the two possible codes allowing them to succeed half of the time. In scenarios where a panic password is used to limit access control, instead of triggering an alarm, it is insufficient to have a single panic password. If the adversary knows the system, a common assumption, then they will simply force the user to a
https://en.wikipedia.org/wiki/Kuratowski%27s%20closure-complement%20problem
In point-set topology, Kuratowski's closure-complement problem asks for the largest number of distinct sets obtainable by repeatedly applying the set operations of closure and complement to a given starting subset of a topological space. The answer is 14. This result was first published by Kazimierz Kuratowski in 1922. It gained additional exposure in Kuratowski's fundamental monograph Topologie (first published in French in 1933; the first English translation appeared in 1966) before achieving fame as a textbook exercise in John L. Kelley's 1955 classic, General Topology. Proof Letting denote an arbitrary subset of a topological space, write for the closure of , and for the complement of . The following three identities imply that no more than 14 distinct sets are obtainable: . (The closure operation is idempotent.) . (The complement operation is an involution.) . (Or equivalently , using identity (2)). The first two are trivial. The third follows from the identity where is the interior of which is equal to the complement of the closure of the complement of , . (The operation is idempotent.) A subset realizing the maximum of 14 is called a 14-set. The space of real numbers under the usual topology contains 14-sets. Here is one example: where denotes an open interval and denotes a closed interval. Let denote this set. Then the following 14 sets are accessible: , the set shown above. Further results Despite its origin within the context of a topological space, Kuratowski's closure-complement problem is actually more algebraic than topological. A surprising abundance of closely related problems and results have appeared since 1960, many of which have little or nothing to do with point-set topology. The closure-complement operations yield a monoid that can be used to classify topological spaces.
https://en.wikipedia.org/wiki/Sylvester%20equation
In mathematics, in the field of control theory, a Sylvester equation is a matrix equation of the form: It is named after English mathematician James Joseph Sylvester. Then given matrices A, B, and C, the problem is to find the possible matrices X that obey this equation. All matrices are assumed to have coefficients in the complex numbers. For the equation to make sense, the matrices must have appropriate sizes, for example they could all be square matrices of the same size. But more generally, A and B must be square matrices of sizes n and m respectively, and then X and C both have n rows and m columns. A Sylvester equation has a unique solution for X exactly when there are no common eigenvalues of A and −B. More generally, the equation AX + XB = C has been considered as an equation of bounded operators on a (possibly infinite-dimensional) Banach space. In this case, the condition for the uniqueness of a solution X is almost the same: There exists a unique solution X exactly when the spectra of A and −B are disjoint. Existence and uniqueness of the solutions Using the Kronecker product notation and the vectorization operator , we can rewrite Sylvester's equation in the form where is of dimension , is of dimension , of dimension and is the identity matrix. In this form, the equation can be seen as a linear system of dimension . Theorem. Given matrices and , the Sylvester equation has a unique solution for any if and only if and do not share any eigenvalue. Proof. The equation is a linear system with unknowns and the same number of equations. Hence it is uniquely solvable for any given if and only if the homogeneous equation admits only the trivial solution . (i) Assume that and do not share any eigenvalue. Let be a solution to the abovementioned homogeneous equation. Then , which can be lifted to for each by mathematical induction. Consequently, for any polynomial . In particular, let be the characteristic polynomial of . Then d
https://en.wikipedia.org/wiki/CobraNet
CobraNet is a combination of software, hardware, and network protocols designed to deliver uncompressed, multi-channel, low-latency digital audio over a standard Ethernet network. Developed in the 1990s, CobraNet is widely regarded as the first commercially successful audio-over-Ethernet implementation. CobraNet was designed for and is primarily used in large commercial audio installations such as convention centers, stadiums, airports, theme parks, and concert halls. It has applications where a large number of audio channels must be transmitted over long distances or to multiple locations. CobraNet is an alternative to analog audio, which suffers from signal degradation over long cable runs due to electromagnetic interference, high-frequency attenuation, and voltage drop. Additionally, the use of digital multiplexing allows audio to be transmitted using less cabling than analog audio. History CobraNet was developed in 1996 by Boulder, Colorado-based Peak Audio. Initial demonstrations were of a 10 Mbit/s point-to-point system with limited channel capacity. The first permanent installation of CobraNet in this early form was to provide background music throughout Disney's Animal Kingdom theme park. The first commercial use of CobraNet was during the halftime show at Super Bowl XXXI in 1997. CobraNet was first introduced as an interoperable standard in collaboration with manufacturer QSC Audio Products. QSC was the first to license the technology from Peak Audio and marketed it under the RAVE brand. At this point CobraNet had graduated to fast Ethernet and used a unique collision avoidance technique to carry up to 64 channels per Ethernet collision domain. CobraNet was subsequently enhanced to support and eventually require a switched Ethernet network. An SNMP agent was added for remote control and monitoring. Support for higher sample rates, increased bit resolutions and lowered latency capabilities were later introduced in an incremental and backward-compatib
https://en.wikipedia.org/wiki/Sum%20%28Unix%29
is a legacy utility available on some Unix and Unix-like operating systems. This utility outputs a 16-bit checksum of each argument file, as well as the number of blocks they take on disk. Two different checksum algorithms are in use. POSIX abandoned sum in favor of cksum. Overview The program is generally only useful for historical interest. It is not part of POSIX. Two algorithms are typically available: a BSD checksum and a SYSV checksum. Both are weaker than the already weak 32-bit CRC used by cksum. The default algorithm on FreeBSD and GNU implementations is the BSD checksum. Switching between the two algorithms is done via command line options. The two commonly used algorithms are as follows. The BSD sum, -r in GNU sum and -o1 in FreeBSD cksum: Initialize checksum to 0 For each byte of the input stream Perform 16-bit bitwise right rotation by 1 bit on the checksum Add the byte to the checksum, and apply modulo 2 ^ 16 to the result, thereby keeping it within 16 bits The result is a 16-bit checksum The above algorithm appeared in Seventh Edition Unix. The System V sum, -s in GNU sum and -o2 in FreeBSD cksum: checksum0 = sum of all bytes of the input stream modulo 2 ^ 32 checksum1 = checksum0 modulo 2 ^ 16 + checksum0 / 2 ^ 16 checksum = checksum1 modulo 2 ^16 + checksum1 / 2 ^ 16 The result is a 16-bit checksum calculated from the initial 32-bit plain byte sum Syntax The utility is invoked from the command line according to the following syntax: sum [OPTION]... [FILE]... with the possible option parameters being: use BSD checksum algorithm, use 1K blocks (defeats ) , use SYSV checksum algorithm, use 512 bytes blocks display the help screen and exit output version information and exit When no file parameter is given, or when FILE is , the standard input is used as input file. Example of use: $ echo Hello > testfile $ sum testfile 36978 1 Example of -s use in GNU sum: $ echo Hello > testfile $ sum -s testfile 510 1 testfile E
https://en.wikipedia.org/wiki/IBM%20JX
The IBM JX (or JXPC) was a personal computer released in 1984 into the Japanese, Australian and New Zealand markets. Designed in Japan, it was based on the technology of the IBM PCjr and was designated the IBM 5511. It was targeted in the Australasian market towards the public education sector rather than at consumers, and was sold in three levels: JX (64 KiB), JX2 (128 KiB) and JX3 (256 KiB). Upgrades were available to both 384 KiB and 512 KiB. The JX was the first IBM PC to use 3.5" floppy drives. IBM Japan expected to sell 200,000 units of JX, but only 40,000 units were produced. The JX was discontinued in 1987, and IBM Japan gave 15,000 units of JX to its employees in honor of the company's 50th anniversary. General The IBM JX's main difference from the PCjr was a professional keyboard (rather than the PCjr's disparaged chiclet keyboard), dual 3.5" floppy drives, as well as options for a 5.25" floppy drive and a hard drive, both of which sat atop the main unit. The JX did not support PCjr-like "sidecar" add-ons for hardware expansion. In common with the PCjr, however, it had no DMA controller. It also supported the otherwise unique-in-the-IBM-PC-world ECGA (Enhanced Color Graphics Adapter—16 simultaneous colors, but only at 320×200 resolution) and the PCjr's 4-channel sound. Support for these two features was utilised by only a handful of software developers—Sierra On-line being the most well-known. Configuration It had several innovative features: Single or twin 3.5" 720 KB (initially only 360 KB) diskette drives Wireless infra-red keyboard 16-color video output Stackable expansion Joystick ports Cartridge slots In Japan, both white and dark gray units were available, but elsewhere all IBM JXs were dark gray—very unusual in the days of the standard color of IBM "beige boxes". All models sold in Japan have a Japanese font stored on 128 KB of ROM, but the basic system only has the capability to display 40×11 Japanese text. The Extended Display Cartridg
https://en.wikipedia.org/wiki/Global%20Ocean%20Sampling%20Expedition
The Global Ocean Sampling Expedition (GOS) is an ocean exploration genome project whose goal is to assess genetic diversity in marine microbial communities and to understand their role in nature's fundamental processes. It was begun as a Sargasso Sea pilot sampling project in August 2003; Craig Venter announced the full expedition on 4 March 2004. The two-year journey, which used Craig Venter's personal yacht, originated in Halifax, Canada, circumnavigated the globe and terminated in the U.S. in January 2006. The expedition sampled water from Halifax, Nova Scotia to the Eastern Tropical Pacific Ocean. During 2007, sampling continued along the west coast of North America. Data analysis The GOS datasets were submitted to both NCBI and Community Cyberinfrastructure for Advanced Marine Microbial Ecology Research and Analysis (CAMERA), a new online resource for marine metagenomics funded by the Gordon and Betty Moore Foundation, developed by JCVI and hosted by UC San Diego's Division of the California Institute for Telecommunications and Information Technology (Calit2). CAMERA's toolset was developed by JCVI, and reflects the tools used in the initial publication of the GOS datasets. Funding The Sorcerer II effort has been funded by: the Gordon and Betty Moore Foundation (sequencing and analysis) the United States Department of Energy, Office of Science (sequencing and analysis) The J. Craig Venter Institute (vessel operation) Moore Foundation seven-year, $24.5 million, grant (CAMERA) Vessel Sorcerer II, a 95-foot sloop, completed a 2-year scientific expedition circumnavigating the globe in mid latitudes collecting samples of microbes in seawater for genetic sequencing and cataloguing. She was designed to be not just a world cruising yacht, but one that would be capable of handling the extremes in latitudes, from equatorial heat and humidity to latitudes between 60 and 70 degrees. SORCERER II's construction is light for performance, but very strong, with her kevl
https://en.wikipedia.org/wiki/Orr%E2%80%93Sommerfeld%20equation
The Orr–Sommerfeld equation, in fluid dynamics, is an eigenvalue equation describing the linear two-dimensional modes of disturbance to a viscous parallel flow. The solution to the Navier–Stokes equations for a parallel, laminar flow can become unstable if certain conditions on the flow are satisfied, and the Orr–Sommerfeld equation determines precisely what the conditions for hydrodynamic stability are. The equation is named after William McFadden Orr and Arnold Sommerfeld, who derived it at the beginning of the 20th century. Formulation The equation is derived by solving a linearized version of the Navier–Stokes equation for the perturbation velocity field , where is the unperturbed or basic flow. The perturbation velocity has the wave-like solution (real part understood). Using this knowledge, and the streamfunction representation for the flow, the following dimensional form of the Orr–Sommerfeld equation is obtained: , where is the dynamic viscosity of the fluid, is its density, and is the potential or stream function. In the case of zero viscosity (), the equation reduces to Rayleigh's equation. The equation can be written in non-dimensional form by measuring velocities according to a scale set by some characteristic velocity , and by measuring lengths according to channel depth . Then the equation takes the form , where is the Reynolds number of the base flow. The relevant boundary conditions are the no-slip boundary conditions at the channel top and bottom and , at and in the case where is the potential function. Or: at and in the case where is the stream function. The eigenvalue parameter of the problem is and the eigenvector is . If the imaginary part of the wave speed is positive, then the base flow is unstable, and the small perturbation introduced to the system is amplified in time. Solutions For all but the simplest of velocity profiles , numerical or asymptotic methods are required to calculate solutions. Some
https://en.wikipedia.org/wiki/Cophenetic
In the clustering of biological information such as data from microarray experiments, the cophenetic similarity or cophenetic distance of two objects is a measure of how similar those two objects have to be in order to be grouped into the same cluster. The cophenetic distance between two objects is the height of the dendrogram where the two branches that include the two objects merge into a single branch. Outside the context of a dendrogram, it is the distance between the largest two clusters that contain the two objects individually when they are merged into a single cluster that contains both. See also Cophenetic correlation
https://en.wikipedia.org/wiki/Double%20inverted%20pendulum
A double inverted pendulum is the combination of the inverted pendulum and the double pendulum. The double inverted pendulum is unstable, meaning that it will fall down unless it is controlled in some way. The two main methods of controlling a double inverted pendulum are moving the base, as with the inverted pendulum, or by applying a torque at the pivot point between the two pendulums. See also Inverted pendulum Inertia wheel pendulum Furuta pendulum Tuned mass damper
https://en.wikipedia.org/wiki/Alois%20Riedler
Alois Riedler (May 15, 1850 - October 25, 1936) was a noted Austrian mechanical engineer, and, as professor in Germany, a vigorous proponent of practically-oriented engineering education. Riedler was born in Graz, Austria, and studied mechanical engineering at the Technische Hochschule (TH) Graz from 1866-1871. After graduation he took on a succession of academic appointments. He first became an assistant at the TH Brünn (1871-1873); then in 1873 moved to the TH Vienna, first as an assistant, then from 1875 onwards as a designer of machines. From 1880 to 1883, Riedler worked as associate professor at the TH Munich. In 1883 he became full professor at the TH Aachen. In 1888 he joined the TH Berlin as Professor for Mechanical Engineering, where he remained until retirement in 1920. From 1899 to 1900, he was appointed the school's principal (rector) and led discussions on how to celebrate its 100th anniversary. As a result, Riedler and Adolf Slaby (1849–1913) convinced Kaiser Wilhelm II (1859–1941) to allow Prussian technical universities to award doctorates. Although the government did not immediately consent, this effort led eventually to the school's reconstitution as today's Technical University of Berlin. Riedler first received international recognition for his reports on the Philadelphia Centennial Exposition (1876) and Paris Exposition Universelle (1878). He was later widely known for his efficient, high-speed pumps widely adopted in waterworks and in draining mines. Riedler was also known for his 1896 book "Das Maschinen-Zeichnen", (Machine Drawing) which introduced modern technical drawing. Riedler was actively involved in the early development of internal combustion engines, both for gasoline and diesel fuel. In 1903 he established the Laboratory for Internal Combustion Engines at the TH Berlin, expanded in 1907 to include investigations of motor vehicles. As laboratory director, Riedler designed a pioneering roller test stand. He also received what was
https://en.wikipedia.org/wiki/Reynolds%20analogy
The Reynolds Analogy is popularly known to relate turbulent momentum and heat transfer. That is because in a turbulent flow (in a pipe or in a boundary layer) the transport of momentum and the transport of heat largely depends on the same turbulent eddies: the velocity and the temperature profiles have the same shape. The main assumption is that heat flux q/A in a turbulent system is analogous to momentum flux τ, which suggests that the ratio τ/(q/A) must be constant for all radial positions. The complete Reynolds analogy* is: Experimental data for gas streams agree approximately with above equation if the Schmidt and Prandtl numbers are near 1.0 and only skin friction is present in flow past a flat plate or inside a pipe. When liquids are present and/or form drag is present, the analogy is conventionally known to be invalid. In 2008, the qualitative form of validity of Reynolds' analogy was re-visited for laminar flow of incompressible fluid with variable dynamic viscosity (μ). It was shown that the inverse dependence of Reynolds number (Re) and skin friction coefficient(cf) is the basis for validity of the Reynolds’ analogy, in laminar convective flows with constant & variable μ. For μ = const. it reduces to the popular form of Stanton number (St) increasing with increasing Re, whereas for variable μ it reduces to St increasing with decreasing Re. Consequently, the Chilton-Colburn analogy of St•Pr2/3 increasing with increasing cf is qualitatively valid whenever the Reynolds’ analogy is valid. Further, the validity of the Reynolds’ analogy is linked to the applicability of Prigogine's Theorem of Minimum Entropy Production. Thus, Reynolds' analogy is valid for flows that are close to developed, for whom, changes in the gradients of field variables (velocity & temperature) along the flow are small. See also Reynolds number Chilton and Colburn J-factor analogy
https://en.wikipedia.org/wiki/Courant%20algebroid
In a field of mathematics known as differential geometry, a Courant geometry was originally introduced by Zhang-Ju Liu, Alan Weinstein and Ping Xu in their investigation of doubles of Lie bialgebroids in 1997. Liu, Weinstein and Xu named it after Courant, who had implicitly devised earlier in 1990 the standard prototype of Courant algebroid through his discovery of a skew symmetric bracket on , called Courant bracket today, which fails to satisfy the Jacobi identity. Both this standard example and the double of a Lie bialgebra are special instances of Courant algebroids. Definition A Courant algebroid consists of the data a vector bundle with a bracket , a non degenerate fiber-wise inner product , and a bundle map subject to the following axioms, where are sections of E and f is a smooth function on the base manifold M. D is the combination with d the de Rham differential, the dual map of , and κ the map from E to induced by the inner product. Skew-Symmetric Definition An alternative definition can be given to make the bracket skew-symmetric as This no longer satisfies the Jacobi-identity axiom above. It instead fulfills a homotopic Jacobi-identity. where T is The Leibniz rule and the invariance of the scalar product become modified by the relation and the violation of skew-symmetry gets replaced by the axiom The skew-symmetric bracket together with the derivation D and the Jacobiator T form a strongly homotopic Lie algebra. Properties The bracket is not skew-symmetric as one can see from the third axiom. Instead it fulfills a certain Jacobi-identity (first axiom) and a Leibniz rule (second axiom). From these two axioms one can derive that the anchor map ρ is a morphism of brackets: The fourth rule is an invariance of the inner product under the bracket. Polarization leads to Examples An example of the Courant algebroid is the Dorfman bracket on the direct sum with a twist introduced by Ševera, (1998) defined as: where X,Y are vector
https://en.wikipedia.org/wiki/Chilton%20and%20Colburn%20J-factor%20analogy
Chilton–Colburn J-factor analogy (also known as the modified Reynolds analogy) is a successful and widely used analogy between heat, momentum, and mass transfer. The basic mechanisms and mathematics of heat, mass, and momentum transport are essentially the same. Among many analogies (like Reynolds analogy, Prandtl–Taylor analogy) developed to directly relate heat transfer coefficients, mass transfer coefficients, and friction factors Chilton and Colburn J-factor analogy proved to be the most accurate. It is written as follows, This equation permits the prediction of an unknown transfer coefficient when one of the other coefficients is known. The analogy is valid for fully developed turbulent flow in conduits with Re > 10000, 0.7 < Pr < 160, and tubes where L/d > 60 (the same constraints as the Sieder–Tate correlation). The wider range of data can be correlated by Friend–Metzner analogy. Relationship between Heat and Mass; See also Reynolds analogy Thomas H. Chilton
https://en.wikipedia.org/wiki/Miss%20Jerry
Miss Jerry is an 1894 American feature-length black-and-white silent pre-film "Picture Play" written and produced by Alexander Black and starring Blanche Bayliss. Miss Jerry was not a film, but a series of posed magic lantern slides projected onto a screen with a dissolving stereopticon, accompanied by narration and music, making it the first example of a feature-length dramatic fiction on screen. Miss Jerry debuted on October 9, 1894 at the Carbon Studio in New York City. It has been described as "the first picture play" and while other early film and peep-show animations produced at this time were short documentaries, Miss Jerry sought to develop what is arguably the first feature of moving pictures. This photoplay attempts to create an impression of movement with the slides changing once every 15 seconds. The Picture Play "In Miss Jerry my purpose has been to test experimentally, in a quiet story, certain possibilities of illusion, with this aim always before me, that the illusion should not, because it need not and could not safely, be that of photographs from an acted play, nor of artistic illustration, but the illusion of reality'." Aware of the progress made by Eadweard Muybridge and other photographers towards the illusion of motion, Black instead set out to present a convincing narrative story in front of an audience, using still photography to present fiction in a convincing way, rather than a perfect illusion of motion. In his 1926 history of the movies, A Million and One Nights the author Terry Ramsaye says: While the motion picture was progressing with mincing steps in the peep show Edison Kinetoscope the sheer force of the evolution of expression presented the world with an interesting paradox – the birth of the photoplay upon the screen. . . Black arrived at a rate of four slides a minute for his presentation. The plan was to make the pictures successively blend into one another in the dissolving stereopticon, avoiding an optical ‘jar’ as much as
https://en.wikipedia.org/wiki/RM%20Nimbus
RM Nimbus was a range of personal computers from British company Research Machines (now RM Education) sold from 1985 until the early 1990s, after which the designation Nimbus was discontinued. The first of these computers, the RM Nimbus PC-186, was not IBM PC compatible, but its successors the PC-286 and PC-386 were. RM computers were predominantly sold to schools and colleges in the United Kingdom for use as LAN workstations in classrooms. Models PC-186 The RM Nimbus PC-186 was a 16-bit microcomputer introduced in 1985. It is one of a small number of computers based on the Intel 80186 processor, a version of the Intel 8086 (as used by the IBM PC) originally intended as a processor for embedded systems. It ran MS-DOS 3.1 but was not IBM PC compatible. The PC-186 could run Windows versions up to and including Windows 3.0, but only in real mode, as protected mode was only available on 286 or higher processors. Most PC-186 systems were used as workstations within a local area network and were supplied without a hard disk. The operating system was started from a floppy disk or via a remote boot ROM on its network interface card, connecting to the LAN's fileserver. Stand-alone workstations were rarer but available with either twin floppy drives or a 20Mbyte hard drive and single floppy drive. I/O Connectors Mouse - 9 pin D-Sub connector, sometimes confused as a serial port, it uses quadrature signalling Keyboard - Earlier models - Proprietary keyboard port, Later models - PS/2 keyboard port Video - 5 pin DIN plug with TTL RGB output compatible with BBC micro monitor connectors - modified CGA output Parallel - Optional via expansion card - usually with BBC micro 'user port' alongside- both with ribbon pin connectors Network -BNC connector - Zilog Z-NET, Ethernet expansion cards were available for later models - both allowed network boot Piconet - a proprietary serial interface for connecting peripherals Power - Standard IEC connector in and out for monitor supply X
https://en.wikipedia.org/wiki/Simple%20transposon
A simple transposon also called "conservative transposon" is an insertion sequence (IS element) that contains its own coding transposase between the short, inverted, repeated sequences that flank (present) its gene coding region. Transposase is responsible for excision and transfer while resolvase is responsible for resolution of the transfer. Simple transposition is also called cut-and-paste transposition because the element is cut out of its original site and pasted into a new one.
https://en.wikipedia.org/wiki/Composite%20transposon
A composite transposon is similar in function to simple transposons and insertion sequence (IS) elements in that it has protein coding DNA segments flanked by inverted, repeated sequences that can be recognized by transposase enzymes. A composite transposon, however, is flanked by two separate IS elements which may or may not be exact replicas. Instead of each IS element moving separately, the entire length of DNA spanning from one IS element to the other is transposed as one complete unit. Composite transposons will also often carry one or more genes conferring antibiotic resistance. Flanked by SINEs in mammalian genomes Two SINEs may act in concert to flank and mobilize an intervening single copy DNA sequence. This was reported for a 710 bp DNA sequence upstream of the bovine beta globin gene. The DNA arrangement forms a composite transposon whose presence has been confirmed by the complete bovine genomic sequence where the mobilized sequence may be found on bovine chromosome 15 in contig NW_001493315.1 nucleotides #1085432–1086142 and the originating sequence may be found on bovine chromosome 2 in contig NW_001501789.2 nucleotides #1096679–1097389. It is likely that similar composite transposons exist in other bovine genomic regions and other mammalian genomes. They could be detected with suitable algorithms. See also Tn10
https://en.wikipedia.org/wiki/Sporeling
A sporeling is a young plant or fungus produced by a germinated spore, similar to a seedling derived from a germinated seed. They occur in algae, fungi, lichens, bryophytes and seedless vascular plants. Sporeling development Most spores germinate by first producing a germ-rhizoid or holdfast followed by a germ tube emerging from the opposite end. The germ tube develops into the hypha, protonema or thallus of the gametophyte. In seedless vascular plants such as ferns and lycopodiophyta, the term "sporeling" refers to the young sporophyte growing on the gametophyte. These sporelings develop via an embryo stage from a fertilized egg inside an archegonium and depend on the gametophyte for their early stages of growth before becoming independent sporophytes. Young fern sporelings can often be found with the prothallus gametophyte still attached at the base of their fronds. See also Conidium (mitospore) Sporogenesis External links British Pteridological Society: An introduction to ferns (contains a picture of a sporeling fern attached to the prothallus) Plant morphology Plant reproduction Fungal morphology and anatomy
https://en.wikipedia.org/wiki/Transforming%20growth%20factor%20beta%20superfamily
The transforming growth factor beta (TGF-β) superfamily is a large group of structurally related cell regulatory proteins that was named after its first member, TGF-β1, originally described in 1983. They interact with TGF-beta receptors. Many proteins have since been described as members of the TGF-β superfamily in a variety of species, including invertebrates as well as vertebrates and categorized into 23 distinct gene types that fall into four major subfamilies: The TGF-β subfamily The bone morphogenetic proteins and the growth differentiation factors The activin and inhibin subfamilies The left-right determination factors A group encompassing various divergent members Transforming growth factor-beta (TGF-beta) is a multifunctional peptide that controls proliferation, differentiation and other functions in many cell types. TGF-beta-1 is a peptide of 112 amino acid residues derived by proteolytic cleavage from the C-terminal of a precursor protein. These proteins interact with a conserved family of cell surface serine/threonine-specific protein kinase receptors, and generate intracellular signals using a conserved family of proteins called SMADs. They play fundamental roles in the regulation of basic biological processes such as growth, development, tissue homeostasis and regulation of the immune system. Structure Proteins from the TGF-beta superfamily are only active as homo- or heterodimer; the two chains being linked by a single disulfide bond. From X-ray studies of TGF-beta-2, it is known that all the other cysteines are involved in intrachain disulfide bonds. As shown in the following schematic representation, there are four disulfide bonds in the TGF-beta's and in inhibin beta chains, while the other members of this superfamily lack the first bond. interchain | +------------------------------------------|+ |
https://en.wikipedia.org/wiki/Radio%20%28magazine%29
Radio magazine, a radio broadcasting trade publication, covers the technology side of radio broadcasting. The publication is targeted at radio broadcast engineers, technology managers and owners of radio stations, networks, and recording studios. It is owned by Future US. History Radio magazine was first published in 1994 under the title BE Radio. It is, essentially, a spin-off of Broadcast Engineering magazine, which began publication in 1959. Prior to 1994, Broadcast Engineering (often known as "BE") covered radio, television, and cable broadcasting. In 1993, the editors of Broadcast Engineering, recognizing a growing divergence in the technical issues faced by radio and television broadcasting, chose to split the scope of the original magazine. This resulted in the creation of BE Radio, to cover only the radio broadcasting industry, and narrowed the scope of Broadcast Engineering to television and cable broadcasting only. For the first year of publication, BE Radio was a supplement mailed with Broadcast Engineering to subscribers working at radio stations. In 1995, BE Radio was mailed as a stand-alone publication. By 1997, the editorial oversight from Broadcast Engineering had ceased, and the tagline "From the editors of Broadcast Engineering" was no longer used. The two publications were completely stand-alone under one publisher. In time, the common lineage of the two magazines was less important, and the connection was not understood by many readers. With that, the "BE" initialism ceased to have relevance for BE Radio, and in 2002 the decision was made to drop the initials from the title, leading to the title of Radio. The March through August 2002 issues of Radio magazine have a small "BE" in the upper corner. This is because the postal permit was registered for BE Radio and not Radio, and the small addition to the cover satisfied postal regulations. In February 2011, Radio was transferred from Penton Media to media and communications publisher NewBay
https://en.wikipedia.org/wiki/Meromelia
Meromelia is a birth defect characterized by the lacking of a part, but not all, of one or more limbs with the presence of a hand or foot. It results in a shrunken and deformed extremity. Cause Such defects are mainly the results of genetic disorders, but some teratogenic (or environmental) factors have been identified, such as the use of thalidomide from 1957 to 1962 for morning sickness (NVP). Diagnosis Meromelia is a birth defect characterized by lacking part of at least one free limb. Treatment Treatment may include prosthetic limbs, surgery, rehabilitation, or LASIK. Etymology The word meromelia comes from the Greek 'part, partial' + 'limb'. See also Amelia (birth defect) Phocomelia Polymelia Amniotic band syndrome
https://en.wikipedia.org/wiki/Quantum%20radar
Quantum radar is a speculative remote-sensing technology based on quantum-mechanical effects, such as the uncertainty principle or quantum entanglement. Broadly speaking, a quantum radar can be seen as a device working in the microwave range, which exploits quantum features, from the point of view of the radiation source and/or the output detection, and is able to outperform a classical counterpart. One approach is based on the use of input quantum correlations (in particular, quantum entanglement) combined with a suitable interferometric quantum detection at the receiver (strongly related to the protocol of quantum illumination). Paving the way for a technologically-viable prototype of a quantum radar involves the resolution of a number of experimental challenges as discussed in some review articles, the latter of which pointed out "inaccurate reporting" in the media. Current experimental designs seem to be limited to very short ranges, of the order of one meter, suggesting that potential applications might instead be for near-distance surveillance or biomedical scanning. Concept behind a microwave-range model A microwave-range model of a quantum radar was proposed in 2015 by an international team and is based on the protocol of Gaussian quantum illumination. The basic concept is to create a stream of entangled visible-frequency photons and split it in half. One half, the "signal beam", goes through a conversion to microwave frequencies in a way that preserves the original quantum state. The microwave signal is then sent and received as in a normal radar system. When the reflected signal is received it is converted back into visible photons and compared with the other half of the original entangled beam, the "idler beam". Although most of the original entanglement will be lost due to quantum decoherence as the microwaves travel to the target objects and back, enough quantum correlations will still remain between the reflected-signal and the idler beams. Using
https://en.wikipedia.org/wiki/String-net%20liquid
In condensed matter physics, a string-net is an extended object whose collective behavior has been proposed as a physical mechanism for topological order by Michael A. Levin and Xiao-Gang Wen. A particular string-net model may involve only closed loops; or networks of oriented, labeled strings obeying branching rules given by some gauge group; or still more general networks. Overview The string-net model is claimed to show the derivation of photons, electrons, and U(1) gauge charge, small (relative to the Planck mass) but nonzero masses, and suggestions that the leptons, quarks, and gluons can be modeled in the same way. In other words, string-net condensation provides a unified origin for photons and electrons (or gauge bosons and fermions). It can be viewed as an origin of light and electron (or gauge interactions and Fermi statistics). However, their model does not account for the chiral coupling between the fermions and the SU(2) gauge bosons in the standard model. For strings labeled by the positive integers, string-nets are the spin networks studied in loop quantum gravity. This has led to the proposal by Levin and Wen, and Smolin, Markopoulou and Konopka that loop quantum gravity's spin networks can give rise to the standard model of particle physics through this mechanism, along with fermi statistics and gauge interactions. To date, a rigorous derivation from LQG's spin networks to Levin and Wen's spin lattice has yet to be done, but the project to do so is called quantum graphity, and in a more recent paper, Tomasz Konopka, Fotini Markopoulou, Simone Severini argued that there are some similarities to spin networks (but not necessarily an exact equivalence) that gives rise to U(1) gauge charge and electrons in the string net mechanism. Herbertsmithite may be an example of string-net matter. Examples Z2 spin liquid Z2 spin liquid obtained using slave-particle approach may be the first theoretical example of string-net liquid. The toric code The toric
https://en.wikipedia.org/wiki/Savaging
Savaging is a term used in the study of ethology that refers to aggressive behaviour displayed by the mother towards the offspring. Aggressive behaviour includes being rough with, injuring, biting, attacking, crushing and killing (maternal infanticide) of the offspring. While savaging behaviour has been seen in multiple species, it is predominantly demonstrated in domestic pigs (Sus scrofa domesticus). As the definition of savaging is so broad, research on the prevalence of savaging behaviour varies with reports of little savaging of offspring to savaging of offspring up to the 20th percentile. Prevalence of aggressive, non-fatal savaging is greater in gilts, or females who have not yet previously farrowed, as piglet-focused aggression is more frequent in young animals than sows, adult females who have previously given birth. Occurrence of savaging demonstrated by sows is greater if the sow has previously savaged her offspring either as a gilt or sow. Savaging behaviour usually occurs during the first two days after parturition. Prevalence of savaging is similar among first and second farrowing cycles. Savaging behaviour has a significant impact on both agricultural economy and animal welfare which is why it is currently a subject of interest in the pig industry. Environmental factors There exist multiple environmental factors that increase the rate of savaging demonstrated by pigs. It has been shown that human attitudes and behaviour can affect the behaviour of domestic pigs. For example, annoyance and transmission of emotions can occur in gilts and sows when a new worker enters the farrowing rooms and makes excessive noise, does not feed the animals in a timely manner and/or is frustrated/annoyed. Research suggests that maternal behaviour may improve when workers and pigs are familiar with their environment and on a set schedule. Another environmental factor that has been examined is the amount of time spent by humans in the farrowing room. Decreased disturbance
https://en.wikipedia.org/wiki/Electrogustometry
Electrogustometry is the measurement of taste threshold by passing controlled anodal current through the tongue. When current passes through the tongue a unique and distinct metallic taste is perceived. Electrogustometry has been in existence since the 1950s. However, not much research has been done in this field. Equipment TR Bull developed an electrogustometer, but it lacked precision and was not widely accepted. In the 1990s the Rion (TR-06) was developed. This was precise and is widely used in the market. The TR-06 is probably the only electrogustometer used now for research and diagnosis. This is manual and needs a trained doctor or nurse to use it. It needs to be manually calculated to arrive at the taste threshold and carry out the false positive tests. Small, light and portable, this hand-held, battery-powered device can be easily placed in any environment. It has an inbuilt false detection test. This is useful for the psychophysical analysis of the subject. Influencing factors Electrogustometric taste threshold depends on pulse duration pulse and area of contact of electrode and tongue. Detailed experiments will be performed to prove and discuss their influence on taste threshold with the automatic electrogustometer. Experiments have been performed with the manual TR-06 on to study the effects of these control factors. Philosophy The measure of the minimum amount of current required to excite the sensation of this taste is called the 'taste threshold'. This project involves the design and fabrication of an automatic instrument to measure this taste threshold. See also Machine olfaction Olfactometer
https://en.wikipedia.org/wiki/The%20Pesticide%20Question
The Pesticide Question: Environment, Economics and Ethics is a 1993 book edited by David Pimentel and Hugh Lehman. Use of pesticides has improved agricultural productivity, but there are also concerns about safety, health and the environment. This book is the result of research by leading scientists and policy experts into the non-technical and social issues of pesticides. In examining the social policies related to pesticides use, they consider the costs as well as the benefits. The book says that Intensive farming cannot completely do without synthetic chemicals, but that it is technologically possible to reduce the amount of pesticides used in the United States by 35-50 per cent without reducing crop yields. The researchers show that to regain public trust, those who regulate and use pesticides must examine fair ethical questions and take appropriate action to protect public welfare, health, and the environment. Anyone concerned with reducing our reliance on chemical pesticides and how human activities can remain both productive and environmentally sound will find this volume a stimulating contribution to a troubling debate. The Pesticide Question builds on the 1962 best seller book Silent Spring by Rachel Carson. Carson did not reject the use of pesticides, but argued that their use was often indiscriminate and resulted in harm to people and the environment. She also highlighted the problem of pests becoming resistant to pesticides. Carson's work is referred to many times in The Pesticide Question, which critically explores many non-technical issues associated with pesticide use, mainly in the United States. The book has 40 contributors, mainly academics from a wide range of disciplines. The Pesticide Question is divided into five main parts: social and environmental effects of pesticides; methods and effects of reducing pesticide use; government policy and pesticide use; history, public attitudes, and ethics in regard to pesticide use; and the ben
https://en.wikipedia.org/wiki/List%20of%203D%20graphics%20libraries
3D graphics have become so popular, particularly in video games, that specialized APIs (application programming interfaces) have been created to ease the processes in all stages of computer graphics generation. These APIs have also proved vital to computer graphics hardware manufacturers, as they provide a way for programmers to access the hardware in an abstract way, while still taking advantage of the special hardware of any specific graphics card. The first 3D graphics framework was probably Core, published by the ACM in 1977. Low-level 3D API These APIs for 3D computer graphics are particularly popular: ANGLE, web browsers graphics engine, a cross-platform translator of OpenGL ES calls to DirectX, OpenGL, or Vulkan API calls. Direct3D (a subset of DirectX) Glide a defunct 3D graphics API developed by 3dfx Interactive. Mantle developed by AMD. Metal developed by Apple. OpenGL and the OpenGL Shading Language OpenGL ES 3D API for embedded devices. OptiX 7.0 and Latest developed by NVIDIA. LibGCM QuickDraw 3D developed by Apple Computer starting in 1995, abandoned in 1998. Vulkan Web-based API WebGL is a JavaScript interface for OpenGL ES API, promoted by Khronos. WebGPU an under-development web standard and JavaScript API for accelerated graphics and compute. High-level 3D API There are also higher-level 3D scene-graph APIs which provide additional functionality on top of the lower-level rendering API. Such libraries under active development include: BGFX ClanLib Crystal Space HOOPS 3D Graphics System Horde3D Irrlicht Engine Java 3D Java FX JMonkey Engine JT Open from Siemens Digital Industries Software magnum Mobile 3D Graphics API (M3G; JSR-184) OGRE OpenGL Performer OpenSceneGraph (now obsolete OSG.JS for WebPlatforms) OpenSG QSDK RAMSES RenderWare Panda3D Zea Engine Unigine VTK JavaScript-based engines There is more interest in web browser based high-level API for 3D graphics engines. Some are: A-Frame Blend4Web Cop