source stringlengths 31 227 | text stringlengths 9 2k |
|---|---|
https://en.wikipedia.org/wiki/Herschel%E2%80%93Bulkley%20fluid | The Herschel–Bulkley fluid is a generalized model of a non-Newtonian fluid, in which the strain experienced by the fluid is related to the stress in a complicated, non-linear way. Three parameters characterize this relationship: the consistency k, the flow index n, and the yield shear stress . The consistency is a simple constant of proportionality, while the flow index measures the degree to which the fluid is shear-thinning or shear-thickening. Ordinary paint is one example of a shear-thinning fluid, while oobleck provides one realization of a shear-thickening fluid. Finally, the yield stress quantifies the amount of stress that the fluid may experience before it yields and begins to flow.
This non-Newtonian fluid model was introduced by Winslow Herschel and Ronald Bulkley in 1926.
Definition
In one dimension, the constitutive equation of the Herschel-Bulkley model after the yield stress has been reached can be written in the form:
where is the shear stress [Pa], the yield stress [Pa], the consistency index [Pas], the shear rate [s], and the flow index [dimensionless]. If the Herschel-Bulkley fluid behaves as a rigid (non-deformable) solid, otherwise it behaves as a fluid. For the fluid is shear-thinning, whereas for the fluid is shear-thickening. If and , this model reduces to that of a Newtonian fluid.
Reformulated as a tensor, we can instead write:
Note that the double underlines indicate a tensor quantity.
Modelling Herschel-Bulkley fluids using regularization
The viscosity associated with the Herschel-Bulkley stress diverges to infinity as the strain rate approaches zero. This divergence makes the model difficult to implement in numerical simulations, so it is common to implement regularized models with an upper limiting viscosity. For instance, the Herschel-Bulkley fluid can be approximated as a generalized Newtonian fluid model with an effective (or apparent) viscosity being given as
Here, the limiting viscosity replaces the diverge |
https://en.wikipedia.org/wiki/Quebracho%20tree | Quebracho is a common name in Spanish to describe very hard (density 0.9–1.3) wood tree species. The etymology of the name derived from quiebrahacha, or quebrar hacha, meaning "axe-breaker". The corresponding English-language term for such hardwoods is breakax or breakaxe.
Species
There are at least three similar commercially important tree species that grow in the Gran Chaco region of South America.
the quebracho
Schinopsis lorentzii (Syn.: Schinopsis marginata Engl., Schinopsis haenkeana Engl.); of the family Anacardiaceae; North Argentina, Paraguay and Bolivia; (the red quebracho, quebracho), quebracho-colorado santiagueño, red quebracho santiagueño, quebracho santiagueño (also known as quebracho cor(o)nillo, cor(o)nillo, quebracho macho, quebracho negro or moro, quebracho rojo, quebracho bolí, horco quebracho, quebracho serrano, quebracho montano, quebracho crespo, quebracho del cerro, quebracho colorado de las sierras o del cerro and quebracho cordobés)
Schinopsis balansae; of the same family; Northeast Argentina, West-Central Brazil, Paraguay; (the willow-leaf red quebracho, red quebracho) quebracho-colorado chaqueño, red quebracho chaqueño, quebracho chaqueño, quebracho vermelho, quebracho vermelho chaqueño (also known as quebracho hembra or femea, quebracho santafesino, quebracho colorado santafesino, quebracho rubio)
Schinopsis aff. heterophylla Ragonese & J.A.Castigl., the quebracho mestizo or quebracho colorado mestizo, horco quebracho; Northeast Argentina, Paraguay
Schinopsis brasiliensis Engl., brazil red quebracho, quebracho-colorado, quebracho crespo; Brazil
Schinopsis cornuta Loes., horned red quebracho, quebracho-colorado; Paraguay, Bolivia
Schinopsis peruviana, quebracho-colorado (boliviano); Peru
the white quebracho or quebracho blanco, quebracho amarillo, Aspidosperma quebracho-blanco of the family Apocynaceae; Paraguay, Argentina, Bolivia, West-Central Brazil
Aspidosperma triternatum Rojas Acosta; North Argentina, Bolivia, Parag |
https://en.wikipedia.org/wiki/Cardiac%20plexus | The cardiac plexus is a plexus of nerves situated at the base of the heart that innervates the heart.
Structure
The cardiac plexus is divided into a superficial part, which lies in the concavity of the aortic arch, and a deep part, between the aortic arch and the trachea. The two parts are, however, closely connected. The sympathetic component of the cardiac plexus comes from cardiac nerves, which originate from the sympathetic trunk. The parasympathetic component of the cardiac plexus originates from the cardiac branches of the vagus nerve.
Superficial part
The superficial part of the cardiac plexus lies beneath the aortic arch, in front of the right pulmonary artery. It is formed by the superior cervical cardiac branch of the left sympathetic trunk and the inferior cardiac branch of the left vagus nerve. A small ganglion, the cardiac ganglion of Wrisberg, is occasionally found connected with these nerves at their point of junction. This ganglion, when present, is situated immediately beneath the arch of the aorta, on the right side of the ligamentum arteriosum.
The superficial part of the cardiac plexus gives branches to:
the deep part of the plexus.
the anterior coronary plexus.
the left anterior pulmonary plexus.
Deep part
The deep part of the cardiac plexus is situated in front of the bifurcation of the trachea (known as the carina), above the point of division of the pulmonary artery, and behind the aortic arch. It is formed by the cardiac nerves derived from the cervical ganglia of the sympathetic trunk, and the cardiac branches of the vagus and recurrent laryngeal nerves.
The only cardiac nerves which do not enter into the formation of the deep part of the cardiac plexus are the superior cardiac nerve of the left sympathetic trunk, and the lower of the two superior cervical cardiac branches from the left vagus nerve, which pass to the superficial part of the plexus.
Right half
The branches from the right half of the deep part of the cardiac plexus p |
https://en.wikipedia.org/wiki/Tensilica | Tensilica Inc. was a company based in Silicon Valley in the semiconductor intellectual property core business. It is now a part of Cadence Design Systems.
Tensilica is known for its customizable Xtensa microprocessor core. Other products include: HiFi audio/voice DSPs (digital signal processors) with a software library of over 225 codecs from Cadence and over 100 software partners; Vision DSPs that handle complex algorithms in imaging, video, computer vision, and neural networks; and the ConnX family of baseband DSPs ranging from the dual-MAC ConnX D2 to the 64-MAC ConnX BBE64EP.
Tensilica was founded in 1997 by Chris Rowen (one of the founders of MIPS Technologies). It employed Earl Killian, who contributed to the MIPS architecture, as director of architecture. On March 11, 2013, Cadence Design Systems announced its intent to buy Tensilica for approximately $380 million in cash. Cadence completed the acquisition in April 2013, with a cash outlay at closing of approximately $326 million.
Cadence Tensilica products
Cadence Tensilica develops SIP blocks to be included on the chip (IC) designs of products of their licensees, such as system on a chip for embedded systems. Tensilica processors are delivered as synthesizable RTL for easy integration into chip designs.
Xtensa configurable cores
Xtensa processors range from small, low-power cache-less microcontroller to high-performance 16-way SIMD processors, 3-issue VLIW DSP cores, or 1 TMAC/sec neural network processors. All Cadence standard DSPs are based on the Xtensa architecture. The Xtensa architecture offers a user-customizable instruction set through automated customization tools that can extend the Xtensa base instruction set, including SIMD instructions, new register files.
Xtensa instruction set
The Xtensa instruction set is a 32-bit architecture with a compact 16- and 24-bit instruction set. The base instruction set has 82 RISC instructions and includes a 32-bit ALU, 16 general-purpose 32-bit registers |
https://en.wikipedia.org/wiki/Celebratory%20gunfire | Celebratory gunfire is the shooting of a firearm into the air in celebration. It sometimes occurs in parts of the Balkans, Russia, the Middle East, South Asia, Latin America, the United States, and Ethiopia, even where illegal.
Common occasions for celebratory gunfire include New Year's Day as well as religious holidays. The practice sometimes results in random death and injury from stray bullets. Property damage is another result of celebratory gunfire; shattered windows and damaged roofs are sometimes found after such celebrations.
Injuries
Depending on the angle it is fired, the speed of a falling bullet changes. A bullet fired nearly vertically will lose the most speed, usually falling at terminal velocity, which is much lower than its muzzle velocity. Despite this, people can still be injured or killed by bullets falling at this speed. If a bullet is fired at other angles, it maintains its angular ballistic trajectory and is far less likely to engage in tumbling motion; it therefore travels at speeds much higher than a bullet in free fall. Dense, small bullets achieve higher terminal velocities than lighter, larger bullets.
Between 1918 and 1920, United States Army Ordnance Corps Julian Hatcher conducted experiments to determine the velocity of falling bullets, and calculated that .30 caliber rounds reach terminal velocities of 90 m/s (300 feet per second or 186 miles per hour). According to computer models, 9mm handgun rounds reach terminal velocities of between 150 and 250 feet per second. A bullet traveling at only 61 m/s (200 feet per second) to 100 m/s (330 feet per second) can penetrate human skin.
Any gunfire can damage hearing of those nearby without ear protection, and blank rounds fired in an unsafe direction can cause injuries or death from muzzle blast at close range, as in the case of actor Jon-Erik Hexum. Birdshot fired from a shotgun disperses and loses energy much faster than slugs, buckshot, or bullets fired from rifles and pistols. Alth |
https://en.wikipedia.org/wiki/Turbonomic | Turbonomic is a resource-simulation software company headquartered in Boston, MA and owned by IBM. The company was originally named VMTurbo.
Reception
In 2011, Gartner named Turbonomic as a Cool Vendor in Cloud Management. In 2016, Turbonomic was listed as the top product for Virtualization Management in a report by IDG and IT Central Station. Turbonomic has made five appearances to the Inc. 5000, and made the Forbes Cloud 100 four times. In 2020, Fast Company named Turbonomic to their Best Workplaces for Innovators List. The company was also deemed a Vendor to Watch in AIOps by Enterprise Management Associates (EMA) for its combination of abstraction, analytics, and automation engine that continually assures performance of a customer’s applications.
History
The company formed partnerships with Cisco and IBM, entering into OEM agreements to bring Application Resource Management to a larger customer base.
Since its founding in 2008 or 2009, Turbonomic had raised more than $250M from venture capital firms including Bain Capital Ventures and Highland Capital Partners.
The company's product was updated in 2017 for use with cloud computing platforms.
The company was originally named VMTurbo and changed its name to Turbonomic in August 2017.
Turbonomic acquired ParkMyCloud and SevOne in 2019.
IBM acquired Turbonomic on June 17, 2021.
The company's product simulates supply and demand forces in order to efficiently allocate resources such as computing, database, memory and storage. |
https://en.wikipedia.org/wiki/Lindeberg%27s%20condition | In probability theory, Lindeberg's condition is a sufficient condition (and under certain conditions also a necessary condition) for the central limit theorem (CLT) to hold for a sequence of independent random variables. Unlike the classical CLT, which requires that the random variables in question have finite variance and be both independent and identically distributed, Lindeberg's CLT only requires that they have finite variance, satisfy Lindeberg's condition, and be independent. It is named after the Finnish mathematician Jarl Waldemar Lindeberg.
Statement
Let be a probability space, and , be independent random variables defined on that space. Assume the expected values and variances exist and are finite. Also let
If this sequence of independent random variables satisfies Lindeberg's condition:
for all , where 1{…} is the indicator function, then the central limit theorem holds, i.e. the random variables
converge in distribution to a standard normal random variable as
Lindeberg's condition is sufficient, but not in general necessary (i.e. the inverse implication does not hold in general).
However, if the sequence of independent random variables in question satisfies
then Lindeberg's condition is both sufficient and necessary, i.e. it holds if and only if the result of central limit theorem holds.
Remarks
Feller's theorem
Feller's theorem can be used as an alternative method to prove that Lindeberg's condition holds. Letting and for simplicity , the theorem states
if , and converges weakly to a standard normal distribution as then satisfies the Lindeberg's condition.
This theorem can be used to disprove the central limit theorem holds for by using proof by contradiction. This procedure involves proving that Lindeberg's condition fails for .
Interpretation
Because the Lindeberg condition implies as , it guarantees that the contribution of any individual random variable () to the variance is arbitrarily small, for sufficiently large val |
https://en.wikipedia.org/wiki/Bluetooth%20advertising | Bluetooth advertising is a method of mobile marketing that utilizes Bluetooth technology to deliver content such as message, information, or advertisement to mobile devices such as cellular phones or tablet computers. Bluetooth advertising can also be received via laptop or personal digital assistants (PDAs).
Bluetooth advertising is permission based advertising, which means that when a mobile device has received a Bluetooth message, the recipient has the choice to either accept or decline the message. The recipient needs to positively indicate that they wish to receive marketing messages.
While not all users of Bluetooth-mobile devices leave their Bluetooth activated, they can interact with a sign to encourage them to turn on their Bluetooth to receive the content. The advertiser is required to explain that those marketing messages may contain information about other companies’ products and services, if appropriate. It is highly recommended that the Direct Marketing Associations guidelines are used.
Bluetooth advertising proximity range
Bluetooth advertising generally is a broadcast function. The average range of Bluetooth advertising in class two is 15 meters to 40 meters for most Bluetooth enabled mobile devices.
As with all wireless transmission, the range and accessibility to most Bluetooth advertising depends on the transmitter power class and the individual portage of the receiver equipment. However, with advances in mobile devices technology, this distance for proper receiving is increasing to reach 250 meters or more in nowadays smart phones, tablet computers and other mobile devices.
The selectivity goes down with extension of range. Hence the transmission power raise as well as receiver sensitivity raise will reduce the contextual connection between actual location of receiver and contents of broadcast message.
There are several major types of Bluetooth advertising solutions. These are generally Bluetooth dongles as transmitter hardware in conjunct |
https://en.wikipedia.org/wiki/Exe2bin | The command-line tool exe2bin is a post-compilation utility program available on MS-DOS and other operating systems.
Overview
Early compilers and linkers for the MS-DOS platform could not produce a COM file executable directly. Instead, the compilers would output an EXE-format file with relocation information. If all 8086 segments were set to be identical in such an EXE file (i.e. the "tiny" memory model was used), then exe2bin could convert it to a COM file.
exe2bin could also be used to convert compiled code to make it suitable to be embedded in ROM as part of BIOS or a device driver.
Availability
The command was included in MS/DOS versions 1 thru 3.1 as part of a standard distribution. For version 3.2, among the changes were
the version included did not permit itself to run on any version except 3.2. For the next version, 3.3, there was no EXE2BIN on the DOS disk. "Instead, IBM sells the program
separately, at an extra cost, with the DOS Technical Reference." IBM also added code to check the version. PC Magazine published a
workaround: just patch it to work with
3.2 or higher.
One way or the other, it was no longer available for the base price after 3.2; for version 6 it was on what was called the Supplemental Disk. The program was also distributed with many language compilers for MS-DOS in the 1980s, and included with certain versions of IBM PC DOS.
PhysTechSoft & Paragon Technology Systems PTS-DOS, Digital Research DR DOS 6.0, and Datalight ROM-DOS, also include an implementation of the command.
The command is also available in FreeDOS. This implementation is licensed under the Sybase Open Watcom Public License.
Windows XP and later versions include exe2bin and other 16-bit commands (nonnative) for the MS-DOS subsystem to maintain MS-DOS compatibility. The 16-bit MS-DOS subsystem commands are not available on 64-bit editions of Windows.
See also
List of DOS commands |
https://en.wikipedia.org/wiki/Mahler%27s%20inequality | In mathematics, Mahler's inequality, named after Kurt Mahler, states that the geometric mean of the term-by-term sum of two finite sequences of positive numbers is greater than or equal to the sum of their two separate geometric means:
when xk, yk > 0 for all k.
Proof
By the inequality of arithmetic and geometric means, we have:
and
Hence,
Clearing denominators then gives the desired result.
See also
Minkowski inequality |
https://en.wikipedia.org/wiki/Gelfand%E2%80%93Zeitlin%20integrable%20system | In mathematics, the Gelfand–Zeitlin system (also written Gelfand–Zetlin system, Gelfand–Cetlin system, Gelfand–Tsetlin system) is an integrable system on conjugacy classes of Hermitian matrices. It was introduced by , who named it after the Gelfand–Zeitlin basis, an early example of canonical basis, introduced by I. M. Gelfand and M. L. Cetlin in 1950s. introduced a complex version of this integrable system. |
https://en.wikipedia.org/wiki/Seismic%20metamaterial | A seismic metamaterial, is a metamaterial that is designed to counteract the adverse effects of seismic waves on artificial structures, which exist on or near the surface of the Earth. Current designs of seismic metamaterials utilize configurations of boreholes, trees or proposed underground resonators to act as a large scale material. Experiments have observed both reflections and bandgap attenuation from artificially induced seismic waves. These are the first experiments to verify that seismic metamaterials can be measured for frequencies below 100 Hz, where damage from Rayleigh waves is the most harmful to artificial structures.
The mechanics of seismic waves
More than a million earthquakes are recorded each year, by a worldwide system of earthquake detection stations. The propagation velocity of the seismic waves depends on density and elasticity of the earth materials. In other words, the speeds of the seismic waves vary as they travel through different materials in the Earth. The two main components of a seismic event are body waves and surface waves. Both of these have different modes of wave propagation.
Towards Seismic Cloaking
Computations showed that seismic waves traveling toward a building, could be directed around the building, leaving the building unscathed, by using seismic metamaterials. The very long wavelengths of earthquake waves would be shortened as they interact with the metamaterials; the waves would pass around the building so as to arrive in phase as the earthquake wave proceeded, as if the building was not there. The mathematical models produce the regular pattern provided by Metamaterial cloaking. This method was first understood with electromagnetic cloaking metamaterials - the electromagnetic energy is in effect directed around an object, or hole, and protecting buildings from seismic waves employs this same principle.
Giant polymer-made split ring resonators combined with other metamaterials are designed to couple at the seismic |
https://en.wikipedia.org/wiki/Automatic%20picture%20transmission | The Automatic Picture Transmission (APT) system is an analog image transmission system developed for use on weather satellites. It was introduced in the 1960s and over four decades has provided image data to relatively low-cost user stations at locations in most countries of the world. A user station anywhere in the world can receive local data at least twice a day from each satellite as it passes nearly overhead.
Transmission
Structure
The broadcast transmission is composed of two image channels, telemetry information, and synchronization data, with the image channels typically referred to as Video A and Video B. All this data is transmitted as a horizontal scan line. A complete line is 2080 pixels long, with each image using 909 pixels and the remainder going to the telemetry and synchronization. Lines are transmitted at 2 per second, which equates to a 4160 words per second, or 4160 baud.
Images
On NOAA POES system satellites, the two images are 4 km / pixel smoothed 8-bit images derived from two channels of the advanced very-high-resolution radiometer (AVHRR) sensor. The images are corrected for nearly constant geometric resolution prior to being broadcast; as such, the images are free of distortion caused by the curvature of the Earth.
Of the two images, one is typically long-wave infrared (10.8 micrometers) with the second switching between near-visible (0.86 micrometers) and mid-wave infrared (3.75 micrometers) depending on whether the ground is illuminated by sunlight. However, NOAA can configure the satellite to transmit any two of the AVHRR's image channels.
Synchronization and telemetry
Included in the transmission are a series of synchronization pulses, minute markers, and telemetry information.
The synchronization information, transmitted at the start of each video channel, allows the receiving software to align its sampling with the baud rate of the signal, which can vary slightly over time. The minute markers are four lines of alternating bla |
https://en.wikipedia.org/wiki/Coupled%20pattern%20learner | Coupled Pattern Learner (CPL) is a machine learning algorithm which couples the semi-supervised learning of categories and relations to forestall the problem of semantic drift associated with boot-strap learning methods.
Coupled Pattern Learner
Semi-supervised learning approaches using a small number of labeled examples with many unlabeled examples are usually unreliable as they produce an internally consistent, but incorrect set of extractions. CPL solves this problem by simultaneously learning classifiers for many different categories and relations in the presence of an ontology defining constraints that couple the training of these classifiers. It was introduced by Andrew Carlson, Justin Betteridge, Estevam R. Hruschka Jr. and Tom M. Mitchell in 2009.
CPL overview
CPL is an approach to semi-supervised learning that yields more accurate results by coupling the training of many information extractors. Basic idea behind CPL is that semi-supervised training of a single type of extractor such as ‘coach’ is much more difficult than simultaneously training many extractors that cover a variety of inter-related entity and relation types. Using prior knowledge about the relationships between these different entities and relations CPL makes unlabeled data as a useful constraint during training. For e.g., ‘coach(x)’ implies ‘person(x)’ and ‘not sport(x)’.
CPL description
Coupling of predicates
CPL primarily relies on the notion of coupling the learning of multiple functions so as to constrain the semi-supervised learning problem. CPL constrains the learned function in two ways.
Sharing among same-arity predicates according to logical relations
Relation argument type-checking
Sharing among same-arity predicates
Each predicate P in the ontology has a list of other same-arity predicates with which P is mutually exclusive. If A is mutually exclusive with predicate B, A’s positive instances and patterns become negative instances and negative patterns for B. For example, |
https://en.wikipedia.org/wiki/Starspot | Starspots are stellar phenomena, so-named by analogy with sunspots.
Spots as small as sunspots have not been detected on other stars, as they would cause undetectably small fluctuations in brightness. The commonly observed starspots are in general much larger than those on the Sun: up to about 30% of the stellar surface may be covered, corresponding to starspots 100 times larger than those on the Sun.
Detection and measurements
To detect and measure the extent of starspots one uses several types of methods.
For rapidly rotating stars – Doppler imaging and Zeeman-Doppler imaging. With the Zeeman-Doppler imaging technique the direction of the magnetic field on stars can be determined since spectral lines are split according to the Zeeman effect, revealing the direction and magnitude of the field.
For slowly rotating stars – Line Depth Ratio (LDR). Here one measures two different spectral lines, one sensitive to temperature and one which is not. Since starspots have a lower temperature than their surroundings the temperature-sensitive line changes its depth. From the difference between these two lines the temperature and size of the spot can be calculated, with a temperature accuracy of 10K.
For eclipsing binary stars – Eclipse mapping produces images and maps of spots on both stars.
For giant binary stars - Very-long-baseline interferometry
For stars with transiting extrasolar planets – Light curve variations.
Temperature
Observed starspots have a temperature which is in general 500–2000 kelvins cooler than the stellar photosphere. This temperature difference could give rise to a brightness variation up to 0.6 magnitudes between the spot and the surrounding surface. There also seems to be a relation between the spot temperature and the temperature for the stellar photosphere, indicating that starspots behave similarly for different types of stars (observed in G–K dwarfs).
Lifetimes
The lifetime for a starspot depends on its size.
For small spots the lifetim |
https://en.wikipedia.org/wiki/Street%20Fighter%206 | Street Fighter 6 is a 2023 fighting game developed by Bandai Namco Studios and published by Capcom. Announced in February 2022, it is the seventh main entry in the Street Fighter franchise, and was released for PlayStation 4, PlayStation 5, Windows and Xbox Series X/S on June 2, 2023, while an arcade version will be published by Taito later in the year. Additionally, a prequel comic book series was unveiled in September 2022.
Developed on the RE Engine, Street Fighter 6 supports cross-platform play and rollback netcode. It offers three overarching game modes and three control options. The game also features a real-time commentary system, providing a tournament-style feel and the option to cheer on the player.
Street Fighter 6 received critical acclaim and sold over 2 million units by July 2023.
Gameplay
Street Fighter 6 features three overarching game modes: Fighting Ground, World Tour, and Battle Hub. Fighting Ground contains local and online versus battles as well as training and arcade modes, all featuring similar 2D fighting gameplay to the previous games in the series, in which two fighters use a variety of attacks and special abilities to knock out their opponent. World Tour is a single-player story mode featuring a customizable player avatar exploring 3D environments, such as Final Fights Metro City and the small, fictional South Asian nation of Nayshall, with action-adventure gameplay. Battle Hub acts as an online lobby mode, using customizable player avatars from the World Tour mode (the first fighting game to implement similar online features was Tecmo's Dead or Alive 4). In the Battle Hub, players can compete in ranked or casual matches, battle using their created avatars, using the skills learned in World Tour mode, participate in special events, or play emulated Capcom arcade titles, using the same emulation technology used in the Capcom Arcade Stadium series, among other features.
The main fighting gameplay of Street Fighter 6 is based around the |
https://en.wikipedia.org/wiki/CDC%20160%20series | The CDC 160 series was a series of minicomputers built by Control Data Corporation. The CDC 160 and CDC 160-A were 12-bit minicomputers built from 1960 to 1965; the CDC 160G was a 13-bit minicomputer, with an extended version of the CDC 160-A instruction set, and a compatibility mode in which it did not use the 13th bit. The 160 was designed by Seymour Cray - reportedly over a long three-day weekend.
It fit into the desk where its operator sat.
The 160 architecture uses ones' complement arithmetic with end-around carry.
NCR joint-marketed the 160-A under its own name for several years in the 1960s.
Overview
A publishing company that purchased a CDC 160-A described it as "a single user machine with no batch processing capability. Programmers and/or users would go to the computer room, sit at the console, load the paper tape bootstrap and start up a program."
The CDC 160-A was a simple piece of hardware, and yet provided a variety of features which were scaled-down capabilities found only on larger systems. It was therefore an ideal platform for introducing neophyte programmers to the sophisticated concepts of low-level input/output (I/O) and interrupt systems.
All 160 systems had a paper-tape reader, and a punch, and most had an IBM Electric typewriter modified to act as a computer terminal. Memory on the 160 was 4096 12-bit words. The CPU had a 12-bit ones' complement accumulator but no multiply or divide. There was a full complement of instructions and several addressing modes. Indirect addressing was almost as good as index registers. The instruction set supported both relative (to the current P register) and absolute. The original instruction set did not have a subroutine call instruction and could only address one bank of memory.
In the 160-A model, a "return jump" and a memory bank-switch instruction was added. Return-jump allowed simple subroutine calls and bank switching allowed other 4K banks of memory to be addressed, albeit clumsily, up to a |
https://en.wikipedia.org/wiki/Lib.ru | Lib.ru, also known as Maksim Moshkow's Library (, started to operate in November 1994) is the oldest electronic library in the Russian Internet segment.
Founded and supported by Maksim Moshkow, it receives contributions mainly from users who send texts they scanned and processed (OCR, proofreading). This method of acquisition provides the library a broad and efficient augmentability, though sometimes it adversely affects the quality (errors, omissions).
The structure of the library includes a section where one can publish his own literary texts ("Samizdat" journal, named after the samizdat of the Soviet era), a project for music publishing ("Music hosting"), a travel notes project ("Foreign countries") and some other sections.
Maksim Moshkow's Library received several Ru-net Awards, including the National Internet Award (2003).
The headline on the site Lib.ru says "With support from the Federal Press and Mass Communications Agency". According to Moshkow, his project received $35,000 from that organization in September 2005, which indicates some level of government support for the online publishing of in-copyright works.
Maksim Moshkow's project could be compared to some Wikimedia Foundation projects and is sometimes referred to as Russia's Project Gutenberg.
KM Online vs. Maksim Moshkow's Library
On 1 April 2004, the "KM Online" media company, which is known for forming its own library by copying texts from the other electronic libraries, issued a lawsuit against Maksim Moshkow's Library in the name of Eduard Gevorkian, Marina Alekseyeva (pen-name "Alexandra Marinina"), Vasili Golovachov and Elena Katasonova. It was later discovered that only Gevorkian had had real claims against Moshkow. Moshkow's lawyer was Andrey Mironov from the Artemy Lebedev Studio, while KM's interests were presented by the so-called "National Society for Digital Technologies (NOCIT)".
This case became a precedent in the Russian legal practice which illustrated pressure on an electro |
https://en.wikipedia.org/wiki/Hard%20Disk%2020SC | The Apple Hard Disk 20SC is Apple's first SCSI based hard drive for the Apple II family as well as the Macintosh and other third party computers using an industry standard SCSI interface.
History
Released in September 1986 along with the Apple IIGS (which required an optional SCSI interface card to use it), it debuted over 9 months after the introduction of the Macintosh Plus, the first to include Apple's SCSI interface. It was a welcome addition, delivering considerably faster data transfer rates (up to 1.25 megabytes per second) than its predecessors, the Hard Disk 20 (62.5 Kilobytes per second) and ProFile.
Hardware
The 20SC originally contained a half height 5.25" Seagate ST-225N 20MB SCSI hard drive, but was later manufactured with a full-height 3.5" MiniScribe 8425SA 20MB SCSI hard drive. The latter drive was the same size as the drive inside the Macintosh Hard Disk 20, but 10 to 15 MB over what had previously been offered by Apple for the II family. The same drive mechanism would also be offered 6 months later as a built-in drive option on the Macintosh II and SE. It had two standard Centronics 50-pin connectors, one for the System and one for daisy-chaining additional SCSI devices and a SCSI ID selection switch. An external terminator was required if it was the only SCSI device connected. The case itself could accommodate a 3.5" or 5.25" full-height hard drive mechanism. Indeed, the case design would be reused unchanged (in Platinum only) for 3 more models introduced the following year: 40SC, 80SC & 160SC (offering respective Megabytes of storage). While the transfer rates were significantly higher due to the faster SCSI bus technology, the actual transfer rate varied from computer to computer thanks to different SCSI implementation based on developing industry standards.
Design
In addition to being the first cross-platform drive offered by Apple it was the first hard drive to use the Snow White design language. Notably, it was the only Snow White product |
https://en.wikipedia.org/wiki/Langmuir%20%28unit%29 | The langmuir (symbol: L) is a unit of exposure (or dosage) to a surface (e.g. of a crystal) and is used in ultra-high vacuum (UHV) surface physics to study the adsorption of gases. It is a practical unit, and is not dimensionally homogeneous, and so is used only in this field. It is named after American physicist Irving Langmuir.
Definition
The langmuir is defined by multiplying the pressure of the gas by the time of exposure. One langmuir corresponds to an exposure of 10−6 Torr during one second. For example, exposing a surface to a gas pressure of 10−8 Torr for 100 seconds corresponds to 1 L.
Similarly, keeping the pressure of oxygen gas at 2.5·10−6 Torr for 40 seconds will give a dose of 100 L.
Conversion
Since both different pressures and exposure times can give the same langmuir (see Definition) it can be difficult to convert Langmuir (L) to exposure pressure × time (Torr·s) and vice versa. The following equation can be used to easily convert between the two: Here, and are any two numbers whose product equals the desired Langmuir value, is an integer allowing different magnitudes of pressure or exposure time to be used in conversion. The units are represented in the [square brackets]. Using the prior example, for a dose of 100 L a pressure of 2.5 × 10−6 Torr can be applied for 40 seconds, thus, , and . However, this dosage could also be gained with 8 × 10−8 Torr for 1250 seconds, here , , . In both scenarios .
Derivation
Exposure of a surface in surface physics is a type of fluence, that is the integral of number flux (JN) with respect to exposed time (t) to give a number of particles per unit area (Φ):
The number flux for an ideal gas, that is the number of gas molecules passing through (in a single direction) a surface of unit area in unit time, can be derived from kinetic theory:
where C is the number density of the gas, and is the mean speed of the molecules (not the root-mean-square speed, although the two are related). The number density of a |
https://en.wikipedia.org/wiki/Sam%20Labs | SAM Labs is a startup that makes app-enabled construction kits, designed for people of all ages to learn STEM, play, and create with technology and the Internet of Things. The company has been featured in WIRED and The Telegraph. Founded by Belgian born CEO Joachim Horn, the company works out of their headquarters in East London.
A month-long Kickstarter campaign initiated in October 2014 raised over $160,000, and helped launch the mass production of the kits. Backers included Jawbone co-founder Alexander Asseily, who pre-ordered the kit.
Kickstarter campaign
On 29 September 2014, SAM Labs launched a month-long Kickstarter campaign to crowdfund the mass manufacture of the SAM kits for the spring of 2015, primarily through pre-orders of the kits. At the conclusion of the campaign on 29 October, SAM Labs had raised over $160,000 from 817 backers.
Products
SAM Labs produces app-enabled construction toys.
The SAM Science Museum Inventor Kit was produced in collaboration with the London Science Museum. Using the SAM Blocks and the free SAM Space app inside of the kit, kids can build awesome projects, games, inventions and hacks. Super quick. Super smart.
The company's hero kit, SAM's Curious Cars, launched in October 2016. Available now at Barnes & Noble across the United States and John Lewis (department store) in the United Kingdom, the Curious Cars kit allows kids to build and program their own cars and games. Winning an IFA Markit Innovation Award, Consumer Electronics Show Award, and a KAPi award, the triple-award-winning kit also comes with its own app: Curious Cars.
Partnerships
SAM Labs has created a habit of working with local design talent. SAM Labs first partnered with MAP Project Office, the London-based creative consultancy that specialises in industrial design, to deliver the Kickstarter campaign. Map also featured SAM Labs in a video collaboration with Honda, where SAM led modules were featured in this promotional video, covering "Technology for Exp |
https://en.wikipedia.org/wiki/Sam%20Spade%20%28software%29 | Sam Spade was the name of a Windows software tool designed to assist in tracking down sources of e-mail spam. It was also the name of a free web service that provides access to similar online tools. The Sam Spade utility was authored by Steve Atkins in 1997. It is named after the fictional character Sam Spade.
Query tools
The main features (query tools) were:
Zone Transfer – ask a DNS server for all it knows about a domain
SMTP Relay Check – check whether a mail server allows third party relaying
Scan Addresses – scan a range of IP addresses looking for open ports
Crawl website – search a website, looking for email addresses, offsite links, etc.
Browse web – browse the web in a raw http format
Check cancels – search your news server for cancel messages
Fast and Slow Traceroute – find the route packets take between you and a remote system
S-Lang command – issue a scripting command; useful for debugging scripts
Decode URL – decipher an obfuscated URL
Parse email headers – read email headers and make a guess about the origin of the email
Website history
The last fully functional version of the website was available 2004-02-26.
Since then it experienced various outage problems due to "blackholing of SamSpade.org by several RIRs and general heavy usage.", and is no longer online. The URL currently redirects to https://tools.wordtothewise.com/ which contains a similar set of web tools under the brand name "wiseTools" and hosted by Atkins's email software business Word to The Wise |
https://en.wikipedia.org/wiki/Reinhold%20Strassmann | Reinhold Strassmann (or Straßmann) (24 January 1893 in Berlin – late October 1944 in Auschwitz concentration camp) was a German mathematician who proved Strassmann's theorem. His Ph.D. advisor at University of Marburg was Kurt Hensel.
Born into a Jewish family, Strassmann refused to leave Nazi Germany, and he was eventually detained and deported to Theresienstadt concentration camp in 1943. On October 23, 1944, he was deported from Theresienstadt to Auschwitz concentration camp, where he was murdered soon after.
He was the son of the forensic pathologist Fritz Strassmann.
Selected publications |
https://en.wikipedia.org/wiki/K%C3%A4ll%C3%A9n%E2%80%93Lehmann%20spectral%20representation | The Källén–Lehmann spectral representation gives a general expression for the (time ordered) two-point function of an interacting quantum field theory as a sum of free propagators. It was discovered by Gunnar Källén and Harry Lehmann independently. This can be written as, using the mostly-minus metric signature,
where is the spectral density function that should be positive definite. In a gauge theory, this latter condition cannot be granted but nevertheless a spectral representation can be provided. This belongs to non-perturbative techniques of quantum field theory.
Mathematical derivation
The following derivation employs the mostly-minus metric signature.
In order to derive a spectral representation for the propagator of a field , one considers a complete set of states so that, for the two-point function one can write
We can now use Poincaré invariance of the vacuum to write down
Next we introduce the spectral density function
.
Where we have used the fact that our two-point function, being a function of , can only depend on . Besides, all the intermediate states have and . It is immediate to realize that the spectral density function is real and positive. So, one can write
and we freely interchange the integration, this should be done carefully from a mathematical standpoint but here we ignore this, and write this expression as
where
.
From the CPT theorem we also know that an identical expression holds for and so we arrive at the expression for the time-ordered product of fields
where now
a free particle propagator. Now, as we have the exact propagator given by the time-ordered two-point function, we have obtained the spectral decomposition. |
https://en.wikipedia.org/wiki/Glazing%20agent | A glazing agent is a natural or synthetic substance that provides a waxy, homogeneous, coating to prevent water loss from a surface and provide other protection.
Natural
Natural glazing agents keep moisture inside plants and insects. Scientists harnessed this characteristic in coatings made of substances classified as waxes. A natural wax is chemically defined as an ester with a very long hydrocarbon chain that also includes a long chain alcohol.
Examples are:
Stearic acid (E570)
Beeswax (E901)
Candelilla wax (E902)
Carnauba wax (E903)
Shellac (E904)
Microcrystalline wax (E905c), Crystalline wax (E907)
Lanolin (E913)
Oxidized polyethylene wax (E914)
Esters of colophonium (E915)
Paraffin
Synthetic
Scientists have produced glazing agents that mimic their natural counterparts. These components are added in different proportions to achieve the optimal glazing agent for a product. Such products include cosmetics, automobiles and food.
Some of the characteristics that are looked for in all of the above industries are:
1. Preservation - the glazing agent must protect the product from degradation and water loss. This characteristic can lead to a longer shelf life for a food or the longevity of a car without rusting.
2. Stability - the glazing agent must maintain its integrity under pressure or heat.
3. Uniform viscosity - this ensures a stronger protective coating that can be applied to the product as a homogeneous layer.
4. Industrial reproduction - because most glazing agents are used on commercial goods and therefore large quantities of glazing agent may be needed.
There are different variations of glazing agents, depending on the product, but they are all designed for the same purpose. |
https://en.wikipedia.org/wiki/Drive%20bay | A drive bay is a standard-sized area for adding hardware to a computer. Most drive bays are fixed to the inside of a case, but some can be removed.
Over the years since the introduction of the IBM PC, it and its compatibles have had many form factors of drive bays. Four form factors are in common use today, the 5.25-inch, 3.5-inch, 2.5-inch or 1.8-inch drive bays. These names do not refer to the width of the bay itself, but rather to the width of the disks used by the drives mounted in these bays.
Form factors
8.0-inch
8.0-inch drive bays were found in early IBM computers, CP/M computers, and the TRS-80 Model II. They were high, wide, and approximately deep, and were used for hard disk drives and floppy disk drives. This form factor is obsolete.
5.25-inch
5.25-inch drive bays are divided into two height specifications, full-height and half-height.
Full-height bays were found in old PCs in the early to mid-1980s. They were high, wide, and up to deep, used mainly for hard disk drives and floppy disk drives. This is the size of the internal (screwed) part of the bay, as the front side is actually . The difference between those widths and the name of the bay size is because it is named after the size of floppy that would fit in those drives, a 5.25-inch-wide square.
Half-height drive bays are high by wide, and are the standard housing for CD and DVD drives in modern computers. They were sometimes used for other things in the past, including hard disk drives (roughly between 10 and 100 MB) and floppy disk drives. As the name indicates, two half-height devices can fit in one full-height bay. Often represented as 5.25-inch, these floppy disk drives are obsolete.
The dimensions of a 5.25-inch floppy drive are specified in the SFF standard specifications which were incorporated into the EIA-741 "Specification for Small Form Factor 133.35 mm (5.25 in) Disk Drives" by the Electronic Industries Association (EIA). Dimensions of 5.25 optical drives are specifi |
https://en.wikipedia.org/wiki/IOS%2014 | iOS 14 is the fourteenth major release of the iOS mobile operating system developed by Apple Inc. for their iPhone and iPod Touch lines. Announced at the company's Worldwide Developers Conference on June 22, 2020 as the successor to iOS 13, it was released to the public on September 16, 2020. It was succeeded by iOS 15 on September 20, 2021.
System features
App Clips
App Clips are a new feature expanding on the functionality of the App Store. Intended as a dynamic feature rather than a permanently installed app, App Clips are extremely pared-back with very few OS permissions. At the time of the announcement, only the use of Apple Pay and Sign in with Apple were shown.
App Clips may be discovered in person via NFC tags (iPhone 7 or newer) or QR codes with App Clips branding. They may also be shared via Messages, or placed on websites or Maps.
CarPlay
CarPlay was updated to allow users to set a built-in wallpaper. Route management in Apple Maps was extended with features alerting the user to available stops, such as parking and food ordering. Additionally, route planning for electric vehicles now considers the location of charging stations.
Car keys
Car keys allow an iPhone to act as a virtual car key using NFC technology with compatible cars. The first compatible car showcased by Apple at the WWDC 2020 was the 2021 BMW 5 Series. Keys are accessible from the Wallet app. Keys may be shared; sharing may be temporary or given restrictions. In the event the iPhone is out of battery, car keys can still be accessed via the power reserve of the iPhone for about five hours. Car keys require an iPhone released in 2018 or later.
Home screen
Unlike previous versions, in which icons on the home screen were rearranged in order and corresponded directly to apps, users may add app icons and newly-introduced app widgets; pages may be added or deleted at will. This allows users to hide infrequently used apps and avoid clutter.
Widgets
To the left of the first page, the Today V |
https://en.wikipedia.org/wiki/Windows%20Phone | Windows Phone (WP) is a discontinued family of mobile operating systems developed by Microsoft for smartphones as the replacement successor to Windows Mobile and Zune. Windows Phone featured a new user interface derived from the Metro design language. Unlike Windows Mobile, it was primarily aimed at the consumer market rather than the enterprise market.
It was first launched in October 2010 with Windows Phone 7. Windows Phone 8 succeeded it in 2012, replacing the Windows CE-based kernel of Windows Phone 7 with the Windows NT kernel used by the PC versions of Windows (and, in particular, a large amount of internal components from Windows 8). Due to these changes, the OS was incompatible with all existing Windows Phone 7 devices, although it still supported apps originally developed for Windows Phone 7. In 2014, Microsoft released the Windows Phone 8.1 update, which introduced the Cortana virtual assistant, and Windows Runtime platform support to create cross-platform apps between Windows PCs and Windows Phone.
In 2015, Microsoft released Windows 10 Mobile, which promoted increased integration and unification with its PC counterpart, including the ability to connect devices to an external display or docking station to display a PC-like interface. Although Microsoft dropped the Windows Phone brand at this time in order to focus more on synergies with Windows 10 for PCs, it was still a continuation of the Windows Phone line from a technical standpoint, and updates were issued for selected Windows Phone 8.1 devices.
While Microsoft's investments in the platform were headlined by a major partnership with Nokia (whose Lumia series of smartphones, including the Lumia 520 in particular, would represent the majority of Windows Phone devices sold by 2013) and Microsoft's eventual acquisition of the company's mobile device business for just over US$7 billion (which included Nokia's then-CEO Stephen Elop joining Microsoft to lead its in-house mobile division), the duopoly of |
https://en.wikipedia.org/wiki/Super%20black | Super black is a surface treatment developed at the National Physical Laboratory (NPL) in the United Kingdom. It absorbs approximately 99.6% of visible light at normal incidence, while conventional black paint absorbs about 97.5%. At other angles of incidence, super black is even more effective: at an angle of 45°, it absorbs 99.9% of light.
Technology
The technology to create super black involves chemically etching a nickel-phosphorus alloy.
Applications of super black are in specialist optical instruments for reducing unwanted reflections. The disadvantage of this material is its low optical thickness, as it is a surface treatment. As a result, infrared light of a wavelength longer than a few micrometers penetrates through the dark layer and has much higher reflectivity. The reported spectral dependence increases from about 1% at 3 µm to 50% at 20 µm.
In 2009, a competitor to the super black material, Vantablack, was developed based on carbon nanotubes. It has a relatively flat reflectance in a wide spectral range.
In 2011, NASA and the US Army began funding research in the use of nanotube-based super black coatings in sensitive optics.
Nanotube-based superblack arrays and coatings have recently become commercially available.
See also
Vantablack
Emissivity
Black hole
Black body |
https://en.wikipedia.org/wiki/Rs16891982 | In genetics, rs16891982, also known as F374L, is the name for a single nucleotide polymorphism found in the SLC45A2 gene. The SNP consists of two alleles: C (cytosine) and G (guanine). It is associated with skin tone and hair/eye color. It is a type of missense mutation.
C allele homozygosity is associated with black hair in people of European descent, although those with this genotype are usually of non-European descent.
C/G allele heterozygosity is associated with black hair in people of European descent
G allele homozygosity is associated with light skin, hair, and eye color (European ancestry), those with this genotype also have a slightly higher susceptibility to melanoma. |
https://en.wikipedia.org/wiki/Thermodynamic%20versus%20kinetic%20reaction%20control | Thermodynamic reaction control or kinetic reaction control in a chemical reaction can decide the composition in a reaction product mixture when competing pathways lead to different products and the reaction conditions influence the selectivity or stereoselectivity. The distinction is relevant when product A forms faster than product B because the activation energy for product A is lower than that for product B, yet product B is more stable. In such a case A is the kinetic product and is favoured under kinetic control and B is the thermodynamic product and is favoured under thermodynamic control.
The conditions of the reaction, such as temperature, pressure, or solvent, affect which reaction pathway may be favored: either the kinetically controlled or the thermodynamically controlled one. Note this is only true if the activation energy of the two pathways differ, with one pathway having a lower Ea (energy of activation) than the other.
Prevalence of thermodynamic or kinetic control determines the final composition of the product when these competing reaction pathways lead to different products. The reaction conditions as mentioned above influence the selectivity of the reaction - i.e., which pathway is taken.
Asymmetric synthesis is a field in which the distinction between kinetic and thermodynamic control is especially important. Because pairs of enantiomers have, for all intents and purposes, the same Gibbs free energy, thermodynamic control will produce a racemic mixture by necessity. Thus, any catalytic reaction that provides product with nonzero enantiomeric excess is under at least partial kinetic control. (In many stoichiometric asymmetric transformations, the enantiomeric products are actually formed as a complex with the chirality source before the workup stage of the reaction, technically making the reaction a diastereoselective one. Although such reactions are still usually kinetically controlled, thermodynamic control is at least possible, in prin |
https://en.wikipedia.org/wiki/Transfer%20orbit | Orbits
Astrodynamics
In orbital mechanics, a transfer orbit is an intermediate elliptical orbit that is used to move a spacecraft in an orbital maneuver from one circular, or largely circular, orbit to another.
There are several types of transfer orbits, which vary in their energy efficiency and speed of transfer. These include:
Hohmann transfer orbit, an elliptical orbit used to transfer a spacecraft between two circular orbits of different altitudes in the same plane
Bi-elliptic transfer, a slower method of transfer, but one that may be more efficient than a Hohmann transfer orbit
Geostationary transfer orbit or geosynchronous transfer orbit is usually also a Hohmann transfer orbit
Lunar transfer orbit is an orbit that touches Low Earth orbit and a lunar orbit. |
https://en.wikipedia.org/wiki/Total%20Recorder | Total Recorder is digital audio editor software from High Criteria, Inc. which is able to record any sound coming through a computer soundcard. In addition to recording through a soundcard, Total Recorder is able to record digital sound directly through its virtual sound driver. This driver provides an advantage of recording audio reproduced by an external program (including Internet broadcasts) directly in digital format, i.e. without digital-analog-digital conversions leading to loss of quality, and even in those cases when a computer soundcard has no loop-back line (e.g. Stereo Mix, "What you hear"). Total Recorder is a shareware program. Evaluation version of Total Recorder is a fully functional version of the program, with the exception that an audible noise will be inserted about every 60 seconds.
Editions
Total Recorder is offered in four editions: Standard, Professional, Developer and VideoPro (beginning with the version 8.0 of Total Recorder). In dependence of edition different functionalities are available.
Supported formats
Video
AVI
WMV (Windows Media Video)
FLV (Flash Video)
MOV (for playback only)
MPEG-4
3GP
Audio
PCM (uncompressed)
RIFF-WAV (Compressed and uncompressed)
MP3
WMA
Ogg Vorbis
FLAC (Monkey's Audio)
APE
AAC audio within MPEG-4
3GP
Features
Recording of audio and video from different sources, sound format conversion, sound editing, sound processing, background recording (ripping), timeshifting, cue-sheet file supporting, splitting, tagging, file name generation, etc.
Add-ons
The following free and trial add-ons are available: Audio Restoration Add-On, Automatic Gain Control and Speech Enhancement Add-On, Spectrum Analyzer and Graphic Equalizer Add-On, Digital Mixer Add-On, Ogg Vorbis Support Add-On, Send to iTunes/iPod Add-On, Streaming Add-on. They can be used for removing distortions such as clicks, crackles, providing high-quality restoration of audio recorded from LPs, tapes and microphones, enhancing the reco |
https://en.wikipedia.org/wiki/Victor%20Negus | Sir Victor Ewings Negus, MS, FRCS (6 February 1887 – 15 July 1974) was a British surgeon who specialised in laryngology and also made fundamental contributions to comparative anatomy with his work on the structure and evolution of the larynx. He was born and educated in London, studying at King's College School, then King's College London, followed by King's College Hospital. The final years of his medical training were interrupted by the First World War, during which he served with the Royal Army Medical Corps. After the war, he qualified as a surgeon and studied with laryngologists in France and the USA before resuming his career at King's College Hospital where he became a junior surgeon in 1924.
In the 1920s, Negus worked on aspects of both throat surgery and the anatomy of the larynx, the latter work contributing to his degree of Master of Surgery (1924). His surgical innovations included designs for laryngoscopes, bronchoscopes, oesophagoscopes, an operating table, and tracheotomy equipment. His major publications were The Mechanism of the Larynx (1929) and his work on the clinical text Diseases of the Nose and Throat, starting with the fourth edition of 1937. Negus was also awarded several lectureships and published many medical papers and other works on comparative anatomy and laryngology. He became a senior surgeon at King's College Hospital in 1940 and a consulting surgeon in 1946.
Negus was one of the founders of the British Association of Otorhinolaryngologists, helping to establish his speciality as a discipline within the newly formed National Health Service. He was a member of numerous international and national otolaryngology organisations, and presided over the Fourth International Congress of Otolaryngology in London in 1949. In this period of his career following the Second World War he also worked on the anatomy of the paranasal sinuses, and played a key role in rebuilding and establishing collections of animal dissections used by comparative a |
https://en.wikipedia.org/wiki/Cirsium%20arvense | Cirsium arvense is a perennial species of flowering plant in the family Asteraceae, native throughout Europe and western Asia, northern Africa and widely introduced elsewhere. The standard English name in its native area is creeping thistle. It is also commonly known as Canada thistle and field thistle.
The plant is beneficial for pollinators that rely on nectar. It also was a top producer of nectar sugar in a 2016 study in Britain, with a second-place ranking due to a production per floral unit of ().
Alternative names
A number of other names are used in other areas or have been used in the past, including: Canadian thistle, lettuce from hell thistle, California thistle, corn thistle, cursed thistle, field thistle, green thistle, hard thistle, perennial thistle, prickly thistle, setose thistle, small-flowered thistle, way thistle, and stinger-needles. Canada and Canadian thistle are in wide use in the United States, despite being a misleading designation (it is not of Canadian origin).
Description
Cirsium arvense is a C3 carbon fixation plant. The C3 plants originated during Mesozoic and Paleozoic eras, and tend to thrive in areas where sunlight intensity is moderate, temperatures are moderate, and ground water is plentiful. plants lose 97% of the water taken up through their roots to transpiration.
Creeping thistle is a herbaceous perennial plant growing up to 150 cm, forming extensive clonal colonies from thickened roots that send up numerous erect shoots during the growing season. It is a ruderal species.
Given its adaptive nature, Cirsium arvense is one of the worst invasive weeds worldwide. Through comparison of its genetic expressions, the plant evolves differently with respect to where it has established itself. Differences can be seen in their R-protein mediated defenses, sensitivities to abiotic stresses, and developmental timing.
Taxonomy
Cirsium arvense is placed in the subtribe Carduinae, tribe Cardueae of the family Asteraceae. Unlike other sp |
https://en.wikipedia.org/wiki/57%20%28number%29 | 57 (fifty-seven) is the natural number following 56 and preceding 58.
In mathematics
Fifty-seven is the sixteenth discrete semiprime (specifically, the sixth distinct semiprime of the form , where is a higher prime). It also forms the fourth discrete semiprime pair with 58.
57 is the third Blum integer since its two prime factors (3 and 19) are both Gaussian primes. 57 has an aliquot sum of 23, which makes it the tenth number to contain a prime aliquot sum. This also makes 57 the first composite member of the 23-aliquot tree (..., 57, 23, 1, 0). The only other numbers to generate an aliquot sum of 57 are 99, 159, 343, 559, and 703; where 343 is the cube of 7, and 703 the sum of the first thirty-seven nonzero integers. Fifty seven is also a repdigit in base-7 (111).
57 is the fifth Leyland number, as it can be written in the form:
57 is the number of compositions of 10 into distinct parts.
57 is the seventh fine number, equivalently the number of ordered rooted trees with seven nodes having root of even degree.
57 is also the number of nodes in a regular octagon when all of its diagonals are drawn, and the first non-trivial icosagonal (20-gonal) number.
In geometry, there are:
57 non-prismatic uniform star polyhedra in 3-space, including four Kepler-Poinsot star polyhedra that are regular.
57 vertices and hemi-dodecahedral facets in the 57-cell, a 4-dimensional abstract regular polytope.
57 uniform prismatic 5-polytopes in the fifth dimension based on four different finite prismatic families, and inclusive of one special non-Wythoffian figure: the grand antiprism prism.
57 uniform prismatic 6-polytopes in the sixth dimension, as prisms of all non-prismatic uniform 5-polytopes.
The split Lie algebra E has a 57-dimensional Heisenberg algebra as its nilradical, and the smallest possible homogeneous space for E8 is also 57-dimensional.
57 lies between prime numbers 53 and 61, which are the only two prime numbers less than 71 that do not divide the order of |
https://en.wikipedia.org/wiki/Elchanan%20Mossel | Elchanan Mossel () is a professor of mathematics at the Massachusetts Institute of Technology. His primary research fields are probability theory, combinatorics, and statistical inference.
Research
Mossel's research spans a number of topics across mathematics, statistics, economics, and computer science, including combinatorial statistics, discrete function inequalities, isoperimetry, game theory, social choice, computational complexity, and computational evolutionary biology.
His work on discrete Fourier analysis and functions with low influence includes important contributions such as the proof of the "Majority is Stablest" conjecture, together with Ryan O'Donnell and Krzysztof Oleszkiewicz, and the proof of the optimality of the Goemans–Williamson MAX-CUT algorithm (assuming the Unique Games Conjecture), with Subhash Khot, Guy Kindler and Ryan O’Donnell.
Mossel has worked on the reconstruction problem on trees.
He connected it to Steel's conjecture in Phylogenetic reconstruction, partially in work with Constantinos Daskalakis and Sébastien Roch.
These result links the extremality of the Ising model on the Bethe lattice to a phase transition in the amount of data required for statistical inference on phylogenetic trees.
With Joe Neeman and Allan Sly he established the role of the reconstruction problem on trees for the problem of detection in block models.
Education and career
Mossel graduated from the Open University of Israel in 1992 with a B.Sc. in mathematics. In 2000, he received his Ph.D. in mathematics from the Hebrew University. Mossel held a postdoctoral position at Microsoft Research and was a Miller Research Fellow at UC Berkeley before becoming a Professor at UC Berkeley, the Weizmann Institute, the University of Pennsylvania and finally MIT.
Mossel is a prolific scholar, with more than 100 coauthors and over 150 papers listed in MathSciNet as of 2022. He has advised 10 graduate students who have subsequently held faculty positions at UCLA, |
https://en.wikipedia.org/wiki/Oil%20skimmer | An oil skimmer is a device that is designed to remove oil floating on a liquid surface from oil spills. The effectiveness of a skimmer is highly dependent on the roughness of the surrounding water that it is working on: the more choppy the surrounding wake and water, the more water the oil skimmer will take in along with the oil, rather than take in oil alone. Skimmers can be self-propelled, used from shore, or operated from vessels, with the best choice being dependent on the specifics for the job at hand. Skimmers can be pressed into use for a number of applications other than oil spills, with the correct type to use again being dependent on the nature of the intended application. Examples of possible uses include making skimmers one component of oily water treatment systems, removing oil from machine tool coolant and aqueous parts washers, and collecting fats, oils, and greases as part of wastewater treatment efforts for food manufacturing industries.
Oil skimmers are different from swimming pool sanitation skimmers, which are designed for a similar but unrelated purpose.
Oil skimmers were used to great effect to assist in the remediation of the Exxon Valdez oil spill in 1989.
Limitation and design factors
There are many different types of oil skimmer. Each type has different design features and therefore results in different applications and use. It is important to understand the design features before employing a particular skimmer type.
Some factors to consider are:
Oil removal flow rate: Alternative Skimmer designs have different oil removal flow rates. Volume removal rates for Oleophilic skimmer types (drum, brush, disc, belt) are comparatively low. Weir type skimmers are capable of very high oil and water removal rates. ASTM F2709 standard establishes the test procedure for determining oil recovery rate (ORR).
Oil removal concentration: It is a common misconception that oil skimmers remove concentrated or pure 'oil'; when in fact they remove a mixture |
https://en.wikipedia.org/wiki/Reinforced%20lipids | Reinforced lipids are lipid molecules in which some of the fatty acids contain deuterium instead of hydrogen. They can be used for the protection of living cells by slowing the chain reaction due to isotope effect on lipid peroxidation. The lipid bilayer of the cell and organelle membranes contain polyunsaturated fatty acids (PUFA) are key components of cell and organelle membranes. Any process that either increases oxidation of PUFAs or hinders their ability to be replaced can lead to serious disease. Correspondingly, use of reinforced lipids that stop the chain reaction of lipid peroxidation has preventive and therapeutic potential.
Examples of reinforced lipids
There is a number of polyunsaturated fatty acids that can be reinforced by deuterisation. They include (the names of the reinforces deuterised versions are separated by a slash):
linoleic acid / D2-linoleic acid (D2-Lin)
α-linolenic acid / D4-α-linolenic acid (D4-Lnn)
arachidonic acid / D6-arachidonic acid (D6-ARA)
eicosapentaenoic acid / D8-eicosapentaenoic acid (D8-EPA)
docosahexaenoic acid / D10-docosahexaenoic acid (D10-DHA)
Mechanism of action
Hydrogen is a chemical element with an atomic number of 1. It has just one proton and one electron. Deuterium is the heavier naturally occurring, non-radioactive, stable isotope of hydrogen. Deuterium contains one proton, one electron, and a neutron, effectively doubling the mass of the deuterium isotope without changing its properties significantly. Substituting deuterium for hydrogen yields deuterated compounds that are similar in size and shape to hydrogen-based compounds.
One of the most pernicious and irreparable types of oxidative damage inflicted by reactive oxygen species (ROS) upon biomolecules involves the carbon-hydrogen bond cleavage (hydrogen abstraction). In theory, replacing hydrogen with deuterium "reinforces" the bond due to the kinetic isotope effect, and such reinforced biomolecules taken up by the body will be more resistant to ROS |
https://en.wikipedia.org/wiki/InterWorking%20Labs | InterWorking Labs is a privately owned company in Scotts Valley, California, in the business of optimizing application performance for applications and embedded systems. Founded in 1993 by Chris Wellens and Marshall Rose, it was the first company formed specifically to test network protocol compliance. Its products and tests allow computer devices from many different companies to communicate over networks.
Products
InterWorking Labs' Products diagnose, replicate, and re-mediate application performance problems.
The company's first product, SilverCreek, tests a Simple Network Management Protocol (SNMP) agent implementation (switch, server, phone) with hundreds of thousands of individual tests, including conformance, stress, robustness, and negative testing. The tests detect and diagnose implementation errors in private and standard MIBs as well as SNMPv1, v2c, and v3 stacks and implementations.
The Maxwell family products emulate real-world networks, with problems such as delays, rerouting, corruption, impaired packets or protocols, Domain Name System delays or limited bandwidth. New impairments are added to Maxwell using C, C++, or Python extensions. It is controlled via graphical, command line, and script interfaces. It supports a set of protocol impairments for TCP/IP, DHCP, ICMP, TLS, and SIP testing.
The Maxwell products are named after Maxwell's Demon, a thought experiment by 19th-century physicist James Clerk Maxwell. Maxwell’s Demon demonstrated that the Second Law of Thermodynamics—which says that entropy increases—is true only on average. In his thought experiment, Maxwell imagined a double chamber with a uniform mixture of hot and cold gas molecules. A demon (some intelligent being) sits between the two chambers operating a trap door. Every time a cold (low-energy) molecule comes by, the demon opens the door and lets the molecule through to the other side. Eventually, the cold gas molecules are all on one side of the chamber and the hot ones all on t |
https://en.wikipedia.org/wiki/Viral%20transport%20medium | Viral transport medium (VTM) is a solution used to preserve virus specimens after collection so that they can be transported and analysed in a laboratory at a later time. Unless stored in an ultra low temperature freezer or in liquid nitrogen, virus samples, and especially RNA virus samples, are prone to degradation. However, such cooling equipment is seldom available in the field due to their cumbersome size, weight, and in the case of freezers, high energy consumption. Hence, there is a need for VTM; a chemical preservative that can be used at ambient temperature. Chemical components may include saline solution, phosphate-buffered saline (PBS), or fetal bovine serum (FBS). VTM must be sterile to avoid introducing contamination to the specimen.
In the United States, the FDA and CDC publish guidelines for VTM production. |
https://en.wikipedia.org/wiki/Astronomia%20nova | Astronomia nova (English: New Astronomy, full title in original Latin: ) is a book, published in 1609, that contains the results of the astronomer Johannes Kepler's ten-year-long investigation of the motion of Mars.
One of the most significant books in the history of astronomy, the Astronomia nova provided strong arguments for heliocentrism and contributed valuable insight into the movement of the planets. This included the first mention of the planets' elliptical paths and the change of their movement to the movement of free floating bodies as opposed to objects on rotating spheres. It is recognized as one of the most important works of the Scientific Revolution.
Background
Prior to Kepler, Nicolaus Copernicus proposed in 1543 that the Earth and other planets orbit the Sun. The Copernican model of the Solar System was regarded as a device to explain the observed positions of the planets rather than a physical description.
Kepler sought for and proposed physical causes for planetary motion. His work is primarily based on the research of his mentor, Tycho Brahe. The two, though close in their work, had a tumultuous relationship. Regardless, in 1601 on his deathbed, Brahe asked Kepler to make sure that he did not "die in vain," and to continue the development of his model of the Solar System. Kepler would instead write the Astronomia nova, in which he rejects the Tychonic system, as well as the Ptolemaic system and the Copernican system. Some scholars have speculated that Kepler's dislike for Brahe may have had a hand in his rejection of the Tychonic system and formation of a new one.
By 1602, Kepler set to work on determining the orbit pattern of Mars, keeping David Fabricius informed of his progress. He suggested the possibility of an oval orbit to Fabricius by early 1604, though was not believed. Later in the year, Kepler wrote back with his discovery of Mars's elliptical orbit. The manuscript for Astronomia nova was completed by September 1607, and was in pr |
https://en.wikipedia.org/wiki/CE%20marking | On commercial products, the presence of the logo means that the manufacturer or importer affirms the goods' conformity with European health, safety, and environmental protection standards. It is not a quality indicator or a certification mark. The CE marking is required for goods sold in the European Economic Area (EEA); goods sold elsewhere may also carry the mark.
The mark indicates that the product may be traded freely in any part of the European Economic Area, regardless of its country of origin. It consists of the CE logo and, if applicable, the four digit identification number of the notified body involved in the conformity assessment procedure. "CE" is the abbreviation of (French for "European conformity").
Meaning
The mark on a product indicates that the manufacturer or importer of that product affirms its compliance with the relevant EU legislation and the product may be sold anywhere in the European Economic Area (EEA). It is a criminal offence to affix a mark to a product that is not compliant or offer it for sale.
For example, most electrical products must comply with the Low Voltage Directive and the EMC Directive, among others; toys must comply with the Toy Safety Directive. (The Low Voltage Directive is about electrical safety; EMC or Electromagnetic Compatibility means the device will work as intended without interfering with, or being affected by, the use or function of any other device.) The mark indicates compliance with as many norms (directives and regulations) as apply at the time of the declaration of compliance (see below). In the case of electrical products, several later norms such as the Restriction of Hazardous Substances Directive (RoHS) and Waste Electrical and Electronic Equipment Directive (WEEE) are relevant in addition to the Low Voltage Directive and EMC Directive. The exact significance of the mark therefore depends on when it was applied to a specific unit.
The marking does not indicate EEA manufacture or that the EU o |
https://en.wikipedia.org/wiki/Cellius | was a Japanese video game developer and publisher headquartered in Shibuya, Tokyo, founded in 2007 as a joint venture between Sony and Bandai Namco Holdings. The aim of the company was to "help take share from Microsoft Corp. and Nintendo Co." Sony hoped that the company would make up for the losses it made during quarter two of its financial year. Ken Kutaragi was announced as CEO. Bandai Namco Holdings held 51% of the company, and Sony held 49%. The company planned to use Sony's Cell microprocessor, the heart of the PlayStation 3, for PlayStation 3 games and games for mobile phones and personal computers. Its only video game project was the poorly-received Ridge Racer on the PlayStation Vita.
Notes |
https://en.wikipedia.org/wiki/Bernhard%20Steffen%20%28computer%20scientist%29 | Bernhard Steffen (born 31 May 1958 in Kiel, West Germany) is a German computer scientist and professor at the TU Dortmund University, Germany. His research focuses on various facets of formal methods ranging from program analysis and verification, to workflow synthesis, to test-based modeling, and machine learning.
After his PhD at the University of Kiel he spent two years as a research fellow at the LFCS (Edinburgh, Scotland) where he co-developed the Edinburgh Concurrency Workbench and authored one of the earliest papers on how to adequately model probabilistic processes, before joining the University of Aarhus in 1989 as a postdoc. From 1990 to 1992 he was associate professor at the RWTH Aachen, before he became full professor at the University of Passau. Since 1997 he holds the chair of programming systems at TU Dortmund University where he was Dean of Computer Science between 2002 and 2006 as well as a member of the Senate in 2006 and 2007. In Dortmund he developed the concept of active automata learning to towards a practical means for model-based testing that does not require any a priori models. Recently his interest shifted towards the application of formal methods for explaining machine learning.
His conceptual background comprises abstract interpretation, computer-aided verification and explanation, automata learning, and the development of domain-specific languages that guarantee properties by design. This is witnessed by receiving the Most Influential PLDI Paper Award for Lazy Code Motion, which is given 10 years in retrospective, and the CAV Artifact Award for the Open-Source LearnLib. Finally, in 2019 he was awarded the title of Honorary Professor of the AMITY School of Engineering and Technology.
Furthermore, Steffen is founding Editor in Chief of STTT, Co-Founder of TACAS, ETAPS, ISoLA, RERS and member of the editorial board of LNCS.
Journal and conference foundations
Bernhard Steffen co-founded the following journals and conferences
Tools |
https://en.wikipedia.org/wiki/Namco%20System%20357 | The Namco System 357 is an arcade system board based on the Sony PlayStation 3. It was released in 2007 as the board for Tekken 6. Unlike its predecessor, it did not see widespread adoption by other manufacturers.
In 2011, Namco released an upgraded version of the arcade board, the System 369, also known as System 359, for use with Tekken Tag Tournament 2. The System 369's specifications are the same as Slim model retail PS3 consoles, and as such feature a smaller chassis with less heat.
Specifications
CPU: Cell Broadband Engine consisting of one 3.2GHz PowerPC-based Power Processing Element (PPE) and six Synergistic Processing Elements (SPEs).
System memory: 256MB XDR DRAM
Graphics: NVIDIA GeForce 7800-based RSX 'Reality Synthesizer' with 256MB GDDR3 RAM
Media: internal HDD used for storage of games.
UI: Sony XMB
List of System 357 / System 359 / System 369 games
Dark Escape 3D (2012)
Dark Escape 4D (2014)
Deadstorm Pirates (2010)
Deadstorm Pirates Special Edition (2014)
Dragon Ball ZENKAI Battle Royale (2011)
Dragon Ball ZENKAI Battle Royale - Super Saiyan Awakening (2012)
Mobile Suit Gundam: Extreme Vs. (2010)
Mobile Suit Gundam: Extreme Vs. Full Boost (2012)
Mobile Suit Gundam: Extreme Vs. Maxi Boost (2014)
Razing Storm (2009)
Taiko no Tatsujin (2011)
Taiko no Tatsujin (C/N: KATSU-DON) (2012)
Taiko no Tatsujin Sorairo ver. (2013)
Taiko no Tatsujin Momoiro ver. (2013)
Taiko no Tatsujin Kimidori ver. (2014)
Taiko no Tatsujin Murasaki ver. (2015)
Taiko no Tatsujin White ver. (2015)
Taiko no Tatsujin Red ver. (2016)
Taiko no Tatsujin Yellow ver. (2017)
Taiko no Tatsujin Blue ver. (2018)
Taiko no Tatsujin Green ver. (2019)
Tekken 6 (2007)
Tekken 6: Bloodline Rebellion (2008)
Tekken Tag Tournament 2 (2011)
Tekken Tag Tournament 2 Unlimited (2012) |
https://en.wikipedia.org/wiki/Mass%20cytometry | Mass cytometry is a mass spectrometry technique based on inductively coupled plasma mass spectrometry and time of flight mass spectrometry used for the determination of the properties of cells (cytometry). In this approach, antibodies are conjugated with isotopically pure elements, and these antibodies are used to label cellular proteins. Cells are nebulized and sent through an argon plasma, which ionizes the metal-conjugated antibodies. The metal signals are then analyzed by a time-of-flight mass spectrometer. The approach overcomes limitations of spectral overlap in flow cytometry by utilizing discrete isotopes as a reporter system instead of traditional fluorophores which have broad emission spectra.
Commercialization
Tagging technology and instrument development occurred at the University of Toronto and DVS Sciences, Inc. CyTOF (cytometry by time of flight) was initially commercialized by DVS Sciences in 2009. In 2014, Fluidigm acquired DVS Sciences to become a reference company in single cell technology. In 2022 Fluidigm received a capitol infusion and changed its name to Standard BioTools. The CyTOF, CyTOF2, Helios (CyTOF3) and CyTOF XT(4th generation) have been commercialized up to now. Fluidigm sells a variety of commonly used metal-antibody conjugates, and an antibody conjugation kit.
Imaging Mass Cytometry (IMC)
Imaging mass cytometry (IMC) is a relatively new imaging technique, emerged from previously available CyTOF technology (cytometry by time of flight), that combines mass spectrometry with UV laser ablation to generate pseudo images of tissue samples. This approach adds spatial resolution to the data, which enables simultaneous analysis of multiple cell markers at subcellular resolution and their spatial distribution in tissue sections. The IMC approach, in the same way as CyTOF, relies on detection of metal-tagged antibodies using time-of-flight mass spectrometry, allowing for quantification of up to 40 markers simultaneously.
Data analysis
Cy |
https://en.wikipedia.org/wiki/Glycosyltransferase | Glycosyltransferases (GTFs, Gtfs) are enzymes (EC 2.4) that establish natural glycosidic linkages. They catalyze the transfer of saccharide moieties from an activated nucleotide sugar (also known as the "glycosyl donor") to a nucleophilic glycosyl acceptor molecule, the nucleophile of which can be oxygen- carbon-, nitrogen-, or sulfur-based.
The result of glycosyl transfer can be a carbohydrate, glycoside, oligosaccharide, or a polysaccharide. Some glycosyltransferases catalyse transfer to inorganic phosphate or water. Glycosyl transfer can also occur to protein residues, usually to tyrosine, serine, or threonine to give O-linked glycoproteins, or to asparagine to give N-linked glycoproteins. Mannosyl groups may be transferred to tryptophan to generate C-mannosyl tryptophan, which is relatively abundant in eukaryotes. Transferases may also use lipids as an acceptor, forming glycolipids, and even use lipid-linked sugar phosphate donors, such as dolichol phosphates in eukaryotic organism, or undecaprenyl phosphate in bacteria.
Glycosyltransferases that use sugar nucleotide donors are Leloir enzymes, after Luis F. Leloir, the scientist who discovered the first sugar nucleotide and who received the 1970 Nobel Prize in Chemistry for his work on carbohydrate metabolism. Glycosyltransferases that use non-nucleotide donors such as dolichol or polyprenol pyrophosphate are non-Leloir glycosyltransferases.
Mammals use only 9 sugar nucleotide donors for glycosyltransferases: UDP-glucose, UDP-galactose, UDP-GlcNAc, UDP-GalNAc, UDP-xylose, UDP-glucuronic acid, GDP-mannose, GDP-fucose, and CMP-sialic acid. The phosphate(s) of these donor molecules are usually coordinated by divalent cations such as manganese, however metal independent enzymes exist.
Many glycosyltransferases are single-pass transmembrane proteins, and they are usually anchored to membranes of Golgi apparatus
Mechanism
Glycosyltransferases can be segregated into "retaining" or "inverting" enzymes according to |
https://en.wikipedia.org/wiki/Soil%20food%20web | The soil food web is the community of organisms living all or part of their lives in the soil. It describes a complex living system in the soil and how it interacts with the environment, plants, and animals.
Food webs describe the transfer of energy between species in an ecosystem. While a food chain examines one, linear, energy pathway through an ecosystem, a food web is more complex and illustrates all of the potential pathways. Much of this transferred energy comes from the sun. Plants use the sun’s energy to convert inorganic compounds into energy-rich, organic compounds, turning carbon dioxide and minerals into plant material by photosynthesis. Plant flowers exude energy-rich nectar above ground and plant roots exude acids, sugars, and ectoenzymes into the rhizosphere, adjusting the pH and feeding the food web underground.
Plants are called autotrophs because they make their own energy; they are also called producers because they produce energy available for other organisms to eat. Heterotrophs are consumers that cannot make their own food. In order to obtain energy they eat plants or other heterotrophs.
Above ground food webs
In above ground food webs, energy moves from producers (plants) to primary consumers (herbivores) and then to secondary consumers (predators). The phrase, trophic level, refers to the different levels or steps in the energy pathway. In other words, the producers, consumers, and decomposers are the main trophic levels. This chain of energy transferring from one species to another can continue several more times, but eventually ends. At the end of the food chain, decomposers such as bacteria and fungi break down dead plant and animal material into simple nutrients.
Methodology
The nature of soil makes direct observation of food webs difficult. Since soil organisms range in size from less than 0.1 mm (nematodes) to greater than 2 mm (earthworms) there are many different ways to extract them. Soil samples are often taken using a metal |
https://en.wikipedia.org/wiki/Mechanica | Mechanica (; 1736) is a two-volume work published by mathematician Leonhard Euler which describes analytically the mathematics governing movement.
Euler both developed the techniques of analysis and applied them to numerous problems in mechanics,
notably in later publications the calculus of variations. Euler's laws of motion expressed scientific laws of Galileo and Newton in terms of points in reference frames and coordinate systems making them useful for calculation when the statement of a problem or example is slightly changed from the original.
Newton–Euler equations express the dynamics of a rigid body. Euler has been credited with contributing to the rise of Newtonian mechanics especially in topics other than gravity. |
https://en.wikipedia.org/wiki/Kronecker%20delta | In mathematics, the Kronecker delta (named after Leopold Kronecker) is a function of two variables, usually just non-negative integers. The function is 1 if the variables are equal, and 0 otherwise:
or with use of Iverson brackets:
For example, because , whereas because .
The Kronecker delta appears naturally in many areas of mathematics, physics, engineering and computer science, as a means of compactly expressing its definition above.
In linear algebra, the identity matrix has entries equal to the Kronecker delta:
where and take the values , and the inner product of vectors can be written as
Here the Euclidean vectors are defined as -tuples: and and the last step is obtained by using the values of the Kronecker delta to reduce the summation over .
It is common for and to be restricted to a set of the form or , but the Kronecker delta can be defined on an arbitrary set.
Properties
The following equations are satisfied:
Therefore, the matrix can be considered as an identity matrix.
Another useful representation is the following form:
This can be derived using the formula for the geometric series.
Alternative notation
Using the Iverson bracket:
Often, a single-argument notation is used, which is equivalent to setting :
In linear algebra, it can be thought of as a tensor, and is written . Sometimes the Kronecker delta is called the substitution tensor.
Digital signal processing
In the study of digital signal processing (DSP), the unit sample function represents a special case of a 2-dimensional Kronecker delta function where the Kronecker indices include the number zero, and where one of the indices is zero. In this case:
Or more generally where:
However, this is only a special case. In tensor calculus, it is more common to number basis vectors in a particular dimension starting with index 1, rather than index 0. In this case, the relation does not exist, and in fact, the Kronecker delta function and the unit sample function are d |
https://en.wikipedia.org/wiki/Caret | Caret is the name used familiarly for the character , (the circumflex and a circumflex accent) provided on most QWERTY keyboards by typing . The symbol has a variety of uses in programming and mathematics. The name "caret" arose from its visual similarity to the original proofreader's caret, a mark used in proofreading to indicate where a punctuation mark, word, or phrase should be inserted into a document. The formal ASCII standard (X3.64.1977) calls it a "circumflex".
History
Typewriters
On typewriters designed for languages that routinely use diacritics (accent marks), there are two possible ways to type these: keys can be dedicated to precomposed characters (with the diacritic included); alternatively a dead key mechanism can be provided. With the latter, a mark is made when a dead key is typed but, unlike normal keys, the paper carriage does not move on and thus the next letter to be typed is printed under the accent. The symbol was originally provided in typewriters and computer printers so that circumflex accents could be overprinted on letters (as in or ).
Transposition into ISO/IEC 646 and ASCII
The incorporation of the circumflex symbol into ASCII is a consequence of this prior existence on typewriters: this symbol did not exist independently as a type or hot-lead printing character. The original 1963 version of the ASCII standard used the code point x5E for an . However, the 1965 ISO/IEC646 standard defined code point x5E as one of five available for national variation, with the circumflex diacritic as the default and the up-arrow as one of the alternative uses. In 1967, the second revision of ASCII followed suit.
Overprinting to add an accent mark was not always supported well by printers, and was almost never possible on video terminals. Instead, precomposed characters were eventually created to show the accented letters. The freestanding circumflex (which had come to be called a caret) quickly became reused for many other purposes, such as |
https://en.wikipedia.org/wiki/List%20of%20mathematical%20topics%20in%20relativity | This is a list of mathematical topics in relativity, by Wikipedia page.
Special relativity
Foundational issues
principle of relativity
speed of light
faster-than-light
biquaternion
conjugate diameters
four-vector
four-acceleration
four-force
four-gradient
four-momentum
four-velocity
hyperbolic orthogonality
hyperboloid model
light-like
Lorentz covariance
Lorentz group
Lorentz transformation
Lorentz–FitzGerald contraction hypothesis
Minkowski diagram
Minkowski space
Poincaré group
proper length
proper time
rapidity
relativistic wave equations
relativistic mass
split-complex number
unit hyperbola
world line
General relativity
black holes
no-hair theorem
Hawking radiation
Hawking temperature
Black hole entropy
charged black hole
rotating black hole
micro black hole
Schwarzschild black hole
Schwarzschild metric
Schwarzschild radius
Reissner–Nordström black hole
Immirzi parameter
closed timelike curve
cosmic censorship hypothesis
chronology protection conjecture
Einstein–Cartan theory
Einstein's field equation
geodesic
gravitational redshift
Penrose–Hawking singularity theorems
Pseudo-Riemannian manifold
stress–energy tensor
worm hole
Cosmology
anti-de Sitter space
Ashtekar variables
Batalin–Vilkovisky formalism
Big Bang
Cauchy horizon
cosmic inflation
cosmic microwave background
cosmic variance
cosmological constant
dark energy
dark matter
de Sitter space
Friedmann–Lemaître–Robertson–Walker metric
horizon problem
large-scale structure of the cosmos
Randall–Sundrum model
warped geometry
Weyl curvature hypothesis
Relativity
Mathematics |
https://en.wikipedia.org/wiki/One%20Small%20Step%3A%20The%20Story%20of%20the%20Space%20Chimps | One Small Step: The Story of the Space Chimps is a 2008 documentary film produced and directed by David Cassidy and Kristin Davy which aired on History Channel UK and CBC Television. The film chronicles the real story behind the early use of chimpanzees in space exploration. The film was released on DVD in April 2008, after several delays. Cassidy is best known for co-producing the 2006 documentary Shut Up and Sing on the Dixie Chicks.
Release
One Small Step: The Story of the Space Chimps is a documentary that screened in over 20 film festivals including the Maryland Film Festival, the Vancouver International Film Festival, and the Hot Springs Documentary Film Festival, before airing on television and being released on DVD.
Plot
Told through archival photos and footage, space historians, testimony from the chimpanzees' trainers, and through the people who fought for the space chimpanzees' peaceful retirement, the film explores the compelling journey of the United States Air Force space from their primate predecessors and early rocket tests to Ham and Enos as they made their ground breaking missions into space.
The story reveals the space chimpanzees' triumphs and tragedies and brings to light the virtually unknown account of how the colony was rewarded for their long and challenging service to NASA, the Air Force, and the United States.
Featured in the documentary are interviews with Dr. Carole Noon who heads up the Save the Chimps sanctuary, Dr. Jane Goodall, and archival footage of President John F. Kennedy's famous 1962 space exploration speech "We choose to go to the Moon". The film also recounts the stories of many early primate missions including those of Able and Baker, and Gordo.
See also
Animals in space
Monkeys and apes in space |
https://en.wikipedia.org/wiki/Schottky%20effect | The Schottky effect or field enhanced thermionic emission is a phenomenon in condensed matter physics named after Walter H. Schottky. In electron emission devices, especially electron guns, the thermionic electron emitter will be biased negative relative to its surroundings. This creates an electric field of magnitude F at the emitter surface. Without the field, the surface barrier seen by an escaping Fermi-level electron has height W equal to the local work-function. The electric field lowers the surface barrier by an amount ΔW, and increases the emission current. It can be modeled by a simple modification of the Richardson equation, by replacing W by (W − ΔW). This gives the equation
where J is the emission current density, T is the temperature of the metal, W is the work function of the metal, k is the Boltzmann constant, qe is the Elementary charge, ε0 is the vacuum permittivity, and AG is the product of a universal constant A0 multiplied by a material-specific correction factor λR which is typically of order 0.5.
Electron emission that takes place in the field-and-temperature-regime where this modified equation applies is often called Schottky emission. This equation is relatively accurate for electric field strengths lower than about 108 V m−1. For electric field strengths higher than 108 V m−1, so-called Fowler–Nordheim (FN) tunneling begins to contribute significant emission current. In this regime, the combined effects of field-enhanced thermionic and field emission can be modeled by the Murphy–Good equation for thermo-field (T-F) emission. At even higher fields, FN tunneling becomes the dominant electron emission mechanism, and the emitter operates in the so-called "cold field electron emission (CFE)" regime.
Thermionic emission can also be enhanced by interaction with other forms of excitation such as light. For example, excited Cs-vapours in thermionic converters form clusters of Cs-Rydberg matter which yield a decrease of collector emitting work fu |
https://en.wikipedia.org/wiki/Alpha%2021064 | The Alpha 21064 is a microprocessor developed and fabricated by Digital Equipment Corporation that implemented the Alpha (introduced as the Alpha AXP) instruction set architecture (ISA). It was introduced as the DECchip 21064 before it was renamed in 1994. The 21064 is also known by its code name, EV4. It was announced in February 1992 with volume availability in September 1992. The 21064 was the first commercial implementation of the Alpha ISA, and the first microprocessor from Digital to be available commercially. It was succeeded by a derivative, the Alpha 21064A in October 1993. This last version was replaced by the Alpha 21164 in 1995.
History
The first Alpha processor was a test chip codenamed EV3. This test chip was fabricated using Digital's 1.0-micrometre (μm) CMOS-3 process. The test chip lacked a floating point unit and only had 1 KB caches. The test chip was used to confirm the operation of the aggressive circuit design techniques. The test chip, along with simulators and emulators, was also used to bring up firmware and the various operating systems that the company supported.
The production chip, codenamed EV4, was fabricated using Digital's 0.75 μm CMOS-4 process. Dirk Meyer and Edward McLellan were the micro-architects. Ed designed the issue logic while Dirk designed the other major blocks. Jim Montanaro led the circuit implementation. The EV3 was used in the Alpha Demonstration Unit (ADU), a multiprocessor system used by Digital to develop software for the Alpha platform before the availability of EV4 parts.
The 21064 was unveiled at the 39th International Solid-State Circuits Conference (ISSCC) in mid-February 1992. It was announced on 25 February 1992, with a 150 MHz sample introduced on the same day. It was priced at $3,375 in quantities of 100, $1,650 in quantities between 100 and 1,000, and $1,560 for quantities over 1,000. Volume shipments began in September 1992.
In early February 1993, the price of the 150 MHz version was reduced to $1,0 |
https://en.wikipedia.org/wiki/Excavation%20%28medicine%29 | In medicine, excavation has two meanings:
the act of hollowing out
the space hollowed out, or a natural cavity or pouch
Examples of the latter include:
Rectouterine pouch or excavation, between the uterus and the rectum
Rectovesical excavation, between the rectum and the male bladder
Vesicouterine excavation, between the bladder and the uterus in a female
Medical terminology |
https://en.wikipedia.org/wiki/Ursula%20Martin | Ursula Hilda Mary Martin (born 3 August 1953) is a British computer scientist, with research interests in theoretical computer science and formal methods. She is also known for her activities aimed at encouraging women in the fields of computing and mathematics. Since 2019, she has served as a professor at the School of Informatics, University of Edinburgh.
From 20142018, Martin was a Professor of Computer Science in the Department of Computer Science at the University of Oxford, and holds an EPSRC Established Career Fellowship. Prior to this she held a chair of Computer Science in the School of Electronic Engineering and Computer Science at Queen Mary, University of London, where she was Vice-Principal of Science and Engineering, 20052009.
Education
Martin was born in London on 3 August 1953 to Anne Louise (née Priestman) and Captain Geoffrey Richard Martin. She was educated at Abbey College at Malvern Wells. In 1975 she graduated with an MA from Girton College, Cambridge, and in 1979 with a PhD from the University of Warwick, both in mathematics.
Career and research
Martin began in mathematics working in group theory, later moving into string rewriting systems. She has held academic posts at University of Illinois at Urbana-Champaign, the University of Manchester and Royal Holloway, University of London. She has made sabbatical visits to Massachusetts Institute of Technology and SRI International (Menlo Park). In 2004 she was a visiting fellow at the Oxford Internet Institute.
From 1992 to 2002, Martin was Professor of Computer Science at the University of St Andrews in Scotland. She was the first female professor at the University since its foundation in 1411.
From 2003 to 2005, Martin was seconded to the University of Cambridge Computer Laboratory part-time and served as the director of the Women@CL project to lead local, national and international initiatives for women in computing, supported by Microsoft Research and Intel Cambridge Research. She was a F |
https://en.wikipedia.org/wiki/Pollination%20of%20orchids | The pollination of orchids is a complex chapter in the biology of this family of plants that are distinguished by the complexity of their flowers and by intricate ecological interactions with their pollinator agents. It has captured the attention of numerous scientists over time, including Charles Darwin, father of the theory of evolution by natural selection. Darwin published in 1862 the first observations of the fundamental role of insects in orchid pollination, in his book The Fertilization of Orchids. Darwin stated that the varied stratagems orchids use to attract their pollinators transcend the imagination of any human being.
Adaptations of orchids to pollination by animals
97% of species of orchids need a pollinator for the transfer of pollen grains from one plant to the pistils of another individual to take place, and thus for fertilization and seeds formation to occur. The pollen of orchids is grouped in compact masses called pollinia (singular: "pollinium"), so that by itself or by wind action the pollen cannot disperse from one flower to another, so pollinators are essential to ensure their sexual reproduction. These pollinators are very varied and, depending on the species in question, may be flies, mosquitos, bees, wasps, butterflies, coleopterans, and birds (especially hummingbirds).
The zoophily that characterizes orchids presupposes that pollinating animals visit the flowers regularly and stop at them long enough; that they brush or touch the anthers and stigma with some frequency and that the pollen remains attached to the visitors so perfectly that it can reach with due safety the stigmas of other flowers. The result of zoophily depends essentially on the animals being able to recognize flowers from a certain distance and being attracted to flowers of the same species. Zoophilous flowers, then, must possess "attractive products" (baits, such as pollen and nectar), "means of attraction" (such as scents and colors) and, in addition, viscous or adh |
https://en.wikipedia.org/wiki/Boilerplate%20code | In computer programming, boilerplate code, or simply boilerplate, are sections of code that are repeated in multiple places with little to no variation. When using languages that are considered verbose, the programmer must write a lot of boilerplate code to accomplish only minor functionality.
The need for boilerplate can be reduced through high-level mechanisms such as metaprogramming (which has the computer automatically write the needed boilerplate code or insert it at compile time), convention over configuration (which provides good default values, reducing the need to specify program details in every project) and model-driven engineering (which uses models and model-to-code generators, eliminating the need for manual boilerplate code).
Origin
The term arose from the newspaper business. Columns and other pieces that were distributed by print syndicates were sent to subscribing newspapers in the form of prepared printing plates. Because of their resemblance to the metal plates used in the making of boilers, they became known as "boiler plates", and their resulting text—"boilerplate text". As the stories that were distributed by boiler plates were usually "fillers" rather than "serious" news, the term became synonymous with unoriginal, repeated text.
A related term is bookkeeping code, referring to code that is not part of the business logic but is interleaved with it in order to keep data structures updated or handle secondary aspects of the program.
Preamble
One form of boilerplate consists of declarations which, while not part of the program logic or the language's essential syntax, are added to the start of a source file as a matter of custom. The following Perl example demonstrates boilerplate:
#!/usr/bin/perl
use warnings;
use strict;
The first line is a shebang, which identifies the file as a Perl script that can be executed directly on the command line on Unix/Linux systems. The other two are pragmas turning on warnings and strict mode, which are |
https://en.wikipedia.org/wiki/Micvac | The MicVac method is a pasteurization method for the food processing industry producing ready meals. Food ingredients are inserted into a package, which is sealed. The food is then cooked and pasteurized in the sealed package. During the heating process a valve on the package opens and releases steam and oxygen. The process time is short since the food is pasteurized with microwaves. When the microwave heating process stops the valve closes. Remaining steam condenses and causes an underpressure in the package. The final result is a cooked, pasteurized and vacuum-packed product. The short cooking time in combination with the absence of oxygen in the pack is unique for the method and has many advantages. The method has been used in the food industry since 2005.
The inventor of the technique is Dr. Joel Haamer and the company MicVac was founded in August 2000. Already in the 1970s, when Haamer tried to find a solution for better taking care of mussels, which are heat-sensitive and easily get rancid, he got the first ideas to the method. In the end of the 1990s, the microwave technology had reached such a progress that Haamer's idea was patented. During 5 years the development work took place and then, in 2005, the method was commercialized. |
https://en.wikipedia.org/wiki/Digital%20firm | The Digital Firm is a kind of organization that has enabled core business relationships through digital networks In these digital networks are supported by enterprise class technology platforms that have been leveraged within an organization to support critical business functions and services. Some examples of these technology platforms are Customer Relationship Management (CRM), Supply Chain Management (SCM), Enterprise Resource Planning (ERP), Knowledge Management System (KMS), Enterprise Content Management (ECM), and Warehouse Management System (WMS) among others. The purpose of these technology platforms is to digitally enable seamless integration and information exchange within the organization to employees and outside the organization to customers, suppliers, and other business partners.
History
Origin of "The Digital Firm"
The term "Digital Firm" originated, as a concept in a series of Management Information Systems (MIS) books authored by Kenneth C. Laudon. It provides a new way to describe organizations that operate differently than the traditional brick and mortar business as a result of broad sweeping changes in technology and global markets. Digital firms place an emphasis on the digitization of business processes and services through sophisticated technology and information systems. These information systems create opportunities for digital firms to decentralize operations, accelerate market readiness and responsiveness, enhance customer interactions, as well as increase efficiencies across a variety of business functions.
Acceleration of technology adoption
Technology adoption has been increasing as digital firms continually look to achieve greater levels cost savings, competitive advantage, and operational performance optimization. As organizations adopt technology, the internal appetite for additional technologies increases and in some cases accelerates. This acceleration of technology adoption by digital firms creates a "digital divide". Emerg |
https://en.wikipedia.org/wiki/Bialgebra | In mathematics, a bialgebra over a field K is a vector space over K which is both a unital associative algebra and a counital coassociative coalgebra. The algebraic and coalgebraic structures are made compatible with a few more axioms. Specifically, the comultiplication and the counit are both unital algebra homomorphisms, or equivalently, the multiplication and the unit of the algebra both are coalgebra morphisms. (These statements are equivalent since they are expressed by the same commutative diagrams.)
Similar bialgebras are related by bialgebra homomorphisms. A bialgebra homomorphism is a linear map that is both an algebra and a coalgebra homomorphism.
As reflected in the symmetry of the commutative diagrams, the definition of bialgebra is self-dual, so if one can define a dual of B (which is always possible if B is finite-dimensional), then it is automatically a bialgebra.
Formal definition
(B, ∇, η, Δ, ε) is a bialgebra over K if it has the following properties:
B is a vector space over K;
there are K-linear maps (multiplication) ∇: B ⊗ B → B (equivalent to K-multilinear map ∇: B × B → B) and (unit) η: K → B, such that (B, ∇, η) is a unital associative algebra;
there are K-linear maps (comultiplication) Δ: B → B ⊗ B and (counit) ε: B → K, such that (B, Δ, ε) is a (counital coassociative) coalgebra;
compatibility conditions expressed by the following commutative diagrams:
Multiplication ∇ and comultiplication Δ
where τ: B ⊗ B → B ⊗ B is the linear map defined by τ(x ⊗ y) = y ⊗ x for all x and y in B,
Multiplication ∇ and counit ε
Comultiplication Δ and unit η
Unit η and counit ε
Coassociativity and counit
The K-linear map Δ: B → B ⊗ B is coassociative if .
The K-linear map ε: B → K is a counit if .
Coassociativity and counit are expressed by the commutativity of the following two diagrams (they are the duals of the diagrams expressing associativity and unit of an algebra):
Compatibility conditions
The four commutative diagrams can be |
https://en.wikipedia.org/wiki/Adjusted%20Peak%20Performance | Adjusted Peak Performance (APP) is a metric introduced by the U.S. Department of Commerce's Bureau of Industry and Security (BIS) to more accurately predict the suitability of a computing system to complex computational problems, specifically those used in simulating nuclear weapons. This is used to determine the export limitations placed on certain computer systems under the Export Administration Regulations 15 CFR.
Further details can be found in the document "Practitioner's Guide To Adjusted Peak Performance".
The (simplified) algorithm used to calculate APP consists of the following steps:
Determine how many 64 bit (or better) floating point operations every processor in the system can perform per clock cycle (best case). This is FPO(i).
Determine the clock frequency of every processor. This is F(i).
Choose the weighting factor for each processor: 0.9 for vector processors and 0.3 for non-vector processors. This is W(i).
Calculate the APP for the system as follows: APP = FPO(1) * F(1) * W(1) + ... + FPO(n) * F(n) * W(n).
The metric was introduced in April 2006 to replace the Composite Theoretical Performance (CTP) metric which was introduced in 1993. APP was itself replaced in November 2007 when the BIS amended 15 CFR to include the December 2006 Wassenaar Arrangement Plenary Agreement Implementation's new metric - Gigaflops (GFLOPS), one billion floating point operations per second, or TeraFLOPS, one trillion floating point operations per second.
The unit of measurement is Weighted TeraFLOPS (WT) to specify Adjusted Peak Performance (APP).
The weighting factor is 0.3 for non-vector processors and 0.9 for vector processors. For example, a PowerPC 750 running at 800 MHz would be rated at 0.00024 WT due to being able to execute one floating point instruction per cycle and not having a vector unit. Note that only 64 bit (or wider) floating point instructions count.
Notes:
Processors without 64 bit (or better) floating point support have an FPO of zero |
https://en.wikipedia.org/wiki/Sea%20state | In oceanography, sea state is the general condition of the free surface on a large body of water—with respect to wind waves and swell—at a certain location and moment. A sea state is characterized by statistics, including the wave height, period, and spectrum. The sea state varies with time, as the wind and swell conditions change. The sea state can be assessed either by an experienced observer (like a trained mariner) or by using instruments like weather buoys, wave radar or remote sensing satellites.
In the case of buoy measurements, the statistics are determined for a time interval in which the sea state can be considered to be constant. This duration has to be much longer than the individual wave period, but shorter than the period in which the wind and swell conditions can be expected to vary significantly. Typically, records of one hundred to one thousand wave periods are used to determine the wave statistics.
The large number of variables involved in creating and describing the sea state cannot be quickly and easily summarized, so simpler scales are used to give an approximate but concise description of conditions for reporting in a ship's log or similar record.
WMO sea state code
The World Meteorological Organization (WMO) sea state code largely adopts the 'wind sea' definition of the Douglas Sea Scale.
The direction from which the swell is coming should be recorded.
Sea states in marine engineering
In engineering applications, sea states are often characterized by the following two parameters:
The significant wave height H1/3 — the mean wave height of the highest third of the waves.
The mean wave period, T1.
In addition to the short-term wave statistics presented above, long-term sea state statistics are often given as a joint frequency table of the significant wave height and the mean wave period. From the long and short-term statistical distributions, it is possible to find the extreme values expected in the operating life of a ship. A ship de |
https://en.wikipedia.org/wiki/Kamp%C3%A9%20de%20F%C3%A9riet%20function | In mathematics, the Kampé de Fériet function is a two-variable generalization of the generalized hypergeometric series, introduced by Joseph Kampé de Fériet.
The Kampé de Fériet function is given by
Applications
The general sextic equation can be solved in terms of Kampé de Fériet functions.
See also
Appell series
Humbert series
Lauricella series (three-variable) |
https://en.wikipedia.org/wiki/Pallister%E2%80%93Killian%20syndrome | The Pallister–Killian syndrome (PKS), also termed tetrasomy 12p mosaicism or the Pallister mosaic aneuploidy syndrome, is an extremely rare and severe genetic disorder. PKS is due to the presence of an extra and abnormal chromosome termed a small supernumerary marker chromosome (sSMC). sSMCs contain copies of genetic material from parts of virtually any other chromosome and, depending on the genetic material they carry, can cause various genetic disorders and neoplasms. The sSMC in PKS consists of multiple copies of the short (i.e. "p") arm of chromosome 12. Consequently, the multiple copies of the genetic material in the sSMC plus the two copies of this genetic material in the two normal chromosome 12's are overexpressed and thereby cause the syndrome. Due to a form of genetic mosaicism, however, individuals with PKS differ in the tissue distributions of their sSMC and therefore show different syndrome-related birth defects and disease severities. For example, individuals with the sSMC in their heart tissue are likely to have cardiac structural abnormalities while those without this sSMC localization have a structurally normal heart.
PKS was first described by Philip Pallister in 1977 and further researched by Maria Teschler-Nicola and Wolfgang Killian in 1981.
Presentation
Individuals with PKS present prenatally or at birth with multiple birth defects. These defects include: brain atrophy, agenesis of the corpus callosum, polymicrogyria of the brain, and/or spot calcifications in the brain's lateral sulcus; deafness and/or blindness; autonomic nervous system dysfunctions such as anhidrosis, hypohidrosis, and/or episodic spells of hyperventilation interspersed with breath-holding; symptoms of spinal cord malformations; profound or less commonly mild to severe intellectual disability; epileptic seizures; heart and/or anal defects; diaphragmatic hernias; marked muscle weakness; supernumerary nipples; abnormal facial features such as frontal bossing, high frontal ha |
https://en.wikipedia.org/wiki/Diamond%20Vision | Diamond Vision (known as Aurora Vision in Japan) displays are large-scale video walls for indoor and outdoor sports venues and commercial applications, produced by the Mitsubishi Electric Corporation. Diamond Vision Systems is a division of Mitsubishi Electric Power Products, Inc. and is headquartered in Warrendale, Pennsylvania, where certain products are designed and assembled for the North American market.
Diamond Vision video screens incorporate technologies developed by Mitsubishi Electric. For wide viewing angles, Diamond Vision screens utilize chip-type LEDs. Mitsubishi Electric also uses patented processing technology in Diamond Vision boards for imagery and color reproduction. Early Diamond Vision displays used Cathode-ray tube technology, similarly to Jumbotrons.
Diamond Vision's awards include an Emmy Award and the Best in Sports Technology award from the Sports Business Journal–Sports Business Daily. In addition, Diamond Vision installations have been recognized five times by Guinness World Records.
Diamond Vision Systems is the Official Large Outdoor Video Display Provider of the PGA Tour.
Historical milestones
Mitsubishi Electric began manufacturing and installing large-scale video screens in 1980, with the introduction of the first Diamond Vision board at Dodger Stadium in Los Angeles for the 1980 Major League Baseball All-Star Game.
Diamond Vision installed 3 screens at the JFK Stadium in Philadelphia, and two in Wembley Stadium during Live Aid. They would be used to show the live feed from Philadelphia to Wembley and Wembley to Philadelphia.
In 2004, Diamond Vision Systems installed North America’s largest indoor high-definition (HD) screen, measuring 34 feet by 110 feet, at the Colosseum at Caesars Palace in Las Vegas, Nevada. The following year, the Diamond Vision LED display at Turner Field (now Center Parc Stadium) in Atlanta was recognized by Guinness World Records as the world’s largest high-definition television screen.
Diamond Visio |
https://en.wikipedia.org/wiki/Decompensation | In medicine, decompensation is the functional deterioration of a structure or system that had been previously working with the help of compensation. Decompensation may occur due to fatigue, stress, illness, or old age. When a system is "compensated," it is able to function despite stressors or defects. Decompensation describes an inability to compensate for these deficiencies. It is a general term commonly used in medicine to describe a variety of situations.
Medical Term
For example, cardiac decompensation may refer to the failure of the heart to maintain adequate blood circulation, after long-standing (previously compensated) vascular disease (see heart failure). Short-term treatment of cardiac decompensation can be achieved through administration of dobutamine, resulting in an increase in heart contractility via an inotropic effect.
Kidney failure can also occur following a slow degradation of kidney function due to an underlying untreated illness; the symptoms of the latter can then become much more severe due to the lack of efficient compensation by the kidney.
Psychology
In psychology, the term refers to an individual's loss of healthy defense mechanisms in response to stress, resulting in personality disturbance or psychological imbalance. |
https://en.wikipedia.org/wiki/Sachs%E2%80%93Wolfe%20effect | The Sachs–Wolfe effect, named after Rainer K. Sachs and Arthur M. Wolfe, is a property of the cosmic microwave background radiation (CMB), in which photons from the CMB are gravitationally redshifted, causing the CMB spectrum to appear uneven. This effect is the predominant source of fluctuations in the CMB for angular scales larger than about ten degrees.
Non-integrated Sachs–Wolfe effect
The non-integrated Sachs–Wolfe effect is caused by gravitational redshift occurring at the surface of last scattering. The effect is not constant across the sky due to differences in the matter/energy density at the time of last scattering.
Integrated Sachs–Wolfe effect
The integrated Sachs–Wolfe (ISW) effect is also caused by gravitational redshift, but it occurs between the surface of last scattering and the Earth, so it is not part of the primordial CMB. It occurs when the Universe is dominated in its energy density by something other than matter. If the Universe is dominated by matter, then large-scale gravitational potential energy wells and hills do not evolve significantly. If the Universe is dominated by radiation, or by dark energy, though, those potentials do evolve, subtly changing the energy of photons passing through them.
There are two contributions to the ISW effect. The "early-time" ISW occurs immediately after the (non-integrated) Sachs–Wolfe effect produces the primordial CMB, as photons course through density fluctuations while there is still enough radiation around to affect the Universe's expansion. Although it is physically the same as the late-time ISW, for observational purposes it is usually lumped in with the primordial CMB, since the matter fluctuations that cause it are in practice undetectable.
Late-time integrated Sachs–Wolfe effect
The "late-time" ISW effect arises quite recently in cosmic history, as dark energy, or the cosmological constant, starts to govern the Universe's expansion. Unfortunately, the nomenclature is a bit confusing |
https://en.wikipedia.org/wiki/Frontiers%20%281989%20TV%20series%29 | Frontiers is an eight-part BBC television series, and accompanying book, that explored the geographic boundaries between countries. Eight writers and journalists in a variety of countries investigated the economic, political, geographical and historical reasons that account for why people are divided. The series was aired in 1989, just a few months before the fall of the Berlin Wall, which was featured in one episode.
Episodes
"Natural Break": Frederic Raphael explored the Pyrenees, the frontier between France and Spain, which at the time was preparing to join the (then) European Economic Community.
"Gone Tomorrow": John Wells covered the Iron Curtain that split East and West Germans.
"Gold and the Gun": Nadine Gordimer visited the war-torn border area between Mozambique and her native South Africa.
"Night and Day": Richard Rodriguez showed how the rich North and poor South converged at the US/Mexican border.
"Long Division": Ronald Eyre looked at the people living on both sides of the border in Ireland that splits the Republic from Ulster.
"Big Brother's Bargain": Nigel Hamilton hiked up the boundary between Russia and Finland.
"Border Run": Jon Swain visited the Thai/Cambodian border where thousands of Cambodian refugees had been stranded for over ten years.
"Cyprus: Stranded in Time": Christopher Hitchens investigated the divided island of Cyprus.
Further reading
Frontiers, published in 1990 by BBC Books,
External links
1989 British television series debuts
1989 British television series endings
1980s British documentary television series
BBC television documentaries
Borders
English-language television shows |
https://en.wikipedia.org/wiki/Stall%20%28engine%29 | A stall is the slowing or stopping of a process and in the case of an engine refers to a sudden stopping of the engine turning, usually brought about accidentally.
It is commonly applied to the phenomenon whereby an engine abruptly ceases operating and stops turning. It might be due to not getting enough air, energy, fuel, or electric spark, fuel starvation, a mechanical failure, or in response to a sudden increase in engine load. This increase in engine load is common in vehicles with a manual transmission when the clutch is released too suddenly.
The ways in which a car can stall are usually down to the driver, especially with a manual transmission. For instance, if a driver takes their foot off the clutch too quickly while stationary then the car will stall; taking the foot off the clutch slowly will stop this from happening. Stalling also happens when the driver forgets to depress the clutch and/or change to neutral while coming to a stop. Stalling can be dangerous, especially in heavy traffic.
A car fitted with an automatic transmission could also have its engine stalled when the vehicle is travelling in the opposite direction to the selected gear. For example, if the selector is in the 'D' position and the car is moving backwards, (on a steep enough hill to overcome the torque from the torque converter) the engine will stall, due to the fact that the engine is forced to turn in the opposite direction to what it is actually doing. This is because, hypothetically, if the car is rolling backward fast enough, the force from the rotating wheels will be transmitted backward through the transmission and act as a sudden load on the engine.
Digital electronics fuel injection and ECU ignition systems have greatly reduced stalling in modern engines.
See also
Stall torque |
https://en.wikipedia.org/wiki/Kialo | Kialo is an online structured debate platform with argument maps in the form of debate trees. It is a collaborative reasoning tool for thoughtful discussion, understanding different points of view, and collaborative decision-making, showing arguments for and against claims underneath user-submitted theses or questions.
The deliberative discourse platform is designed to present hundreds of supporting or opposing arguments in a dynamic argument tree and is streamlined for rational civil debate on topics such as philosophical questions, policy deliberations, entertainment, ethics, science questions, and unsolved problems or subjects of disagreement in general.
Argument-boxes are structured into hierarchical branches where the root is the main thesis (or theses) of the debate, enabling deliberation and navigable debates between opposing perspectives. A debate is divided into Pro (supporting) and Con (refuting or devaluing) columns where registered users can add arguments and rate the of the parent claim. The arguments .
Its argument tree structure enables detailed scrutiny of claims at all levels of the tree and allows users to for example quickly understand why a decision was made or which of the aggregated arguments swayed it this way. Newcomers can join a debate at any time and look back at the structured discussion history, and then weigh in at the right place with their new argument or their comment on a specific argument. The design presets a structure on debates "that allows participants to easily see, process, and ultimately assess the many facets of competing claims".
The word Kialo is Esperanto for "reason". The platform is the world's largest argument mapping and structured debate site.
Overview
Users can comment on every Pro or Con, for example for requesting sources or expansions. Recent activities of a debate are shown in a panel on the right side of the respective debate. Debates can be found through the search or on the Explore page through their |
https://en.wikipedia.org/wiki/Semi-infinite | In mathematics, semi-infinite objects are objects which are infinite or unbounded in some but not all possible ways.
In ordered structures and Euclidean spaces
Generally, a semi-infinite set is bounded in one direction, and unbounded in another. For instance, the natural numbers are semi-infinite considered as a subset of the integers; similarly, the intervals and and their closed counterparts are semi-infinite subsets of . Half-spaces and half-lines are sometimes described as semi-infinite regions.
Semi-infinite regions occur frequently in the study of differential equations. For instance, one might study solutions of the heat equation in an idealised semi-infinite metal bar.
A semi-infinite integral is an improper integral over a semi-infinite interval. More generally, objects indexed or parametrised by semi-infinite sets may be described as semi-infinite.
Most forms of semi-infiniteness are boundedness properties, not cardinality or measure properties: semi-infinite sets are typically infinite in cardinality and measure.
In optimization
Many optimization problems involve some set of variables and some set of constraints. A problem is called semi-infinite if one (but not both) of these sets is finite. The study of such problems is known as semi-infinite programming. |
https://en.wikipedia.org/wiki/Born%E2%80%93von%20Karman%20boundary%20condition | Born–von Karman boundary conditions are periodic boundary conditions which impose the restriction that a wave function must be periodic on a certain Bravais lattice. Named after Max Born and Theodore von Kármán, this condition is often applied in solid state physics to model an ideal crystal. Born and von Karman published a series of articles in 1912 and 1913 that presented one of the first theories of specific heat of solids based on the crystalline hypothesis and included these boundary conditions.
The condition can be stated as
where i runs over the dimensions of the Bravais lattice, the ai are the primitive vectors of the lattice, and the Ni are integers (assuming the lattice has N cells where N=N1N2N3). This definition can be used to show that
for any lattice translation vector T such that:
Note, however, the Born–von Karman boundary conditions are useful when Ni are large (infinite).
The Born–von Karman boundary condition is important in solid state physics for analyzing many features of crystals, such as diffraction and the band gap. Modeling the potential of a crystal as a periodic function with the Born–von Karman boundary condition and plugging in Schrödinger's equation results in a proof of Bloch's theorem, which is particularly important in understanding the band structure of crystals.
However, since any real crystal always has a finite size, the electronic states in the crystal do not satisfy the Born–von Karman boundary condition. Consequently, the conventional theory of electronic states in crystals based on the Bloch's theorem has some fundamental difficulties. |
https://en.wikipedia.org/wiki/Transplastomic%20plant | A transplastomic plant is a genetically modified plant in which genes are inactivated, modified or new foreign genes are inserted into the DNA of plastids like the chloroplast instead of nuclear DNA.
Currently, the majority of transplastomic plants are a result of chloroplast manipulation due to poor expression in other plastids. However, the technique has been successfully applied to the chromoplasts of tomatoes.
Chloroplasts in plants are thought to have originated from an engulfing event of a photosynthetic bacteria (cyanobacterial ancestor) by a eukaryote. There are many advantages to chloroplast DNA manipulation because of its bacterial origin. For example, the ability to introduce multiple genes (operons) in a single step instead of many steps and the simultaneous expression of many genes with its bacterial gene expression system. Other advantages include the ability to obtain organic products like proteins at a high concentration and the fact that production of these products will not be affected by epigenetic regulation.
The reason for product synthesis at high concentrations is because a single plant cell can potentially carry up to 100 chloroplasts. If all these plastids are transformed, all of them can express the introduced foreign genes. This is may be advantageous compared to transformation of the nucleus, because the nucleus typically contains only one or two copies of the gene.
The advantages provided by chloroplast DNA manipulation has seen growing interest into this field of research and development, particularly in agricultural and pharmaceutical applications. However, there are some limitations in chloroplast DNA manipulation, such as the inability to manipulate cereal crop DNA material and poor expression of foreign DNA in non- green plastids as mentioned before. In addition, the lack of post- translational modification capability like glycosylation in plastids may make some human- related protein expression difficult. Nevertheless, much pro |
https://en.wikipedia.org/wiki/OpenIndiana | OpenIndiana is a free and open-source illumos distribution descended from UNIX System V Release 4 via the OpenSolaris operating system. Forked from OpenSolaris after OpenSolaris was discontinued by Oracle Corporation, OpenIndiana takes its name from Project Indiana, the internal codename for OpenSolaris at Sun Microsystems before Oracle’s acquisition of Sun in 2010.
Created by a development team led by Alasdair Lumsden, the OpenIndiana project is now stewarded by the illumos Foundation, which develops and maintains the illumos operating system. The project aims to make OpenIndiana "the de facto OpenSolaris distribution installed on production servers where security and bug fixes are provided free of charge."
History
Origins
Project Indiana was originally conceived by Sun Microsystems, to construct a binary distribution around the OpenSolaris source code base. Project Indiana was led by Ian Murdock, founder of the Debian Linux distribution.
OpenIndiana was conceived after negotiations of a takeover of Sun Microsystems by Oracle were proceeding, in order to ensure continued availability and further development of an OpenSolaris-based OS, as it is widely used. Uncertainty among the OpenSolaris development community led some developers to form tentative plans for a fork of the existing codebase.
These plans came to fruition following the announcement of discontinuation of support for the OpenSolaris project by Oracle.
Initial reaction
The formal announcement of the OpenIndiana project was made on September 14, 2010, at the JISC Centre in London. The first release of the operating system was made available publicly at the same time, despite being untested. The reason for the untested release was that the OpenIndiana team set a launch date ahead of Oracle OpenWorld in order to beat the release of Solaris 11 Express.
The announcement of OpenIndiana was met with a mainly positive response; over 350 people viewed the online announcement, the ISO image was downloa |
https://en.wikipedia.org/wiki/Deformation%20%28engineering%29 | In engineering, deformation refers to the change in size or shape of an object. Displacements are the absolute change in position of a point on the object. Deflection is the relative change in external displacements on an object. Strain is the relative internal change in shape of an infinitesimally small cube of material and can be expressed as a non-dimensional change in length or angle of distortion of the cube. Strains are related to the forces acting on the cube, which are known as stress, by a stress-strain curve. The relationship between stress and strain is generally linear and reversible up until the yield point and the deformation is elastic. The linear relationship for a material is known as Young's modulus. Above the yield point, some degree of permanent distortion remains after unloading and is termed plastic deformation. The determination of the stress and strain throughout a solid object is given by the field of strength of materials and for a structure by structural analysis.
Engineering stress and engineering strain are approximations to the internal state that may be determined from the external forces and deformations of an object, provided that there is no significant change in size. When there is a significant change in size, the true stress and true strain can be derived from the instantaneous size of the object.
In the figure it can be seen that the compressive loading (indicated by the arrow) has caused deformation in the cylinder so that the original shape (dashed lines) has changed (deformed) into one with bulging sides. The sides bulge because the material, although strong enough to not crack or otherwise fail, is not strong enough to support the load without change. As a result, the material is forced out laterally. Internal forces (in this case at right angles to the deformation) resist the applied load.
The concept of a rigid body can be applied if the deformation is negligible.
Types of deformation
Depending on the type of materia |
https://en.wikipedia.org/wiki/Libellula | Libellula is a genus of dragonflies, commonly called skimmers, in the family Libellulidae, distributed throughout the temperate zone of the Northern Hemisphere. Most species are found in the United States, where they are the best-known large dragonflies, often seen flying over freshwater ponds in summer. Many have showy wing patterns.
Overview
The taxa Ladona (corporals) and Plathemis (whitetails) have been considered as synonyms of Libellula, subgenera, or separate genera by different authorities. Recent phylogenetic analysis has supported their status as either subgenera or full genera.
Species
List of species.
Extant species
Ladona
Plathemis
Fossils
Libellula brodieri†
Libellula calypso†
Libellula doris †
Libellula eusebioi†
Libellula kieseli†
Libellula martini†
Libellula melobasis†
Libellula pannewitziana†
Libellula perse†
Libellula sieboldiana†
Libellula thetis†
Libellula thoe†
Libellula ukrainensis† |
https://en.wikipedia.org/wiki/Conservation%20psychology | Conservation psychology is the scientific study of the reciprocal relationships between humans and the rest of nature, with a particular focus on how to encourage conservation of the natural world. Rather than a specialty area within psychology itself, it is a growing field for scientists, researchers, and practitioners of all disciplines to come together and better understand the Earth and what can be done to preserve it. This network seeks to understand why humans hurt or help the environment and what can be done to change such behavior. The term "conservation psychology" refers to any fields of psychology that have understandable knowledge about the environment and the effects humans have on the natural world. Conservation psychologists use their abilities in "greening" psychology and make society ecologically sustainable. The science of conservation psychology is oriented toward environmental sustainability, which includes concerns like the conservation of resources, conservation of ecosystems, and quality of life issues for humans and other species.
One common issue is a lack of understanding of the distinction between conservation psychology and the more-established field of environmental psychology, which is the study of transactions between individuals and all their physical settings, including how people change both the built and the natural environments and how those environments change them. Environmental psychology began in the late 1960s (the first formal program with that name was established at the City University of New York in 1968), and is the term most commonly used around the world. Its definition as including human transactions with both the natural and built environments goes back to its beginnings, as exemplified in these quotes from three 1974 textbooks: "Environmental psychology is the study of the interrelationship between behavior and the built and natural environment" and "...the natural environment is studied as both a problem area, wit |
https://en.wikipedia.org/wiki/Desulfovibrio%20paquesii | Desulfovibrio paquesii is a bacterium. It is sulfate-reducing and hydrogenotrophic. The type strain is SB1(T) (=DSM 16681(T)=JCM 14635(T)). |
https://en.wikipedia.org/wiki/Year%20loss%20table | A Year Loss Table (YLT) is a list of historical or simulated years, with financial losses for each year. YLTs are widely used in catastrophe modeling, as a way to record and communicate historical or simulated losses from catastrophes. The use of lists of years with historical or simulated financial losses is discussed in many references on catastrophe modelling and disaster risk management, but it is only more recently that the name YLT has become standard.
Overview
Year of interest
In a simulated YLT, each year of simulated loss is considered a possible loss outcome for a single year, the year of interest, which is usually in the future. In insurance industry catastrophe modelling, the year of interest is often this year or next year, due to the annual nature of many insurance contracts. However, the year can also be defined to be any year in the past or the future.
Events
Many YLTs are event based i.e., they are constructed from historical or simulated catastrophe events, each of which has an associated loss.
Each event is allocated to one or more years in the YLT and there may be multiple events in a year. The events may have an associated frequency model, that specifies the distribution for the number of different types of events per year, and an associated severity distribution, that specifies the distribution of loss for each event.
Events in an event-based YLT may all be of one peril-type (such as hurricane) or may be a mixture of peril-types (such as hurricane and earthquake).
Period Loss Tables (PLTs)
YLTs represent the possible losses in a period of one year, but can be generalized to represent the possible losses in any length of time, in which case they may be referred to as Period Loss Tables (PLTs).
Use in insurance
YLTs are widely used in the insurance industry, because they are a flexible way to store samples from a distribution of possible losses. Two properties, in particular, make them useful:
The number of events within a year ca |
https://en.wikipedia.org/wiki/Alternation%20of%20generations | Alternation of generations (also known as metagenesis or heterogenesis) is the predominant type of life cycle in plants and algae. In plants both phases are multicellular: the haploid sexual phase – the gametophyte – alternates with a diploid asexual phase – the sporophyte.
A mature sporophyte produces haploid spores by meiosis, a process which reduces the number of chromosomes to half, from two sets to one. The resulting haploid spores germinate and grow into multicellular haploid gametophytes. At maturity, a gametophyte produces gametes by mitosis, the normal process of cell division in eukaryotes, which maintains the original number of chromosomes. Two haploid gametes (originating from different organisms of the same species or from the same organism) fuse to produce a diploid zygote, which divides repeatedly by mitosis, developing into a multicellular diploid sporophyte. This cycle, from gametophyte to sporophyte (or equally from sporophyte to gametophyte), is the way in which all land plants and most algae undergo sexual reproduction.
The relationship between the sporophyte and gametophyte phases varies among different groups of plants. In the majority of algae, the sporophyte and gametophyte are separate independent organisms, which may or may not have a similar appearance. In liverworts, mosses and hornworts, the sporophyte is less well developed than the gametophyte and is largely dependent on it. Although moss and hornwort sporophytes can photosynthesise, they require additional photosynthate from the gametophyte to sustain growth and spore development and depend on it for supply of water, mineral nutrients and nitrogen. By contrast, in all modern vascular plants the gametophyte is less well developed than the sporophyte, although their Devonian ancestors had gametophytes and sporophytes of approximately equivalent complexity. In ferns the gametophyte is a small flattened autotrophic prothallus on which the young sporophyte is briefly dependent for its n |
https://en.wikipedia.org/wiki/Lament | A lament or lamentation is a passionate expression of grief, often in music, poetry, or song form. The grief is most often born of regret, or mourning. Laments can also be expressed in a verbal manner in which participants lament about something that they regret or someone that they have lost, and they are usually accompanied by wailing, moaning and/or crying. Laments constitute some of the oldest forms of writing, and examples exist across human cultures.
History
Many of the oldest and most lasting poems in human history have been laments. The Lament for Sumer and Ur dates back at least 4000 years to ancient Sumer, the world's first urban civilization. Laments are present in both the Iliad and the Odyssey, and laments continued to be sung in elegiacs accompanied by the aulos in classical and Hellenistic Greece. Elements of laments appear in Beowulf, in the Hindu Vedas, and in ancient Near Eastern religious texts. They are included in the Mesopotamian City Laments such as the Lament for Ur and the Jewish Tanakh, (which Christians refer to as the Old Testament).
In many oral traditions, both early and modern, the lament has been a genre usually performed by women: Batya Weinbaum made a case for the spontaneous lament of women chanters in the creation of the oral tradition that resulted in the Iliad The material of lament, the "sound of trauma" is as much an element in the Book of Job as in the genre of pastoral elegy, such as Shelley's "Adonais" or Matthew Arnold's "Thyrsis".
The Book of Lamentations or Lamentations of Jeremiah figures in the Old Testament. The Lamentation of Christ (under many closely variant terms) is a common subject from the Life of Christ in art, showing Jesus' dead body being mourned after the Crucifixion. Jesus himself lamented over the prospective fall of Jerusalem as he and his disciples entered the city ahead of his passion.
A lament in the Book of Lamentations or in the Psalms, in particular in the Lament/Complaint Psalms of the Tanak |
https://en.wikipedia.org/wiki/Polysaccharide%20A | Polysaccharide A (PSA) is a polysaccharide produced by the Gram-negative bacterium Bacteroides fragilis. B.fragilis produces eight identified distinct capsular polysaccharides, identified by the letters "A" through "H".
PSA colonization of B. fragilis in the gut mucosa induces regulatory T cells and suppresses pro-inflammatory T helper 17 cells. PSA has been shown to protect animals from experimental diseases like colitis, asthma, or pulmonary inflammation. Nonetheless, PSA can be pro-inflammatory as well as anti-inflammatory. |
https://en.wikipedia.org/wiki/Lancaster%20University%20School%20of%20Mathematics | Lancaster University School of Mathematics, also known as LUSoM, is a maths school located in Preston, Lancashire, England. As a maths school, it is a specialist mathematics free school sixth form college.
The school was set up by the Rigby Education Trust, a single-academy trust set up in partnership between the Lancaster University and Cardinal Newman College for the purpose of opening and operating the school. It opened to students in September 2022 and is located in a purpose-built £8.5 million school building on London Road, Preston.
The school is highly selective, with prospective students expected to have GCSE mathematics qualification at grade 8 or 9 and required to sit an admissions assessment. The course structure at LUSoM requires all students to study A-level Mathematics and Further Mathematics and a third A-level from either Physics, Chemistry or Computer Science. In addition, students may select a fourth subject from those three, or choose any other A-level subject to be taught at Cardinal Newman College, which is located less than half a mile from the school site. |
https://en.wikipedia.org/wiki/Associative%20interference | Associative interference is a cognitive theory established on the concept of associative learning, which suggests that the brain links related elements. When one element is stimulated, its associates can also be activated. The most known study demonstrating the credibility of this concept was Pavlov's experiment in 1927 which was later developed into the learning procedure known as classical conditioning.
However, whilst classical conditioning and associative learning both explore how the brain utilizes this cognitive association to benefit us, studies have also shown how the brain can mistakenly associate related, but incorrect elements together, and this is known as associative interference. A simple example of this would be when one was asked a series of multiplication questions. A study conducted in 1985 showed that over 90% of the mistakes subjects made were actually answers to other questions with a common multiplicand. That is, questions such as 4 x 6 = 24 and 3 x 8 = 24 were very likely to promote errors (8 x 4 = 24) due to associative interference.
Associative interference was widely investigated and researchers realized there were different types of interference, namely retroactive interference which investigates how new memories disrupts the retrieval of old memories, and proactive interference which investigates how old memories disrupts the retrieval of new memories. These two were subsequently known as the interference theory.
Therefore, associative interference is a fundamental theory which the interference theory draws upon. The essential difference between these two is time. Both retroactive and proactive interference are concerned with when the interfering elements, or memories were obtained. However, associative interference does not encompass time, as shown by the previous example. The chronological acquisition of the four times table in relation to the three times table is independent as to why subjects made an error, highlighting the differe |
https://en.wikipedia.org/wiki/Event%20generator | Event generators are software libraries that generate simulated high-energy particle physics events.
They randomly generate events as those produced in particle accelerators, collider experiments or the early universe.
Events come in different types called processes as discussed in the Automatic calculation of particle interaction or decay article.
Despite the simple structure of the tree-level perturbative quantum field theory description of the collision and decay processes in an event, the observed high-energy process usually contains significant amount of modifications, like photon and gluon bremsstrahlung or loop diagram corrections, that usually are too complex to be easily evaluated in real calculations directly on the diagrammatic level. Furthermore, the non-perturbative nature of QCD bound states makes it necessary to include information that is well beyond the reach of perturbative quantum field theory, and also beyond present ability of computation in lattice QCD. And in collisional systems more complex than a few leptons and hadrons (e.g. heavy-ion collisions), the collective behavior of the system would involve a phenomenological description that also cannot be easily obtained from the fundamental field theory by a simple calculus.
Use in simulations
As said above, the experimental calibration involves processes that usually are too complicated to be easily evaluated in calculations directly, so any realistic test of the underlying physical process in a particle accelerator experiment, therefore, requires an adequate inclusion of these complex behaviors surrounding the actual process. Based on the fact that in most processes a factorization of the full process into individual problems is possible (which means a negligible effect from interference), these individual processes are calculated separately, and the probabilistic branching between them are performed using Monte Carlo methods.
The final-state particles generated by event generators can be f |
https://en.wikipedia.org/wiki/Chandrasekhar%27s%20variational%20principle | In astrophysics, Chandrasekhar's variational principle provides the stability criterion for a static barotropic star, subjected to radial perturbation, named after the Indian American astrophysicist Subrahmanyan Chandrasekhar.
Statement
A baratropic star with and is stable if the quantity
is non-negative for all real functions that conserve the total mass of the star .
where
is the coordinate system fixed to the center of the star
is the radius of the star
is the volume of the star
is the unperturbed density
is the small perturbed density such that in the perturbed state, the total density is
is the self-gravitating potential from Newton's law of gravity
is the Gravitational constant |
https://en.wikipedia.org/wiki/Elaphoglossoideae | Elaphoglossoideae is a subfamily of the fern family Dryopteridaceae. It has previously been regarded as the family Elaphoglossaceae. As circumscribed by the Pteridophyte Phylogeny Group in their 2016 classification (PPG I), the subfamily excludes the Polybotryoideae, which are kept separate. It can be divided into three clades: the bolbitidoid ferns (Arthrobotrya, Bolbitis, Elaphoglossum, Lomagramma, Mickelia, Teratophyllum), genus Pleocnemia, sister clade to the bolbititoids, and the lastreopsid ferns (Lastreopsis, Megalastrum, Parapolystichum, Ruhmora), sister to the combination of the first two clades. |
https://en.wikipedia.org/wiki/Malapterurus%20stiassnyae | Malapterurus stiassnyae is a species of electric catfish native to Guinea, Liberia and Sierra Leone. Members of this species grow to a length of SL. |
https://en.wikipedia.org/wiki/Overlap%20fermion | In lattice field theory, overlap fermions are a fermion discretization that allows to avoid the fermion doubling problem. They are a realisation of Ginsparg–Wilson fermions.
Initially introduced by Neuberger in 1998, they were quickly taken up for a variety of numerical simulations. By now overlap fermions are well established and regularly used in non-perturbative fermion simulations, for instance in lattice QCD.
Overlap fermions with mass are defined on a Euclidean spacetime lattice with spacing by the overlap Dirac operator
where is the ″kernel″ Dirac operator obeying , i.e. is -hermitian. The sign-function usually has to be calculated numerically, e.g. by rational approximations. A common choice for the kernel is
where is the massless Dirac operator and is a free parameter that can be tuned to optimise locality of .
Near the overlap Dirac operator recovers the correct continuum form (using the Feynman slash notation)
whereas the unphysical doublers near are suppressed by a high mass
and decouple.
Overlap fermions do not contradict the Nielsen–Ninomiya theorem because they explicitly violate chiral symmetry (obeying the Ginsparg–Wilson equation) and locality. |
https://en.wikipedia.org/wiki/Ant%20Attack | Ant Attack is a ZX Spectrum computer game by Sandy White, published by Quicksilva in 1983. It was converted to the Commodore 64 in 1984.
While Zaxxon and Q*bert previously used isometric projection, Ant Attack added an extra degree of freedom (ability to go up and down instead of just north, south, east and west), and it may be the first isometric game for personal computers. The same type of isometric projection was used in Sandy White's later Zombie Zombie. It was also one of the first games to allow players to choose their gender.
Gameplay
The player chooses whether to control the character of "Girl" or "Boy", who then enters the walled city of Antescher to rescue the other, who has been captured and immobilised somewhere in the city.
The city is inhabited by giant ants which chase and attempt to bite the player. The player can defend themselves by throwing grenades at the ants, but these can also harm the humans. Once the hostage is rescued, the two must escape the city. The game then starts again with the hostage located in a different, harder-to-reach part of the city.
Development
Almost all of the game code was written by hand on paper using assembler mnemonics, then manually assembled, with the resulting hexadecimal digits typed sequentially into an external EEPROM emulator device (aka SoftROM or "softie") attached to a host Spectrum. Similarly, the character graphics and other custom sprites were all hand-drawn on squared paper and manually converted to strings of hex data. Additionally, some minor add-on routines such as high score registration were added on to the core game using regular Sinclair BASIC.
The game's setting of "Antescher" is a reference to the artist M. C. Escher.
Reception
Ant Attack was well received by gaming press. The game was nominated in the 1983 Golden Joystick Awards for Best Original Game of the Year, eventually coming second to Ah Diddums. The ZX Spectrum version was rated number 14 in the Your Sinclair'''s Official Top |
https://en.wikipedia.org/wiki/Herpetosiphonales | Herpetosiphonales is an order of bacteria in the class Chloroflexia.
See also
List of bacterial orders |
https://en.wikipedia.org/wiki/Hugging%20Face | Hugging Face, Inc. is a French-American company and open-source community that develops tools and resources to build, deploy, and train machine learning models. Based in New York City, the company is most notable for its Transformers library built for natural language processing and emphasis on community collaboration and accessibility. Its platform allows users to share machine learning models and datasets and showcase their work.
History
The company was founded in 2016 by French entrepreneurs Clément Delangue, Julien Chaumond, and Thomas Wolf in New York City, originally as a company that developed a chatbot app targeted at teenagers. After open-sourcing the model behind the chatbot, the company pivoted to focus on being a platform for machine learning.
In March 2021, Hugging Face raised US$40 million in a Series B funding round.
On April 28, 2021, the company launched the BigScience Research Workshop in collaboration with several other research groups to release an open large language model. In 2022, the workshop concluded with the announcement of BLOOM, a multilingual large language model with 176 billion parameters.
On December 21, 2021, the company announced its acquisition of Gradio, a software library used to make interactive browser demos of machine learning models.
On May 5, 2022, the company announced its Series C funding round led by Coatue and Sequoia. The company received a $2 billion valuation.
On May 13, 2022, the company introduced its Student Ambassador Program to help fulfill its mission to teach machine learning to 5 million people by 2023.
On May 26, 2022, the company announced a partnership with Graphcore to optimize its Transformers library for the Graphcore IPU.
On August 3, 2022, the company announced the Private Hub, an enterprise version of its public Hugging Face Hub that supports SaaS or on-premises deployment.
In February 2023, the company announced partnership with Amazon Web Services (AWS) which would allow Hugging Face's |
https://en.wikipedia.org/wiki/Wigner%20rotation | In theoretical physics, the composition of two non-collinear Lorentz boosts results in a Lorentz transformation that is not a pure boost but is the composition of a boost and a rotation. This rotation is called Thomas rotation, Thomas–Wigner rotation or Wigner rotation. If a sequence of non-collinear boosts returns an object to its initial velocity, then the sequence of Wigner rotations can combine to produce a net rotation called the Thomas precession.
The rotation was discovered by Émile Borel in 1913, rediscovered and proved by Ludwik Silberstein in his 1914 book 'Relativity', rediscovered by Llewellyn Thomas in 1926, and rederived by Wigner in 1939. Wigner acknowledged Silberstein.
There are still ongoing discussions about the correct form of equations for the Thomas rotation in different reference systems with contradicting results. Goldstein:
The spatial rotation resulting from the successive application of two non-collinear Lorentz transformations have been declared every bit as paradoxical as the more frequently discussed apparent violations of common sense, such as the twin paradox.
Einstein's principle of velocity reciprocity (EPVR) reads
We postulate that the relation between the coordinates of the two systems is linear. Then the inverse transformation is also linear and the complete non-preference of the one or the other system demands that the transformation shall be identical with the original one, except for a change of to
With less careful interpretation, the EPVR is seemingly violated in some models. There is, of course, no true paradox present.
Let it be u the velocity in which the lab reference frame moves respect an object called A and let it be v the velocity in which another object called B is moving, measured from the lab reference frame. If u and v are not aligned the relative velocities of these two bodies will not be opposite, that is since there is a rotation between them
The velocity that A will measure on B will be:
The Lorentz f |
https://en.wikipedia.org/wiki/Mir-299%20microRNA%20precursor%20family | In molecular biology mir-299 microRNA is a short RNA molecule. MicroRNAs function to regulate the expression levels of other genes by several mechanisms.
See also
MicroRNA |
https://en.wikipedia.org/wiki/Modesty%20in%20medical%20settings | Modesty in medical settings refers to the practices and equipment used to preserve patient modesty in medical examination and clinics.
Tools for modesty
Prior to the invention of the stethoscope, a physician who wanted to perform auscultation to listen to heart sounds or noise inside a body would have to physically place their ear against the body of the person being examined. In 1816, male physician René Laennec invented the stethoscope as a way to respect the modesty of a female patient, as it would have been awkward for him to put his ear on her chest.
Hospital gowns increase modesty as compared to the patient presenting nude, but in the past have been odd clothing which exposes the body. Some contemporary changes to the design of hospital gowns are proposed.
Society and culture
In places with more cultural diversity it becomes more likely that people will make new and different requests for modesty in health care.
Special populations
Sometimes women do not access healthcare because of modesty concerns.
Muslims in non-Muslim societies sometimes make requests for modesty. |
https://en.wikipedia.org/wiki/OpenID | OpenID is an open standard and decentralized authentication protocol promoted by the non-profit OpenID Foundation. It allows users to be authenticated by co-operating sites (known as relying parties, or RP) using a third-party identity provider (IDP) service, eliminating the need for webmasters to provide their own ad hoc login systems, and allowing users to log in to multiple unrelated websites without having to have a separate identity and password for each. Users create accounts by selecting an OpenID identity provider, and then use those accounts to sign on to any website that accepts OpenID authentication. Several large organizations either issue or accept OpenIDs on their websites.
The OpenID standard provides a framework for the communication that must take place between the identity provider and the OpenID acceptor (the "relying party"). An extension to the standard (the OpenID Attribute Exchange) facilitates the transfer of user attributes, such as name and gender, from the OpenID identity provider to the relying party (each relying party may request a different set of attributes, depending on its requirements). The OpenID protocol does not rely on a central authority to authenticate a user's identity. Moreover, neither services nor the OpenID standard may mandate a specific means by which to authenticate users, allowing for approaches ranging from the common (such as passwords) to the novel (such as smart cards or biometrics).
The final version of OpenID is OpenID 2.0, finalized and published in December 2007. The term OpenID may also refer to an identifier as specified in the OpenID standard; these identifiers take the form of a unique Uniform Resource Identifier (URI), and are managed by some "OpenID provider" that handles authentication.
Adoption
, there are over 1 billion OpenID-enabled accounts on the Internet (see below) and approximately 1,100,934 sites have integrated OpenID consumer support: AOL, Flickr, Google, Amazon.com, Canonical (provider |
https://en.wikipedia.org/wiki/Chiral%20model | In nuclear physics, the chiral model, introduced by Feza Gürsey in 1960, is a phenomenological model describing effective interactions of mesons in the chiral limit (where the masses of the quarks go to zero), but without necessarily mentioning quarks at all. It is a nonlinear sigma model with the principal homogeneous space of a Lie group as its target manifold. When the model was originally introduced, this Lie group was the SU(N) , where N is the number of quark flavors. The Riemannian metric of the target manifold is given by a positive constant multiplied by the Killing form acting upon the Maurer–Cartan form of SU(N).
The internal global symmetry of this model is , the left and right copies, respectively; where the left copy acts as the left action upon the target space, and the right copy acts as the right action. Phenomenologically, the left copy represents flavor rotations among the left-handed quarks, while the right copy describes rotations among the right-handed quarks, while these, L and R, are completely independent of each other. The axial pieces of these symmetries are spontaneously broken so that the corresponding scalar fields are the requisite Nambu−Goldstone bosons.
The model was later studied in the two-dimensional case as an integrable system, in particular an integrable field theory. Its integrability was shown by Faddeev and Reshetikhin in 1982 through the quantum inverse scattering method. The two-dimensional principal chiral model exhibits signatures of integrability such as a Lax pair/zero-curvature formulation, an infinite number of symmetries, and an underlying quantum group symmetry (in this case, Yangian symmetry).
This model admits topological solitons called skyrmions.
Departures from exact chiral symmetry are dealt with in chiral perturbation theory.
Mathematical formulation
On a manifold (considered as the spacetime) and a choice of compact Lie group , the field content is a function . This defines a related field , a -v |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.