id int64 39 79M | url stringlengths 31 227 | text stringlengths 6 334k | source stringlengths 1 150 ⌀ | categories listlengths 1 6 | token_count int64 3 71.8k | subcategories listlengths 0 30 |
|---|---|---|---|---|---|---|
44,220,881 | https://en.wikipedia.org/wiki/Pigment%20Red%20178 | Pigment Red 178 is an organic compound that is used as a pigment. Structurally, it is a derivative of perylene, although it is produced from perylenetetracarboxylic dianhydride by derivatization with 4-aminoazobenzene.
References
Perylene dyes
Vat dyes
Imides
Azo compounds | Pigment Red 178 | [
"Chemistry"
] | 74 | [
"Functional groups",
"Organic compounds",
"Imides",
"Organic compound stubs",
"Organic chemistry stubs"
] |
44,220,917 | https://en.wikipedia.org/wiki/Pigment%20Red%20179 | Pigment Red 179 is an organic compound used as a pigment. Structurally, it is a derivative of perylene, produced from perylenetetracarboxylic dianhydride through derivatization with methylamine.
It is used in watercolor paints, usually labeled as Perylene Maroon.
References
Perylene dyes
Vat dyes
Imides | Pigment Red 179 | [
"Chemistry"
] | 77 | [
"Functional groups",
"Organic compounds",
"Imides",
"Organic compound stubs",
"Organic chemistry stubs"
] |
44,223,073 | https://en.wikipedia.org/wiki/Japan%20Trench%20Fast%20Drilling%20Project | The Japan Trench Fast Drilling Project (JFAST) was a rapid-response scientific expedition that drilled oceanfloor boreholes through the fault-zone of the 2011 Tohoku earthquake. JFAST gathered important data about the rupture mechanism and physical properties of the fault that caused the huge earthquake and tsunami which devastated much of northeast Japan.
Background
The 2011 Tohoku-oki earthquake, with a moment magnitude of 9.0, was the largest in Japan's history, and severely damaged regions of northeast Honshu, with over 15,000 deaths and economic losses of $US 200 to 300 billion. Because of the huge societal impact, there was an urgency among scientists to respond with information and research results to explain the disastrous event. Soon after the earthquake, researchers of the Integrated Ocean Drilling Program (IODP) began planning the Japan Trench Fast Drilling Project (JFAST) to investigate the earthquake with ocean floor boreholes to the plate boundary fault.
This ambitious project drilled boreholes through the fault that slipped during the earthquake in order to understand the unprecedented huge slip (40 to 60 meters) that occurred on the shallow portion of the megathrust fault and was the primary source of the large tsunami that devastated much of the coast of northeast Honshu. There was much public interest in this high-profile scientific project with considerable Japanese and English media coverage of the operations and results
Specific science objectives included,
Estimation of the stress (mechanics) state in the region of the shallow fault from borehole breakouts.
Retrieval of core sample from the plate boundary fault zone to see geologic structures and measure physical properties of the fault zone. Before this project, no one had directly seen a fault zone that recently moved tens of meters in an earthquake.
Measurement of the temperature across the fault zone to estimate the level of dynamic friction during the earthquake. These thermal observations needed to be done quickly after the earthquake and was the main reason for the rapid mobilization of JFAST.
The site for the offshore drilling was located about 220 km east of Sendai in the region of very large fault slip during the earthquake near the Japan Trench.
Deep water drilling operations
The D/V Chikyu, operated by the Japan Agency for Marine-Earth Science and Technology (JAMSTEC) sailed on IODP Expedition 343 from the port of Shimizu, Shizuoka on April 1, 2012, within 13 months after the earthquake. Chikyu is the only research vessel with the capabilities for the necessary drilling in very deep water of over 6900 meters. Two months of operations from April 1 to 24 May 24, 2012 were scheduled for drilling several boreholes to carry out Logging while drilling (LWD), install temperature sensors and retrieve core samples. The extreme water depths caused many technical challenges that had to be considered, such as the strength of the long pipe string, onboard handling of the pipe sections, and instrument operations at very high water pressure. These aspects needed careful planning and new tools on the ship. Various equipment had not been previously used in such deep water and caused many problems and delays during the first month at sea. Eventually difficult engineering problems were overcome enabling retrieval of borehole core and installation of a temperature observatory across the fault zone at a depth of about 820 meters below the sea floor. New records were set for scientific drilling including, longest drilling string (7740 m) from the ocean surface and deepest core from the ocean surface (7752 m).
Because of delays due to technical difficulties and bad weather, the temperature observatory could not be deployed during the main expedition. However, during the supplementary Expedition 343T from July 5 to 19 a new borehole was quickly drilled and the temperature sensors were installed
Retrieval of the temperature data was scheduled for cruise KR13-04 during February 11 to 20, 2013 using the JAMSTEC ship R/V Kairei and Remotely Operated Vehicle (ROV) Kaiko-7000II. Kaiko-7000II is one of the few vehicles that can operate at 7000 meters water depth. Because of inclement weather and navigation problems the instruments could not be retrieved at this time. However, during the subsequent cruise KR13-08 from April 21 to May 9, 2013, the temperature instruments were successfully recovered on April 26.
Scientific results
Borehole stress
Fractures in the borehole wall (borehole breakouts) were used to estimate the stress field in the region close to the fault zone. These fractures can be observed in the wall resistivity records obtained from the LWD data. From the orientations and crack widths of the fractures, the direction and magnitude of the stress can be calculated. The results of these analyses show that the region has changed from a thrust fault regime before the earthquake to a normal fault regime after the earthquake. The horizontal stress became close to zero, indicating that almost all of the stress was released during the earthquake. This confirms previous suggestions that the earthquake had a complete stress drop, which is different from most other large earthquakes.
Fault zone
From the core samples, geologic structure data and measurements of physical properties, a single plate-boundary fault zone was identified with a high level of confidence at a depth of about 820 meters below the seafloor. The fault is localized in a thin layer of highly deformed pelagic clays. The entire section of the fault zone was not retrieved, but from the amount of the recovered and unrecovered sections, the total width of the fault zone is determined to be less than 5 meters. This is a considerably simpler and thinner plate-boundary fault than has been observed at other locations, such as the Nankai Trough . The actual slip surface for the 2011 earthquake may not have been recovered, but it is assumed that the structures and physical properties of the core are representative of the entire fault zone.
Fault friction
One of the main objectives of JFAST was to estimate the level of friction on the fault during the earthquake. To determine the frictional strength, high-speed laboratory experiments were carried out on samples from the plate boundary fault zone. The measured shear stress strength for permeable and impermeable conditions yielded values of 1.32 and 0.22 MPa, respectively, with the equivalent values for the coefficient of friction of 0.19 and 0.03, respectively. These results show that the fault slipped with very low levels of friction, which are lower than observed from other subduction zones, such as the Nankai Trough. The very low frictional strength for the material from the Japan Trench fault zone is much lower than typically observed for other types of rocks. The low friction properties are largely caused by the high content of the clay mineral smectite. Examination of the microstructures in the laboratory samples, suggests that fluids are important in the faulting process and contribute to the low friction properties, possibly through thermal pressurization.
The temperature measurements were also designed to estimate the frictional heat on the fault by measuring the thermal anomaly at the fault zone. A temperature signal was clearly observed in the data about 4 months after instrument installation which is 18 month after the earthquake. At that time the temperature in the fault zone was about 0.3 °C above the geothermal gradient. This is interpreted to represent the frictional heat produced at the time of the earthquake. Analyses of these data showed that the coefficient of friction on the fault at the time of the earthquake was about 0.08 and the average shear stress on the fault was estimated to be 0.54 MPa. The temperature measurements give independent and similar results to the laboratory friction experiments, and confirm the very low frictional properties of the fault. The low friction properties likely contributed to the very large slip during the earthquake.
Summary
JFAST is considered to be a successful rapid scientific response to a natural hazard event that had a great societal impact. Technical challenges associated with drilling in very deep water of about 6900 meters were overcome enabling borehole stress measurements, recovery of valuable core samples of the plate-boundary fault zone and collection of unique temperature measurements. The results of the scientific investigations show that the huge slip during the 2011 Tohoku earthquake occurred on a simple and thin fault zone composed of pelagic sediments with a high smectite content. Both laboratory experiments on the fault zone material and temperature measurements across the fault zone, show that the friction level was very low during the earthquake. The localized fault zone, low friction properties of its material and complete stress drop during the earthquake, are important characteristics that likely contributed to the huge slip during the earthquake.
See also
Drilling Vessel Chikyū
Integrated Ocean Drilling Program
References
Media coverage
Scientific American, Drilling Ship to Probe Fault Zone that Caused Fukushima Quake, October 31, 2011.
Discovery Channel, Daily Planet feature story on March 9, 2012.
NHK TV News, Japan, news story on April 14, 2012 (in Japanese).
TBS TV News 23, Japan, featured story on May 3, 2012 (in Japanese).
Otago Daily Times, New Zealand, Researchers drill deep into fault, May 2, 2012
NHK TV Science Zero, 30 minute TV program about JFAST, June 17, 2012 (in Japanese)
phys.org, New report illuminates stress change during the 2011 Tohoku-Oki earthquake, February 7, 2013
Physics Today, Scientists dig deep in Tohoku fault to crack earthquake's secrets, August 2013.
Christian Science Monitor, How slippery clay helps Japan's 2011 mega tsunami, December 6, 2013.
External links
JFAST website
IODP
2011 Tōhoku earthquake and tsunami
Earthquake engineering
Science and technology in Japan | Japan Trench Fast Drilling Project | [
"Engineering"
] | 1,929 | [
"Earthquake engineering",
"Civil engineering",
"Structural engineering"
] |
44,223,676 | https://en.wikipedia.org/wiki/African%20Wildlife%20Defence%20Force | The African Wildlife Defence Force (AWDF), Kikosi cha ulinzi ya wanyama pori barani Afrika (Swahili), is a private park ranger and anti-poaching organization based in Dungu, in the north-east of the Democratic Republic of the Congo. AWDF uses direct action tactics to protect wildlife and rainforests. The organization was founded in 2012 by Congolese-Belgian philanthropist Jean Kiala-Inkisi. It is proposed as an alternative to regular park ranger organizations who struggle with corruption, and seeks to eliminate the increasing levels of violence which poachers face.
History
The AWDF was found in 2012 after founder Jean Kiala-Inkisi travelled through Africa. He did not believe the rest of the world was doing enough to help the parks in central Africa. After he conducted ground surveys on the border of the Democratic Republic of the Congo and South Sudan, and had been contracted on a private ranch in South Africa, he decided to train a team of Congolese rangers.
In April 2014, the AWDF began with the selection of candidate rangers in the Democratic Republic of the Congo, South Sudan, Kenya and South Africa for training of Advanced Force (AFR) and Special Force (SFR) rangers. Kiala-Inkisi approached a former French Legionnaire in aim of providing training.
Organization
The AWDF is a non-profit private ranger organization. 80% of the organization's revenue is spent on its programs and 20% on administration and fundraising. It is supported by private and corporate donations, internet advertising and grants. The group is operated by both paid rangers and volunteers. The organization chooses to operate with a few Special Operations Task Forces only. It provides services including anti-poaching, wildlife management, forestry management and agroforestry consulting. It is also involved in ranger training, close quarter training and specialist rural security services.
The AWDF is open to African citizens except those from countries located north of the Sahel. They refuse African expats from outside Africa for ranger functions. In general, foreigners from outside Africa can only work as an instructor or scientist.
AWDF rangers wear their insignia on the left side of the beret, to distinguish themselves from the regular park ranger organizations.
As a private ranger services contractor, the AWDF focuses on wildlife conservation and rainforest conservation. The AWDF has expertise to also intervene in mangroves, lakes and waterways but does not work at sea. Their major working field is central Africa in the parks located in the border region of Democratic Republic of the Congo, South Sudan, Uganda and Central African Republic.
Departments
A working group, Convention on African Trade in Endangered Species of Wild Fauna, was formed in protest against what the AWDF calls the disastrous policies of CITES. The group works on wildlife law enforcement and promote non-conventional livestock farming including insect farming, crocodile farming and game farming. They also examine the pros and cons of rhino horn farming.
Rangers are trained in Basic Military Training and Wildlife Management courses. Advanced Force Rangers consist of parachute/commando units, operating as a support-reconnaissance unit. Special Force Rangers are a special forces unit selected from the AFR units, trained in three specialities. These are free fall from high altitude HAHO/HALO, underwater fighting skills and operating in mountainous terrain. The Special Operations Affiliate Ranger Group are a special forces unit, mostly US veterans working pro-bono for a short or long-term basis.
The AWDF focuses on the conservation of the rainforest from the basin of the Congo which has 70% of Africa's plant cover. The AWDF plans to start a nursery for Wild Edible Plants (WEP) & Non-Timber Forest Products (NTFP)'' and Tropical Hardwood. The AWDF also created a list of 150 tree species for multiplication to avoid the alienation and destruction of the rainforest.
Actions
On 13 August 2014, Kiala-Inkisi flew to Burbank, California to do a hunger strike. The goal was to get media attention in Hollywood for his cause. After sending 2500 emails without receiving any reply, he talked with workers of Animal Defenders International (ADI) in Los Angeles. He was also invited by rhino zookeeper and Anti-Poaching Ranger Mike Daniels to visit the San Diego Zoo Safari Park and to look behind the scenes. On invitation of Matt Rossell, (ADI) Campaigns Director, a dinner took place with actress and animal activist Georja Umano. Later he also met the actress and editor Moon Hi Hanson.
Since 2012, a team of rangers is deployed near the border of Bengangai Game Reserve, Bire Kpatous Game Reserve and Mbarizunga Game reserve of neighboring country South Sudan to quest the Lord's Resistance Army (LRA) of Joseph Kony.
Projects
Nutrecul Agroforestry Project
References
External links
Environmental organizations established in 2012
Ecology organizations
Wildlife rehabilitation and conservation centers
Wildlife conservation organizations
Environmental treaties
Endangered species
Wildlife smuggling
Military units and formations established in 2012
Business services companies established in 2012
Private military contractors
Security consulting firms
Animal rights organizations
Ranches
Animal welfare organisations based in the Democratic Republic of the Congo | African Wildlife Defence Force | [
"Biology"
] | 1,042 | [] |
44,224,167 | https://en.wikipedia.org/wiki/Process%20validation | Process validation is the analysis of data gathered throughout the design and manufacturing of a product in order to confirm that the process can reliably output products of a determined standard. Regulatory authorities like EMA and FDA have published guidelines relating to process validation. The purpose of process validation is to ensure varied inputs lead to consistent and high quality outputs. Process validation is an ongoing process that must be frequently adapted as manufacturing feedback is gathered. End-to-end validation of production processes is essential in determining product quality because quality cannot always be determined by finished-product inspection. Process validation can be broken down into 3 steps: process design (Stage 1a, Stage 1b), process qualification (Stage 2a, Stage 2b), and continued process verification (Stage 3a, Stage 3b).
Stage 1: Process Design
In this stage, data from the development phase are gathered and analyzed to define the commercial manufacturing process. By understanding the commercial process, a framework for quality specifications can be established and used as the foundation of a control strategy. Process design is the first of three stages of process validation. Data from the development phase is gathered and analyzed to understand end-to-end system processes. These data are used to establish benchmarks for quality and production control.
Design of experiment (DOE)
Design of experiments is used to discover possible relationships and sources of variation as quickly as possible. A cost-benefit analysis should be conducted to determine if such an operation is necessary.
Quality by design (QBD)
Quality by design is an approach to pharmaceutical manufacturing that stresses quality should be built into products rather than tested in products; that product quality should be considered at the earliest possible stage rather than at the end of the manufacturing process. Input variables are isolated in order to identify the root cause of potential quality issues and the manufacturing process is adapted accordingly.
Process analytical technology (PAT)
Process analytical technology is used to measure critical process parameters (CPP) and critical quality attributes (CQA). PAT facilitates measurement of quantitative production variables in real time and allows access to relevant manufacturing feedback. PAT can also be used in the design process to generate a process qualification.
Critical process parameters (CPP)
Critical process parameters are operating parameters that are considered essential to maintaining product output within specified quality target guidelines.
Critical quality attributes (CQA)
Critical quality attributes (CQA) are chemical, physical, biological, and microbiological attributes that can be defined, measured, and continually monitored to ensure final product outputs remain within acceptable quality limits. CQA are an essential aspect of a manufacturing control strategy and should be identified in stage 1 of process validation: process design. During this stage, acceptable limits, baselines, and data collection and measurement protocols should be established. Data from the design process and data collected during production should be kept by the manufacturer and used to evaluate product quality and process control. Historical data can also help manufacturers better understand operational process and input variables as well as better identify true deviations from quality standards compared to false positives. Should a serious product quality issue arise, historical data would be essential in identifying the sources of errors and implementing corrective measures.
Stage 2: Process Performance Qualification
In this stage, the process design is assessed to conclude if the process is able to meet determined manufacturing criteria. In this stage all production processes and manufacturing equipment is proofed to confirm quality and output capabilities. Critical quality attributes are evaluated, and critical process parameters taken into account, to confirm product quality. Once the process qualification stage has been successfully accomplished, production can begin. Process Performance Qualification is the second phase of process validation.
Stage 3: Continued Process Verification
Continued process verification is the ongoing monitoring of all aspects of the production cycle. It aims to ensure that all levels of production are controlled and regulated. Deviations from prescribed output methods and final product irregularities are flagged by a process analytics database system. The FDA requires production data be recorded (FDA requirements (§ 211.180(e)). Continued process verification is stage 3 of process validation.
The European Medicines Agency defines a similar process known as ongoing process verification. This alternative method of process validation is recommended by the EMA for validating processes on a continuous basis. Continuous process verification analyses critical process parameters and critical quality attributes in real time to confirm production remains within acceptable levels and meets standards set by ICH Q8, Pharmaceutical Quality Systems, and Good manufacturing practice.
See also
Cleaning validation
Process qualification
Verification and validation
References
External links
FDA – U.S. Food and Drug Administration
EMA – European Medicines Agency
Parental Drug Association
AAPS Process Validation
AAPS Technology Transfer Drug Product Manufacturing Process
Software testing
Formal methods
Software quality
Enterprise modelling
Business process management | Process validation | [
"Engineering"
] | 932 | [
"Systems engineering",
"Software testing",
"Enterprise modelling",
"Software engineering",
"Formal methods"
] |
44,224,302 | https://en.wikipedia.org/wiki/Glossary%20of%20levelling%20terms | This is a glossary of levelling terms. Levelling is a surveying method used to find relative height, one use of which is to ensure ground is level during construction, for example, when excavating to prepare for laying a foundation for a house.
Levelling terms
Automatic level – variant of the dumpy level, that makes use of a compensator that ensures that the line of sight remains horizontal once the operator has roughly leveled the instrument. The surveyor sets the instrument up quickly and doesn't have to relevel it carefully each time he sights on a rod on another point. It also reduces the effect of minor settling of the tripod. Three adjustment screws are used to level the instrument.
Back sight (BS) – short for "back sight reading", the first staff reading taken by the surveyor after the levelling instrument is set up and levelled. B.S is generally taken on the point of known reduced level as on the benchmark or a change point.
Benchmark (surveying) – fixed reference point of known elevation with respect to which RL of other points is determined. Benchmarks can be arbitrary or permanent, the former is used for calculation of reduced levels for small survey works and the latter is used to calculate the elevations of significantly important locations and points. Arbitrary benchmarks are assumed to be equal to 100 meters generally and then the elevations with respect to assumed benchmark is determined. It is commonly practiced by engineering students. For GTS surveys of the country, surveyors use permanent benchmarks to calculate the elevations of different points.
Datum surface – reference plane with respect to which RL of the other survey points is determined. The datum surface may be real or imaginary location with a nominated elevation of zero. The commonly used datum is mean sea level.
Dumpy level – optical instrument used to establish or check points in the same horizontal plane. It is used in surveying and building with a vertical staff to measure height differences and to transfer, measure and set heights. Also called a builder's level or leveling instrument.
Fore sight (FS) – short for "fore sight reading", the last staff reading taken before changing the instrument to the other position. It is the staff reading taken on point whose RL is to determined. This sight is considered as negative and deduced from Height of Instrument to determine RL of the point.
Intermediate sights – all readings taken between back sight and fore sight. These are the points whose RL is determined by the method already mentioned above in FS. Also called inter-sight readings.
Laser level – control tool consisting of a laser beam projector that can be affixed to a tripod, and which projects a fixed red or green beam along the horizontal and/or vertical axis. A rotary laser level is a more advanced laser level in that it spins the beam of light fast enough to give the effect of a complete 360 degree horizontal or vertical plane, thus illuminating not just a fixed line, but a horizontal plane.
Levelling – measurement of geodetic height using an optical levelling instrument and a level staff or rod having a numbered scale. Common levelling instruments include the spirit level, the dumpy level, the digital level, and the laser level.
Levelling staff – specialized measuring stick or vertical staff used with the dumpy level, held by a second person while the operator of the level looks through it and takes readings off of the staff. Also call a rod.
Reduced level (RL) – equating elevations of survey points with reference to a common assumed datum. The elevation is positive or negative according as point lies above or below datum.
Spirit level – instrument designed to indicate whether a surface is horizontal (level) or vertical (plumb). While used by surveyors, different types of spirit levels may be used by carpenters, stonemasons, bricklayers, other building trades workers, millwrights and other metalworkers, and in some photographic or videographic work.
See also
Surveying
:Prismatic compass (surveying)
References
Civil engineering
Geomatics engineering
Levelling Terms
Surveying
Wikipedia glossaries using unordered lists | Glossary of levelling terms | [
"Engineering"
] | 844 | [
"Construction",
"Surveying",
"Civil engineering"
] |
44,224,558 | https://en.wikipedia.org/wiki/Transmission%20line%20loudspeaker | A transmission line loudspeaker is a loudspeaker enclosure design which uses the topology of an acoustic transmission line within the cabinet, compared to the simpler enclosures used by sealed (closed) or ported (bass reflex) designs. Instead of reverberating in a fairly simple damped enclosure, sound from the back of the bass speaker is directed into a long (generally folded) damped pathway within the speaker enclosure, which allows far greater control and use of speaker energy and the resulting sound.
Inside a transmission line (TL) loudspeaker is a (usually folded) pathway into which the sound is directed. The pathway is often covered with varying types and depths of absorbent material, and it may vary in size or taper, and may be open or closed at its far end. Used correctly, such a design ensures that undesired resonances and energies, which would otherwise cause undesirable auditory effects, are instead selectively absorbed or reduced ("damped") due to the effects of the duct, or alternatively only emerge from the open end in phase with the sound radiated from the front of the driver, enhancing the output level ("sensitivity") at low frequencies. The transmission line acts as an acoustic waveguide, and the padding both reduces reflection and resonance, and also slows the speed of sound within the cabinet to allow for better tuning.
Transmission line loudspeakers designs are more complex to implement, making mass production difficult, but their advantages have led to commercial success for a number of manufacturers such as IMF, TDL, and PMC. As a rule, transmission line speakers tend to have exceptionally high fidelity low frequency response far below that of a typical speaker or subwoofer, reaching into the infrasonic range (British company TDL's studio monitor range from the 1990s quoted their frequency responses as starting from as low as 17 Hz depending upon model with a sensitivity of 87 dB for 1 W @ 1 meter), without the need for a separate enclosure or driver. Acoustically, TL speakers roll off more slowly (less steeply) at low frequencies, and they are thought to provide better driver control than standard vented-box cabinet designs, are less sensitive to positioning, and tend to create a very spacious soundstage. Modern TL speakers were described in a 2000 review as "match[ing] reflex cabinet designs in every respect, but with an extra octave of bass, lower LF distortion and a frequency balance which is more independent of listening level".
Although more complex to design and tune, and not as easy to analyze and calculate as other designs, the transmission line design is valued by several smaller manufacturers, as it avoids many of the major disadvantages of other loudspeaker designs. In particular, the basic parameters and equations describing sealed and reflex designs are fairly well understood, the range of options involved in a transmission line design mean that the general design can be somewhat calculated but final transmission line tuning requires considerable attention and is less easy to automate.
Purpose and design overview
A transmission line is used in loudspeaker design to reduce time, phase, and resonance related distortions, and in many designs to gain exceptional bass extension to the lower end of human hearing, and in some cases the near-infrasonic (below 20 Hz). TDL's 1980s reference speaker range (now discontinued) contained models with frequency ranges of 20 Hz upwards, down to 17 Hz upwards, without needing a separate subwoofer. Irving M. Fried, an advocate of TL design, stated that:
"I believe that speakers should preserve the integrity of the signal waveform and the Audio Perfectionist Journal has presented a great deal of information about the importance of time domain performance in loudspeakers. I’m not the only one who appreciates time- and phase-accurate speakers but I have been virtually the only advocate to speak out in print in recent years. There’s a reason for that."
"It is difficult and costly to design and manufacture a time- and phase-accurate speaker system. Few of today’s high-end loudspeakers are time- and phase-accurate designs. The audio magazines need to appeal to a broad spectrum of advertisers including many who make speaker systems which are time incoherent. The magazines, and the reviewers who write for them, have ignored or downplayed the issue of time- and phase-accuracy in order to maximize advertising revenue. I am not alone in recognizing this situation."
Some proponents of TL loudspeakers consider that using a TL is the theoretical ideal manner in which to load a moving-coil drive unit. However, it is also one of the more complex of constructions. The most common and practical implementation is to fit a drive unit to the end of a long duct that is usually open at the far end. In practice, the duct is folded inside a conventional shaped cabinet, so that the open end of the duct appears as a vent on the speaker cabinet. There are many ways in which the duct can be folded, and the line is often tapered in cross section to avoid parallel internal surfaces that encourage standing waves. Some speaker designs also use a spiral or elliptic spiral shaped duct, usually with one speaker element in the front or two speaker elements arranged one on each side of the cabinet. Depending upon the drive unit, and quantity and various physical properties of absorbent material, the amount of taper will be adjusted during the design process to tune the duct to remove irregularities in its response. The internal partitioning provides substantial bracing for the entire structure, reducing cabinet flexing and colouration. The inside faces of the duct or line, are treated with an absorbent material to provide the correct termination with frequency to load the drive unit as a TL. The enclosure behaves like an infinite baffle, potentially absorbing most or all of the speaker unit's rear energies. A theoretically perfect TL would absorb all frequencies entering the line from the rear of the drive unit but remains theoretical, as it would have to be infinitely long. The physical constraints of the real world, demand that the length of the line must often be less than 4 meters before the cabinet becomes too large for any practical applications, so not all the rear energy can be absorbed by the line. In a realized TL, only the upper bass is TL loaded in the true sense of the term (i.e. fully absorbed); the low bass is allowed to freely radiate from the vent in the cabinet. The line therefore effectively works as a low pass filter, another crossover point in fact, achieved acoustically by the line and its absorbent filling. Below this “crossover point” the low bass is loaded by the column of air formed by the length of the line. The length of the line is specified so as to reverse the phase of the rear output of the drive unit as it exits the vent. This acoustic energy combines with the output of the bass unit, extending its response and effectively creating a second driver.
Essentially, the goal of the transmission line is to minimize acoustical or mechanical impedance at frequencies corresponding to the fundamental free-air resonance of the bass driver. This simultaneously reduces stored energy in the driver's motion, reduces distortion, and critically damps the driver by maximizing acoustic output (maximal acoustical loading or coupling) at the terminus. This also minimizes the negative effects of acoustic energy that would otherwise (as with a sealed enclosure) be reflected back to the driver in a sealed cavity.
Transmission line loudspeakers employ this tube-like resonant cavity, with the length set between 1/6 and 1/2 the wavelength of the fundamental resonant frequency of the loudspeaker driver being used. The cross-sectional area of the tube is typically comparable to the cross-sectional area of the driver's radiating surface area. This cross section is typically tapered down to approximately 1/4 of the starting area at the terminus or open end of the line. While not all lines use a taper, the standard classical transmission line employs a taper from 1/3 to 1/4 area (ratio of terminus area to starting area directly behind driver). This taper serves to dampen the buildup of standing waves within the line, which can create sharp nulls in response at the terminus output at even multiples of the driver's Fs.
In a transmission line speaker, the transmission line itself can be open ("vented") or closed at the far end. Closed designs typically have negligible acoustic output from the enclosure except from the driver, while open ended designs exploit the low-pass filter effect of the line, and the resultant low bass energy emerges to reinforce the output from the driver at low frequencies. Well designed transmission line enclosures have smooth impedance curves, possibly from a lack of frequency-specific resonances, but can also have low efficiency if poorly designed.
One key advantage of transmission lines is their ability to conduct the back wave behind the transducer more effectively away from it – reducing the chance for reflected energy permeating back through the diaphragm out of phase with the primary signal. Not all transmission lines designs do this effectively. Most offset transmission line speakers place a reflective wall fairly close behind the transducer within the enclosure – posing a problem for internal reflections emanating back through the transducer diaphragm. Older descriptions explained the design in terms of "impedance mismatch", or pressure waves "reflected" back into the enclosure; these descriptions are now considered outdated and inaccurate as technically the transmission line works through selective production of standing waves and constructive and destructive interference (see below).
A second benefit is that the resulting music is time coherent (i.e., in phase). Fried quoted in 2002, a listening test performed and reported in December 2000's Hi-Fi News (as he believed) in which a high-quality recording was obtained using reputable but non-time-coherent loudspeakers and this recording was then time phase corrected; an expert listening panel "voted unanimously for the superior realism and accuracy of the time corrected output" for high quality sound reproduction.
One of the significant and common problems with a transmission line loudspeaker system is the unwanted phase-cancellation effects of higher line harmonics bleeding from the transmission line and adversely affecting the overall sound field. For example, in the PMC PMC6 mid-sized transmission line monitoring loudspeaker, there is a dip around 300 Hz that is caused by the fifth harmonic of the transmission line’s resonant frequency. This type of problem is quite common, and it was readily apparent in other transmission line loudspeakers. For example, the large IMF TLS80 MkII from 1977 also had an anomaly, but this time at the lower frequency of about 140 Hz, consisting of an almost one-octave-wide deleterious 2-dB dip in the on-axis response. Another problem is that the sound radiation from the exit of the line is spread over a quite broad frequency range caused by the hump of the quarter-wave transmission line resonance, whereas the high-Q port resonance of a vented-box loudspeaker rolls off much more quickly and extends over a much narrower frequency band. These sorts of issues with transmission line loudspeakers can lead to tonal accuracy problems that cannot be resolved.
A transmission line speaker employs, essentially, two distinct forms of bass loading, which historically and confusingly have been amalgamated in the TL description. Separating the upper and lower bass analysis reveals why such designs have so many potential advantages and disadvantages over reflex and infinite baffle designs. Measurements indicate that the upper bass is only partially absorbed by the line, making a clean and neutral response somewhat difficult if not impossible to achieve. The lower bass is extended and distortion is lowered by the line's control over the drive unit's excursion. One of the exclusive benefits of a TL design is its ability to produce very low frequencies even at low monitoring levels – TL speakers can routinely produce full range sound usually requiring a subwoofer, and do so to very high levels of low-frequency accuracy. The main disadvantage of the design is that it is more labor-intensive to create and tune a high quality and consistent transmission line, compared to building a simple vented-box or closed-box enclosure. One PMC employee was quoted as saying that optimising a transmission line loudspeaker is "like juggling water". A 2010 Hifi Avenue TL speaker review commented that "One thing I have noticed about transmission line designs is that they create a rather big soundstage and seem to handle crescendoes with ease".
History of transmission line loudspeakers
Invention and early use
The concept was innovated within acoustic enclosure design, and originally termed an "acoustical labyrinth", by acoustic engineer and later Director of Research, Benjamin Olney, who developed the concept at the Stromberg-Carlson Telephone Co. in the early 1930s while studying the effect of enclosure shape and size on speaker output, including the effect of "extreme length in a box baffle". A patent was filed in 1934. The design was used in their console radios beginning in 1936. A loudspeaker enclosure based on the concept was proposed in October 1965 by Dr A.R. Bailey in Wireless World magazine, referencing a production version of an acoustic-line enclosure design from Radford Electronics Ltd. The article postulated that energy from the rear of a driver unit could be essentially absorbed, without damping the cone's motion or superimposing internal reflections and resonance, so Bailey and Radford reasoned that the rear wave could be channelled down a long pipe. If the acoustic energy was absorbed, it would not be available to excite resonances. A pipe of sufficient length could be tapered, and stuffed so that the energy loss was almost complete, minimizing output from the open end. No broad consensus on the ideal taper (expanding, uniform cross-section, or contracting) has been established.
"Classic" era transmission line loudspeakers
Source for much of this section: Loudspeakers: for music recording and reproduction (Newell & Holland, 2007)
The birth of the modern transmission line speaker design came about in 1965 with the publication of A.R. Bailey's article in Wireless World, “A Non-resonant Loudspeaker Enclosure Design”, detailing a working Transmission Line. Bailey followed up his first article with a second one in 1972. Radford Electronics Ltd took up this innovative design and briefly manufactured the first commercial Transmission Line loudspeaker. Although acknowledged as the father of the Transmission Line, Bailey's work drew on the work on labyrinth design, dating back as early as the 1930s. His design, however, differed significantly in the way in which he filled the cabinet with absorbent materials. Bailey hit upon the idea of absorbing all the energy generated by the bass unit inside the cabinet, providing an inert platform for the drive unit to work from; unchecked, this energy produces spurious resonances in the cabinet and its structure, adding distortion to the original signal.
Shortly thereafter the design entered mainstream Hi-Fi, through the works of Irving M. "Bud" Fried in the United States, and a British trio: John Hayes, John Wright, and David Brown. Dave D'Lugos describes the period that followed (approximately 35 years until the start of the 21st Century) as a period when the "classical designs" were created.
Fried was exposed during his time at Harvard University to high fidelity audio reproduction, and later became an importer of audiophile items. Under the trademark "IMF" (his initials), from 1961, he eventually became involved with many advancements in audiophile equipment: cartridges (IMF – London, IMF – Goldring), tonearms (SME, Gould, Audio and Design), amplifiers (Quad, Custom Series), loudspeakers (Lowther, Quad, Celestion, Bowers and Wilkins, Barker, etc.). In 1968 he met John Hayes and John Wright, who had already designed an award-winning tonearm in the UK and had brought along a transmission line speaker designed by John Wright — described by Hayes as "fanatical regarding quality" — in order to promote and demonstrate the tonearm at a New York hifi show. Fried unexpectedly received a number of orders for the unnamed speaker, which he dubbed the "IMF". The British pair, along with Hayes' colleague David Brown, agreed to form a UK company to design and manufacture speakers which would be sold by Fried in the United States. John Hayes later wrote that:
Of course, Bud, had called it the IMF, and therefore, perhaps mistakenly we registered IMF and formed an IMF company... At no time did Bud Fried have any input on the designs. We sold him speakers and he was the US Distributor... [...] Bud Fried was never a Director or shareholder of IMF Electronics. IMF electronics were the only company manufacturing the transmission line speakers. The name IMF was adopted because Bud Fried had demonstrated the first prototype speakers at the New York hi fi show, and because of the publicity and the fact that he had used his name on the then unnamed speakers, we stuck with the name which was a mistake on our part. It was never his company. After our lawsuit he called his speakers Fried.
The relationship broke down acrimoniously when Fried began to make his own, poorer quality speakers, also marketed as "IMF", and refused to cease until a court agreed that the UK business had the right to the trademark IMF for loudspeakers. Following the split, Fried in the USA (under the brandname "Fried") and the three founders of IMF Electronics in the UK (via a joint venture with driver manufacturer Elac under the name TDL), both became well known in audiophile circles for many years as major advocates of transmission line speaker design. TDL closed after John Wright's gradual failing health and death in 1999 from cancer. He was described in his 1999 obituary as "one of the most important figures on the British hi-fi scene since the mid-1960s... best remembered for his transmission-line loudspeaker designs". The brand was acquired by Audio Partnerships (part of retailer group Richer Sounds). Fried died six years later, in 2005.
21st century
In the early 21st century, mathematical models that seemed to approximate the behavior of real-world TL speakers and cabinets began to emerge. According to the website t-linespeakers.org, this led to an understanding that what he termed the "classical" speakers, designed largely by "trial and error", were a "good job" and the best that was reasonably possible at those time, but that better designs were now achievable based on modeled responses.
Design principles
Phase inversion is achieved by selecting a length of line that is equal to the quarter wavelength of the target lowest frequency. The effect is illustrated in Fig. 1, which shows a hard boundary at one end (the speaker) and the open-ended line vent at the other. The phase relationship between the bass driver and vent is in phase in the pass band until the frequency approaches the quarter wavelength, when the relationship reaches 90 degrees as shown. However, by this time the vent is producing most of the output (Fig. 2). Because the line is operating over several octaves with the drive unit, cone excursion is reduced, providing higher SPLs and lower distortion levels, as compared with bass reflex and infinite baffle loudspeaker enclosure designs.
The complex loading of the bass drive unit demands specific Thiele-Small driver parameters to realise the full benefits of a TL design. Most drive units in the marketplace are developed for the more common reflex and infinite baffle designs and are usually not suitable for TL loading. High efficiency bass drivers with extended low frequency ability, are usually designed to be extremely light and flexible, having very compliant suspensions. Whilst performing well in a reflex design, these characteristics do not match the demands of a TL design. The drive unit is effectively coupled to a long column of air which has mass. This lowers the resonant frequency of the drive unit, negating the need for a highly compliant device. Furthermore, the column of air provides greater force on the driver itself than a driver opening onto a large volume of air (in simple terms it provides more resistance to the driver's attempt to move it), so to control the movement of air requires an extremely rigid cone, to avoid deformation and consequent distortion.
The introduction of the absorption materials reduces the velocity of sound through the line, as discovered by Bailey in his original work. Bradbury published his extensive tests to determine this effect in an AES journal article in 1976, and his results agreed that heavily damped lines could reduce the velocity of sound by as much as 50%, although 35% is typical in medium damped lines. The behaviour of various damping materials has also been studied by Lusztak and Bujacz. Bradbury's tests were carried out using fibrous materials, typically longhaired wool and glass fibre. However, these kinds of materials produce highly variable effects that are not consistently repeatable for production purposes. They are also liable to produce inconsistencies due to movement, climatic factors and effects over time. High specification acoustic foams, developed by manufacturers such as PMC, with similar characteristics to longhaired wool, provide repeatable results for consistent production. The density of the polymer, the diameter of the pores and the sculptured profiling are all specified to provide the correct absorption for each speaker model. The quantity and position of the foam is critical to engineer a low-pass acoustic filter that provides adequate attenuation of the upper bass frequencies, whilst allowing an unimpeded path for the low bass frequencies. Although the result may require a lot of modeling and testing, the starting point is usually based on one of three basic principles. Filling the entire tube treats the TL as a damper, aiming at completely eliminating the rear wave. Filling half the cross section throughout the line's entire length treats the TL as an infinite baffle, basically damping high frequencies and wall-to-wall resonances. Filling the tube from the driver to half the tube's length aims at a quarter-wave resonator, leaving the fundamental tone with its velocity maxima at the open end of the tube intact, while damping all the overtones.
Mathematical equations, modelling, and design process
The external links section of this article links to a number of resources that detail the mathematical principles, models, and DIY calculations, as well as extended practical design material, related to transmission line speakers.
For most of the 20th century, transmission line design remained more of an art than a science, requiring much trial and error. Jon Risch states in an article on classic transmission line design, that the hard part was finding the best stuffing density along the line's length, because "the line stuffing affects both the total apparent line length AND the total apparent box volume simultaneously". He summarized the state of design at the time as:
"The classic transmission line bass enclosure has never been completely and successfully modeled such that it can be built from a pat set of equations. Some claim to have done this, but it doesn't seem to allow a first time build without adjustments, so the models have enough wrong to require a fudge factor..."
Dave D'Lugos, founder of fan site t-linespeakers.org, comments that this reflects the "classical" designs from the 1960s until Risch's writing, during which period "TL design was seat of the pants".
However, from the 21st Century, Martin King and George Augspurger (both separately and referencing each other's works), produced models which show these to be "generally less than optimal" designs which "did a good job of approaching what was possible in their day". Audio engineer Augspurger had modeled TLs using an electrical analogy, and found it to agree closely with King's existing work, based on a mechanical analogy. D'Lugos concluded in his overview of TL modeling and design theory: "I think that using modern drivers and tools such as King's software you can build a better TL easier today".
More recently, Andrea Rubino has developed a sophisticated simulation model based on electrical circuit theory and published a series of articles in the Italian electroacoustic journal AUDIOreview. Many resources are available on his website: transmissionlinespeakers.com
In addition to these more sophisticated models a number of approximation algorithms exist. One such is to design a closed-box loudspeaker enclosure, then building a transmission line of the same volume tuned to the closed-box loudspeaker's resonance frequency. Another is to design a bass reflex loudspeaker, again building a transmission line of the same volume, tuned to the frequency of the Helmholtz resonator.
Prominent individuals and companies
Pioneers:
Benjamin Olney – originated the idea of a duct in speaker enclosure design, which he termed an "acoustic labyrinth", while working for Stromberg-Carlson as an acoustic engineer and studying the effect of enclosure size on output sound.
Bailey and Radford – worked together and developed the concept for loudspeakers (1965). Their design was a significant development from the earlier work. Bailey's name was on the article and Radford built the first commercial TL speaker.
John Wright together with business partner John Hayes and (later) David Brown, and their company IMF Electronics Ltd (later: TDL) – Wright, a "fanatical" pursuer of quality, had designed an award winning tonearm and to demonstrate it, brought to New York a non-commercial TL speaker he had also designed. The speaker gained considerable attention and Wright, Hayes and colleague Brown formed a company that specialized in TL speakers, and won numerous awards (1968). TDL disbanded following Wright's death in 1999 and the brand—as a shell—was bought by Richer Sounds.
Irving M. "Bud" Fried – American audiophile and TL advocate, who encountered Wright and Hayes in 1968, recognized the potential of Wright's unnamed speaker, and began marketing their TL speakers in the United States. Later set up a TL company of his own to design speakers.
Bo Hansson – Swedish designer of HiFi equipment and founder of Opus3 Record Company created the "Rauna Njord" concrete speaker as a transmission line design.
Martin King and George Augspurger – researchers and designers who succeeded in modeling realistic TL speaker designs in the early 21st century.
Other companies and individuals who have produced or researched TL speakers:
Lentek
Newtronics (Temperance line)
Gini B+ (Bass Extenders line)
Quadral
T+A Electronics (Criterion line)
J M Reynaud,
PMC
Salk Sound
Rega (their Naos then RS7)
Adelaide Speakers
TBI Audio Systems LLC (subcontracted by Asis to research and design smaller TL speakers suitable for embedding into laptops)
Marantz (Karoke range)
Merkel Acoustic Research/Jeff Merkel
Albedo (Helmholine range)
Transmission Audio
Audio Reference (Acoustic Zen line)
Radford
DIY kit manufacturers:
UK – IPL Acoustics (article), Falcon Acoustics "Thor"
USA – GR Research "N3"
USA – New York Acoustics, loosely associated with New York Audio Labs , kits and plans for 8" and 10" driver TL speaker cabinets, active in the mid 1980s.
See also
Acoustic suspension – a method of loudspeaker cabinet design and utilisation that uses one or more loudspeaker drivers mounted in a sealed box or cabinet.
Bass reflex – a type of loudspeaker enclosure that uses a port (hole) or vent cut into the cabinet and a section of tubing or pipe affixed to the port.
Frequency response
Loudspeaker acoustics
Loudspeaker enclosure
Loudspeaker measurement
Passive radiator
References
External links
Transmission Line Speakers Pages – TL projects, history and more.
Quarter-wave.com – by Martin J King, developer of TL modeling software; also includes design calculations for professional and DIY TL speaker creation.
http://www.perrymarshall.com/articles/industrial/transmission-line/ - mathematics of the TL speaker, Perry Marshall
Brines Acoustics Articles (Archived 2009-10-24) – Application, tips, essays.
Loudspeaker Handbook and Lexicon, Windslow Burhoe, 1978 (revised 1995/95/97) – has a sizeable section on TL speakers.
Papers
Papers and documents related to Olney's original "Acoustic Labyrinth":
Acoustic Society of America paper, 1936
Followup paper, 1937
1934/1936 "Labyrinth" patent, filed by Stromberg-Carlson Telephone Manufacturing Co.
Archive of Olney's papers at Rush Rhees Library, University of Rochester
Loudspeaker technology
Audio engineering | Transmission line loudspeaker | [
"Engineering"
] | 6,021 | [
"Electrical engineering",
"Audio engineering"
] |
44,224,665 | https://en.wikipedia.org/wiki/Felix%20Chayes%20Prize | The Felix Chayes Prize is presented in alternate years for Excellence in Research in Mathematical Petrology by the International Association for Mathematical Geosciences (IAMG). The cash prize, named after American geologist and petrographer Felix Chayes, was established in 1997
Recipients
Source:
1997 Committee on Data Bases for Petrology
1999 Hugh R. Rollinson
2001 James Nicholls
2003 Antonella Buccianti
2005 Eric Grunsky
2007 Hilmar von Eynatten
2009 not awarded for this year
2011 Istvan Dunkl
2013 Raimon Tolosana-Delgado
2015 Yongzhang Zhou
2017 Clifford R. Stanley
2019 Peter Filzmoser
2021 Grethe Hystad
2023 Pieter Vermeesch
2025 Dario Grana
See also
List of geology awards
List of geophysics awards
List of mathematics awards
References
Awards of the International Association for Mathematical Geosciences
Awards established in 1997 | Felix Chayes Prize | [
"Technology"
] | 181 | [
"Science and technology awards",
"Science award stubs"
] |
44,225,070 | https://en.wikipedia.org/wiki/Reviews%20on%20Environmental%20Health | Reviews on Environmental Health is a quarterly peer-reviewed review journal covering the field of environmental health. It was established in 1972 and was published by Freund Publishing House until 2011, when it moved to Walter de Gruyter. The editors-in-chief are David O. Carpenter and Peter Sly.
Abstracting and indexing
The journal is abstracted and indexed in:
References
External links
De Gruyter academic journals
Quarterly journals
Academic journals established in 1972
Environmental health journals
Review journals | Reviews on Environmental Health | [
"Environmental_science"
] | 98 | [
"Environmental science journals",
"Environmental health journals"
] |
44,226,082 | https://en.wikipedia.org/wiki/Employee%20experience%20design | Employee experience design (EED or EXD) is the application of experience design in order to intentionally design HR products, services, events, and organizational environments with a focus on the quality of the employee experience whilst providing relevant solutions for an organization.
Overview
EED can be described as the "intentional design of the active or passive use of HR products or services", and employee experiences in general, that affect employees' emotional reaction and therefore their particular behaviors and loyalty.
The underlying assumption is that best (customer/employee) relationships are emotional in nature and achieved when companies succeed in not only satisfying certain needs (e.g. compensation), but also making interactions pleasurable.
The goal is to yield better customer experience through increased employee engagement and employee empowerment. Following Krippendorf, EED focuses on creating meaningful and sense-making opportunities for engagement, and addressing aspirational and fundamental psychological needs of an employee, such as autonomy, competence and relatedness.
Methods
Related to design strategy, EED is a participatory systems approach to workplace improvements that applies methods and principles of experience design, such as design thinking, co-creation and empathic design and new digital tools and technologies. It also uses tools and techniques that are typical to customer experience management and service design, e.g. employee experience journey mapping or touchpoint analysis.
Primary design object is the employee experience, which – when successful – an employee finds unique, memorable and sustainable over time, would want to repeat and build upon, and enthusiastically promotes via word of mouth. It is suspected to encourage loyalty by creating an emotional connection through engaging, compelling, and consistent context. The categories for employee experience design context are products, processes, artefacts, content, space and interactions.
While employee experience design is beneficial to create positive customer experiences, it is also beneficial for non-customer-facing roles.
Many elements can make up a successful employee experience, including office environment, reward and benefits, flexible working and casual dress policies.
Stakeholders
Human resource management, operating across hierarchies and departments, plays a central role in design, distribution and delivery of EED. As co-creation is an important design principle, it is a shared task and joint responsibility of leadership, HR professionals and employees. Following the logic of the service-profit chain, beneficiaries are also customers, as the recipients of improved service quality and the organization itself through increased profits.
References
Human resource management
Design
Workplace
Employee relations
Strategic management
Customer experience | Employee experience design | [
"Engineering"
] | 500 | [
"Design"
] |
44,226,817 | https://en.wikipedia.org/wiki/Construction%20partnering | Construction partnering is a type of business partnering used in the architecture, engineering and construction industry. Partnering is intended to assist project teams with setting goals, resolving disputes and improving project outcomes. The construction partnering team is made up of the project’s owner (client), the consulting engineers and/or architects, the contractor(s) and other key project stakeholders. Construction partnering has been used both in the United States and elsewhere since the early 1980s as a methodology to reduce litigation and improve productivity.
Objective
The objective of construction partnering is to reduce project costs and schedules, eliminate change orders and claims, and improve communication by developing mutually agreed upon project and partnership success goals and by monitoring the achievement of these goals for the duration of the project. The construction partnering team will also develop an agreed upon process for resolving disputes should they arise, called a dispute resolution ladder.
History
Litigation in the construction industry exploded in the 1980s. Settling disputes in court added time, cost and energy to projects. Partnering was developed as a response to this problem.
In 1987, the Construction Industry Institute (CII) at The University of Texas at Austin formed a task force to explore a process to address the issues that brought the people involved in construction projects to litigation. The task force’s objectives were to examine the risks and benefits of partnering, to provide guidelines on the process and to define the relationship between partnering and the construction contract. Their results were published in January 1991 entitled, “In Search of Partnering Excellence.”
The United States Army Corps of Engineers (Corps) was part of the CII task force. In the late 1980s the Corps used the partnering process on two of their construction projects with great success. As a result, the Corps established a partnering program in 1991. They developed a pamphlet which described the partnering process, the reasons for using it, and the potential benefits of partnering the Corps had experienced in using the process.
The Associated General Contractors of America (AGC) endorsed partnering in January 1991. AGC’s president announced that one of AGC’s objectives for the year was building construction quality through partnering. They later developed an annual award, the Marvin M. Black Award, to honor the AGC contractor(s) involved in the year’s most collaborative project.
In the past 30 years, partnering has grown. Twenty-four of the 50 state transportation programs have adopted Partnering as an important process to improve outcomes. Several state transportation departments (including Caltrans, Maryland SHA, ODOT, ADOT and others) now require a partnering process by specification.
Awards
Construction partnering has been embraced by the construction industry to such an extent that organizations have started offering awards to the top partnered projects. The current awards available are:
AGC’s Marvin M. Black Partnering Competition (https://web.archive.org/web/20141013112037/http://www.agc.org/cs/about_agc/recognition_programs/marvin_m_black_partnering_awards)
AGC’s Excellence in Partnering Award (https://web.archive.org/web/20141015013431/http://www.agc-ca.org/about.aspx?id=60)
Arizona Transportation’s Partnering Excellence Awards https://www.azdot.gov/business/programs-and-partnerships/partnering/2013-arizona-transportation-partnering-excellence-awards
Caltrans’ Excellence in Partnering (http://www.dot.ca.gov/hq/construc/partnering.php)
International Partnering Institute’s Annual Partnering Awards http://partneringinstitute.org/awards/
Kansas Department of Transportation’s KCA/KDOT Partnering Awards http://www.ksdot.org/bureaus/divoperat/partnering/awards.asp
Nevada Department of Transportation’s Excellence in Partnering Awards https://web.archive.org/web/20141022055048/http://www.nevadadot.com/Doing_Business/Partnering_Program.aspx
Ohio DOT’s Don Conway Partnering Award (http://www.dot.state.oh.us/Divisions/ConstructionMgt/Pages/Partnering.aspx)
References
Construction industry | Construction partnering | [
"Engineering"
] | 879 | [
"Construction",
"Construction industry"
] |
44,227,105 | https://en.wikipedia.org/wiki/List%20of%20protein%20secondary%20structure%20prediction%20programs | List of notable protein secondary structure prediction programs
See also
List of protein structure prediction software
Protein structure prediction
Structural bioinformatics software
Protein structure
Protein methods | List of protein secondary structure prediction programs | [
"Chemistry",
"Biology"
] | 31 | [
"Biochemistry methods",
"Protein methods",
"Protein biochemistry",
"Structural biology",
"Protein structure"
] |
44,228,628 | https://en.wikipedia.org/wiki/Emilio%20Baiada | Emilio Baiada (January 12, 1914 in Tunis – May 14, 1984 in Modena) was an Italian mathematician.
Education and career
He studied at the Scuola Normale Superiore in Pisa, where he graduated with highest honors in June 1937 along with Leonida Tonelli, with whom he worked as an assistant from 1938 to 1941, when he left for the war. In 1945 he began to teach analysis, theory of functions, calculus and rational mechanics at the Scuola Normale. In 1948 he obtained a degree in Analysis; his Ph.D. thesis was written under the direction of Tonelli and Marston Morse.
In 1949 he moved first to University of Cincinnati, where he worked with scientists like Otto Szász and Charles Napoleon Moore, and then to Princeton University, where he worked with Morse.
In 1952 he obtained the chair of analysis of the University of Palermo, where he taught until 1961 before transferring to the University of Modena, where he re-launched the Institute of Mathematics and developed its Library and Mathematical Seminar.
Contributions
He published more than 60 papers on differential equations, Fourier series and the series expansion of orthonormal functions, topology of varieties, real analysis, calculus of variations and the theory of functions.
Recognition
Baiada won the Michel prize for the best thesis in Pisa and the Whiting Award in 1940 for "contributions on subjects of calculus of variations".
Notes
1984 deaths
1914 births
20th-century Italian mathematicians
People from Tunis
Scuola Normale Superiore di Pisa alumni
Academic staff of the University of Palermo
Academic staff of the University of Modena and Reggio Emilia | Emilio Baiada | [
"Mathematics"
] | 327 | [
"Mathematical analysis",
"Mathematical analysts"
] |
44,229,008 | https://en.wikipedia.org/wiki/Cologix | Cologix, a network neutral interconnection and data center company, runs 40+ interconnection locations across 11 North American markets. The edge markets that Cologix operates in are: Columbus, Ohio; Dallas, Texas; Jacksonville, Florida; Lakeland, Florida; Minneapolis, Minnesota; Montreal, Quebec; Silicon Valley, California; Toronto, Ontario; Northern New Jersey; and Vancouver, British Columbia. The company supports five Internet exchanges.
Network locations
United States
Columbus
Cologix acquired DataCenter.BZ in February 2014, procuring 86,000 square feet across two data center facilities at 555 Scherers Court in Columbus, Ohio. The site is located at the intersection of two trunks of nationwide fiber optic network that carries most Internet transmissions. Cologix is currently building a 160K SQF (half of that raised floor space), 18+MW data center on its existing 8 acre campus in Columbus. The new $130M+ facility will become Columbus's largest neutral data center and will be directly linked to Cologix's existing data centers, offering connectivity to 45+ network service providers, 20+ cloud service providers and the Ohio-IX Internet Exchange. The data center is built based on a concurrently maintainable design with 2N power and N+1 cooling. Further attributes include an EF-4 tornado rating, K-rated perimeter fence and 24x7 guards. The redundancy, scale, security and connectivity enables Cologix to address growing market demand ranging from individual cabinets to multi-megawatt deployments.
Dallas
Dallas is situated at the major crossroads of network connectivity in the South Central United States. Cologix operates 40,000 square feet across two data centers in the Dallas Infomart, including a meet-me-room. The Dallas INFOMART building, is the region's preeminent carrier hotel located at 1950 North Stemmons Freeway. The INFOMART has the largest number of carriers out of any single building in a 900-mile radius, with more than 8,700 strands of fiber.
Jacksonville
Cologix operates two data centers in Jacksonville, Florida, including one site, JAX 1, at 421 West Church Street (the region's carrier hotel) and a second data center, JAX 2, at 4800 Spring Park Road which provides enterprise colocation and disaster recovery. Cologix maintains a meet-me-room in the 421 West Church Street building, which is wired with submarine communications cable to connect to Central and South America. Cologix connects the NAP of the Americas in Miami to Cologix's Jacksonville data center and meet-me-room at 421 West Church Street. The 421 West Church Street site facilitates connections at the intersection of metro, long-haul and subsea fiber routes.
Lakeland
Cologix operates 100,000 square feet of data center space in Lakeland, Florida offering disaster recovery and colocation services from their data center.
Minneapolis
Cologix operates data center space and the meet-me-room at the carrier hotel at the 511 Building (Minneapolis) in downtown Minneapolis, Minnesota providing access to more than 80+ network providers. Cologix also provides the Midwest Internet Cooperative Exchange (MICE) with space and power in the 511 Building.
Northern New Jersey
Cologix's Northern New Jersey data centers are located about 30 miles outside of Manhattan and 280+ feet above sea level. Cologix operates four New Jersey data centers and business continuity sites, within more than 230,000 SQF of usable floor space, provide enterprise‐grade colocation, cloud and managed services.
Canada
Cologix operates 11 network neutral data centers in Montreal, Toronto and Vancouver.
Montreal
Cologix operates more than 100,000 square feet of colocation space across seven data centers in Montreal, including in the carrier hotel at 1250 Rene Levesque West. A dedicated fiber ring connects the seven Cologix Montreal sites to share connectivity. Additionally, the Montreal Internet Exchange (also known as QIX), deployed a core node with Cologix at 1250 René-Lévesque West and 625 René-Lévesque West.
Toronto
Cologix operates two downtown Toronto data centers at 151 Front Street (the area's carrier hotel) and 905 King West in Toronto. The sites share connectivity through the use of a diverse metro fiber ring. Cologix Toronto offers access to more than 150+ networks and provides a direct on-ramp to the Toronto Internet Exchange (TORIX).
Vancouver
Cologix operates three data centers in Vancouver, British Columbia in the Harbour Centre at 555 West Hastings Street (VAN1), 1050 West Pender Street (VAN2) and 2828 Natal Street (VAN3). The Vancouver Internet Exchange (VANIX), an open and participant-run non-profit Internet exchange, has nodes in VAN2 and in VAN3. Cologix is partnered with VANIX, contributing space and fiber optic network connections.
References
Technology companies of the United States
Internet technology companies of the United States
Data centers
Companies based in Colorado | Cologix | [
"Technology"
] | 1,029 | [
"Data centers",
"Computers"
] |
44,229,916 | https://en.wikipedia.org/wiki/Cloud%20Application%20Management%20for%20Platforms | Cloud Application Management for Platforms (CAMP) is a specification for managing applications in the context of a platform as a service (PaaS) system. CAMP is designed to address the needs of a high-level PaaS system; one in which the consumer (generally a developer or application administrator) provides application artifacts (code, data, graphics, etc.) and specifies which provider-supplied services are required to realize these artifacts as an application. The details of the infrastructure (compute, storage, and networking) used to support these services are hidden from the consumer by the provider of the PaaS system.
Overview
CAMP defines the following:
A domain-specific language that describes the artifacts that make up an application, the services that are required to execute or utilize those artifacts, and the relationship of the artifacts to those services.
A resource model for representing applications and their constituent components as well the services used by those components along with runtime status information, configuration information, and metadata that describes the PaaS system.
A REST protocol for manipulating these resources and, by so doing, changing the state of the underlying application.
History
CAMP 1.0
CAMP 1.0 was produced as a collaboration between CloudBees, Cloudsoft Corporation, Huawei, Oracle, Rackspace, Red Hat, and Software AG. It was published in August, 2012.
CAMP 1.1
In August, 2012 CAMP 1.0 was submitted to OASIS CAMP Technical Committee with the intent of producing an OASIS Standard. This Technical Committee has produced an OASIS Committee Specification. As per its charter, the CAMP TC is awaiting evidence of two, interoperable implementations of CAMP v1.1 before asking OASIS to approve the specification as an OASIS Standard.
Motivation
Most PaaS systems provide some form of application management API. These APIs are used to upload applications into the cloud, configure which services will be used to run the application, start the application, monitor the status and performance of the application, stop the application, etc. These APIs are usually hidden behind a web application and/or a command line tool. This kind of API is a "me too" technology; its existence is a necessary prerequisite to providing a workable PaaS system, but there is little advantage in providing a better management API than your competitors. No one ever selected a PaaS offering solely on the strength of its application management API.
Disadvantages
The fact that every PaaS system provides a custom application management API creates a number of issues:
Any monitoring or management systems, continuous integration systems, etc. written to consume such APIs will need to be re-written if a customer wishes to change or add additional PaaS systems. This increases the cost of switching between PaaS providers which, in turn, decreases the value of using PaaS.
Integrated development environments that wish to target PaaS environments must do so on an individual, case-by-case basis (for example, by providing custom connectors for each PaaS system). This both increases the initial development effort as well as the accumulated "technical debt" of maintaining each of these connectors.
Because the quality of the application management API is not a differentiator, time spent designing/tweaking the management API is not a good investment. Platform providers can save time and resources by implementing a basic, consensus API. Value-added features can be implemented as extensions to this basic API.
CAMP Implementations
nCAMP
Developed in tandem with the work of the OASIS CAMP Technical Committee, nCAMP is a proof-of-concept implementation of the CAMP v1.1 specification. nCAMP was not intended to be a useful PaaS system but, instead, to act as a vehicle to test the concepts and constructs of the CAMP specification. nCAMP presents a simple system that uses Tomcat and MySQL to support Java Servlet based web applications that can use MySQL as a database.
Project Solum
Solum is an OpenStack Related Stackforge project designed to make cloud services easier to consume and integrate into the developers application development process. Solum's resource model and plan schema are based on CAMP, but not fully CAMP compliant. Work is currently ongoing to provide an additional, CAMP-compliant API in addition to the native Solum API.
Apache Brooklyn
Apache Brooklyn is a framework for modeling, monitoring, and managing applications through autonomic blueprints. Apache Brooklyn blueprints conform to CAMP v1.1 Public Review Draft 01.
References
External links
CAMP in 7 Slide: Episode 2
Cloud Mining Overview
IPv4 Cloud VPS
Cloud standards | Cloud Application Management for Platforms | [
"Technology"
] | 927 | [
"Computer standards",
"Cloud standards"
] |
44,230,614 | https://en.wikipedia.org/wiki/C%2B%2B17 | C++17 is a version of the ISO/IEC 14882 standard for the C++ programming language. C++17 replaced the prior version of the C++ standard, called C++14, and was later replaced by C++20.
History
Before the C++ Standards Committee fixed a 3-year release cycle, C++17's release date was uncertain. In that time period, the C++17 revision was also called C++1z, following C++0x or C++1x for C++11 and C++1y for C++14. The C++17 specification reached the Draft International Standard (DIS) stage in March 2017. This DIS was unanimously approved, with only editorial comments, and the final standard was published in December 2017. Few changes were made to the C++ Standard Template Library, although some algorithms in the <algorithm> header were given support for explicit parallelization and some syntactic enhancements were made.
New features
C++17 introduced many new features. The following lists may be incomplete.
Language
Making the text message for optional
Allow (as an alternative to ) in a template template parameter
New rules for deduction from braced-init-list
Nested namespace definitions, e.g., instead of
Allowing attributes for namespaces and enumerators
New standard attributes , and
UTF-8 () character literals (UTF-8 string literals have existed since C++11; C++17 adds the corresponding character literals for consistency, though as they are restricted to a single byte they can only store "Basic Latin" and C0 control codes, i.e. ASCII)
Hexadecimal floating-point literals
Use of as the type for a non-type template parameter
Constant evaluation for all non-type template arguments
Fold expressions, for variadic templates
A compile-time static with the form
Structured binding declarations, allowing
Initializers in and statements
copy-initialization and direct-initialization of objects of type from prvalue expressions of type (ignoring top-level cv-qualifiers) shall result in no copy or move constructors from the prvalue expression. See copy elision for more information.
Some extensions on over-aligned memory allocation
(CTAD), introducing constructor deduction guides, e.g. allowing instead of requiring explicit constructor arguments types or an additional helper template function .
Inline variables, which allows the definition of variables in header files without violating the one definition rule. The rules are effectively the same as inline functions
, allowing the availability of a header to be checked by preprocessor directives
Value of changed to
Exception specifications were made part of the function type
Lambda expressions can capture "*this" by value
Library
Most of Library Fundamentals TS I, including:
std::string_view, a read-only non-owning reference to a character sequence or string-slice
std::optional, for representing optional objects, a data type that may not always be returned by a given algorithm with support for non-return
std::any, for holding single values of any type
std::uncaught_exceptions, as a replacement of std::uncaught_exception in exception handling
New insertion functions try_emplace and insert_or_assign for std::map and std::unordered_map key-value associative data structures
Uniform container access: std::size, std::empty and std::data
Definition of "contiguous iterators"
A file system library based on boost::filesystem
Parallel versions of STL algorithms
Additional mathematical special functions, including elliptic integrals and Bessel functions
std::variant, a tagged union container
std::byte, allowing char to be replaced for data types intending to model a byte of data as a byte rather than a character
Logical operator traits: std::conjunction, std::disjunction and std::negation
<memory_resource> header, for polymorphic memory resources
Removed features
This revision of C++ not only added new features but also removed a few.
Trigraphs were removed.
Some deprecated types and functions were removed from the standard library, including std::auto_ptr, std::random_shuffle, and old function adaptors. These were superseded in C++11 by improved facilities such as std::unique_ptr, std::shuffle, std::bind, and lambdas.
The (formerly deprecated) use of the keyword as a storage class specifier was removed. This keyword is still reserved but now unused.
Compiler support
GCC has had complete support for C++17 language features since version 8.
Clang 5 and later supports all C++17 language features.
Visual Studio 2017 15.8 (MSVC 19.15) and later supports all C++17 language features.
Library support
libstdc++ since version 9.1 has complete support for C++17 (8.1 without Parallelism TS and referring to C99 instead of C11)
libc++ as of version 9 has partial support for C++17, with the remainder "in progress"
Visual Studio 2017 15.8 (MSVC 19.15) Standard Library and later supports all C++17 library features except for "Elementary String Conversions" and referring to C99 instead of C11. "Elementary String Conversions" is added in Visual Studio 2019 16.4
See also
C++ compilers
C11 (C standard revision)
C17 (C standard revision)
References
C++
C++ programming language family
Programming language standards
Articles with example C++ code
IEC standards
ISO standards | C++17 | [
"Technology"
] | 1,216 | [
"Computer standards",
"Programming language standards",
"IEC standards"
] |
47,325,142 | https://en.wikipedia.org/wiki/Brazilian%20Chemical%20Society | The Brazilian Chemical Society () is a not for profit organization that supports Chemistry in Brazil. It was founded in 1977.
It publishes the Journal of the Brazilian Chemical Society.
References
External links
Chemistry societies
1977 establishments in Brazil
Learned societies of Brazil | Brazilian Chemical Society | [
"Chemistry"
] | 49 | [
"Chemistry societies",
"nan",
"Chemistry organization stubs"
] |
47,325,951 | https://en.wikipedia.org/wiki/Quintessence%3A%20The%20Search%20for%20Missing%20Mass%20in%20the%20Universe | Quintessence: The Search for Missing Mass in the Universe is the fifth non-fiction book by the American theoretical physicist Lawrence M. Krauss. The book was published by Basic Books on December 21, 2000. This text is an update of his 1989 book The Fifth Essence. It was retitled Quintessence after the now widely accepted term for dark energy.
Overview
Krauss focuses on theoretical physics and has published researches on a number of topics within that field. His primary contribution is to cosmology as one of the first physicists to suggest that most of the mass and energy of the universe resides in empty space, an idea now widely known as "dark energy". Furthermore, Krauss has formulated a model in which the universe could have potentially come from "nothing," as outlined in his later book A Universe from Nothing.
Whether our universe is ever-expanding depends on the amount and properties of matter, but there is too little visible matter around us to explain the behavior we can see—over 90% of the universe consists of the missing mass or dark matter, which Krauss termed "the fifth essence." In this book Krauss demonstrates how the dark matter problem is now connected with two widely discussed areas in the modern cosmology: the ultimate fate of the universe and the cosmological constant. He also discusses an antigravity force that may explain recent observations of a permanently expanding universe.
See also
Dark matter in fiction
Exotic matter
Mirror matter
Negative mass
Quintessence (physics)
Scalar field dark matter
Self-interacting dark matter
Unparticle physics
The 4 Percent Universe
References
External links
Popular physics books
2000 non-fiction books
Books by Lawrence M. Krauss
Dark energy | Quintessence: The Search for Missing Mass in the Universe | [
"Physics",
"Astronomy"
] | 352 | [
"Unsolved problems in astronomy",
"Physical quantities",
"Concepts in astronomy",
"Unsolved problems in physics",
"Energy (physics)",
"Dark energy",
"Wikipedia categories named after physical quantities"
] |
47,326,005 | https://en.wikipedia.org/wiki/Susan%20M.%20Gasser | Susan M. Gasser (born 1955) is a Swiss molecular biologist. From 2004 to 2019 she was the director of the Friedrich Miescher Institute for Biomedical Research in Basel, Switzerland, where she also led a research group from 2004 until 2021. She was in parallel professor of molecular biology at the University of Basel until April 2021. Since January 2021, Susan Gasser is director of the ISREC Foundation, which supports translational cancer research. She is also professor invité at the University of Lausanne in the department of fundamental microbiology. She is an expert in quantitative biology and studies epigenetic inheritance and genome stability. Recipient of multiple swiss and European awards, she was named member of the US Academy of Sciences in 2022.
Early career
Susan Gasser received her doctorate from the University of Basel in biochemistry in 1982 at the Biozentrum of the University of Basel, after a BA at the University of Chicago with an honors thesis in biophysics (1979). For her PhD she developed an in vitro system for the import of mitochondrial proteins, demonstrating its energy-dependence and identifying compartment-specific processing pathways for protein import, advised by Gottfried (Jeff) Schatz. As a post-doctoral fellow at the University of Geneva with Ulrich K. Laemmli, she established roles for topoisomerase II in metaphase chromosome structure (Gasser et al., JMB 1986) and for A/T-rich sequences in long-range chromatin folding (Gasser and Laemmli, Cell 1986). She established her own laboratory at the ISREC Swiss Institute for Experimental Cancer Research in Epalinges in 1986.
Career
Gasser led a research group at the Swiss Institute for Experimental Cancer Research in Epalinges sur Lausanne until 2001. There she pioneered live fluorescence imaging of telomeres and repressed chromatin in budding yeast, coupled with biochemical and genetic approaches to understand chromosome structure and nuclear organization. She was named professor of molecular biology at the University of Geneva in 2001, prior to moving to lead the Friedrich Miescher Institute for Biomedical Research in Basel, as its director, in 2004. At the FMI she led a research group studying the spatial organization of double-strand break repair and checkpoint activation in budding yeast, and established and exploited means to study the spatial organization and genetic control of heterochromatin in C. elegans. From 2005 - 2021 she also held the position of full professor at the University of Basel. Since February 2021, she is the director of the ISREC Foundation, which built and maintains the new Agora institute of translational cancer research in Lausanne, Switzerland. Susan Gasser also holds the position of guest professor at the University of Lausanne.
Initially studying chromatin organization in budding yeast, her laboratory combined genetic, biochemical and fluorescence microscopy approaches, developing quantitative live imaging tools to study the subnuclear dynamics of DNA loci in living cells. Her work elucidated roles of histone modifications and turnover in genome stability and in the spatial organization of chromatin in the interphase nucleus, with an emphasis on the function of subnuclear compartments both in yeast, and in C. elegans during tissue differentiation.
Gasser has served on review boards and advisory councils throughout Switzerland, Europe. and Japan, and is currently chair of the strategic advisory board of the Helmholtz Society Health Program of Germany. She serves on the ETH board (Rat der Eidgenossosichen Technischen Hochschulen) and the Swiss Science Council. She served on the Gairdner Foundation Medical Prize committee, the EMBL Science Advisory Council (SAC), and numerous foundation boards in Switzerland. She was vice chairperson and then chairperson of the EMBO Council (2000-2004) and a member of the President's Science and Technology Advisory Council (PSTAC; 2012-2014) for the European Commission. Susan led the Gender Committee of the Swiss National Science Foundation from 2014- 2019, and initiated the PRIMA program for the promotion of women in academia, actively promoting the careers of women scientists in Switzerland. In Japan, she co-founded the Women in Science Japan organization. Susan Gasser was elected to the National Academy of Science (US), the Académie de France, Leopoldina, EMBO, AAAS, as well as the Swiss Academy of Medical Sciences, and received the INSERM International Prize in 2011, the FEBS | EMBO Women in Science Award in 2012, the Weizmann Institute Women in Science award, the Otto Naegeli prize in medical research (2006). She holds honorary doctorates from the University of Lausanne, the University of Fribourg, the University of Geneva and the Charles University in Prague. In Switzerland she was also recipient of the Friedrich Miescher Award (1991(, the National Latsis Prize (1992), the Otto Naegeli Prize in Medical Sciences (2006) and the Lelio Orci Award (2023).
She has published over 300 articles and reviews in leading journals (see Google Scholar entry) and serves on editorial boards of Molecular Cell, Genes & Development among other journals.
Career history
1977-1979: University of Chicago, Illinois, United States, Bachelor of Arts with honours in biophysics
1979-1982: University of Basel (Biozentrum) PhD Magna cum laude in biochemistry, thesis advisor: Prof. G. Schatz
1983-1986: maître assistant, Prof. U. K. Laemmli, University of Geneva, Department of Molecular Biology
1986-1990: junior group leader, Swiss Institute for Experimental Cancer Research
1991-2001: senior group leader, Swiss Institute for Experimental Cancer Research
2001-2004: full professor, department of molecular biology, University of Geneva, Geneva
2004-2019: director of the Friedrich Miescher Institute for Biomedical Research, Basel
2005-2021: full professor in molecular biology, University of Basel
2021–present: director of the ISREC Foundation, Lausanne and guest professor in fundamental microbiology, University of Lausanne, Switzerland
External appointments (selected)
1993-2002 Member, research council of the Swiss National Science Foundation, Bern
2000-2009 Member, selection committee, SNF Career Professorship Program, Bern
2002-2003 Member, selection jury, Institut Universitaire de France (IUF), Paris
2000-2005 Member, vice-chair and chair of the EMBO Council, Heidelberg, Germany
2004-2007 Member, search committee, Körber-Foundation, Hamburg
2006-2016 Board of directors, Gebert Rüf Foundation, Basel
2005-2010 Advisory Board Cancer Research UK, London Research Institutes, London
2006-2012 Advisory Group on Health Research FP7, European Commission, Brussels
2006-2008 advisory board, Max Planck Institute for Dev. Biology, Tübingen
2006-2011 advisory board, Center for Integrative Genomics, Lausanne
2007-2015 advisory board for Wellcome Trust Centre for Research, Dundee, Scotland
2007-2014 advisory board for Max Planck Institute for Immunology, Freiburg
2008-2018 Member of the Nestlé Nutrition Council, Nestlé SA, Vevey, Switzerland
2012-2017 Member, Wissenschafliches Beirat of the Institute of Advanced Study, Berlin
2016-now Member, Crick Institute Review Board, Scientific Advisory Board, London 2012,
2012-2016 advisory board for Max Planck Institut for Biophysical Chemistry, Göttingen, German
2013-2014 President's Science and Technology Advisory Council (STAC) of the European Commission
2012-2015 Selection committee for the Royal Netherlands Academy of Arts and Science Prize
2003-2018 Member, Human Frontiers Science Organization Grant Review and Career Development Committees, Strasbourg
2014-2018 Member, executive board, Genomics Institute of the Novartis Research Foundation
2014-2020 Chairwoman, Gender Commission of the Swiss National Science Foundation, Bern
2015-2022 Member, scientific advisory council, European Molecular Biology Laboratories, Heidelberg, Germany
since 2017 Member, scientific advisory board, University of Tokyo Institute of Quantitative Biology, Tokyo, Japan
2015-2022 Member, Gairdner Prize Award committee, Toronto, Canada
since 2016 Member, Swiss Wissenschaftsrat (Swiss Science Council, SSC), Bern
since 2018 Member, ETH Board (Governing Board of the ETH Domain), Switzerland
2019-2027 Chair, strategic board of the Helmholtz Society Health Program, Germany
2019 Chair, review board of the systems biology department, Weizmann Inst. Rehovot, Israel
since 2020 Member, scientific advisory board, Biozentrum der Universität Basel
2020-2022 Board of trustees, Institute of Science and Technology Austria (ISTA), Austria
since 2020 Chair, international review panel for ONCODE network, NL
since 2021 Review board VIB, then scientific advisory board member, Belgium
since 2021 Board of directors, UCB, Brussels, Belgium
Awards
2022 Honorary doctorate, University of Geneva, Switzerland
2022 Elected international member of the National Academy of Sciences USA
2021 Honorary doctorate, University of Fribourg, Switzerland
2016 Honorary doctorate, Charles University, Prague, Czech Republic
2016 Lee Hartwell Award of the Genetics Society of America
2014 Doctorat honoris causa, University of Lausanne
2013 Member of the EC Presidents Science and Technology advisory Council (PSTAC)
2013 Weizmann Institute, Women in Science Award
2012 FEBS/EMBO Women in Science Award
2011 Prix International de l'INSERM, France
2009 Election to American Association for the Advancement of Science
2009 London Royal Society of Chemistry "Nucleic Acid" Award
2007 Election to German Academy of Science, Leopoldina
2006 Election to Academy of Medical Sciences, Switzerland
2006 Gregor Mendel Medal, Czech Academy of Science
2006 Otto Naegeli Prize for Biomedical Research, Switzerland
2005 Foreign member, Académie des Sciences, Institut de France
1999 Medal of Honor, 3rd Medical Faculty of Charles University, Prague
1998 Election to Academia Europaea
1994 Friedrich Miescher Prize, Swiss Society for Biochemistry
1993 Member, EMBO
1991 National Latsis Prize, Swiss National Science Foundation
Selected publications
2020 The LSM2-8 complex and XRN-2 mediate RNA decay at H3K27me3-marked genes in C. elegans. Nature Cell Biology, doi: 10.1038/s41556-020-0504-1 Mattout A, Gaidatzis D, Padeken J, Schmid CD, Aeschlimann F, Kalck V and Gasser SM
2019 Active chromatin marks drive spatial sequestration of heterochromatin in differentiated cells. Nature 569, 734 - 739. doi: 10.1038/s41586-019-1243-y Cabianca D, Munoz Jimenez C, Kalck V, Gaidatzis D, Padeken J, Askjaer P and Gasser SM
2017 Histone degradation in response to DNA damage enhances chromatin dynamics and recombination rates. Nature Struct. Mol. Biology, 24, 99 – 107. doi: 10.1038/nsmb.3347v Hauer MH, Seeber A, Singh V, Thierry R, Sack R, Amitai A, Kryzhanovska M, Eglinger J, Holcman D, Owen-Hughes T and Gasser SM
2016 Histone H3K9 methylation is dispensable for C. elegans development, but suppresses RNA-DNA hybrid-associated repeat instability. Nature Genetics, 48, 1385 - 1395. doi: 10.1038/ng.3672 Zeller P, Padeken J, van Schendel R, Kalck V, Tijsterman M and Gasser SM
2015 Perinuclear Anchoring of H3K9-Methylated Chromatin Stabilizes Induced Cell Fate in C. elegans embryos. Cell, 163, 1333 – 1347 Gonzalez-Sandoval A, Towbin BD, Kalck V, Cabianca DS, Gaidatzis D, Hauer MH, Geng L, Wang L, Yang T, Wang X, Zhao K and Gasser SM
2013 TORC2 signaling pathway guarantees genome stability in face of DNA strand breaks. Mol Cell, 51:829-839, Shimada K, Fillipuzzi I, Stahl M, Helliwell SB, Seeber A, Loewith R, Movva R, Gasser SM
2013 The shelterin protein POT-1 anchors C. elegans telomeres through SUN-1 at the nuclear periphery. J Cell Biol, 203:727-35, Ferreira HC, Towbin BD, Jegou T, Gasser SM
2013 Checkpoint kinases and nucleosome remodelers enhance global chromatin mobility in response to DNA damage. Genes Dev, 27:1999-2008. doi:10.1101/gad.222992.113, Seeber A, Dion V, Gasser SM
2013 Cohesin and the nucleolus constrain the mobility of spontaneous repair foci. EMBO Rep. 14:984-991, Dion V, Kalck V, Seeber A, Schleker T, Gasser SM
2013 SIR proteins and the assembly of silent chromatin in budding yeast. Annu Rev Genet. 47:275-306, Kueng S, Oppikofer M, Gasser SM
2012 Step-wise methylation of histone H3K9 positions chromosome arms at the nuclear periphery in C. elegans embryos. Cell, 150:934-947, Towbin BD, Gonzalez-Aguilera C, Sack R, Gaidatzis D, Kalck V, Meister P, Askjaer P, Gasser SM
2012 Increased dynamics of double strand breaks requires Mec1, Rad9 and the homologous recombination machinery. Nat. Cell Biol. 14:502-509, Dion V, Kalck V, Horigome C, Towbin BD, Gasser SM
2012 Targeted INO80 enhances subnuclear chromatin movement and ectopic homologous recombination. Genes Dev 26:369-38 Neumann FR, Dion V, Gehlen L, Tsai-Pflugfelder M, Schmid R, Taddei A, Gasser SM
2008 Functional Targeting of DNA Damage to a Nuclear Pore-associated SUMO-dependent Ubiquitin ligase. Science, 322, 597 - 602 Nagai S, Dubrana K, Tsai-Pflugfelder M, Davidson MB, Roberts TM, Brown GW, Varela E, Hediger F, Gasser SM* and Krogan NJ
2005 Automatic tracking of individual fluorescence particles - Application to the study of chromosome dynamics. IEEE Transactions on Image Processing, 14, 1372 – 1383 Sage D, Neumann FR, Hediger F, Gasser SM and Unser M
2004 INO80 recruitment by H2A phosphorylation links ATP-dependent chromatin remodeling with DNA double-strand break repair. Cell, 119, 777 – 788 van Attikum H, Fritsch O, Hohn B and Gasser SM
Notes and references
External links
Research Europe
Gasser's Lab
Molecular biologists
Swiss medical researchers
Living people
Women biologists
Women medical researchers
Academic staff of the University of Basel
University of Basel alumni
20th-century Swiss biologists
21st-century Swiss biologists
20th-century women scientists
21st-century Swiss women scientists
1955 births | Susan M. Gasser | [
"Chemistry"
] | 3,220 | [
"Biochemists",
"Molecular biology",
"Molecular biologists"
] |
47,326,615 | https://en.wikipedia.org/wiki/Adamantyl-THPINACA | Adamantyl-THPINACA (ATHPINACA, AD-THPINACA) is an indazole-based synthetic cannabinoid, which was first reported to Europol in Slovenia in January 2015. It is known as both the 1-adamantyl and 2-adamantyl isomers (SGT-40 and SGT-194 respectively), which can be distinguished by GC-EI-MS. It is banned in Sweden and Russia. Both the 1-adamantyl and 2-adamantyl isomers are specifically listed as illegal drugs in Japan. Given the known metabolic liberation (and presence as an impurity) of amantadine in the related compound APINACA, it is suspected that metabolic hydrolysis of the amide group of Adamantyl-THPINACA may also release amantadine.
See also
5F-ADB
5F-AMB
5F-APICA
AB-CHMINACA
AB-FUBINACA
AB-CHFUPYCA
AB-PINACA
ADB-CHMINACA
ADB-FUBINACA
ADB-PINACA
APICA
APINACA
APP-FUBINACA
CUMYL-THPINACA
MDMB-CHMINACA
MDMB-FUBINACA
PX-3
References
Adamantanes
Cannabinoids
Amines
Designer drugs
Indazolecarboxamides | Adamantyl-THPINACA | [
"Chemistry"
] | 273 | [
"Amines",
"Bases (chemistry)",
"Functional groups"
] |
47,327,207 | https://en.wikipedia.org/wiki/Ophiostoma%20breviusculum | Ophiostoma breviusculum is a species of fungus in the family Ophiostomataceae, associated with bark beetles infesting larch in Japan. It belongs in the Ophiostoma piceae complex. It was also found in insect galleries of Larix kaempferi. The length of its perithecial necks and synnemata are shorter than in O. piceae. The synnemata also differ from the latter morphologically.
References
Further reading
Yamaoka, Yuichi, et al. "Constant association of ophiostomatoid fungi with the bark beetle Ips subelongatus invading Japanese larch logs." Mycoscience 50.3 (2009): 165–172.
Paciura, Dina, et al. "Characterisation of synnematous bark beetle-associated fungi from China, including Graphium carbonarium sp. nov." Fungal Diversity40.1 (2010): 75–88.
Grobbelaar, Joha W., et al. "Discovery of Ophiostoma tsotsi on Eucalyptus wood chips in China." Mycoscience 52.2 (2011): 111–118.
External links
MycoBank
Fungi described in 2006
Fungal tree pathogens and diseases
Ophiostomatales
Fungus species | Ophiostoma breviusculum | [
"Biology"
] | 274 | [
"Fungi",
"Fungus species"
] |
47,328,766 | https://en.wikipedia.org/wiki/Beer%20chemistry | The chemical compounds in beer give it a distinctive taste, smell and appearance. The majority of compounds in beer come from the metabolic activities of plants and yeast and so are covered by the fields of biochemistry and organic chemistry. The main exception is that beer contains over 90% water and the mineral ions in the water (hardness) can have a significant effect upon the taste.
Four main ingredients
Four main ingredients are used for making beer in the process of brewing: carbohydrates (from malt), hops, yeast, and water.
Carbohydrates (from malt)
The carbohydrate source is an essential part of the beer because unicellular yeast organisms convert carbohydrates into energy to live. Yeast metabolize the carbohydrate source to form a number of compounds including ethanol. The process of brewing beer starts with malting and mashing, which breaks down the long carbohydrates in the barley grain into more simple sugars. This is important because yeast can only metabolize very short chains of sugars. Long-carbohydrates are polymers, large branching linkages of the same molecule over and over. In the case of barley, we mostly see polymers called amylopectin and amylose which are made of repeating linkages of glucose. On very large time-scales (thermodynamically) these polymers would break down on their own, and there would be no need for the malting process. The process is normally sped up by heating up the barley grain. This heating process activates enzymes called amylases. The shape of these enzymes, their active site, gives them the unique and powerful ability to speed these degradation reactions to over 100,000 times faster. The reaction that takes place at the active site is called a hydrolysis reaction, which is a cleavage of the linkages between the sugars. Repeated hydrolysis breaks the long amylopectin polymers into simpler sugars that can be digested by the yeast.
Hops
Hops are the flowers of the hops plant Humulus lupulus. These flowers contain over 440 essential oils, which contribute to the aroma and non-bitter flavors of beer. However, the distinct bitterness especially characteristic of pale ales comes from a family of compounds called alpha-acids (also called humulones) and beta-acids (also called lupulones). Generally, brewers believe that α-acids give the beer a pleasant bitterness whereas β-acids are considered less pleasant. α-acids isomerize during the boiling process in the reaction pictured. The six-member ring in the humulone isomerizes to a five-member ring, but it is not commonly discussed how this affects perceived bitterness.
Yeast
In beer, the metabolic waste products of yeast are a significant factor. In aerobic conditions, the yeast will use in the glycolysis the simple sugars obtained from the malting process, and convert pyruvate, the major organic product of glycolysis, into carbon dioxide and water via the cellular respiration. Many homebrewers use this aspect of yeast metabolism to carbonate their beers. However, under industrial anaerobic conditions yeasts cannot use pyruvate, the end products of glycolysis, to generate energy in cellular respiration. Instead, they rely on a process called fermentation. Fermentation converts pyruvate into ethanol through the intermediate acetaldehyde.
Water
Water can often play, directly or indirectly, a very important role in the way a beer tastes, as it is the main ingredient. The ion species present in water can affect the metabolic pathways of yeast, and thus the metabolites one can taste. For example, calcium and iron ions are essential in small amounts for yeast to survive, because these metal ions are usually required cofactors for yeast enzymes.
Beer carbonation
In aerobic conditions, yeast turns sugars into pyruvate then converts pyruvate into water and carbon dioxide. This process can carbonate beers. In commercial production, the yeast works in anaerobic conditions to convert pyruvate into ethanol, and does not carbonate beer. Beer is carbonated with pressurized CO2. When beer is poured, carbon dioxide dissolved in the beer escapes and forms tiny bubbles. These bubbles grow and accelerate as they rise by feeding off of nearby smaller bubbles, a phenomenon known as Ostwald ripening. These larger bubbles lead to “coarser” foam on top of poured beer.
Nitro beer (CO2 replaced by N2 gas)
Beers can be carbonated with CO2 or made sparkling with an inert gas such as nitrogen (N2), argon (Ar), or helium (He). Inert gases are not as soluble in water as carbon dioxide, so they form bubbles that do not grow through Ostwald ripening. This means that the beer has smaller bubbles and a more creamy and stable head. These less soluble inert gases give the beer a different and flatter texture. In beer terms, the mouthfeel is smooth, not bubbly like beers with normal carbonation. Nitro beer (for nitrogen beer) could taste less acidic than normal beer.
Aromatic compounds
Beers contain many aromatic substances. Up to now, chemists using advanced analytical instruments such as gas and high performance liquid chromatographs coupled to mass spectrometers, have discovered over 7,700 different chemical compounds in beers.
Foam stabilizers
The beer foam stability depends amongst other on the presence of transition metal ions (, , , ...), macromolecules such as polysaccharides, proteins, and isohumulone compounds from hops in the beer. Foam stability is an important concern for the first perception of the beer by the consumer and is therefore the object of the greatest care by the brewers and the barmen in charge to serve draft beer, or to properly pour beer into a glass from the bottle (with a good head retention and without overfoaming, or gushing when opening the bottle).
Many patents for various types of beer foam stabilizers have been filed by breweries and the agro-chemical industry in the last decades. Cobalt salts added at low concentration (1 – 2 ppm) were popular in the sixties, but raised the question of cobalt toxicity in case of undetected accidental overdosage during beer production. As an alternative, organic foam stabilizers are produced by hydrolysis of recovered by-products of beer manufacture, such as spent grains or hops residues.
Amongst the large spectrum of purified, or modified, natural food additives available on the market, soluble carboxymethyl hydroxyethyl cellulose, propylene glycol alginate (PGA, food additive with E number E405), pectins and gellan gum have also been investigated as foam stabilizer.
Cobalt salts
In 1957, two brewing chemists, Thorne and Helm, discovered that the cation was able to stabilize beer foam and to avoid beer overfoaming and gushing. The addition of a tiny amount of cobalt ions in the range 1 – 2 mg/L (ppm) was effective. Higher concentrations would be toxic and lower ones ineffective.
Cobalt is a transition metal whose atomic orbitals are able to interact with ligands, or functional groups (–OH, –COOH, –NH2), attached to organic molecules naturally present in the beer, making macromolecular coordination complexes stabilizing the beer foam. Cobalt could behave as an inter- or intra-molecular bridge between different polysaccharide molecules (changing their shape or size), or cause some conformational changes of different types of molecules present in solution, affecting their absolute configuration and thus the foam molecular structure and its behavior.
Thorne and Helm (1957) also formulated the hypothesis that cobalt, by being complexed with certain nitrogenous constituents of the beer (e.g., amino acids from malt proteins), might produce surface-active substances inactivating the gaseous nuclei responsible for overfoaming and gushing.
Gushing is a specific problem also studied into more details by Rudin and Hudson (1958). These authors discovered that gushing is also promoted by other transition metal ions such as these of nickel and iron, but not by cobalt ions. Isohumulone (an iso-alpha acid responsible for the bitter taste of hops) and its combinations with Ni, or Fe, also favor gushing, while pure Co ions or their combination with isohumulone do not exhibit gushing and overfoaming. This explains why cobalt salts were specifically selected at a concentration of 1 – 2 mg/L as anti-gushing agent for beer. Rudin and Hudson (1958) and other authors also found that Co, Ni and Fe ions preferentially concentrate in the foam itself.
In the sixties, after approval by the US FDA, cobalt sulfate was commonly used at low concentration in the USA as an additive to stabilize beer foam and to prevent gushing after beer is exposed to vibrations during its transport or handling.
Although cobalt is an essential micronutrient needed for vitamin B12 synthesis, excess levels of cobalt in the body can lead to cobalt poisoning and must be avoided. It triggered the development of qualitative and quantitative analysis methods to accurately assay cobalt in beer in order to prevent accidental overdosage and cobalt poisoning.
Too high levels of cobalt are known to be responsible for the beer drinker's cardiomyopathy. The first issues mentioned in the literature were reported in Canada in the middle of the sixties after an accidental overdosage in the Dow Breweries in Quebec City.
In August 1965, a person presented to a hospital in Quebec City with symptoms suggestive of alcoholic cardiomyopathy. Over the next eight months, fifty more cases with similar findings appeared in the same area with twenty of these being fatal. It was noted that all were heavy drinkers who mostly drank beer and preferred the Dow brand; thirty out of those drank more than 6 litres (12 pints) of beer per day. Epidemiological studies found that the Dow Breweries had been adding cobalt sulfate to the beer for foam stability since July 1965 and that the concentration added in the Quebec city brewery was ten times that of the same beer brewed in Montreal where there were no reported cases.
Storage and degradation
A particular problem with beer is that, unlike wine, its quality tends to deteriorate as it ages. A cat urine smell and flavor called ribes, named for the genus of the black currant, tends to develop and peak. A cardboard smell then dominates which is due to the release of 2-nonenal. In general, chemists believe that the "off-flavors" that come from old beers are due to reactive oxygen species. These may come in the form of oxygen free radicals, for example, which can change the chemical structures of compounds in beer that give them their taste. Oxygen radicals can cause increased concentrations of aldehydes from the Strecker degradation reactions of amino acids in beer.
Beer is unique when compared to other alcoholic beveragess because it is unstable in the final package. There are many variables and chemical compounds that affect the flavor of beer during the production steps, but also during the storage of beer. Beer will develop an off-flavor during storage because of many factors, including sunlight and the amount of oxygen in the headspace of the bottle. Other than changes in taste, beer can also develop visual changes. Beer can become hazy during storage. This is called colloidal stability (haze formation) and is typically caused by the raw materials used during the brewing process. The primary reaction that causes beer haze is the polymerization of polyphenols and binding with specific proteins. This type of haze can be seen when beer is cooled below 0 degrees Celsius. When the beer is raised to room temperature, the haze dissolves. But if a beer is stored at room temperature for too long (about 6 months) a permanent haze will form. A study done by Heuberger et al. (2012) concludes that storage temperature of beers affects the flavor stability. They found that the metabolite profile of room temperature and cold temperature stored beer differed significantly from fresh beer. They also have evidence to support significant beer oxidation after weeks of storage, which also has an effect on the flavor of beer.
The off-flavour in beer, such as a cardboard or green apple taste, is often associated with the appearance of staling aldehydes. The Strecker aldehydes responsible for the flavor change are formed during storage of the beers. Philip Wietstock et al. performed experiments to test what causes the formation of Strecker aldehydes during storage. They found that only amino acid concentration (leucine (Leu), isoleucine (Ile), and phenylalanine (Phe), specifically) and dissolved oxygen concentration caused Strecker aldehyde formation. They also tested carbohydrate and Fe2+ additions. A linear relationship was found between Strecker aldehydes formed and total packaged oxygen. This is important for brewers to know so that they can control the taste of their beer. Wietstock concludes that capping beers with oxygen barrier crown corks will diminish Strecker aldehyde formation.
In another study done by Vanderhaegen et al. (2003), different aging conditions were tested on a bottled beer after 6 months. They found a decrease in volatile esters was responsible for a reduced fruity flavor. They also found an increase in many other compounds including carbonyl compounds, ethyl esters, Maillard compounds, dioxolanes, and furanic ethers. The carbonyl compounds, as stated previously in the Wietstock experiments, will create Strecker aldehydes, which tend to cause a green apple flavor. Esters are known to cause fruity flavors such as pears, roses, and bananas. Maillard compounds will cause a toasty, malty flavor.
A study done by Charles Bamforth and Roy Parsons (1985) also confirms that beer staling flavors are caused by various carbonyl compounds. They used thiobarbituric acid (TBA) to estimate the staling substances after using an accelerated aging technique. They found that beer staling is reduced by scavengers of the hydroxyl radical (•OH), such as mannitol and ascorbic acid. They also tested the hypothesis that soybean extracts included in the fermenting wort enhance the shelf life of beer flavor.
See also
Barm
Beer head
Beer faults
Brewing
Bright beer
Carbonation
Lightstruck beer
Mouthfeel
Isohumulone
Yeast in Beermaking
References
Citations
Sources
External links
Tapping into the Chemistry of Beer and Brewing—an online lecture by Charles Bamforth, Professor of Malting & Brewing Sciences at the University of California
Chemistry of Beer—an online course in the subject at the University of Oklahoma
Beer
Alcohol chemistry | Beer chemistry | [
"Chemistry"
] | 3,094 | [
"Food chemistry",
"Alcohol chemistry"
] |
47,329,392 | https://en.wikipedia.org/wiki/Suillus%20flavidus | Suillus flavidus is a bolete mushroom in the genus Suillus native to Europe. It is considered endangered in the Czech Republic and Switzerland. It is also considered edible, but with an unpleasant taste.
References
External links
flavidus
Fungi described in 1815
Fungi of Europe
Fungus species | Suillus flavidus | [
"Biology"
] | 59 | [
"Fungi",
"Fungus species"
] |
47,329,410 | https://en.wikipedia.org/wiki/Suillus%20flavogranulatus | Suillus flavogranulatus is a bolete mushroom in the genus Suillus native to North America. It was described as new to science in 1965 by mycologists Alexander H. Smith, Harry Delbert Thiers, and Orson K. Miller.
See also
List of North American boletes
References
External links
flavogranulatus
Fungi described in 1965
Fungi of North America
Fungus species | Suillus flavogranulatus | [
"Biology"
] | 84 | [
"Fungi",
"Fungus species"
] |
47,329,429 | https://en.wikipedia.org/wiki/Suillus%20cavipoides | Suillus cavipoides is a bolete mushroom in the genus Suillus native to China. Initially described as Boletinus cavipoides, it was transferred by Wang and Yao to the genus Suillus in a 2004 nomenclatural revision of several Chinese boletes.
References
External links
cavipoides
Fungi described in 1892
Fungi of China
Fungus species | Suillus cavipoides | [
"Biology"
] | 77 | [
"Fungi",
"Fungus species"
] |
47,329,480 | https://en.wikipedia.org/wiki/One-to-one%20%28data%20model%29 | In systems analysis, a one-to-one relationship is a type of cardinality that refers to the relationship between two entities (see also entity–relationship model) A and B in which one element of A may only be linked to one element of B, and vice versa. In mathematical terms, there exists a bijective function from A to B.
For instance, think of A as the set of all human beings, and B as the set of all their brains. Any person from A can and must have only one brain from B, and any human brain in B can and must belong to only one person that is contained in A.
In a relational database, a one-to-one relationship exists when one row in a table may be linked with only one row in another table and vice versa. It is important to note that a one-to-one relationship is not a property of the data, but rather of the relationship itself. A list of mothers and their children may happen to describe mothers with only one child, in which case one row of the mothers table will refer to only one row of the children table and vice versa. The real-world relationship that the data models is not one-to-one, because mothers may have more than one child, thus forming a one-to-many relationship.
See also
One-to-many (data model)
Many-to-many (data model)
External links
Design pattern: many-to-many (order entry), Tomjewett.com
Data modeling | One-to-one (data model) | [
"Engineering"
] | 312 | [
"Data modeling",
"Data engineering"
] |
47,329,491 | https://en.wikipedia.org/wiki/One-to-many%20%28data%20model%29 | In systems analysis, a one-to-many relationship is a type of cardinality that refers to the relationship between two entities (see also entity–relationship model). For example, take a car and an owner of the car. The car can only be owned by one owner at a time or not owned at all, and an owner could own zero, one, or multiple cars. One owner could have many cars, one-to-many.
In a relational database, a one-to-many relationship exists when one record is related to many records of another table. A one-to-many relationship is not a property of the data, but rather of the relationship itself. One-to-many often refer to a primary key to foreign key relationship between two tables, where the record in the first table can relate to multiple records in the second table. A foreign key is one side of the relationship that shows a row or multiple rows, with one of those rows being the primary key already listed on the first table. This is also called a foreign key constraint, which is important to keep data from being duplicated and have relationships within the database stay reliable as more information is added.
Many-to-many relationships are not able to be used in relational databases and must be converted to one-to-many relationships. Both one-to-many and one-to-one relationships are common in relational databases but are normally created majorly with one-to-many relationships.
The opposite of one-to-many is many-to-one. The transpose of a one-to-many relationship is a many-to-one relationship.
Entity relationship diagram (ERD) notations
One notation as described in Entity Relationship modeling is Chen notation or formally Chen ERD notation created originally by Peter Chen in 1976 where a one-to-many relationship is notated as 1:N where N represents the cardinality and can be 0 or higher.
A many-to-one relationship is sometimes notated as N:1.
See also
One-to-one (data model)
Many-to-many (data model)
References
Data modeling | One-to-many (data model) | [
"Engineering"
] | 436 | [
"Data modeling",
"Data engineering"
] |
47,330,034 | https://en.wikipedia.org/wiki/Suillus%20triacicularis | Suillus triacicularis is a species of bolete fungus in the family Suillaceae. Described as new to science in 2014, it is found in the northwestern Himalayas, India, where it grows in association with Pinus roxburghii.
References
External links
tridentinus
Fungi described in 2014
Fungi of India
Fungus species | Suillus triacicularis | [
"Biology"
] | 67 | [
"Fungi",
"Fungus species"
] |
47,330,055 | https://en.wikipedia.org/wiki/NGC%207759 | NGC 7759 is a lenticular galaxy in the constellation Aquarius. It is located about 340 million light-years (100 megaparsecs) away from the Sun. It was discovered independently by American astronomers Lewis A. Swift and Francis Preserved Leavenworth.
References
Aquarius (constellation)
7759
Lenticular galaxies | NGC 7759 | [
"Astronomy"
] | 66 | [
"Constellations",
"Aquarius (constellation)"
] |
47,330,613 | https://en.wikipedia.org/wiki/Suillus%20pallidiceps | Suillus pallidiceps is a species of bolete fungus in the family Suillaceae. It was first described scientifically by American mycologists Alexander H. Smith and Harry D. Thiers in 1964.
See also
List of North American boletes
References
External links
pallidiceps
Fungi described in 1964
Fungi of North America
Fungus species | Suillus pallidiceps | [
"Biology"
] | 73 | [
"Fungi",
"Fungus species"
] |
47,330,614 | https://en.wikipedia.org/wiki/Suillus%20ponderosus | Suillus ponderosus is a species of bolete fungus in the family Suillaceae. It was first described scientifically by American mycologists Alexander H. Smith and Harry D. Thiers in 1964.
See also
List of North American boletes
References
External links
ponderosus
Fungi described in 1964
Fungi of North America
Taxa named by Harry Delbert Thiers
Taxa named by Alexander H. Smith
Fungus species | Suillus ponderosus | [
"Biology"
] | 86 | [
"Fungi",
"Fungus species"
] |
47,330,617 | https://en.wikipedia.org/wiki/Suillus%20acerbus | Suillus acerbus is a species of bolete fungus in the family Suillaceae. It was first described scientifically by American mycologists Alexander H. Smith and Harry D. Thiers in 1964.
See also
List of North American boletes
References
External links
acerbus
Fungi described in 1964
Fungi of North America
Fungus species | Suillus acerbus | [
"Biology"
] | 71 | [
"Fungi",
"Fungus species"
] |
47,330,619 | https://en.wikipedia.org/wiki/Suillus%20appendiculatus | Suillus appendiculatus is a species of bolete fungus in the family Suillaceae. It was first described scientifically in 1896 as a species of Boletinus by American mycologist Charles Horton Peck. Harry D. Thiers and Alexander H. Smith transferred it to the genus Suillus in 1964.
See also
List of North American boletes
References
External links
appendiculatus
Fungi described in 1896
Fungi of North America
Taxa named by Charles Horton Peck
Fungus species | Suillus appendiculatus | [
"Biology"
] | 100 | [
"Fungi",
"Fungus species"
] |
47,330,620 | https://en.wikipedia.org/wiki/Suillus%20brunnescens | Suillus brunnescens is a species of bolete fungus in the family Suillaceae. It was first described scientifically by American mycologists Alexander H. Smith and Harry D. Thiers in 1964.
See also
List of North American boletes
References
External links
brunnescens
Fungi described in 1964
Fungi of North America
Fungus species | Suillus brunnescens | [
"Biology"
] | 73 | [
"Fungi",
"Fungus species"
] |
47,330,621 | https://en.wikipedia.org/wiki/Suillus%20caerulescens | Suillus caerulescens, commonly known as the douglas-fir suillus is an edible species of bolete fungus in the family Suillaceae. It was first described scientifically by American mycologists Alexander H. Smith and Harry D. Thiers in 1964. It can be found growing with Douglas fir trees. Its stem bruises blue, which sometimes takes a few minutes.
The cap is yellowish to reddish brown, sometimes with streaks from its darker center. It ranges from in diameter, shaped convex to flat, and viscid when wet, sometimes with veil remnants on the edge. The flesh is yellowish, as are the pores. The stalk is yellowish to brown, darkening with age, 2–8 cm tall and 1–3 cm wide, and bruises bluish at the base; it sometimes has a faint ring.
While edible, it is considered of poor quality.
Suillus lakei is fairly similar.
See also
List of North American boletes
References
External links
caerulescens
Edible fungi
Fungi described in 1964
Fungi of North America
Taxa named by Harry Delbert Thiers
Taxa named by Alexander H. Smith
Fungus species | Suillus caerulescens | [
"Biology"
] | 232 | [
"Fungi",
"Fungus species"
] |
47,331,006 | https://en.wikipedia.org/wiki/Translation%20of%20axes | In mathematics, a translation of axes in two dimensions is a mapping from an xy-Cartesian coordinate system to an x'y'-Cartesian coordinate system in which the x' axis is parallel to the x axis and k units away, and the y' axis is parallel to the y axis and h units away. This means that the origin O' of the new coordinate system has coordinates (h, k) in the original system. The positive x' and y' directions are taken to be the same as the positive x and y directions. A point P has coordinates (x, y) with respect to the original system and coordinates (x', y') with respect to the new system, where
or equivalently
In the new coordinate system, the point P will appear to have been translated in the opposite direction. For example, if the xy-system is translated a distance h to the right and a distance k upward, then P will appear to have been translated a distance h to the left and a distance k downward in the x'y'-system . A translation of axes in more than two dimensions is defined similarly. A translation of axes is a rigid transformation, but not a linear map. (See Affine transformation.)
Motivation
Coordinate systems are essential for studying the equations of curves using the methods of analytic geometry. To use the method of coordinate geometry, the axes are placed at a convenient position with respect to the curve under consideration. For example, to study the equations of ellipses and hyperbolas, the foci are usually located on one of the axes and are situated symmetrically with respect to the origin. If the curve (hyperbola, parabola, ellipse, etc.) is not situated conveniently with respect to the axes, the coordinate system should be changed to place the curve at a convenient and familiar location and orientation. The process of making this change is called a transformation of coordinates.
The solutions to many problems can be simplified by translating the coordinate axes to obtain new axes parallel to the original ones.
Translation of conic sections
Through a change of coordinates, the equation of a conic section can be put into a standard form, which is usually easier to work with. For the most general equation of the second degree, which takes the form
it is always possible to perform a rotation of axes in such a way that in the new system the equation takes the form
that is, eliminating the xy term. Next, a translation of axes can reduce an equation of the form () to an equation of the same form but with new variables (x', y') as coordinates, and with D and E both equal to zero (with certain exceptions—for example, parabolas). The principal tool in this process is "completing the square." In the examples that follow, it is assumed that a rotation of axes has already been performed.
Example 1
Given the equation
by using a translation of axes, determine whether the locus of the equation is a parabola, ellipse, or hyperbola. Determine foci (or focus), vertices (or vertex), and eccentricity.
Solution: To complete the square in x and y, write the equation in the form
Complete the squares and obtain
Define
and
That is, the translation in equations () is made with The equation in the new coordinate system is
Divide equation () by 225 to obtain
which is recognizable as an ellipse with In the x'y'-system, we have: center ; vertices ; foci
In the xy-system, use the relations to obtain: center ; vertices ; foci ; eccentricity
Generalization to several dimensions
For an xyz-Cartesian coordinate system in three dimensions, suppose that a second Cartesian coordinate system is introduced, with axes x', y' and z' so located that the x' axis is parallel to the x axis and h units from it, the y' axis is parallel to the y axis and k units from it, and the z' axis is parallel to the z axis and l units from it. A point P in space will have coordinates in both systems. If its coordinates are (x, y, z) in the original system and (x', y', z') in the second system, the equations
hold. Equations () define a translation of axes in three dimensions where (h, k, l) are the xyz-coordinates of the new origin. A translation of axes in any finite number of dimensions is defined similarly.
Translation of quadric surfaces
In three-space, the most general equation of the second degree in x, y and z has the form
where the quantities are positive or negative numbers or zero. The points in space satisfying such an equation all lie on a surface. Any second-degree equation which does not reduce to a cylinder, plane, line, or point corresponds to a surface which is called quadric.
As in the case of plane analytic geometry, the method of translation of axes may be used to simplify second-degree equations, thereby making evident the nature of certain quadric surfaces. The principal tool in this process is "completing the square."
Example 2
Use a translation of coordinates to identify the quadric surface
Solution: Write the equation in the form
Complete the square to obtain
Introduce the translation of coordinates
The equation of the surface takes the form
which is recognizable as the equation of an ellipsoid.
See also
Change of basis
Translation (geometry)
Notes
References
Functions and mappings
Euclidean geometry
Linear algebra
Transformation (function) | Translation of axes | [
"Mathematics"
] | 1,135 | [
"Functions and mappings",
"Mathematical analysis",
"Transformation (function)",
"Mathematical objects",
"Mathematical relations",
"Geometry",
"Linear algebra",
"Algebra"
] |
47,331,609 | https://en.wikipedia.org/wiki/Pelargonium%20inquinans | Pelargonium inquinans, the scarlet geranium, is a species of plant in the genus Pelargonium (family Geraniaceae). It is a shrub endemic to South Africa, ranging from Mpumalanga to KwaZulu-Natal and Eastern Cape provinces. It is one of the ancestors of the hybrid line of horticultural pelargoniums, referred to as the zonal group. They can easily be propagated by seeds and cuttings.
Etymology and history
The generic name Pelargonium in scientific Latin derives from the Greek pelargós (πελαργός), which means the stork and the shape of their fruit evoking the beak of the wader. The specific epithet "messy" derives from the Latin verb inquino "dirty, soil" because the leaves leave a brown trace on the fingers when touched.
The Pelargonium inquinans was grown in the garden of the Bishop of London, Henry Compton, an admirer of exotic plants. In 1713, when he died, Pelargonium inquinans was found in his collection. The first illustration from 1732 was made from a plant growing in the garden of British botanist James Sherard. Many hybrids have been derived from this species, but the true wild species can be recognized by its red glandular hairs.
Description
In the wild, Pelargonium inquinans is a small shrub, about 2 m tall, branched, with young succulent twigs becoming woody with age, bearing red glandular hairs.
The evergreen leaves, borne by long petioles, are orbicular (like Pelargonium × hortorum but without dark markings), incised in 5 to 7 crenate lobes, with a viscous pubescence, giving a cottony appearance to both sides. To the touch, the leaves stain the fingers brown rust.
The scarlet red flowers, sometimes pink or white, are grouped by 10 to 20 in pseudo-umbels. They are bilateral symmetry (zygomorph) with the 2 upper petals may be a little smaller than the 3 lower petals. Stamens and style are exerted. The filaments of the seven fertile stamens join over most of their length.
In South Africa flowering is spread throughout the year.
The pericardial fruit is composed of 5 capsules terminated by a long, hairy, twisted curl at maturity.
Distribution
The pelargonium with scarlet flowers grows in the Eastern Cape, Uitenhage, Albany and Caffirland, south of Kwazulu-Natal, South Africa.
It grows on clay soils, like Pelargonium × hortorum.
Hybrid
Pelargonium inquinans and Pelargonium zonale are generally considered the two main wild ancestors of the zonal group of horticultural pelargoniums, commonly referred to as "florist geraniums" or "zoned leaf hybrid pelargoniums". In botany, the name Pelargonium × hortorum L.H. Bailey is accepted.
These two species were introduced in the great gardens of Europe at the beginning of the eighteenth century.
Uses
Indigenous people use crushed leaves for headache and influenza. They are also used as a body deodorant.
Gallery
References
Endemic flora of South Africa
Flora of the Cape Provinces
Flora of KwaZulu-Natal
Flora of the Northern Provinces
inquinans
Plants described in 1753
Plants used in traditional African medicine
Plants that can bloom all year round | Pelargonium inquinans | [
"Biology"
] | 720 | [
"Plants that can bloom all year round",
"Plants"
] |
47,331,849 | https://en.wikipedia.org/wiki/Pill%20reminder | A pill reminder is any device that reminds users to take medications. Traditional pill reminders are pill containers with electric timers attached, which can be preset for certain times of the day to set off an alarm. More sophisticated pill reminders can also detect when they have been opened, and therefore when the user is away during the time they were supposed to take their medication, they will be reminded of it when they return. This reminder can be in the form of a light, which also helps for deaf or hearing-impaired users.
A newer type of pill reminder is a mobile app that reminds the owner to take the medication. Some of these applications might effectively support adherence to taking medications.
See also
Pill dispenser
Pill organizer
References
Clinical pharmacology
Pharmacy
Application software
Consumer goods | Pill reminder | [
"Chemistry"
] | 160 | [
"Pharmacology",
"Clinical pharmacology",
"Pharmacy"
] |
47,332,276 | https://en.wikipedia.org/wiki/Amanita%20austroviridis | Amanita austroviridis, commonly known as the Australian verdigris lepidella, is a species of agaric fungus in the family Amanitaceae native to Australia.
External links
austroviridis
Fungi native to Australia
Fungi described in 1992
Fungus species | Amanita austroviridis | [
"Biology"
] | 57 | [
"Fungi",
"Fungus species"
] |
47,332,291 | https://en.wikipedia.org/wiki/Amanita%20chepangiana | Amanita chepangiana, commonly known as the Chepang slender Caesar, is a species of agaric fungus in the family Amanitaceae native to China and southern Asia.
In parts of Yunnan, China, the species is traditionally consumed. However, toxicity analysis found out at least one type of amatoxin and phallotoxin each within the species. Since it is difficult to distinguish from other lethal species, human consumption is generally not recommended.
References
External links
chepangiana
Fungi of China
Fungi described in 1992
Fungus species | Amanita chepangiana | [
"Biology"
] | 113 | [
"Fungi",
"Fungus species"
] |
47,332,303 | https://en.wikipedia.org/wiki/Amanita%20armeniaca | Amanita armeniaca is a species of agaric fungus in the family Amanitaceae native to Australia.
References
External links
armeniaca
Fungi native to Australia
Fungi described in 1997
Fungus species | Amanita armeniaca | [
"Biology"
] | 41 | [
"Fungi",
"Fungus species"
] |
47,333,617 | https://en.wikipedia.org/wiki/Penicillium%20rasile | Penicillium rasile is a species of fungus in the genus Penicillium.
References
rasile
Fungi described in 1979
Fungus species | Penicillium rasile | [
"Biology"
] | 30 | [
"Fungi",
"Fungus species"
] |
47,333,806 | https://en.wikipedia.org/wiki/C21H21NO6 | {{DISPLAYTITLE:C21H21NO6}}
The molecular formula C21H21NO6 (molar mass: 383.39 g/mol, exact mass: 383.1369 u) may refer to:
Hydrastine
Rhoeadine (rheadine)
Molecular formulas | C21H21NO6 | [
"Physics",
"Chemistry"
] | 66 | [
"Molecules",
"Set index articles on molecular formulas",
"Isomerism",
"Molecular formulas",
"Matter"
] |
47,333,891 | https://en.wikipedia.org/wiki/All%20American%20Turbo-Cat | The Turbo-Cat was a jet aircraft catapult launch system powered by six jet engines. It was invented by Don Doolittle, and manufactured by the All American Engineering Co. of Wilmington, Delaware, USA. Six Allison J33 engines in a capstan arrangement provided 50,000 horsepower directly to the launching cable.
It was developed for expeditionary use by the U.S. Marine Corps, and it was to be air transportable.
Six jet engines are arranged in a circle with the exhausts facing the center where the mass flow of hot exhaust gasses is vented through a diffuser into a pair of large launching turbines. The launching turbines are connected by a main shaft to a cable drum from which an endless tensioned cable shunts underground to a launching track. Then down the track and around a pulley and back to the cable drum. By diverting the flow of exhaust gasses into a launching turbine, the cable drum and its tensioned cable can be accelerated to very high speeds.
See also
Index of aviation articles
Rocket sled
Rocket sled launch
Dolly (trailer)
Ground carriage
References
Jet Cat, FLYING magazine, May 1958, pp. 32–33, 74
External links
Turbo-Cat Circa 1958
Aviation in the United States
Jet engine technology
Jet engines
Types of take-off and landing | All American Turbo-Cat | [
"Technology"
] | 264 | [
"Jet engines",
"Engines"
] |
57,583,484 | https://en.wikipedia.org/wiki/Basel%20Convention%20Coordinating%20Centre%20for%20the%20African%20Region%20in%20Nigeria | The Basel Convention Coordinating Centre for the African Region in Nigeria (BCCC Africa) is a regional centre of the Basel Convention. It is located along Ijoma Road, inside the main campus of the University of Ibadan, Ibadan, Nigeria.
It has the vision of strengthening countries in the African Region in the Environmentally Sound Management (ESM) of Hazardous Waste. The Coordinating Centre in Nigeria aims at fast-tracking the implementation of the Basel Convention in the African Region.
The Centre is aimed at African countries implement the Basel Convention and its amendments. It also tries to inform these countries about chemicals and hazardous wastes issues. The Centre receives financial assistance from the Nigerian Government, and the Basel Convention Trust fund.
References
Organizations based in Nigeria
Hazardous waste | Basel Convention Coordinating Centre for the African Region in Nigeria | [
"Technology"
] | 151 | [
"Hazardous waste"
] |
57,583,649 | https://en.wikipedia.org/wiki/Rochester%20Ramjet | The Rochester Ramjet is an automotive fuel injection system developed by the Rochester Products Division of General Motors and first offered as a high-performance option on the Corvette and GM passenger cars in 1957. It was discontinued partway through 1965 in favor of the Chevrolet Big Block as a performance option. Unlike electronic fuel injection systems that would become common decades later, the Ramjet is purely mechanical and relies on vacuum and pressure signals to measure airflow and meter fuel.
History
In the early 1950s, fuel injection was the topic of a significant amount of research by the auto industry in the US and internationally. Ed Cole, who had become the chief engineer of Chevrolet in 1952, pushed for Chevrolet to be the first GM brand to offer a fuel injection option on a production car. Much of the development of the Ramjet was done by engineer John Dolza, with supporting effort from Zora Arkus-Duntov. Dynamometer-based tests of a 265 cid small block engine running with fuel injection were performed as early as 1955.
Between 1957 and 1965, the Ramjet was offered as the top performance option on the Corvette. When it was first introduced, it passed the threshold of one horsepower per cubic inch of engine displacement, and this fact was used in marketing material. In addition to the Corvette, the system was offered on Chevrolet and Pontiac passenger cars; each of which used a slightly different configuration with respect to the air cleaner and other components. Within the Chevrolet brand in 1957, Ramjet was covered by RPO-578 on the passenger cars and RPO-579 on the Corvette.
Although offered on Corvettes until 1965, the Ramjet system was available on Chevy passenger cars from 1957 through 1959.
Technical details
The Ramjet is a continuous-flow port-injection system. Unlike later fuel injection systems that used electronics, this one is based on purely mechanical principles. The two main sub-assemblies of the system are the air meter and the fuel meter. The air meter measures airflow into the engine and manages thermostatic warmup enrichment, fuel shutoff on overrun, and idle settings. These measurements are sent via pressure and vacuum signals to the fuel meter, which contains the high-pressure fuel pump and controls delivery of fuel to the injector nozzles.
There are three easily identifiable revisions of the Ramjet, each spanning three years of the system's nine year production life. The first of these (1957-1959) are identified by the "finned top" sand cast plenum, followed by the "flat top" plenum (1960-1962), and finally the die-cast plenum (1963-1965). All versions of the system required a special distributor that provided a cable drive to the high-pressure fuel pump.
Limitations and issues
Although the Ramjet system allowed a significant increase in engine performance, its innovative design had several inherent issues that affected reliability and tuning. On the early systems (1957 through 1961), one of these problems affected the cold-start fuel enrichment: it would tend to discretely toggle on or off, rather than using a gradual series of steps. Later systems used electric chokes or exhaust-heat chokes for cold enrichment, although these had issues of their own. Additionally, the vapor pressure of modern pump gasoline may not be appropriate to prevent percolation in the Ramjet fuel distribution spider.
The cranking signal valve was also the source of reliability issues. It was intended to pass a certain maximum vacuum signal that would be used for fueling during cranking (i.e. before engine start). The valve would sometimes fail in a partially open position and cause the fueling to run much richer than stoichiometric.
References
Engine components
Fuel injection systems | Rochester Ramjet | [
"Technology"
] | 756 | [
"Engine components",
"Engines"
] |
57,585,009 | https://en.wikipedia.org/wiki/Lunar%20Ultraviolet%20Cosmic%20Imager | Lunar Ultraviolet Cosmic Imager (LUCI) is a small planned telescope that will be landed on the Moon to scan the sky in near UV wavelengths. It is a technology demonstrator developed by the Indian Institute of Astrophysics, and it was planned to be one of several small payloads to be deployed by the commercial Z-01 lander developed by TeamIndus in partnership with OrbitBeyond. The mission was planned to be launched in 2020 as part of NASA's Commercial Lunar Payload Services (CLPS). On 29 July 2019 OrbitBeyond announced that it would drop out of the CLPS contract with NASA, meaning that the 2020 launch was canceled and it is unknown whether the mission will ever take place.
Science objectives
The science objectives of LUCI telescope are primarily to search for transient astronomic events such as supernovae, novae, tidal disruption events by massive black holes, and more exotic energetic sources such as superluminous supernovae and flashes from cosmic collisions which can be very energetic on all scales.
LUCI will also look for faint asteroids and comets in the Solar System, especially for near-Earth objects (NEO) and potentially hazardous objects. The aims are focused on UV sources not accessible by the more sensitive large space missions.
Overview
The Earth's atmosphere absorbs and scatters UV photons, preventing observations of the active Universe. Placing a telescope on the surface of the Moon is advantageous because of its absence of atmosphere and ionosphere offers an unobstructed view of the space in all wavelengths. The Moon surface provides not just a stable platform, but an inexpensive and long-term access to observations in wavelengths not normally used by large orbital telescope missions. The only UV astronomical observations from the Moon to date were made by Apollo 16 team in 1972 and theLunar-based Ultraviolet telescope aboard the Chang'e 3 lunar lander in 2018. LUCI project started in 2013 and is funded by India's Department of Science and Technology. The telescope team is headed by Jayant Murthy.
The telescope has been completed as of March 2019, and was awaiting integration to the Z-01 lander. It was planned be launched in Q3 2020 on a Falcon 9 rocket and land at Mare Imbrium (29.52º N 25.68º W). On 29 July 2019 OrbitBeyond, the builder of the lander, announced that it will withdraw from the launch and the mission. Thus the mission is effectively dead. OrbitBeyond and NASA agreed that OrbitBeyond will be released from the NASA CLPS contract in general. However, OrbitBeyond remains eligible to bid for future NASA CLPS contracts.
Design
LUCI is a small technology demonstrator without 3-axis pointing freedom, so it will rely on the motion of the lunar sky. The optical system is a two-spherical mirror configuration and a double-pass corrector lens. Its primary lens is all-spherical measuring 80 mm transmitting light through the system to a photon-counting charge-coupled device (CCD) detector which is sensitive to ultraviolet wavelengths. The detector is an 8 mm UV-sensitive CCD with the response between 200 - 900 nm, so the engineers placed a solar blind filter before the CCD to restrict the bandpass to 200 - 320 nm.
LUCI is planned to be mostly contained within the lander, and it will be lowered back into its storage bay during the cold lunar nights. The baseline for LUCI's operation is "a few months".
See also
International Lunar Observatory, a planned lunar-based telescope
Ultraviolet astronomy
References
Ultraviolet astronomy
Ultraviolet telescopes
Space telescopes
Missions to the Moon | Lunar Ultraviolet Cosmic Imager | [
"Chemistry",
"Astronomy"
] | 739 | [
"Space telescopes",
"Astronomical sub-disciplines",
"Ultraviolet astronomy",
"Ultraviolet radiation"
] |
57,585,124 | https://en.wikipedia.org/wiki/Pyoluteorin | Pyoluteorin is a natural antibiotic that is biosynthesized from a hybrid nonribosomal peptide synthetase (NRPS) and polyketide synthase (PKS) pathway. Pyoluteorin was first isolated in the 1950s from Pseudomonas aeruginosa strains T359 and IFO 3455 and was found to be toxic against oomycetes, bacteria, fungi, and against certain plants. Pyoluteorin is most notable for its toxicity against the oomycete Pythium ultimum, which is a plant pathogen that causes a global loss in agriculture. Currently, pyoluteorin derivatives are being studied as an Mcl-1 antagonist in order to target cancers that have elevated Mcl-1 levels.
Biosynthesis
Pyoluteorin is synthesized from an NRPS/PKS hybrid pathway. The resorcinol ring is derived from a type I PKS while the dichloropyrrole moiety is derived from a type II NRPS. Pyoluteorin biosynthesis begins with the activation of L-proline to prolyl-AMP by the adenylation domain PltF. With prolyl-AMP still in the active site, the active form of the peptidyl carrier protein PltL binds to PltF. Then PltF catalyzes the aminoacylation of PltL by attaching L-proline to the thiol of the 4’phosphopantetheine arm of PltL. Next, the dehydrogenase PltE desaturates the prolyl moiety on PltL to create pyrrolyl-PltL. The halogenation domain PltA then dichlorinates the pyrrole moiety first at position 5 and then at position 4 in a FADH2 dependent manner. The dichloropyrroyl residue is then transferred to the type I PKS PltB and PltC, however, the mechanism of transfer is unknown. The addition of 3 malonyl-CoA monomers, cyclization, and release by the thioesterase PltG gives pyoluteorin.
References
Antibiotics
Biosynthesis
Halogen-containing natural products
Pyrroles | Pyoluteorin | [
"Chemistry",
"Biology"
] | 477 | [
"Biotechnology products",
"Antibiotics",
"Biosynthesis",
"Chemical synthesis",
"Biocides",
"Metabolism"
] |
57,585,472 | https://en.wikipedia.org/wiki/Maklamicin | Maklamicin is a spirotetronate-class polyketide natural product. Isolated from Micromonospora sp. GMKU326 found in the root of Maklam phueak, it displays antibiotic activity against Gram-positive bacterial strains Micrococcus luteus, Bacillus subtilis, Bacillius cereus, Staphylococcus aureus, and Enterococcus faecalis.
Biosynthesis
Maklamicin arises from a type I modular polyketide synthase (PKS) system. The structure of its polyketide chain extension has been shown to contain thirteen discrete PKS modules (a loading module and twelve extension modules).
After passing through the extension domains the linear polyketide precursor to Maklamicin undergoes an intramolecular Diels-Alder reaction (IMDA) to form its trans-decalin motif. 1,3-bisphosphoglycerate is converted to glyceryl S-ACP and condenses onto the decalin-bearing polyketide intermediate to furnish an intermediate containing the premature tetronate moiety. This intermediate then acetylated and subsequently dehydrated to form the exocyclic olefin of the mature tetronate moiety. The intermediate, now containing both the tetronate and trans-decalin motifs of Maklamicin is then systematically reduced to afford a diene. A second IMDA then occurs between the tetronate and newly formed dienophile to yield the ultimate intermediate which is oxidized to afford Maklamicin.
References
Polyketides
Heterocyclic compounds with 5 rings
Oxygen heterocycles
Ketones
Eleven-membered rings
Lactones
Triols | Maklamicin | [
"Chemistry"
] | 375 | [
"Biomolecules by chemical classification",
"Natural products",
"Ketones",
"Functional groups",
"Polyketides"
] |
57,585,688 | https://en.wikipedia.org/wiki/Paul%20Chirik | Paul James Chirik (born June 13, 1973) is an American chemist known for his work in sustainable chemistry using Earth-abundant metals like iron, cobalt, and nickel to surpass the performance of more exotic elements traditionally used in catalysis. He is the Edwards S. Sanford Professor of Chemistry and chair of the chemistry department at Princeton University.
Academic career
Chirik received his B.S. in chemistry from Virginia Tech, studying organometallic chemistry under the advisement of Joseph Merola. He earned his Ph.D. in 2000 at the California Institute of Technology studying polymerization and hydrometallation chemistry with John E. Bercaw.
After his postdoctoral work at the Massachusetts Institute of Technology, Chirik joined the chemistry faculty at Cornell University until 2011, when he was named the Edwards S. Sanford Professor of Chemistry at Princeton University.
Research
Chirik’s multidisciplinary research seeks to transform traditional catalysis, which relies on exotic metals like platinum and rhodium to drive chemical reactions. Instead, Chirik uses alternative, Earth-abundant metals like iron and cobalt, developing techniques that allow these metals to mimic or surpass the performance of exotics.
An important example of this research was published in 2021 in Nature Chemistry, in which Chirik detailed a route to recyclable plastics through a molecule he discovered called oligocyclobutane. This molecule can be “unzipped” back to its original monomer by employing an iron catalyst in a process known as depolymerization.
Another major focus of Chirik’s lab is improving the process surrounding iron- and cobalt-based catalysis cross-coupling for carbon-carbon bond formation, an essential technology used by the pharmaceutical industry to develop new therapies. Chirik publishes regularly on this technology, with recent papers on cobalt-catalyzed cross-coupling and the addition of halides to a reduced-iron pincer complex to create an improved pathway for a desired end product.
In 2022, Chirik was among the first chemists in the nation to receive a Gordon and Betty Moore Foundation exploration-phase grant in green chemistry based on his proposal for iron catalysts for a biorenewable hydrocarbon future.
He has been the editor-in-chief of the journal Organometallics since 2015.
Awards and honors
Named Fellow in the American Association for the Advancement of Science, 2023
Gabor Somorjai Award for Creative Research in Catalysis, 2021
The Linus Pauling Award, 2020
Paul N. Rylander Award, 2020
The Arthur C. Cope Scholar Award, American Chemical Society, 2009
David and Lucile Packard Fellowship in Science and Engineering, 2004
Editor-in-Chief, Organometallics, 2015 to present
References
Living people
Virginia Tech alumni
California Institute of Technology alumni
American organic chemists
21st-century American chemists
1973 births
Scientists from Philadelphia
Princeton University faculty
Cornell University faculty | Paul Chirik | [
"Chemistry"
] | 596 | [
"Organic chemists",
"American organic chemists"
] |
57,585,781 | https://en.wikipedia.org/wiki/List%20of%20Canadian%20bioethics%20programs | This following list of Canadian bioethics undergraduate and graduate programs was developed by the Canadian Task-force of the Association of Bioethics Program Directors in March 2017, based on a 2012 list developed by the Canadian Bioethics Society.
References
Postgraduate schools
Bioethics | List of Canadian bioethics programs | [
"Technology"
] | 55 | [
"Bioethics",
"Ethics of science and technology"
] |
57,586,391 | https://en.wikipedia.org/wiki/Trametes%20cubensis | Trametes cubensis is a poroid bracket fungus in the family Polyporaceae. It was first described in 1837 as Polyporus cubensis by Camille Montagne. Pier Andrea Saccardo transferred it to the genus Trametes in 1891.
References
Polyporaceae
Fungi described in 1837
Fungi of North America
Fungi of South America
Taxa named by Camille Montagne
Fungus species | Trametes cubensis | [
"Biology"
] | 77 | [
"Fungi",
"Fungus species"
] |
57,587,843 | https://en.wikipedia.org/wiki/List%20of%20search%20appliance%20vendors | A search appliance is a type of computer which is attached to a corporate network for the purpose of indexing the content shared across that network in a way that is similar to a web search engine. It may be made accessible through a public web interface or restricted to users of that network. A search appliance is usually made up of: a gathering component, a standardizing component, a data storage area, a search component, a user interface component, and a management interface component.
Vendors of search appliances
Fabasoft
Google
InfoLibrarian Search Appliance™
Maxxcat
Searchdaimon
Thunderstone
Former/defunct vendors of search appliances
Black Tulip Systems
Google Search Appliance
Index Engines
Munax
Perfect Search Appliance
References
External links
7 Enterprise Search Appliances That Can Save the Day
Computer hardware
Information retrieval systems
Internet search
Computing-related lists | List of search appliance vendors | [
"Technology",
"Engineering"
] | 175 | [
"Information retrieval systems",
"Computer engineering",
"Computing-related lists",
"Computer hardware",
"Computer systems",
"Computer science",
"Information technology",
"Computers"
] |
57,588,368 | https://en.wikipedia.org/wiki/Department%20of%20Atomic%20Energy%20%28Malaysia%29 | The Department of Atomic Energy Malaysia (Atom Malaysia), or formerly known as Atomic Energy Licensing Board or in Malay known as Lembaga Perlesenan Tenaga Atom (AELB), is a government agency under the Ministry of Science, Technology & Innovation (MOSTI) that is responsible for the regulation of atomic energy activities in Malaysia as stipulated in the Atomic Energy Licensing Act 1984 (Act 304).
The Public Service Department (JPA) of Malaysia approved the name change from AELB to Atom Malaysia in June 2022. The name change took effect on that date, rebranding the agency from AELB to the Department of Atomic Energy (Atom Malaysia).
Atom Malaysia is led by a Director General, current Director General is Hjh. Noraishah binti Pungut. Atom Malaysia is a regulatory authority that operates from its headquarters in Dengkil, Selangor and its four branches around Malaysia, namely in the Penang state (has relocated to Kedah state since October 2022), Johor state, Terengganu State and Sarawak, all of which are responsible in licensing processes and enforcement activities related to radioactivity and atomic energy.
Establishment
Control over the use of radioactive substances began in 1968, when Parliament passed the Radioactive Substances Act 1968. Due to the rapid development and utilization of atomic energy activities in Malaysia, in which requires more effective regulatory control, inspection and enforcement, the Atomic Energy Licensing Bill was drafted. This bill was then passed in Parliament in April 1984, as the Atomic Energy Licensing Act (Act 304). In line with Section 3 of the Act 304, Atomic Energy Licensing Board (AELB) was established under the Prime Minister's Department on 1 February 1985. AELB acts as an enforcement authority for the implementation of the Act. However, on 27 October 1990, AELB was then placed under the Ministry of Science, Technology and Innovation.
Today in 2022, AELB is renamed to be the Department of Atomic Energy (Atom Malaysia) due to the fact that since its establishment in 1985, this department is not known to have a name of its own. The Atomic Energy Licensing Board, whereby the "Board" by all definition only refers to the 5 board members. At the same time the Director General of this department also acts as an Executive Secretary to this Board.
Objectives
To ensure all atomic energy activities are safe, secured and safeguarded (3S's) for protecting the public, workers and environment.
Mission
To regulate the use of atomic energy for nation's wealth and well being.
Vision
An effective regulatory body for safety, security and safeguards in atomic energy activities.
Functions
Atom Malaysia's functions as set forth in Act 304 are as follows:
To advise the Ministers and the Government of Malaysia on matters relating to the Atomic Energy Licensing Act 1984 and the progress of its development, particularly on the implications of the development for Malaysia.
- controlling and monitoring the activity of atomic energy and the matters incidental thereto;
- to establish, maintain and develop scientific cooperation with any body, institution or other organization relating to nuclear matters or atomic energy as the Board deems fit for the purposes of the Atomic Energy Licensing Act 1984;
- where so directed by the Government of Malaysia, to implement or provide for the execution of obligations arising out of agreements, conventions or treaties relating to nuclear or atomic matters of which Malaysia is a party if the treaty, convention or treaty is connected with the purpose - the Atomic Energy Licensing Act 1984; and
- to do other matters arising or arising out of the functions of this Board under the Atomic Energy Licensing Act 1984 which is not contrary to the meaning of the Act, whether directed or not directed by the Minister.
References
External links
Official Website of Department of Atomic Energy Malaysia
Official Facebook of Atomic Energy Licensing Board, Malaysia (Atom Malaysia)
Ministry of Energy, Technology, Science, Climate Change and Environment (Malaysia)
Nuclear regulatory organizations | Department of Atomic Energy (Malaysia) | [
"Engineering"
] | 782 | [
"Nuclear regulatory organizations",
"Nuclear organizations"
] |
57,589,493 | https://en.wikipedia.org/wiki/Xiaomi%20Mi%208 | The Xiaomi Mi 8 is a flagship Android smartphone developed by Xiaomi Inc. It was launched at an event held in Shenzhen, China as the successor to the Xiaomi Mi 6. The naming of the Xiaomi Mi 8 (skipping the Mi 7) is in celebration of Xiaomi Inc's eighth anniversary. The Mi 8 draws parallels to the iPhone X, as both the rear and front of the phone are replicated. This design was later carried on to the mid-range Redmi Note 6 Pro and Mi A2 Lite.
Specifications
Hardware
The Xiaomi Mi 8 is powered by the Qualcomm Snapdragon 845 processor, with 6 GB LPDDR4X RAM and Adreno 630 GPU. It has a FullHD plus AMOLED display. Storage options include 64 GB or 128 GB. The handset features a fingerprint scanner on the rear or on the screen under the display, in the Explorer Edition. It features a 3,400 mAh battery with a USB-C reversible connector which supports Quick Charge 4.0+. It has Gorilla Glass 5. It does not feature a 3.5mm headphone jack and comes with a USB-C to 3.5mm headphone jack adapter provided in the box. The Mi 8 includes a dual camera setup with a 12 MP wide angle lens sensor and a 12 MP telephoto lens sensor. The front camera has a 20 MP sensor with an aperture of f/2.0. The Mi 8 camera has an overall score of 99 and a photo score of 105 on DxOMark. Explorer Edition also introduces a 3D optical facial recognition with the standard IR Sensor for dark condition and a dual band GNSS which allows reception of L1 and L5 signals simultaneously.
Software
It runs on Android 10, with Xiaomi's custom MIUI 11 skin which is upgradeable to MIUI 12.
References
Android (operating system) devices
Mobile phones introduced in 2018
Mobile phones with multiple rear cameras
Mobile phones with 4K video recording
Discontinued flagship smartphones
Xiaomi smartphones | Xiaomi Mi 8 | [
"Technology"
] | 422 | [
"Mobile technology stubs",
"Discontinued flagship smartphones",
"Flagship smartphones",
"Mobile phone stubs"
] |
57,590,109 | https://en.wikipedia.org/wiki/Wireline%20QA/QC | Wireline quality assurance and quality control (wireline QA/QC) is a set of requirements and operating procedures which take place before, during, and after the wireline logging job. The main merits of wireline QA/QC include accuracy and precision of recorded data and information. Accuracy is a measure of the correctness of the result and is generally depended on how well the systematic errors (a reproductible inaccuracy introduced by faulty design, inadequate calibration or a change in borehole) are controlled and compensated for. Precision is depended on how well random errors (errors that cannot be reproduced and are mostly related to the physics of a measurement) are analysed and overcame .
Introduction
Wireline logging is a part of exploration geophysics and is mainly used to detect the presence of economically useful hydrocarbons in the Earth’s sub-terrain. Products of wireline logging are wireline logs or well logs.: (Fig. 1. Ll1LOG) In the oil industry the logs are “a recording against depth of any of the characteristics of the rock formations traversed by a measuring apparatus in the well-bore”. The well logs are obtained by lowering the measuring tools on a wireline (cable) into the well (borehole). The integral parts of wireline logging operations are quality assurance and quality control procedures. Quality control is the “process that defines how well the solution for a specific problem is known”. Well log quality control is a “set of methods that identifies and analyses data deviations from established standards and allows the design of a remedy”.
Unlike measurements in well-known and controlled laboratory conditions, logging is performed in-situ and can be affected by different possible failure sources and susceptible to systematic and random errors. The objective of wireline logging job is to obtain a permanent continuous record of the rock properties penetrated by the wellbore with the result of wireline logs, fluid, and rock samples. Of all the well data sets recorded and collected, well logs are the most valuable as they are vital for reservoir and formation evaluation.
Wireline (well) logs are then combined with drilling data, mud logs, and measurements while drilling (MWD) and coring information in order to choose correct testing and completion intervals to properly evaluate the production potential of the well. There are two categories of well log data: the original data (e.g. Gamma Ray log as a result of gamma rays measurements in the borehole, through an iodine crystal scintillation detector, calibrated as per normal field operational procedure) or derived data (data resulting from the processing of the original data, e. g. calculating the volume of shale).
Main components of the well log data record are logs themselves (main curves for the main and repeat sections with relevant information - calibration information, parameter tables... and additional curves of the main and repeat passes including down-logs and images of quality control logs) and contextual information, such as data acquisition plans, other job reports, witness reports, and tool specification. Contextual data is of a great value for exploration but can be lost or difficult to access.
With the best quality wireline logging data acquired, subsequent steps in well production and completion are more precise and effective. High quality data and conscientious data management enables the prognosis of potential problems and failures so the whole process (the whole rig or oilfield lifetime) can be adapted and configured in order to prevent possible consequences in regard to human safety, environmental preservation, and infrastructural integrity. Acquisition of geotechnical and petrophysical data is costly, but necessary. Data of dubious quality cannot yield reliably good decisions. The poorer the quality of data, the higher risk associated with decisions based upon them. The less knowledge there is about the quality of recorded and collected data, the higher is the uncertainty of said decisions and can lead to false interpretations and evaluations. A good evaluation is only possible with good quality data so it is essential to acquire the best quality data.
Factors affecting the data quality
The data quality may be compromised to a greater or lesser extent by bad hole conditions, wireline equipment failures, human errors, or even extreme weather conditions. Within the bounds of the rig time costs and preservation of the personnel safety, equipment and the well, the logging engineer, operational geologists and wireline log witnesses have to ensure that the best quality data is acquired. The objective of the wireline logging services is to provide the best possible quality data in the minimum possible rig-time. Quality of the data is directly dependent on available rig-time. Factors affecting data quality and time efficiency are operational procedures and environmental conditions, but mainly problems caused by them. During wireline job planning, some factors are accountable for specific types of equipment used for specific borehole geometry and conditions (temperature, pressures and mud type used) and others are wireline responsibility on site and during the wireline logging job.
Data quality and time efficiency factors
Environmental conditions:
Drilling mud – hole diameter affects log readings, mud type and density affect type of tools used and measurements such as conductivity and resistivity, invasion of drilling, and mud fluids into reservoir rock, forming of mud-cake, wash-outs in soft sediment, formation of stress-related break-offs.
Borehole geometry – e. g. affects type of logging and tools used (wireline, pipe conveyed or LWD).
Borehole environment – temperature, pressures and possible hostile (H2S) environments affect tools used and their operating limits.
Operational procedures:
Equipment calibrations and verifications – tools are calibrated by adjusting their response to read some predetermined value in a situation for witch the response is known, and the only time we know for certain that the tool is working properly is during calibration and verification. They should be checked as often as possible to ensure the correct data is recorded.
Tool positioning and configuration (geometry) – tools have to be centralised or decentralised depending on the type of measurement recorded, they have different depths of recording intervals depending on their position in tool-string, different vertical resolution (e. g. difference in vertical resolution in short and long spacing resistivity tools), spacing of sensors on tools influences the volume of rock recorded. Different tools have different depths of penetration into formation, making the need for accuracy very important.
Depth mismatch/accuracy – elemental requirement for consistency of data is to know on what depth the measurement is taken, otherwise it is difficult to compare different logs needed for further data processing (interpretation).
Logging speed – logging speed limits directly affect rig time efficiency as they are not the same for all types of logs (and connected tools), and service companies provide maximum speed values for chosen operating tool.
Types of Log Data Quality control
Log data quality control is a method to assure a certain level of well log data quality and accuracy in the oil company’s system. The most effective way to be certain of the quality is to make checks at the time of the data acquisition while the wireline operation is ongoing or just finished. The best practice is for all three types of log quality checks to be done on-site during a wireline job.
According to Storey (2016) there are three types of log data quality control (LQC):
Type 1 - Acquisition LQC
Quality control applies to original data as they are recorded by the wireline service company and is performed by the wireline service companies engineers and/or wireline log witnesses. The main objectives are to monitor operational procedures, that the logging program is followed, communication about progress and problems, recording of any notable events, verifying the precision, accuracy and completeness of acquired data (logs and contextual information), and equipment verifications and calibrations.
Type 2 - Acceptance LQC
Quality control during the acceptance of data by the oil/gas company (log data recorded and collected by the wireline service company and/or 3rd party). Objective is to verify the data and record for any deviations from original job plans and register them without delay. The purpose is to assure completeness and accuracy, to remedy all unacceptable deviations from the original plan and record all acceptable deviations, to clarify any significant questions or problems, to communicate progress, problems and concerns to subject matter experts, to record the log quality control summary in writing and to accept the data formally. The main concern is the completeness and accuracy of data, and consistency of different components of the data and information set.
Type 3 – Pre-exploitation LQC
Verifying that the Type 1 and Type 2 LQCs have been done before data exploitation. This control can happen during operations/job to help decide on the next operation, or after the operation/job to construct, constrain, and refine formation and reservoir models, appraise uncertainties and decide on follow-up action.
Wireline QA/QC service providers
The well site is the front line of well log quality control. Only while the equipment is on-site and the well bore is open is it possible to investigate in-situ formations, and, if necessary, influence decisions which can result in better, higher data quality. If the problem is detected it can be readily addressed by the personnel around. Quality executed wireline logging job from start to finish makes subsequent steps in oil and gas extraction (the main goal of well site operations) more straightforward and thus ultimately less expensive. It makes better prediction possible and adapting to unwanted surprises and problems which usually cause delays (down-time), additional interventions and higher costs.
Service companies
Wireline service companies are responsible for delivering complete, accurate and consistently recorded all-encompassing data and information which has been acquired by correctly calibrated and verified instruments, correctly used during the logging runs (standard division of logging job) by providing relevant documentation in hard and soft copy data. Wireline service companies provide the personnel and equipment needed and agreed upon, in coordination with the oil company, for any specific logging job on the specific well site (FORASERV, Schlumberger, Baker Hughes, Halliburton, Archer well to name a few).
Wireline logging QA/QC witness
Oil companies, to ensure that the highest quality data is recorded, engage an experienced expert to supervise wireline logging job on site that coordinates an entire operation (with assistance of logging engineer and operational geologist) from start to finish and has knowledge of every step of the operation in intimate detail. They are called wireline log witnesses and are either individual consultants or part of the wireline QA/QC consulting firms. Their objective is to keep track of the wireline performance of logging provider, their efficiency and procedures. Some of the notable wireline QA/QC consulting firms are QO Inc., Gaia Earth Sciences Limited, OGEC, one & zero and STAG. Complexity of the modern wireline and LWD logs, with several pages of logging job parameters and calibrations, makes this duty very demanding and even specialized training courses for log witnesses are available (for example Petroknowledge and Opus Kinetic).
Roles and responsibilities of wireline log witness
The role of wireline log witness is all-encompassing and includes coordination, participation, and supervision of all pre-job, during job (on site), and post-job logging activities:
They are responsible for communication with on-site personnel, for reporting all logging objectives, and operations to the contractor (usually the oil company) during the job and a final report after the job.
Their answerability includes following up the predetermined logging program and objectives (between the oil/gas company and wireline service company) and approving and recording of all subsequent modifications within the limits of service company abilities and the oil company wants.
They need to know what tools are required and what tools are available.
They need to make sure the logging crew has checked all the equipment (to ensure all the logging tools are in working condition), that no equipment required for the job is missing, and all tool calibrations and verifications are valid.
They are responsible for repeating the tool calibrations and verifications several times during the job to check them.
They are obliged to review and discus job risk assessment and potential safety issues.
They are in charge of implementing "Lessons learned" from previous jobs and taking the notes of "Lessons learned" on present job.
Well log QA/QC procedures
Foundations of a good well log data quality are data consistency, tool calibration, tool reliability, and wireline service company performance. In collecting right and good quality data, proper and valid calibrations and verifications of all the equipment, consistent complementary readings checks (e. g. density and neutron logs or porosity logs) and repeatability of recorded logs are all essential.
Basic steps in log quality control are:
Having a well-planned and detailed logging program which takes into account site and safety conditions, information about the equipment, acquisition parameters, borehole environmental conditions, and planned operational procedures. The program should be coordinated by the oil/gas company, wireline service company and wireline log witness.
Paramount to the on-site job is documentation of all the fieldwork – field logbooks, data collection sheets, and field notes, as well as completely recorded instrument digital data.
The highest priority before the job and during the job are equipment calibrations and verifications. Routine checks of the equipment should be made on periodic bases and after each problem and repair. An operational check of equipment, along with test measurements, should be carried out before the start of each job and before starting each run.
Standard corrections and changes in logging programs made by field engineers or data managers for operational procedures, borehole conditions, previously unknown site conditions and main depth shifts should all be documented. The rationale for why the change was made, compromises and consequences that change may represent should also be documented.
Recording the conditions which affect the survey and measurements. When recognized they should be documented to provide guidance for later projects ("Lessons learned").
All the equipment problems, and steps taken to correct these problems, should be documented with insights how the corrections may affect the data.
It is important to review electronically recorded (digital) data (well logs) to ensure that the data recorded and their values are consistent with the setting. The logs should be depth matched (final depth control), compared, and any borehole geometry effect taken into account. Overall data quality and accuracy should be documented.
Log Quality Control (LQC) Digitalization
Data management is more commonly done using a variety of software packages. Most equipment manufacturers have data transfer (real-time data transmission) or download software provided for their instruments which permits data editing and limited data manipulation.
Digitalization and digital solutions are an important part of the oil and gas industry. Harnessing new technologies, as Devold said, ''“is a critical business need”. Digitalization and finding new software solutions is a mandate in logging because the largest errors come from inadequate knowledge of the position and orientation of the measurement sensors and the biggest problems come from human error. This can be partly eliminated by employing smart software solutions. No matter how careful, people make mistakes, equipment fails, accidents happen. Effectiveness of a quality wireline logging management system is the ability of the wireline service companies, their engineers and logging witnesses to predict, eliminate, overcome, and compensate for possible problems and limitations.
Major steps are taken in the field of log data interpretation and visualisation (e. g. Interactive Petrophysics by Lloyd's Register, Log Studio by Logtek softwares, Petra® by IHS, GeoSoftware by CGG, JewelSuite by Baker and Hughes, Delphi by Schlumberger, Petrosys, Geolog®, GEOSuite7 to name a few) with some software development in the area of well operational performance management and optimisation (RIGIQ® by Trigpoint Solutions, EnergySys), but log data and contextual information management (including wireline program optimisations) is sadly underrepresented. Some of the wireline service QA/QC companies have developed their own internal software solution to this problem (e. g. one & zero consulting company), but almost none of the software is open-sourced or commercially available. At present there can be found one commercial software - RIGPRO by the same company – that offers solutions for digitally supervising all logging data, including the contextual information, in a format that enables quality QA/AC.
References
Quality assurance
Geophysics | Wireline QA/QC | [
"Physics"
] | 3,388 | [
"Applied and interdisciplinary physics",
"Geophysics"
] |
57,591,866 | https://en.wikipedia.org/wiki/National%20Initiative%20for%20Cybersecurity%20Education | The National Initiative for Cybersecurity Education (NICE) is a partnership between government, academia, and the private sector focused supporting the country's ability to address current and future cybersecurity education and workforce challenges through standards and best practices. NICE is led by the National Institute of Standards and Technology (NIST) in the U.S. Department of Commerce.
History
The Comprehensive National Cybersecurity Initiative (CNCI), established by President George W. Bush in January 2008, included over twelve Initiatives, one of which, Initiative 8, was aimed at making the Federal cybersecurity workforce better prepared to handle cybersecurity challenges.
In May 2009, the Cyberspace Policy Review, directed by President Barack Obama, elevated the CNCI Initiative 8, which had initially been focused on improving the Federal cybersecurity workforce's ability to perform cybersecurity work. The scope was expanded beyond the Federal workforce to include the private sector workforce, truly making it a national charge.
In March 2010, the Obama administration declassified limited material regarding the CNCI, making Initiative 8 public: Initiative #8. Expand cyber education. While billions of dollars are being spent on new technologies to secure the U.S. Government in cyberspace, it is the people with the right knowledge, skills, and abilities to implement those technologies who will determine success. However there are not enough cybersecurity experts within the Federal Government or private sector to implement the CNCI, nor is there an adequately established Federal cybersecurity career field. Existing cybersecurity training and personnel development programs, while good, are limited in focus and lack unity of effort. In order to effectively ensure our continued technical advantage and future cybersecurity, we must develop a technologically-skilled and cyber-savvy workforce and an effective pipeline of future employees. It will take a national strategy, similar to the effort to upgrade science and mathematics education in the 1950s, to meet this challenge. Additionally, the CNCI described training, education, and professional development programs as lacking “unity of effort”.
Cybersecurity Enhancement Act of 2014 Title IV established the “National cybersecurity awareness and education program”, which is now known as the National Initiative for Cybersecurity Education (NICE).
Organization
NICE is headquartered at NIST facilities in Gaithersburg, Maryland. The NICE Program Office activities are organized into three categories: government engagement, industry engagement, and academic engagement.
See also
List of computer security certifications
Cyber security standards
References
External links
Federal website for NICE
Workforce
Computer network security
Initiatives in the United States
2010 establishments in the United States
National Institute of Standards and Technology | National Initiative for Cybersecurity Education | [
"Engineering"
] | 544 | [
"Cybersecurity engineering",
"Computer networks engineering",
"Computer network security"
] |
39,961,693 | https://en.wikipedia.org/wiki/Double-lumen%20endobronchial%20tube | A double-lumen endotracheal tube (also called double-lumen endobronchial tube or DLT) is a type of endotracheal tube which is used in tracheal intubation during thoracic surgery and other medical conditions to achieve selective, one-sided ventilation of either the right or the left lung.
Indications
There are several conditions that may make one-sided lung ventilation necessary. Absolute indications include separation of the right from the left lung to avoid spillage of blood or pus from an infected or bleeding side to the unaffected side. Relative indications include the collapse of one lung and the selective ventilation of the remaining lung in order to facilitate exposure of the anatomical structures to be operated on in thoracic surgeries, such as the repair of a thoracic aortic aneurysm, pneumonectomy or lobectomy.
Development and description
A DLT is made up of two small-lumen endotracheal tubes of unequal length fixed side by side. The shorter tube ends in the trachea while the longer one is placed in either the left or right bronchus in order to selectively ventilate the left or right lung respectively. The first double-lumen tube used for bronchospirometry and later for one-lung anaesthesia in humans was introduced by Carlens in 1949. Modifications to the original Carlens tube have been introduced by White, Robertshaw and others. The most commonly used DLTs today are the Carlens and the Robertshaw tubes. These allow single-lung ventilation while the other lung is collapsed to make Thoracic surgery easier or possible. This may be necessary so as to facilitate the surgeon's view and access to relevant structures within the thoracic cavity. The deflated lung is re-inflated as surgery finishes to check for leakages or other injuries.
These tubes are typically coaxial, with two separate channels and two separate openings. They incorporate an endotracheal lumen which terminates in the trachea and an endobronchial lumen, the distal tip of which is positioned 1–2 cm into the right or left mainstem bronchus.
Proper placement of DLTs requires considerable clinical experience, various techniques for their insertion having been developed. And there is a small simulator to help in the training of Carlens tube rotation maneuvers.
Placement has been found to be easier with the aid of fiber optical equipment such as a bronchoscope. Currently, flexible fiberoptic bronchoscopy examination is recommended before, during placement, and at the conclusion of the use of DLTs.
Alternatives
Other methods of achieving a one sided lung ventilation are the Univent tube, which has a single tracheal lumen and blocker, and other endobronchial blockers.
The approach to ventilating each lung via a separate ventilator is called the DuoVent approach. This system operates by connecting both ventilators to a master control unit, allowing for synchrony between the two ventilators.
See also
Combitube
Endotracheal tube
Airtraq
Laryngeal tube
References
Further reading
Alt URL
Medical equipment
Anesthetic equipment | Double-lumen endobronchial tube | [
"Biology"
] | 667 | [
"Medical equipment",
"Medical technology"
] |
39,961,773 | https://en.wikipedia.org/wiki/Application%20Portability%20Profile | The Application Portability Profile (APP) is a 1990s framework for Open-System Environment designed by the NIST for use by the U.S. Government. It contains a selected suite of specifications that defines the interfaces, services, protocols, and data formats for a particular class or domain of applications.
The Application Portability Profile offers structure to "integrate US federal, national and international, and other specifications to provide the functionality necessary to accommodate the broad range of US federal information technology requirements."
Overview
In the second half of the 20th century information systems initially developed from isolated islands of computing. Through progressive changes, these individual systems became connected by common users and common information needs. Late 20th century these systems were well on the way to migrating toward computing environments that consist of distributed, heterogeneous, networked applications, databases, and hardware. The concept emerged of a federal computing environment, that is built on an infrastructure defined by open, consensus-based standards which serve as de facto means of organizing these systems. The NIST developed such an infrastructure, and named it Open System Environment (OSE).
An Open System Environment (OSE) encompasses the functionality needed to provide interoperability, portability, and scalability of computerized applications across networks of heterogeneous, multi-vendor hardware/software/communications platforms. The Open System Environment forms an extensible framework that allows services, interfaces, protocols, and supporting data formats to be defined in terms of nonproprietary specifications that evolve through open (public), consensus-based forums.
Complementary to the Open System Environment is the Application Portability Profile standard. This standard can covers a broad range of application software domains of interest to many US federal agencies, but it does not include every domain within the U.S. Government's application inventory. The individual standards and specifications in the APP define data formats, interfaces, protocols, or a mix of these elements.
APP topics
APP and the NIST Enterprise Architecture Model
The "Application Portability Profile (APP) - The U.S. Government’s Open System Environment Profile Version 3.0" provides recommendations on a set of industry, Federal, national, international and other specifications that define interfaces, services, protocols, and data formats to support an Open System Environment (OSE).
The APP addresses the lowest architecture in the NIST Enterprise Architecture Model, i.e., the Delivery System Architecture. On this level the hardware of the computer architecture, the software and the communications are being specified. Based on these specification recommendations, various services and agencies have defined detailed technical reference models.
APP service areas
The services defined in the Application Portability Profile fall into the following broad spectrum of service areas:
Operating system services (OS)
Human/computer interface services (HCI)
Data management services (DM)
Data interchange services (DI)
Software engineering services (SWE)
Graphics services (GS)
Network services (NS)
Each of the Application Portability Profile service areas addresses specific components around which interface, data format, or protocol specifications have been or will be defined. Security and management services are common to all of the
service areas and pervade these areas in one or more forms.
Applications
In the 1990s the NIST's Application Portability Profile has been applicated in several Enterprise Information Architecture frameworks, such as:
Information Architecture framework for the U.S. Patent and Trademark Office (PTO) of the Department of Commerce (DoC), and
Department of Defense (DoD) in its Technical Architecture Framework for Information Management (TAFIM)
Further reading
Department of Defense (1996). Technical Architecture Framework for Information Management. Vol. 2, Technical Reference Model.
Gary Fisher (1993). Application Portability Profile (APP) : The U.S. Government’s Open System Environment Profile OSE/1 Version 2.0. NIST Special Publication 500-210, June 1993.
Joseph Hungate (1995) "Conference Report: Application Portability Profile and Open System Environment Users Forum Gaithersburg, MD May 9–10, 1995" in: Journal of Research of the National Institute of Standards and Technology. Volume 100, Number 6, November–December 1995
IEEE P1003.22 Draft Guide for POSIX Open Systems Environment—A Security Framework
References
Reference models
Enterprise modelling | Application Portability Profile | [
"Engineering"
] | 872 | [
"Systems engineering",
"Enterprise modelling"
] |
39,962,134 | https://en.wikipedia.org/wiki/San%20Rafael%20River%20train%20wreck | The San Rafael River train disaster occurred on August 10, 1989, when the Rio Bamoa Bridge collapsed under an 11-car train operated by Ferrocarriles Nacionales de México, traveling from Mazatlan to Mexicali. Several cars fell into the San Rafael River. The bridge's supports had been damaged by heavy rains, causing them to fail. Of the approximately 330 people on the train, 112 perished, most by drowning, and 205 were injured, making it Mexico's second deadliest peacetime rail disaster.
References
Railway accidents in 1989
Derailments in Mexico
1989 in Mexico
Bridge disasters caused by scour damage
Bridge disasters in Mexico | San Rafael River train wreck | [
"Technology"
] | 135 | [
"Railway accidents and incidents",
"Rail accident stubs"
] |
39,962,200 | https://en.wikipedia.org/wiki/Analytically%20irreducible%20ring | In algebra, an analytically irreducible ring is a local ring whose completion has no zero divisors. Geometrically this corresponds to a variety with only one analytic branch at a point.
proved that if a local ring of an algebraic variety is a normal ring, then it is analytically irreducible. There are many examples of reduced and irreducible local rings that are analytically reducible, such as the local ring of a node of an irreducible curve, but it is hard to find examples that are also normal. gave such an example of a normal Noetherian local ring that is analytically reducible.
Nagata's example
Suppose that K is a field of characteristic not 2, and K is the formal power series ring over K in 2 variables. Let R be the subring of K generated by x, y, and the elements zn and localized at these elements, where
is transcendental over K(x)
.
Then R[X]/(X 2–z1) is a normal Noetherian local ring that is analytically reducible.
References
Commutative algebra | Analytically irreducible ring | [
"Mathematics"
] | 229 | [
"Fields of abstract algebra",
"Commutative algebra"
] |
39,963,337 | https://en.wikipedia.org/wiki/Dmitrii%20Ivanovich%20Zhuravskii | Dmitrii Ivanovich Zhuravskii (1821–1891) was an engineer who was one of the pioneers of bridge construction and structural mechanics in the Russian Empire.
Zhuravskii attended the Nezhin lycée and entered the St. Petersburg Institute of the Corps of Railroad Engineers where he was influenced by the academician Mikhail Ostrogradsky. He graduated from the institute as first in his class in 1842.
In the beginning of his career he took part in the surveying and planning of the Moscow – Saint Petersburg Railway. In 1857-58 he led the reconstruction of the Peter and Paul Cathedral in Saint Petersburg. In 1871–76 he took part in the reconstruction of the Mariinsky Canal System
He was awarded the prestigious Demidov Prize in 1855 by the Russian Academy of Sciences.
The Zhuravskii Shear Stress formula is named after him (derived it in 1855):
where
V = total shear force at the location in question;
Q = statical moment of area;
t = thickness in the material perpendicular to the shear;
I = Moment of Inertia of the entire cross sectional area.
Legacy
In 1897 a marble bust of Zhuravskii has been installed in PGUPS.
There are streets in the cities of Omsk, Donetsk and Nizhyn are named after him.
References
1821 births
1891 deaths
Mechanical engineers from the Russian Empire
Railway civil engineers
Demidov Prize laureates
Structural engineers | Dmitrii Ivanovich Zhuravskii | [
"Engineering"
] | 289 | [
"Structural engineering",
"Structural engineers"
] |
39,963,351 | https://en.wikipedia.org/wiki/Ways%20That%20Are%20Dark | Ways That Are Dark: The Truth About China is a 1933 non-fiction book by Ralph Townsend which presents Townsend's observations on the state of then-contemporary China. The book is considered an anti-Chinese polemic by scholars.
A harsh critique of Chinese society and culture, Ways That Are Dark was written at a time when China was in the grip of considerable civil strife. Townsend claimed that the source of China's problems lay in fundamental defects in the ethnic characteristics of the Chinese people. Although the book was a bestseller in the United States, it met with highly polarized reactions from its supporters and detractors. Though praised by some periodicals, it was denounced by missionaries and sinologists, including Owen Lattimore who condemned it as "a general indictment of a whole race". It was banned by the government of China.
After World War II, the book fell into obscurity. It was reprinted in 1997 by the white nationalist publisher Barnes Review and subsequently gained renewed popularity in Japan in 2004 when a Japanese translation was published.
Background
Ways That Are Dark is based on its author's experience of living in China for more than one year. Townsend had worked as a journalist and educator in New York before joining the United States Foreign Service on 16 December 1930. He served as a vice-consul in Shanghai between 10 December 1931 and 9 January 1932 and then in Fuzhou until 1 March 1933. The book was released on 10 November 1933 by G. P. Putnam's Sons.
Summary of contents
In his introduction, Townsend describes the book as "an honest attempt to present the facts as they are, however unpleasant" and a counterbalance to the "maudlin sentiment" and "misinformation" of other writers on China who have made "fatuous attempts to sprinkle bright hopes over dark facts." He notes that while China's virtues will speak for themselves, "in appraising a stranger with whom we are to deal, it is important to know his shortcomings".
In the first two chapters he describes the atrocious conditions he witnessed in China. Shanghai is depicted as a squalid, noisy, and polluted den of crime, poverty, and disease, and yet still comparatively wealthy compared to the rest of the country. The interior of China is difficult to access due to lack of infrastructure, is mostly unsafe for travel, and is wracked constantly by famine and starvation.
Townsend asserted in chapter one that the cause of China's present misery lies in the fundamental defects that exist in the ethnic characteristics of the Chinese people. In chapters three and four he explains what he believes these defects are. Townsend states that dishonesty is "the most prominent characteristic in the Chinese mentality as opposed to our own". Townsend gives many examples of him being lied to by Chinese employees, coolies, shopkeepers, and government officials, and notes that many other consuls were driven out of the service by this relentless and "aimless lying, with each lie merely a pretext for another".
According to Townsend, the other highly salient trait of the Chinese is their "indifference to fellow suffering". Through a large number of personal and second-hand anecdotes, Townsend argues that the Chinese may be the only people in the world who are completely unable to comprehend the basic human impulses of sympathy or gratitude toward other people. Because the Chinese feel no empathy toward others, they behave in an unbelievably sadistic and cruel fashion toward one another, and they view altruistic foreigners as targets to be mercilessly taken advantage of.
Other traits Townsend identifies as being typically Chinese are cowardice, lust for money, lack of a sense of personal hygiene, lack of critical thinking skills, insincerity, and obsession with hollow rites. Townsend believes that these personality traits are as notable among China's leaders and educated strata, as much as they are in the poor masses. Townsend's analysis of historical documents leads him to believe that they are not a recent product of the present chaos, but rather are deeply ingrained traits of China's national character. He concludes that the "outstanding characteristics" of the Chinese people "neither enable other peoples to deal satisfactorily with them, nor enable the Chinese to deal satisfactorily with themselves."
In chapters five and six, Townsend discusses US-funded charitable organizations in China, especially missions. He points out that although the United States had given at least $160 million in philanthropy to China, these charities in China, even secular hospitals and schools, are generally preyed upon by the very Chinese people who the charitable workers seek to help. These charities are subject to rampant looting, arson, and murderous mob violence by Chinese people, who their government refuses to prosecute. Much of this violence is incited by the anti-foreigner propaganda of the KMT, a party described as being "worse than the Ku Klux Klan at its most degenerate stage", and through US aid to China Townsend suggests that America is actually bankrolling propaganda against itself.
Though the missionaries are sincere and hard-working, Townsend finds that they suffer from extreme delusion over the ultimate futility of their goal of reforming the Chinese. The missionaries, he states, have willfully ignored overwhelming evidence that no degree of care or education can uproot the intrinsic sickness of Chinese culture. He calls for an end to all missionary and charitable work in China.
In chapter seven, Townsend details the sheer horror of China's ongoing civil war. Among the factions competing for power in China, Townsend believes that none of them, neither the leaders nor their men, have any fixed loyalties or higher motivations apart from desire for loot. With every man only out for himself and "China's microscopically few good men... too weak to be felt", Townsend predicts no end to the chaos.
In chapter eight, much of the violence, Townsend explains, is fuelled by opium, the addiction of one out of every eight Chinese. Peasants are often compelled to plant opium by local administrators and warlords to pay for their armies, to such an extent that many districts are more heavily planted with opium than food, and all the while KMT officials lie incessantly to the international community about their efforts to suppress the trade.
In chapter nine, in contrast to the depredations of the KMT and other warlords, Townsend praises the positive influence of the Japanese in China. Townsend considered the Japanese to be a loyal, brave, reliable, honest, and cleanly people, and thus the polar opposites of the Chinese. Townsend provides his own firsthand account of the Shanghai Incident of 1932, which he claims was probably provoked by Chinese aggression, and similarly sees the Japanese decision to invade Manchuria as a fitting response to the "foredoomed contest of covert violence against the Japanese" waged by Zhang Xueliang (Chang Hsueh-liang). Townsend lavishes praise on the puppet state of Manchukuo as "a blessing to the thirty million or so Chinese living there" which has achieved "stability and well-being for millions". Townsend concludes that informed observers are grateful for Japan's role in dealing with an unruly China.
In chapter ten, Townsend affirms that the "backward Chinese" are America's "only legitimate problem in Asia" and asks what can be done to deal with a nation that spends aid money corruptly, does not respect its loans, mistreats and attacks foreigners, ignores international drug laws, will not protect foreign investment, and does not engage in productive diplomacy with other nations. He warns Americans that the Chinese see kindness only as weakness and thus can never respond to any type of positive reinforcement. "For every Chinese, from highest to lowest," he argues, "all the acts of life are concentrated upon extracting, from those who mean nothing to him, what he can for the benefit of himself and his clan." By contrast, he believes that the Chinese do understand force and respect strength. Therefore, he advocates that the United States forgo naive "sentimentalist" thinking and adopt a policy of "stern insistence upon our rights without cruel abuse of our strength", including withholding further loans without strict conditions and holding on to foreign concessions and extraterritoriality.
Reception
Ways That Are Dark hit the bestsellers list in the United States, and The Robesonian noted that the book "was given lavish praise and bitter abuse by some of the leading newspapers in America". It was advertised as doing "for China what Katherine Mayo did for Mother India", and Foreign Affairs magazine described it as "a sensationally unorthodox and unvarnished picture of the Chinese".
Praise
E. Francis Brown, writing for Current History, approved of the book's comprehensive and frank discussion of conditions within China. Though the book takes a strongly negative stance towards China, Brown argued that "this very unfriendliness makes the book a welcome antidote to much that has been written in recent years and some of its conclusions might be well pondered by those who shape America's Far Eastern Policy."
Willis J. Abbot of the Christian Science Monitor especially praised Townsend's study of the social life and customs of the Chinese and claimed that "Any capable observer with a few weeks at his disposal in China will corroborate much that appears in this volume." His praise was echoed by Douglas Jerrold of The English Review who found the work "brilliant and outspoken".
Negative criticism
By contrast, Lewis S. Gannett of The Nation criticized Townsend for writing an "apology for Japan" and painting the Chinese as "all alike, all generically different from Japanese and Westerners". A similarly negative assessment published in The China Weekly Review observed that "A Chinese might easily write a similar book and by emphasizing the activities of the Capones and Dillingers, the bootleggers, kidnappers and racketeers, prove to his own satisfaction at least that the Americans constituted a degenerate branch of the white race." The Republican concluded its review of the book with "Throughout the presentation of his observations runs a thread of ill-considered bias which taints his words. We must look to other... more discerning writers for the truth about China."
Prominent sinologists were also critical of the work. In the pages of The New York Times, Owen Lattimore condemned Ways That Are Dark as "a general indictment of a whole race" which lacks insight, contains factual errors, and relies on second-hand accounts. He stated that the book would "only convince people who are convinced already."
Nathaniel Peffer denounced the book as "a rehash of all the old patter of the outport hotel lobbies, with all its half-truths, inaccuracies, provincialism, ignorance and sometimes crassness... [Townsend] has not managed to observe accurately the most simple and superficial things." JOP Bland, though deeming Townsend's conclusions "as a whole... unconvincing", at least found the chapter on opium "particularly instructive."
Among the book's detractors were missionaries, whom Townsend had criticized. They reviewed the book negatively in a variety of periodicals, including The Chinese Recorder, which accused Townsend of having "gathered a lot of stories and put them together in a clever, cynical and unusually warped way". Other negative reviews appeared in The Missionary Review of the World, The China Christian Year Book, and The Missionary Herald at Home and Abroad.
Ways That Are Dark continues to be noted for its Sinophobic viewpoint. In 1985, the historian Frank P. Mintz called it "a classic in the literature of Sinophobia." In 2000, the scholar Yong Chen referred to Townsend's attitude towards the ethnic characteristics of the Chinese people as being derivative of "the anti-Chinese prejudice that nineteenth-century writers had propagated." In 2004, the writer Gregory Clark described it as "a viciously anti-China book" that "contrasts an allegedly dirty, devious Chinese nation with the trustworthy, hardworking Japanese".
Though the book was originally to be called "Chinese Merry-Go-Round", the title under which it was ultimately published is a quote from Bret Harte's poem "The Heathen Chinee". Though "The Heathen Chinee" had been used a rallying cry by opponents of Chinese immigration to the United States, Harte had intended it as a parody of the anti-Chinese bigotry prevalent in the United States of the nineteenth century.
Political responses
The Japanese Army and Navy strongly approved of the book. Several thousand copies of it were bought by the War Ministry and Naval Ministry who by the beginning of 1934 were distributing the books for free to foreign journalists and officials. By contrast, the Nationalist Government of China responded by banning Townsend's book throughout China from 1935. Even so, the journalist George Moorad reported that in 1946 Chinese communists distributed contraband copies of the book to American China Marines in the hopes of disillusioning them about conditions in China.
Revival in Japanese translation
Ways That Are Dark was already in its fifth printing by 1937, but in 1942 Townsend was imprisoned for having accepted money from the Japanese Committee on Trade and Information between 1937 and 1940 without registering as a foreign agent. The book was not re-printed again until the white supremacist magazine Barnes Review published a new edition in 1997.
In 2004, the Barnes Review edition was translated into Japanese by a pair of Japanese translators, Hideo Tanaka and Kenkichi Sakita. It became an instant success, selling out 10 successive re-printings before 2007, when it was reissued as a mass market paperback. In 2004,Gregory Clark noted that the Japanese edition of Ways That Are Dark, which was entitled Ankoku Tairiku Chūgoku no Shinjitsu ("The Truth About the Dark Continent China"), had garnered widespread popularity among members of the Japanese right-wing.
Tanaka, one of the translators of the book, praised Townsend in Shokun! magazine for his "penetrating insight" in reporting "starkly and vividly on the true nature of the Chinese that he had witnessed firsthand". Sakita, the other translator, likewise lauded the book in an article written for Nobukatsu Fujioka's Association for Advancement of Unbiased View of History. Sakita called the book "essential reading to understand what China really is" and argued that Townsend's ideas continue to offer important lessons in conducting Sino-Japanese relations today. In 2004, the book was also reviewed positively by the newspaper Yomiuri Shimbun.
See also
America First
Second Sino-Japanese War
References
Bibliography
External links
free copy of Ways That Are Dark from Hathitrust
1933 non-fiction books
Books about China
Old Right (United States)
Propaganda in the United States
Anti-Chinese sentiment in the United States
Far-right publications in the United States
Propaganda books and pamphlets
Polemic
Anti-Chinese sentiment in Japan
Japan–United States relations | Ways That Are Dark | [
"Biology"
] | 3,078 | [
"Behavior",
"Aggression",
"Polemic"
] |
39,963,415 | https://en.wikipedia.org/wiki/Overlay%20transport%20virtualization | Overlay transport virtualization (OTV) is a Cisco proprietary protocol for relaying layer 2 communications between layer 3 computer networks.
See also
Distributed Overlay Virtual Ethernet (DOVE)
Generic Routing Encapsulation (GRE)
IEEE 802.1ad, an Ethernet networking standard, also known as provider bridging, Stacked VLANs, or simply QinQ.
NVGRE, a similar competing specification
Virtual Extensible LAN (VXLAN)
Virtual LAN (VLAN)
External links
Cisco Overlay Transport Virtualization overview
Network protocols
Tunneling protocols | Overlay transport virtualization | [
"Technology",
"Engineering"
] | 115 | [
"Computing stubs",
"Computer networks engineering",
"Tunneling protocols",
"Computer network stubs"
] |
39,963,732 | https://en.wikipedia.org/wiki/Ingate%20Systems | Ingate Systems AB is a Swedish company that sells data network security and telecommunication equipment. The company primarily provides SIP Trunking of IP PBX:s on the US market. It is associated with sister company Intertex Data AB.
History
Ingate Systems was founded in 2001 as a joint venture between the Swedish companies Intertex Data AB, which had developed the first SIP proxy-based SIP aware firewall for the SOHO and SMB market, and Cendio Systems AB, which had developed the Fuego enterprise firewall. Ingate developed SIP-capable enterprise firewalls and SIParators®, Ingate's brand of Session Border Controllers (SBC:s). Intertex became sister company to Ingate, with joint development, and continued its development of its embedded products, Ethernet and DSL access routers for telecom service provider's volume deployment of SIP services to the home and SMB LAN:s – products that include firewall, SBC and IP PBX functionality.
In the fall of 2012, the group companies Ingate and Intertex merged into current Ingate Systems, to join their efforts for global unified communication beyond traditional telephony using both the SIP standard and the emerging WebRTC standard and to cover the product range from home and SOHO to enterprise and carrier usage.
Since 2006, Ingate has hosted the SIP Trunking-Unified Communications Seminars at the IXEXPO conferences.
Ingate also participates in the SIP Trunking Community by TMCnet.
Products
Ingate has a line of enterprise Session border controllers ranging from home and SOHO usage to enterprise and telecom service provider usage of up to 20,000 concurrent telephone calls.
The SBC:s are often used for SIP Trunking, i.e. connecting enterprise IP PBX:s to telecom service provider's SIP based Voice over IP (VoIP) connections. A wide range of IP PBX:s and service providers are supported by the Ingate startup configuration tool. Unified Communications (UC) solutions such as the Microsoft Lync and Cloud computing solutions such as the Microsoft Office 365 are also supported.
The Ingate products are all based on an SBC architecture where an IETF RFC 3261-compliant SIP proxy controls the NAT and firewall engine to route all types of media between parties on both public and private IP networks. All products include a routing SIP Proxy and a SIP registrar, and when required a SIP Back-to-back user agent (B2BUA), and can in addition to security at the enterprise edge, also do accounting by reporting usage in Call Detail Records. The Ingate products are not limited to VoIP but support full UC or multimedia real-time communication, and can route such communication over the Internet with or without service provider involvement.
Products are being developed to also support and enable the upcoming WebRTC standard, an initiative by Google to bring high quality multimedia real-time communication directly into the web browser.
References
External links
Internet+, A Model for Global Unified Communication website
Computer companies of Sweden
Computer hardware companies
Networking hardware companies
Telecommunications equipment vendors
Companies established in 2001
Technology companies of Sweden
Companies based in Stockholm | Ingate Systems | [
"Technology"
] | 642 | [
"Computer hardware companies",
"Computers"
] |
39,963,800 | https://en.wikipedia.org/wiki/Arise%2C%20awake%2C%20and%20stop%20not%20till%20the%20goal%20is%20reached | "Arise, awake and stop not till the goal is reached" is a slogan popularized in the late 19th century by Indian Hindu monk and philosopher Swami Vivekananda, who took inspiration in a sloka of Katha Upanishad. It was his message to the world to get out of their hypnotized state of mind and discover their true nature. This shloka is the basis of the title of the 1944 book The Razor's Edge and its 1946 film adaptation, its 1984 film adaptation, and of various music albums in the west by bands like AC/DC, Dave Holland, etc.
In Katha Upanishad
Nachiketa, the child protagonist of Katha Upanishad, was sent to Yama, the Hindu god of death, by his father Vajashrava. In the abode of Yama, he answered Nachiketa's questions and taught him Self-knowledge and the methods of Yoga. The words "Arise, awake..." can be found in the 1.3.14 chapter of the book, where Yama is advising Nachiketa—
In Swami Vivekananda's teachings
The inspirational sloka was Swami Vivekananda's message to the Indians to get out of their hypnotized state of mind. The sloka was meant as a call to his countrymen to awaken their "sleeping soul" and propagate the message of peace and blessings given by the "ancient Mother" to the world. "Awake" also denotes the awakening of one's real nature and the consequent ushering in of prosperity.
On 24 April 1897 Vivekananda wrote a letter to Sarala Ghoshal. In that letter, he stressed giving the public only positive education, because of his belief that negative thoughts weaken men. In that letter, he also reiterated this sloka.
Vivekananda quoted this sloka in several lectures and discourses. In a lecture delivered on 12 November 1896 at Lahore, he said:
Therefore, young men of Lahore, raise once more that mighty banner of Advaita, for on no other ground can you have that wonderful love until you see that the same Lord is present everywhere. Unfurl that banner of love! "Arise, awake, and stop not till the goal is reached." Arise, arise once more, for nothing can be done without renunciation.
In her essay Reminiscences of Swami Vivekananda, Sister Christine wrote that Vivekananda wanted to see men striving to find the Supreme. She wrote:
All else might be false, this alone was true. He realized it. After his own great realization, life held but one purpose—to give the message with which he was entrusted, to point out the path and to help others on the road to the same supreme goal. "Arise, awake, and stop not till the goal is reached."
"Arise" was meant as a passionate call for national awakening to obtain political freedom for the country from colonialism, and to not to "stop" until the "goal" was achieved. This was essential in the social, economic and political fields. "Arise" was also intended to mean to get out of the state of helplessness. His emphasis was on freedom to the nation, as in the US on 4 July 1776. Swamiji also urged people to learn from Hindu sacred scriptures, which he felt contained all the instructions to arise out of the "hypnotism of weakness" and which indicated that no individual is inherently weak.
Influence
The 1998 film Swami Vivekananda, directed by G. V. Iyer, ends with this quote where Mammootty gives a brief speech on Vivekananda and his ideals and concludes the speech with this quote. On 12 January 2013, on the 150 birth anniversary, then Gujarat Chief Minister Narendra Modi and now prime minister of India, wrote a blog post on his personal website to pay tribute to Vivekananda. He named the post "Commemorating Swami Vivekananda: Arise, Awake and stop not till the goal is reached". The sloka is inscribed on the main stage of an auditorium of Ramakrishna Mission Institute of Culture, Kolkata, a branch of Ramakrishna Math and Ramakrishna Mission. Dr. Sanjeev Kumar, an Indian author, called this a "life-transforming line" and wrote a book named Stop Not Till the Goal is Reached in 2010.
Usage
Higher education
The phrase is used as the motto of many universities, colleges like IIEST, Shibpur, IIT (ISM) Dhanbad.
See also
Awake: The Life of Yogananda, (2014 film)
References
Citations
Sources
External links
Swami Vivekananda Quotes: Arise, awake, and stop not till the goal is reached
Upanishadic concepts
Swami Vivekananda
Motivation
Slogans
1890s neologisms
1890s quotations
Quotations from religion | Arise, awake, and stop not till the goal is reached | [
"Biology"
] | 1,012 | [
"Ethology",
"Behavior",
"Motivation",
"Human behavior"
] |
39,965,553 | https://en.wikipedia.org/wiki/Neocities | Neocities is a commercial web hosting service for static pages. It offers 1 GB of storage space for free sites and no server-side scripting for both paid and free subscriptions. The service's expressed goal is to "revive the support of free web hosting of the now-defunct GeoCities". Neocities was launched in 2013 by Kyle Drake. As of April 2024, it hosted more than 765,600 sites. The service is powered by an open-source backend provided under the FreeBSD license.
History
Neocities was created by Kyle Drake on May 23, 2013, and launched on June 28, 2013, offering 10 megabytes of file storage for every user. It initially served as an archive for sites previously hosted on GeoCities before the latter's shutdown.
On May 8, 2014, Neocities announced that it would limit the bandwidth speed of the FCC headquarters to early dial-up modem speeds as a protest against FCC's stance on net neutrality. This protest received wide attention and lasted until February 2, 2015.
The service hosted about 55,000 to 57,000 sites in 2015, which had risen to over 460,000 by 2022, and 615,700 by 2023.
As of currently, Neocities allows 1 GB of storage to free users, and 50 GB of storage to "supporters".
Usage
Neocities allows users to create their own websites using HTML, CSS, and JavaScript, and the development tool comes with a built-in debugger for these languages. The intention is for users to create personal websites reminiscent of GeoCities.
Neocities has 2 options for users to store their data. A free plan, which has 1 gigabyte of data storage and slower transfer speeds, and a paid plan, which allows 50 gigabytes of storage and faster transfer speeds. The paid plan costs $5.00 per month, and funds go to server expenses.
The files that free users can host on Neocities are restricted to HTML files, CSS files, Javascript files, Markdown files, XML files, text files, fonts and images. By upgrading to their paid plan, this restriction is removed. This restriction is in place to prevent it from becoming a "file dump".
References
External links
Free web hosting services
Internet properties established in 2013
Free software websites | Neocities | [
"Technology"
] | 488 | [
"Computing websites",
"Free software websites"
] |
39,965,610 | https://en.wikipedia.org/wiki/Martin%20Schr%C3%B6der%20%28chemist%29 | Martin Schröder in an inorganic chemist. He is Vice President and Dean for the Faculty of Science and Engineering and Professor of Chemistry in the Department of Chemistry at the University of Manchester since June 2015. He served previously as Executive Dean of the Faculty of Science from 2011 to 2015 and Professor of Inorganic Chemistry at the University of Nottingham from 1995 to 2015.
Early life and education
Martin Schröder was born in Taplow, Buckinghamshire of Estonian refugee parents in 1954, and was educated at Montem Primary School and Slough Grammar School. He is first in family to attend university, and was awarded a Bachelor of Science degree in chemistry from the University of Sheffield in 1975 and a PhD from Imperial College London in 1978 where his research on oxo complexes of osmium and ruthenium was supervised by William P. Griffith.
Career and research
After postdoctoral fellowships at the ETH, Zürich with Albert Eschenmoser, funded by a Royal Society-Swiss National Foundation Fellowship, and at the University of Cambridge with Jack Lewis, he was appointed to a senior demonstratorship at the University of Edinburgh in 1982. He was subsequently promoted to lecturer, reader and then professor, and in 1995 was appointed to the University of Nottingham as head and professor of inorganic chemistry. He served as head of the School of Chemistry at the University of Nottingham from 1999 to 2005, and as executive dean of the Faculty of Science (2011–2015). In 2015 he moved to his current position as vice-president and dean of the Faculty of Science and Engineering and professor of chemistry at the University of Manchester.
He has been a visiting professor at the University of Toronto, Canada, the University of Otago, Dunedin, New Zealand and the Université Louis Pasteur, Strasbourg, France, and has published over 540 publications and patents.
His early independent research focussed on the chemistry of transition metal thioether and aza macrocyclic complexes with particular focus on the stabilisation of unusual oxidation state species. This work led to the isolation and characterisation of unique mononuclear M(I)/(III) (M = Ni, Pd, Pt) and M(II) (M = Ag, Au, Rh, Ir) complexes. His current research focuses on the development of new advanced functional materials, particularly metal-organic framework materials for selective fuel and toxic gas capture, purification and catalysis.
Controversy
In 2021, Schröder sent an email to Christopher Jackson in his capacity as a Vice President of the University of Manchester, linking to a right wing website GB News and disputing the presence of institutional racism at the University of Manchester. Jackson has subsequently left the institution, and Schröder has declined to apologise.
Awards and honours
In 1994 he was elected Fellow of the Royal Society of Edinburgh (FRSE) and Fellow of the Royal Society of Chemistry (FRSC), and in 2016 he was elected Member of Academia Europaea (MAE). He is currently a Member of Council of the Engineering and Physical Sciences Research Council (EPSRC). He has held a Leverhulme Trust Senior Research Fellowship, and has Honorary Degrees from Tallinn Technical University and from Nikolaev Institute of Inorganic Chemistry, Russian Academy of Sciences. In 2020, he was awarded the Nyholm Prize for Inorganic Chemistry by the Royal Society of Chemistry.
References
1954 births
Alumni of the University of Sheffield
Alumni of Imperial College London
Fellows of the Royal Society of Chemistry
Fellows of the Royal Society of Edinburgh
Academics of the University of Nottingham
Academics of the University of Edinburgh
Members of Academia Europaea
People educated at Upton Court Grammar School
Living people
British people of Estonian descent
British chemists
Inorganic chemists | Martin Schröder (chemist) | [
"Chemistry"
] | 735 | [
"Inorganic chemists"
] |
39,967,918 | https://en.wikipedia.org/wiki/Analytically%20normal%20ring | In algebra, an analytically normal ring is a local ring whose completion is a normal ring, in other words a domain that is integrally closed in its quotient field.
proved that if a local ring of an algebraic variety is normal, then it is analytically normal, which is in some sense a variation of Zariski's main theorem. gave an example of a normal Noetherian local ring that is analytically reducible and therefore not analytically normal.
References
Commutative algebra | Analytically normal ring | [
"Mathematics"
] | 103 | [
"Fields of abstract algebra",
"Commutative algebra"
] |
39,968,360 | https://en.wikipedia.org/wiki/Vaccine-naive | Vaccine-naive is a lack of immunity, or immunologic memory, to a disease because the person has not been vaccinated. There are a variety of reasons why a person may not have received a vaccination, including contraindications due to preexisting medical conditions, lack of resources, previous vaccination failure, religious beliefs, personal beliefs, fear of side-effects, phobias to needles, lack of information, vaccine shortages, physician knowledge and beliefs, social pressure, and natural resistance.
Effect on herd immunity
Communicable diseases, such as measles and influenza, are more readily spread in vaccine-naive populations, causing frequent outbreaks. Vaccine-naive persons threaten what epidemiologists call herd immunity. This is because vaccinations provide not just protection to those who receive them, but also provide indirect protection to those who remain susceptible because of the reduced prevalence of infectious diseases. Fewer individuals available to transmit the disease reduce the incidence of it, creating herd immunity.
See also
Immune system
Outbreak
Pulse vaccination strategy
Vaccine
References
External links
Centers for Disease Control and Prevention (CDC) immunization schedules
CDC Vaccine contraindications
Epidemiology
Vaccination | Vaccine-naive | [
"Biology",
"Environmental_science"
] | 250 | [
"Epidemiology",
"Vaccination",
"Environmental social science"
] |
39,968,404 | https://en.wikipedia.org/wiki/Degenerative%20chain%20transfer | In polymer chemistry, degenerative chain transfer (also called degenerate chain transfer) is a process that can occur in a radical polymerization where the active site is transferred from one site along the polymer chain to another site, without changing the active site's reactivity (hence the term "degenerate," signifying that the pre- and post-transfer active sites have the same energy or reactivity). Thus, the prevalence of degenerative chain transfer in relation to other chain transfer mechanisms has a significant influence on the molecular weight distribution of the resulting product.
In chain polymerization processes it is observed that during the chemical reactions an active centre on a growing chain is transferred from a growing macromolecule - P• - or oligomer to another molecule (transfer agent XR) or to another site on the same molecule.
P• + XR → PX + R•
This transfer involves termination of the initially growing chain to the completed macromolecule PX, where X denotes one end-group of the macromolecule. The example shows that the growing macromolecule as well as the transfer agent are consumed during this process. However, there are also chain transfer reactions that generate new chain carriers and new chain transfer agents at the same time with significant consequences for the distribution of the (average) molecular weight distribution, the dispersity Đ and the (average) degree of polymerization of the product. These chain transfer reactions are called degenerative chain transfer reactions and are observed, for example in RAFT-, ITRP-, or TERP- processes. RAFT = reversible addition-fragmentation chain transfer polymerization; ITRP = iodine-transfer polymerization; TERP = telluride-mediated polymerization. These polymerization techniques belong to the class of reversible deactivation radical polymerizations (RDRP) that show some characteristics of a living polymerization, however, they must not be addressed as living polymerizations because true living polymerizations are characterized by the absence of any termination reaction.
In this sense, a chain-transfer agent RX is a substance that is able to react with a chain carrier by a reaction in which the original chain P• is deactivated and a new chain carrier R• is generated, as shown above.
References
Polymerization reactions | Degenerative chain transfer | [
"Chemistry",
"Materials_science"
] | 474 | [
"Polymerization reactions",
"Polymer chemistry"
] |
55,838,407 | https://en.wikipedia.org/wiki/Esmeralda%20Mallada | Esmeralda Herminia Mallada Invernizzi (born 10 January 1937 – 12 September 2023) was a Uruguayan astronomer and professor who, for her contributions to that scientific discipline, had been honored with the designation of her name to an asteroid.
Career
Mallada was a student of cosmography with Professor Alberto Pochintesta. At the University of the Republic's , she was a colleague of Gladys Vergara, who helped her prepare for the cosmography professorship competition. She became a professor of cosmography and mathematics in secondary education at age 21. She also taught at the university's Faculty of Sciences, where she graduated with a licentiate in astronomy. She is currently retired.
On 16 October 1952, at the invitation of Pochintesta, she was one of the founders of the Association of Amateur Astronomers (AAA) in Uruguay, and in 2015 was made its honorary president. In 2015, the Minor Planet Center of the International Astronomical Union designated an asteroid that orbits between Mars and Jupiter with her name, 16277 Mallada. It is the first asteroid to bear the name of a Uruguayan woman astronomer.
Publications
Some of the works published by Mallada together with Julio A. Fernández are:
"Distribution of Binding Energies in Wide Binaries"
"Potential sources of terrestrial water close to Jupiter"
"Dynamical Evolution of Wide Binaries"
References
External links
Esmeralda Mallada at autores.uy
Interview on El Observador TV at YouTube
1937 births
20th-century astronomers
20th-century Uruguayan educators
21st-century astronomers
21st-century Uruguayan educators
2023 deaths
University of the Republic (Uruguay) alumni
Academic staff of the University of the Republic (Uruguay)
Uruguayan astronomers
Uruguayan educators
Uruguayan women educators
Women astronomers | Esmeralda Mallada | [
"Astronomy"
] | 356 | [
"Women astronomers",
"Astronomers"
] |
55,838,656 | https://en.wikipedia.org/wiki/Culture%20of%20violence%20theory | The culture of violence theory addresses the pervasiveness of specific violent patterns within a societal dimension. The concept of violence being ingrained in Western society and culture has been around for at least the 20th century. Developed from structural violence, as research progressed the notion that a culture can sanction violent acts developed into what we know as culture of violence theory today. Two prominent examples of culture legitimizing violence can be seen in rape myths and victim blaming. Rape myths lead to misconstrued notions of blame; it is common for the responsibility associated with the rape to be placed on the victim rather than the offender.
Furthermore, the culture of violence theory potentially accounts for inter-generational theories of violence and domestic violence. Childhood exposure to violence in the household may later lead to similar patterns in marital relations. Similarly, early experience with domestic violence is likely to increase an individual's potential for development of clinical symptoms. Additionally, presence of a preexisting mental disorder may heighten the chances of becoming involved in an abusive relationship.
There are many factors which contribute to the persistence of violence among individuals and on a societal level; gender is one relevant factor to understanding the culture of violence theory. In the United States, a majority of reported rapes involve female victims. However, there is a growing body of evidence to support the notion that women can perpetuate relational cycles of violence. While a culture of violence has an impact on people as a whole, for individuals who have experienced trauma in their lives the impact can be much larger.
Development of the theory
As mentioned previously the culture of violence theory addresses the pervasiveness of specific violent patterns within a societal dimension. Specifically, culture of violence theory explains how cultures and societies can sanction violent acts. While related to structural violence, cultural violence theory is different by explaining why direct acts of violence or violence built into systems of society exists and how they are legitimized. Research suggests that cultures can encourage and permit violence to exist as a response to various environmental obstacles, such as widespread resource impoverishment. This can be seen within various aspects of culture, such as film, television, music, language, art, and propaganda.
Austrian peace researcher Franz Jedlicka has made an attempt to measure the culture of violence in different countries of the world with his "Culture of Violence Scale 2023".
Relation to cultural values in the United States
Rape myths
Rape myths refer to the inaccurate views and stereotypes of forced sexual acts, and the victims and perpetuators of them. These notions are prevalent among the general population and often suggest that the victims of non-consensual sexual acts have bad reputations, are promiscuous, dress provocatively, or are fabricating assault when they regret the consensual acts after the fact. These views are often legitimized by the status quo of men dominating women across domains such as family, education, work, and many others. Rape myth acceptance can lead to poor assault/rape prevention measures, decrease in reporting of assaults/rapes, increases of assaults/rapes, and re-victimization.
Violence in relationships
Violence in relationships, commonly referred to as intimate partner violence (IPV), is impacted by various factors including the presence of mental illness or use of substances. Specifically, individuals with depression, generalized anxiety (GAD), or panic disorder are potentially at risk for physical violence towards a partner; findings are consistent for both men and women regarding the connection between psychiatric diagnoses and perpetuation of relationship violence. Additionally, propensity to engage in specific behaviors such as gambling or endorsement of violent pornography have also been associated with increased risk for relationship violence occurrence. Individual factors have also been suggested to be associated with relationship violence including anger, aggressiveness, and adverse emotional internalization. Contrarily, exposure to relationship violence is also linked to the later development of mental health symptoms or diagnoses.
Violence in pop culture and media
The prevalence of legitimization of violence may be facilitated by its presence in various media. There is evidence to suggest that sex-related crimes account for nearly 10% of all dialogue on television related to sex, most of which is found on fictional programs. Additionally, research has also found a positive relationship between pornography consumption and attitudes supporting violence against women, especially when the pornography in question is violent in nature. However, consideration of individual differences is necessary in evaluating exposure to violent media and overall outcomes. Factors which influence media content exposure and subsequent outcomes include gender and personality traits. Individuals who are male, hostile, impulsive, and are low on empathy are more likely to be susceptible to violent media exposure.
Public justification of violence
Public justification of violence arise when those not necessarily directly involved in the violent act will not react negatively to the violence because they believe it is warranted. Examples of public justification of violence are most evident in rape myths and victim blaming, as discussed above. However, the common belief regarding legitimate violence tends to place responsibility on victims or potential victims of violence. Another example that is not as often noted, is the pervasive notion of the "chosen one," within some extremist religious language and various nationalism propaganda that will function as a means to perpetuate the undermining of the other and allowance of violence against the other.
In the Ottoman Empire and Turkey
Turkish sociologist Fatma Müge Göçek has argued that the Armenian genocide and other violent repression in the Ottoman Empire was caused by the Committee of Union and Progress' adoption of a "culture of violence"; she argues that this culture of violence remains embedded in Turkish political culture.
References
Violence | Culture of violence theory | [
"Biology"
] | 1,138 | [
"Behavior",
"Aggression",
"Human behavior",
"Violence"
] |
55,839,872 | https://en.wikipedia.org/wiki/PS1-10adi | PS1-10adi is an unusual and highly energetic optical transient discovered by the Pan-STARRS survey on 15 August 2010. The explosion or transient event emitted 2.3×1052 ergs (2.3×1045 Joules), exceeding ASASSN-15lh. It may be a superluminous supernova or a stellar disruption event. The magnitude of the explosion challenges the limits of the current models for theoretical physics.
The event happened near the centre of J204244.74+153032.1 (a point in Delphinus, specifically right ascension in usual sexagesimal and then declination in decimal terms, with the point/spacing moved four places to the right in both cases). The optical transient peaked at two magnitudes brighter than its host galaxy, and remained brighter for 1000 days. Its distance can be implied by its galaxy's red shift of 0.203. Balmer emission lines due to hot hydrogen showed relative motions of up to 900 km/s. The lines also had a red shoulder. There was also a hot black body continuum spectrum showing temperatures dropping from 11,000K. No X-rays or radio waves were detected from the source.
Similar optical transients include PS1-13jw, CSS100217, J094806, J094608, and J233454.
References
Hypernovae
Delphinus | PS1-10adi | [
"Astronomy"
] | 297 | [
"Astronomical events",
"Hypernovae",
"Delphinus",
"Constellations"
] |
55,840,419 | https://en.wikipedia.org/wiki/Browning%20in%20red%20wine | Oxidation and reduction in red wines can lead to a particularly undesirable brick red color in red wines (or an "orangey" color in white wines). This process is sometimes referred to as browning. In chemical terms, this is called a redox reaction because first the color of the wine deepens after fermentation through oxidation, and then the color begins to brown after bottling through reduction. Browning is not strictly considered a bad thing in every wine that is produced, as sometimes the oxidation process can contribute to the style of the wine in a positive and desirable way.
Generally speaking, however, browning is a sign of the wine going stale from too much exposure to oxygen. Although wine that has gone bad is typically associated with the smell and taste of vinegar or unwanted effervescence, oxidation itself can actually lead to "nutty", "applesauce", and "burnt marshmallow" aromas. The sharp vinegar component comes from acetic acid formed through bacterial processes. A lower pH level is typically preferred in wines because it decreases the overall risk of spoilage. The lower pH reduces the effect of browning and yields better colors, particularly in red wines.
Significance of color
The color of a red wine will have many variables that influence it besides its exposure to oxygen that pertain to other viticultural aspects such as different growing climates, cultivars, and production methods. For example, wines made from hybrid grapes can range anywhere from light pink to purple due to their chemical composition. Dr. Leo McCloskey, best known for his contributions in developing the Enologix software, points out that "Chemical ecology says that a wine's flavor, color and fragrance are expressions of its ecosystem."
However, when people perceive wines, the visual cues received from the color of the wine have a strong impact on their opinion of the wine. In a primitive sense, visual cues will generally have a stronger impact on humans than aromatic components since humans have evolved to identify things with eyesight, rather than through scent relative to other evolved species that rely on different sensory capacities more than they do on eyesight. This idea of a particular significance of color in the perception of wine is especially true in red wines.
Browning is therefore an important issue both for winemakers and for wine consumers, and is generally avoided. A well-known exception to this is sherry, which is often purposely developed with oxidation as part of the production process; however, it can be considered a white wine since it is traditionally made from Palomino grapes.
Odor effects
Through the aging of wine, its aroma compounds are affected as well. The esters in the wine that were initially created during fermentation become reduced as they break back down into alcohol and carboxyl groups so the floral, fruity smells that are associated with freshness in a wine decrease. Meanwhile, the nutty, cooked smells increase through carbohydrate degradation.
See also
Wine fault
Maderisation
References
Wine
Chemical reactions | Browning in red wine | [
"Chemistry"
] | 608 | [
"nan"
] |
55,841,017 | https://en.wikipedia.org/wiki/Mathspace | Mathspace is an online mathematics program designed for students in primary/elementary, secondary, and higher education. It is designed for students aged between 7 and 18, and is used by schools in Australia, New Zealand, the United States, Canada, the United Kingdom, Hong Kong and India.
Mathspace uses an adaptive learning model to personalize the software experience for each student. The questions presented to a user are chosen by an algorithm that responds to past performance, and student input is evaluated to provide feedback on their progress within each problem. Additional support is offered in the form of hints and video tutorials to guide them to the solution. The software assesses each student's performance and generates accompanying report statistics.
Partnerships
Mathspace partnered with Eddie Woo in 2017. Together they created a video hub to categorize Woo's education videos in the various state curricula across Australia.
Awards
External links
References
Educational math software
Education companies of Australia
Companies based in Sydney | Mathspace | [
"Mathematics"
] | 194 | [
"Educational math software",
"Mathematical software"
] |
55,841,611 | https://en.wikipedia.org/wiki/Calcium%20signaling%20in%20cell%20division | Calcium plays a crucial role in regulating the events of cellular division. Calcium acts both to modulate intracellular signaling as a secondary messenger and to facilitate structural changes as cells progress through division. Exquisite control of intracellular calcium dynamics are required, as calcium appears to play a role at multiple cell cycle checkpoints.
The major downstream calcium effectors are the calcium-binding calmodulin protein and downstream calmodulin-dependent protein kinases I / II. Evidence points to this signaling cascade as a major mediator of calcium signaling in cell division.
Meiosis
Historically, one of the most well characterized roles of intracellular calcium is activation of the ovum after sperm fertilization. In deuterosome eggs (mammals, fish, amphibians, ascidians, sea urchins, etc.), successful sperm entry leads to a distinct rise in intracellular calcium ions (Ca2+), with mammals and ascidians displaying a series of intracellular calcium spikes required for completion of meiosis.... Unfertilized vertebrate eggs arrest development after meiosis I. This developmental pause is attributed to the vaguely defined cytostatic factor (CSF). Current researches suggest “CSF” is actually multiple pathways working together to halt division at metaphase of meiosis II. Upon sperm entry into the egg, Ca2+ is released from intracellular stores, leading to inhibition of the CSF-arrest mechanism. Calmodulin dependent kinase II was shown to be the protein responsible for converting the Ca2+ influx signal into inhibition of CSF and activation of cyclin degradation machinery to degrade cyclin B, resulting in progression through meiosis II.
In mammals, this rise in Ca2+ was shown to be driven by IP3 stimulation induced by PLCζ provided by the sperm. In general, PLC enzymes stimulate calcium release by internal stores through the breakdown of PIP2 into IP3 and DAG.
Mitosis
Signaling
Beyond the events of meiosis, changes in Ca2+ levels are observed in a variety of organisms at different stages of division, such as nuclear membrane breakdown and the metaphase-anaphase transition. Further, recent work has shown mechanically induced rapid entry into mitosis of cells paused in G2. Further, progression through division requires the presence of calcium (G1/S, G2/M, and metaphase/anaphase), suggesting checkpoints require a calcium-dependent signaling mechanism
G1/S
Entry into S-phase is calcium dependent. Depleting internal calcium stores inhibits initiation of DNA synthesis. One possible mechanism is that cyclin A synthesis is inhibited, preventing cdk2 activity which is required for initiation of DNA synthesis.
G2/M
Cell cycle progression is regulated by multiple pathways. It was shown using human cancer cell lines, that the G2/M checkpoint is regulated by CaMKII and MAPK crosstalk. Here, CaMKII activates MEK/ERK, which degrades the cell cycle arresting p27 protein
Disease
In general, transformed cells proliferate in a calcium-independent manner, whereas non-transformed cells show high sensitivity to extra-cellular calcium concentration, suggesting oncogenic growth may include disruption of calcium signaling.
Chromatin Structure
Condensation of chromatin is a vital step in cell division, allowing cells to equally distribute chromosomes to the daughter cells. Recent work has suggested that Ca2+ is required for enabling chromatin condensation in prometaphase. Calcium was found to concentrate on condensed DNA to much higher levels compared to normal cytosolic calcium concentration. The role of calcium in condensation was independent of CAMK function, suggesting a purely structural role of Ca2+ in chromatin compaction. Further, this result was demonstrated in vitro with extracted chromatin, emphasizing that the mere presence of Ca2+ can influence the structure of chromatin
References
Cell signaling
Calcium
Cell cycle
Cellular processes | Calcium signaling in cell division | [
"Biology"
] | 812 | [
"Cell cycle",
"Cellular processes"
] |
55,841,783 | https://en.wikipedia.org/wiki/Dejean%27s%20theorem | Dejean's theorem (formerly Dejean's conjecture) is a statement about repetitions in infinite strings of symbols. It belongs to the field of combinatorics on words; it was conjectured in 1972 by Françoise Dejean and proven in 2009 by Currie and Rampersad and, independently, by Rao.
Context
In the study of strings, concatenation is seen as analogous to multiplication of numbers. For instance, if is any string,
then the concatenation of two copies of is called the square of , and denoted . This exponential notation may also be extended to fractional powers: if has length , and is a non-negative rational number of the form , then denotes the string formed by the first characters of the infinite repetition .
A square-free word is a string that does not contain any square as a substring. In particular, it avoids repeating the same symbol consecutively, repeating the same pair of symbols, etc. Axel Thue showed that there exists an infinite square-free word using a three-symbol alphabet, the sequence of differences between consecutive elements of the Thue–Morse sequence. However, it is not possible for an infinite two-symbol word (or even a two-symbol word of length greater than three) to be square-free.
For alphabets of two symbols, however, there do exist infinite cube-free words,
words with no substring of the form . One such example is the Thue–Morse sequence itself; another is the Kolakoski sequence. More strongly, the Thue–Morse sequence contains no substring that is a power strictly greater than two.
In 1972, Dejean investigated the problem of determining, for each possible alphabet size, the threshold between exponents for which there exists an infinite -power-free word, and the exponents for which no such word exists. The problem was solved for two-symbol alphabets by the Thue–Morse sequence, and Dejean solved it as well for three-symbol alphabets.
She conjectured a precise formula for the threshold exponent for every larger alphabet size; this formula is Dejean's conjecture, now a theorem.
Statement
Let be the number of symbols in an alphabet.
For every , define , the repeat threshold, to be the infimum of exponents such that there exists an infinite -power-free word on a -symbol alphabet. Thus, for instance, the Thue–Morse sequence shows that , and an argument based on the Lovász local lemma can be used to show that is finite for all .
Then Dejean's conjecture is that the repeat threshold can be calculated by the simple formula
except in two exceptional cases:
and
Progress and proof
Dejean herself proved the conjecture for .
The case was proven by Jean-Jacques Pansiot in 1984.
The next progress was by Moulin Ollagnier in 1992, who proved the conjecture for all alphabet sizes up to .
This analysis was extended up to in 2007 by Mohammad-Noori and Currie.
In the other direction, also in 2007, Arturo Carpi showed the conjecture to be true for large alphabets, with . This reduced the problem to a finite number of remaining cases, which were solved in 2009 and published in 2011 by Currie and Rampersad and independently by Rao.
Dejean words
An infinite string that meets Dejean's formula (having no repetitions of exponent above the repetition threshold) is called a Dejean word.
Thus, for instance, the Thue–Morse sequence is a Dejean word.
References
Combinatorics on words | Dejean's theorem | [
"Mathematics"
] | 730 | [
"Mathematical theorems",
"Theorems in combinatorics",
"Combinatorics",
"Theorems in discrete mathematics",
"Conjectures that have been proved",
"Mathematical problems",
"Combinatorics on words"
] |
55,842,224 | https://en.wikipedia.org/wiki/Haloplanus%20salinarium | Haloplanus salinarum is a halophilic Archaeon in the family of Halobacteriaceae. It was isolated from the Gomso solar saltern in Buan County, South Korea.
References
Euryarchaeota
Archaea described in 2017 | Haloplanus salinarium | [
"Biology"
] | 55 | [
"Archaea",
"Archaea stubs"
] |
55,843,837 | https://en.wikipedia.org/wiki/Automated%20machine%20learning | Automated machine learning (AutoML) is the process of automating the tasks of applying machine learning to real-world problems. It is the combination of automation and ML.
AutoML potentially includes every stage from beginning with a raw dataset to building a machine learning model ready for deployment. AutoML was proposed as an artificial intelligence-based solution to the growing challenge of applying machine learning. The high degree of automation in AutoML aims to allow non-experts to make use of machine learning models and techniques without requiring them to become experts in machine learning. Automating the process of applying machine learning end-to-end additionally offers the advantages of producing simpler solutions, faster creation of those solutions, and models that often outperform hand-designed models.
Common techniques used in AutoML include hyperparameter optimization, meta-learning and neural architecture search.
Comparison to the standard approach
In a typical machine learning application, practitioners have a set of input data points to be used for training. The raw data may not be in a form that all algorithms can be applied to. To make the data amenable for machine learning, an expert may have to apply appropriate data pre-processing, feature engineering, feature extraction, and feature selection methods. After these steps, practitioners must then perform algorithm selection and hyperparameter optimization to maximize the predictive performance of their model. If deep learning is used, the architecture of the neural network must also be chosen manually by the machine learning expert.
Each of these steps may be challenging, resulting in significant hurdles to using machine learning. AutoML aims to simplify these steps for non-experts, and to make it easier for them to use machine learning techniques correctly and effectively.
AutoML plays an important role within the broader approach of automating data science, which also includes challenging tasks such as data engineering, data exploration and model interpretation and prediction.
Targets of automation
Automated machine learning can target various stages of the machine learning process. Steps to automate are:
Data preparation and ingestion (from raw data and miscellaneous formats)
Column type detection; e.g., Boolean, discrete numerical, continuous numerical, or text
Column intent detection; e.g., target/label, stratification field, numerical feature, categorical text feature, or free text feature
Task detection; e.g., binary classification, regression, clustering, or ranking
Feature engineering
Feature selection
Feature extraction
Meta-learning and transfer learning
Detection and handling of skewed data and/or missing values
Model selection - choosing which machine learning algorithm to use, often including multiple competing software implementations
Ensembling - a form of consensus where using multiple models often gives better results than any single model
Hyperparameter optimization of the learning algorithm and featurization
Neural architecture search
Pipeline selection under time, memory, and complexity constraints
Selection of evaluation metrics and validation procedures
Problem checking
Leakage detection
Misconfiguration detection
Analysis of obtained results
Creating user interfaces and visualizations
Challenges and Limitations
There are a number of key challenges being tackled around automated machine learning. A big issue surrounding the field is referred to as "development as a cottage industry". This phrase refers to the issue in machine learning where development relies on manual decisions and biases of experts. This is contrasted to the goal of machine learning which is to create systems that can learn and improve from their own usage and analysis of the data. Basically, it's the struggle between how much experts should get involved in the learning of the systems versus how much freedom they should be giving the machines. However, experts and developers must help create and guide these machines to prepare them for their own learning. To create this system, it requires labor intensive work with knowledge of machine learning algorithms and system design.
Additionally, some other challenges include meta-learning challenges and computational resource allocation.
See also
Artificial intelligence
Artificial intelligence and elections
Neural architecture search
Neuroevolution
Self-tuning
Neural Network Intelligence
ModelOps
Hyperparameter optimization
References
Further reading
Ferreira, Luís, et al. "A comparison of AutoML tools for machine learning, deep learning and XGBoost." 2021 International Joint Conference on Neural Networks (IJCNN). IEEE, 2021. https://repositorium.sdum.uminho.pt/bitstream/1822/74125/1/automl_ijcnn.pdf
Feurer, M., Klein, A., Eggensperger, K., Springenberg, J., Blum, M., & Hutter, F. (2015). Efficient and robust automated machine learning. Advances in neural information processing systems, 28. https://proceedings.neurips.cc/paper_files/paper/2015/file/11d0e6287202fced83f79975ec59a3a6-Paper.pdf
Machine learning
Artificial intelligence
Computational neuroscience | Automated machine learning | [
"Engineering"
] | 996 | [
"Artificial intelligence engineering",
"Machine learning"
] |
55,846,382 | https://en.wikipedia.org/wiki/Cut-eye | Cut-eye is a visual gesture using one's eyes and face to communicate displeasure or disapproval, and in some cases hostility. The gesture is usually performed by looking at someone out of the corners of one's eyes, then turning the eyes away quickly down towards the foot opposite the eye of the person the gesture is being performed at. The main focus of this gesture is the "cut" of the eyes, and can be performed by moving them in one direction or several sharp up and down movements, meant to convey a feeling of anger or disapproval. This gesture has evolved into popular culture and has taken on new monikers while remaining similar to the original African gesture. The cut-eye gesture now is also identified with: stink-eye, evil-look, eye-roll, or death stare. The gesture also is similar in nature to the evil-eye gesture of the Christian and Islamic theologies.
Uses
The action of performing cut-eye is commonly associated with the suck-teeth gesture and a neck roll for maximum effect. In media depictions, the people who perform the neck roll are usually also portrayed as loud, tough, argumentative, and even combative. Cut-eye is used primarily to depict power over another person who has done something wrong or to the user's displeasure. The use of the cut-eye gesture is meant to send a feeling of anger or disgust without verbally communicating the same sentiment. The cut-eye gesture can be done with or without the knowledge of the recipient and has different meanings in each context. When done directly to the recipient of the gesture, the cut-eye conveys direct anger, disappointment, and disapproval. When done without the knowledge of the recipient, the cut-eye gesture can convey the same message but with the desire to have the sentiment remain unknown due to fear of the recipient or general lack of desire to verbally or blatantly portray anger.
Cut-eye makes its full effect when the eyes of the person performing the gesture are turned the opposite direction from the person to which it is intended for. This sends the message that they aren't worth the attention. The gesture is used in many cases, most typically when a person is talking and is interrupted by another individual. In this case, a sharp cut-eye is given in order to show the level of anger and disrespect felt by the user. In addition, it shows the perceived irrelevancy of the individual who interrupted. It also gives the person to which the gesture is intended the same feeling of disrespect of anger the user of the gesture felt. By "rummaging" one's eyes over a person, one violates their personal space and sense of confidence. The victim cannot do anything in this situation to prevent this violation.
Origins
Africa
In Africa, the origin of the cut-eye gesture is the region in and around the modern-day countries of Ghana and Nigeria. In this region, the gesture began as a show of anger and conflict between two individuals and could be used with the intent of the onset of combat or without said intent.
Europe and the Middle East
In Europe, the origin of the cut-eye gesture is centered around the gesture of evil-eye in the context of its biblical references. The original form in Europe and the Middle East was far more serious than the nature of the gesture in the African setting as in the theologies of Europe and the Middle East, the gesture was believed to carry a supernatural power. In the Eurasian context, the cut-eye or evil-eye gesture would bring bad omens upon the recipient and was believed to bring sickness, curses, and possibly death; as a result the use of the evil-eye gesture could often lead to banishment or Eurasian version of the gesture is most represented in modern culture by the potential offense upon receiving the cut-eye, but also shows in the large range of severity the gesture can carry.
Translations
The visual gesture of cut-eye is found in many Ghanaian languages, including translations in Akan, Ga, and Ewe.
Akan – anikyibuo / anikyie 'the art of breaking the back of the eye'
Ga – o-kpâ-mi 'you cut-eye me'
Ewe – treåku 'seal eyes'
See also
List of gestures
Nonverbal communication
References
Gestures
Human eye | Cut-eye | [
"Biology"
] | 885 | [
"Behavior",
"Gestures",
"Human behavior"
] |
55,846,592 | https://en.wikipedia.org/wiki/Miriam%20Balaban | Miriam Balaban (born in Philadelphia) is a publisher, editor and scientist, recognized for her work in science communication and desalination. She has founded international organizations (European Association of Science Editors; International Federation of Science Editors), conferences (International Conference of Scientific Editors) and journals (Desalination, editor 1966–2009; Desalination and Water Treatment, editor 2009-; Symbiosis 1985-) and has edited numerous journals, conference proceedings and books. She also publishes the Desalination Directory, an international online database.
Balaban was a research associate for science communication at Boston University from 1975 to 2008. She became secretary general of the organizing committee of the International Federation of Scientific Editors' Associations (IFSEA) in 1978, and the founder and first president of its successor organization, the International Federation of Science Editors (IFSE) in 1990. She was Professor and founding Dean of the School for Science Communication at the Mario Negri Sud Institute for Biomedical and Pharmacological Research in Italy from 1988 to 1992. She has been the secretary general of the European Desalination Society (EDS) since 1993. Among other awards, Balaban received the Order of the Star of Italian Solidarity from the President of Italy in 2012.
Early life
Miriam Balaban was born in Philadelphia, Pennsylvania, and graduated from the Philadelphia High School for Girls in 1945.
She attended the University of Pennsylvania where she graduated with a B.Sc. degree in chemistry in 1949.
Science communication
Soon after graduating, Balaban moved to Jerusalem where she found work editing the quarterly scientific bulletin of the Government Research Council.
The first volume of the Bulletin of the Research Council of Israel was published by Balaban in both English and Hebrew.
Under her direction, publication of the bulletin was greatly expanded, eventually splitting into five separate publications focusing on various areas of science, published through the Weizmann Science Press. In 1958, Balaban established the Israel Program for Scientific Translation, contracting with the United States National Science Foundation to fund the translation of foreign language publications.
In the United States, the Council of Biology Editors (CBE), later the Council of Science Editors (CSE), was formed in 1957 following discussions at the Conference of Biological Editors in New Orleans. CBE had the support of the National Science Foundation and the American Institute of Biological Sciences.
In 1960, the organization published its first Style Manual for Biological Journals.
At the 1964 CBE conference in Ann Arbor, Michigan, a special meeting was held for both American and European editors.
This inspired the formation of similar organizations in Europe.
Miriam Balaban helped to start the European Association of Editors of Biological Periodicals (EAEBP). EAEBP was formed in April 1967 in Amsterdam, and renamed European Life Science Editors (ELSE) at its first General Assembly, held at the Royal Society in London in 1970.
In 1982, ELSE joined with the European Association of Earth Science Editors (Editerra) to form the European Association of Science Editors (EASE). Balaban served multiple terms as treasurer of the organization. Concerns of the organization included the development of international standards for science journals, guidelines for authors whose first language was not English, republishing of articles in multiple languages, and the improvement of science communication generally.
From 1975 to 2008, Balaban was a research associate at the Center for Philosophy and History of Science at Boston University, focusing on science communication.
In 1977-1978 Balaban helped to found the International Federation of Scientific Editors' Associations (IFSEA).
The idea for such an organization was put forward at the First International Conference of Scientific Editors, April 24–29, 1977, which Balaban organized. The idea was further discussed at a UNESCO-supported consultation meeting, chaired by Balaban, which was held June 5–6, 1978 in Paris, France. The formation of IFSEA was agreed upon and Balaban was named to the organizing committee as secretary-general.
Balaban organized 12 international conferences for scientific editors, beginning in 1977 with the first International Conference of Scientific Editors in Jerusalem.
In 1990, IFSEA opened its membership to individuals as well as associations, and became IFSE, the International Federation of Science Editors (variously the International Federation of Scientific Editors), with Balaban as its founder and first president.
Balaban is president, publisher and editor of the scientific publishing house International Science Services. Based in Rehovot, Israel and L'Aquila, Italy. ISS publishes in a variety of scientific disciplines.
Balaban has supported the development of journals both within and outside her own field. In 1985, she worked with lichenologist Margalith Galun to create the journal Symbiosis. The first 48 issues were published by Balaban International Science Services. From October 2009 onwards, the journal has been published by Springer.
She established the School for Science Communication at the Mario Negri Sud Institute for Biomedical and Pharmacological Research in Italy where she served as Professor and Dean from 1988 to 1992.
Desalination
The focus of Balaban's research career has been desalination. In 1966, Balaban founded Desalination, the first international journal for desalting and purification of water, serving as its editor-in-chief from 1996 to 2009. She was succeeded as editor by Nidal Hilal.
In 2009 Balaban established and became editor-in-chief of the monthly Desalination and Water Treatment Journal, to accommodate the expanding field. She has reviewed and edited more than 20,000 papers and several books from over 100 countries.
She is the editor and publisher of the Desalination Directory. The international online database connects over 30,000 individuals and 5,000 academic and government institutions and companies involved in desalination and water conservation.
Balaban has been a member of the International Desalination Association since 1975 and has served as a board member and officer. She is a member of the Scientific Program Committee for the International Desalination Workshop.
Since 1993, Balaban has been the secretary general of the European Desalination Society (EDS), located at the Università Campus Bio-Medico, Rome, Italy. Balaban organizes international courses, conferences and workshops in desalination, traveling and speaking internationally. She has been referred to as "the soul of the European Desalination Society".
In her position with the EDS, she served on the Committee to Review the Desalination and Water Purification Technology Roadmap, a document prepared by Sandia National Laboratories and the U.S. Department of Interior in 2003. The committee's review was published by the National Research Council and the National Academy of Sciences in 2004. It recommended that a more critical focus be taken, examining the steps needed to reach desired long-term objectives and the environmental, economic, and social costs.
Balaban is involved with the desalination program at the Center for Clean Water and Energy in the Department of Mechanical Engineering at Massachusetts Institute of Technology (MIT).
She is on the Scientific Advisory Council of the Sharing Knowledge Foundation.
Awards
2015, named #7 of Top 25 Water Industry Leaders by Water & Wastewater International
2014, Sidney Loeb Award, European Desalination Society
2012, Order of the Star of Italian Solidarity (Stella della solidarietà italiana) from the President of Italy
2009, Lifetime Achievement award, International Desalination Association, Dubai
Honorary member of the European Membrane Society (EMS)
References
21st-century American women scientists
21st-century American scientists
American editors
American women editors
Jewish American scientists
Living people
Year of birth missing (living people)
University of Pennsylvania School of Arts and Sciences alumni
Water desalination
21st-century American women writers
American science communicators | Miriam Balaban | [
"Chemistry"
] | 1,550 | [
"Water treatment",
"Water technology",
"Water desalination"
] |
55,846,595 | https://en.wikipedia.org/wiki/Shadow%20board | A shadow board is a type of tool board for organizing a set of tools; the board defines where particular tools should be placed when they are not in use. Shadow boards have the outlines of a work station's tools marked on them, allowing operators to identify quickly which tools are in use or missing. The boards are commonly located near the work station where the tools are used. Shadow boards are often used in the manufacturing environment to improve a facility's lean six sigma capabilities.
Shadow boards reduce time spent looking for tools and also reduce losses. They improve work station safety because tools are replaced safely after use, rather than becoming potential hazards.
See also
Knolling
5S (methodology)
Peg board
References
Tools
Industrial equipment
Containers
Ordering
Lean manufacturing | Shadow board | [
"Engineering"
] | 150 | [
"Lean manufacturing",
"nan"
] |
55,847,287 | https://en.wikipedia.org/wiki/Benita%20Mehra | Benita Mehra FIET, FWES, FIHEEM is a British engineer based in London. She was President of the Women's Engineering Society from 2015 to 2018.
Education and career
Mehra studied Electrical & Electronic Engineering at the City University London before completing an MSc in Construction Management at Heriot-Watt University. Upon graduation she joined British Airports Authority (BAA) the airports operator and with their support attained chartered status, having been seconded from BAA into a manufacturing company.
In 2005, Mehra achieved an MBA in Henley Business School, allowing her to manage large groups of engineers.
In 2015 she became a Fellow of the Institute of Healthcare Engineering and Estate Management, and the Institution of Engineering and Technology in 2016.
Grenfell Tower Inquiry
The British prime minister Boris Johnson appointed Mehra to the Grenfell Tower inquiry, having undertaken a (non-transparent) due diligence of Benita Mehra.
Given that Benita Mehra had run an organisation (the Women's Engineering Society) that had received a £71,000 grant from the Arconic Foundation - the philanthropic arm of the very manufacturer who created the cladding that played such a pivotal role in the Grenfell fire - her appointment and willingness to take up the position sparked controversy with many in the Grenfell community while Mehta insisted on there being no conflict of interest. Unite the Union assistant general secretary for legal services Howard Beckett, said "“Benita Mehra has a clear conflict of interest and she should play no role in the inquiry". He further said,“Benita Mehra’s claim that there is no conflict of interest is simply not plausible". Only after sustained outcry by the Grenfell community, a handful of major media outlets, and Grenfell United, did Benita Mehra resign from the Grenfell Tower inquiry.
Prime Minister accepts resignation of Grenfell Tower Inquiry panel member. The Prime Minister said: "I can confirm that Benita Mehra wrote to me yesterday to offer her resignation from the Grenfell Tower Inquiry panel and I have accepted. I would like to thank Benita for her commitment and I am very grateful for her sensitivity to the work of the Inquiry."
In her letter she had hoped to draw on her experience and knowledge of the construction industry, of community engagement and of governance within housing management contribute to the vital work of the Inquiry in discovering how and why the devastating fire at Grenfell Tower happened.
However, it is apparent that her former role as President of the Women's Engineering Society (WES), which in 2017 accepted a charitable donation from the Arconic Charitable Foundation to support the mentoring of women engineers and apprentices, has caused serious concerns to a number of the bereaved, survivors and resident Core Participants. Within her letter to the Prime Minister she made it quite clear she had not had any contact with either the Arconic Charitable Foundation or Arconic and her involvement in the donation was to review the initial proposal.
As President of WES, her role was unpaid and the grant from the Arconic Charitable Foundation was ring-fenced to fund the mentoring scheme alone, and for these reasons, she did not link any aspect of her former role as President to the panel member for the Grenfell Tower Inquiry.
She did not feel to cause any further distress to those who had already suffered such unimaginable losses and trauma, nor to create any distractions from the vital work of the public inquiry and felt no option but to tender her resignation as a panel member of the Independent Inquiry into Grenfell Tower.
Many in the Grenfell community, as well as MP's including Labour Deputy Leader and MP for Hackney North and Stoke Newington, Diane Abbott, saw her appointment and her acceptance of the position as unacceptable.
Diversity
In 2016, Mehra became the President of the Women's Engineering Society. Under her guidance, National Women in Engineering Day (NWED) became an international celebration. She has campaigned for women to be more readily welcomed back into the engineering workforce after time taken out for maternity leave, and encouraging small-to-medium enterprises to explore job-sharing for mid career workers.
Mehra works to inspire the next generation, working with teachers and girls to highlight careers in engineering. She was a judge of the annual Top 50 Women in Engineering campaign run by the Women's Engineering Society in partnership with The Telegraph newspaper. She received an Honorary Doctorate for Science for her diversity work from Chichester University in 2017. and as a diversity champion was awarded an honorary degree.
References
Year of birth missing (living people)
Living people
Alumni of City, University of London
Alumni of Henley Management College
Alumni of Heriot-Watt University
Alumni of the University of Reading
21st-century British women engineers
Fellows of the Institution of Engineering and Technology
Presidents of the Women's Engineering Society
Fellows of the Women's Engineering Society
21st-century Scottish women engineers
21st-century Scottish engineers | Benita Mehra | [
"Engineering"
] | 1,002 | [
"Institution of Engineering and Technology",
"Fellows of the Institution of Engineering and Technology"
] |
55,848,255 | https://en.wikipedia.org/wiki/Sticky%20rice%20mortar | Sticky rice mortar was invented in ancient China utilizing organic materials in inorganic mortar. Hydraulic mortar was not available in ancient China, possibly due to a lack of volcanic ash.
Around 500 CE, sticky rice soup was mixed with slaked lime to make an inorganic−organic composite mortar that had more strength and water resistance than lime mortar. Sticky rice played a major role in maintaining the durability of the Great Wall, as well as tombs, pagodas, and city walls. Sticky rice mortar had high adhesive strength, sturdiness, waterproofing capability, and prevented weeds from growing as crude mortar made of sticky rice and burnt lime created a seal between bricks that would rival modern cement in strength.
During the Ming dynasty (1368–1644 AD), brick-making techniques improved significantly in terms of quantity and quality of production. Since then, Great Wall sections were widely built with bricks, with lime mortar and sticky rice used to reinforce the bricks strongly enough to resist earthquakes and modern bulldozers while keeping the building intact.
Modern chemists, through their research, identified amylopectin, a type of polysaccharide, or complex carbohydrate, found in rice and other starchy foods to appear to be responsible for the sticky rice mortar's strength and durability.
References
Bricks
Building materials
Cement
Chinese inventions
Concrete
History of the Great Wall of China
Rice
Traditional East Asian architecture | Sticky rice mortar | [
"Physics",
"Engineering"
] | 289 | [
"Structural engineering",
"Building engineering",
"Materials stubs",
"Construction",
"Materials",
"Building materials",
"Concrete",
"Matter",
"Architecture"
] |
55,848,604 | https://en.wikipedia.org/wiki/Moment%20distance%20index | The moment distance index (MDI) is a shape-based metric or shape index that can be used to analyze spectral reflectance curves and waveform LiDAR, proposed by Dr. Eric Ariel L. Salas and Dr. Geoffrey M. Henebry (Salas and Henebry, 2014). In the case of spectral data, the shape of the reflectance curve should unmask fine points of the spectra usually not considered by existing band-specific indices. It has been used to identify spectral regions for chlorophyll and carotenoids, detect greenhouses using WorldView-2, Landsat, and Sentinel-2 satellite data, identify greenhouse crops, compute canopy heights, estimate green vegetation fraction, and optimize Fourier-transform infrared (FTIR) scans for soil spectroscopy.
Various approaches have been devised to analyze medium and fine spectral resolution data and maximize their use to extract specific information for vegetation biophysical and biochemical properties. Combinations of spectral bands, called indices, have been used to diminish the effects of soil background and/or atmospheric conditions while highlighting specific spectral features associated with plant or canopy properties. Vegetation indices (VIs) use the concept of band ratio and differences or weighted linear combinations to take advantage of the visible and NIR bands, two important spectral bands for vegetation studies, in measuring the photosynthetic activity of the plant and explore vegetation dynamics. There is an extensive list of such indices, including the normalized difference vegetation index (NDVI), ratio-based indices such as the modified simplerRatio, soil-distance-based indices such as the modified soil adjusted vegetation index (MSAVI), and many others. Whereas most indices incorporate two-band or three-band relations – slope-based, distance-based on soil line or optimized (slope-based and distance-based concepts combined) – no approach deals with the raw shape of the spectral curve. MDI, however, investigates the shape of the reflectance curve using multiple spectral bands not considered by other indices, which could carry additional spectral information useful for vegetation monitoring.
A full-waveform Light Detection And Ranging (LiDAR) system has the ability to record many returns per emitted pulse, as a function of time, to reveal the vertical structure of the illuminated object, showing position of the individual targets, and finer details of the signature of intercepted surfaces or the proportion of the canopy complexity. Information associated with the illuminated object can be decoded from the generated backscattered waveform, as key features of the waveform such as the shape, area, and power are directly related to the geometry of the illuminated object. The richness of the LiDAR waveform holds a promise to address the challenge of characterizing in detail the geometric and reflection characteristics of vegetation structure, e.g., the vertical canopy volume distribution. MDI utilizes the raw waveform and place importance on its shape and its return power. MDI departs from the usual Gaussian modeling in detecting peaks (canopy and ground) for example in canopy height estimation and focus more on the full geometry (raw shape) and radiometry (raw power) of the LiDAR waveform to retain richness of the data.
The moment distance is a matrix of distances computed from two reference locations (pivots) to each spectral or waveform point within the specified range.
Assume that a curve (reflectance or absorption curve or backscattered waveform) is displayed in Cartesian coordinates with the abscissa displaying the wavelength λ or time lapse t and the ordinate displaying the reflectance ρ or the backscattered power p. Let the subscript LP denote the left pivot (located in a shorter wavelength for the spectral curve and earlier temporal reference point for the waveform) and subscript RP denote the right pivot (located in a longer wavelength for the spectral curve and later temporal reference point for the waveform). Let λLP and λRP be the wavelength locations observed at the left and right pivots for a reflectance data, respectively, where left (right) indicates a shorter (longer) wavelength. Let tLP and tRP be the time value observed at the left and right pivots for a waveform data, respectively, where left (right) indicates an earlier (later) time. The proposed MD approach can be described in a set of equations.
For spectral data, the index is given as:
where is the moment distance from the right pivot, is the moment distance from the left pivot, is the wavelength location at left pivot, is the wavelength location at right pivot, is the spectral reflectance at a given wavelength, and is successive wavelength location.
For waveform LiDAR data, the index is given as:
where the moment distance from the left pivot (MDLP) is the sum of the hypotenuses constructed from the left pivot to the power at successively later times (index from tLP to tRP): one base of each triangle is difference from the left pivot ( − tLP) along the abscissa and the other base is simply the backscattered power at . Similarly, the moment distance from the right pivot (MDRP) is the sum of the hypotenuses constructed from the right pivot to the power at successively earlier times (index from tRP to tLP): one base of each triangle is the difference from the right pivot (tRP − ) along the abscissa and the other base is simply the backscattered power at .
MDI is an unbounded metric. It increases or decreases as a nontrivial function of the number of spectral bands or bins considered and the shape of the spectrum or waveform that spans those contiguous bands or bins. The number of bands or bins is a function of the spectral resolution of the imaging spectrometer or the temporal resolution of the LiDAR (digitization rate) and the length of the reference range (i.e., full extent or subsets of the curve) being analyzed.
References
Remote sensing
Biogeography | Moment distance index | [
"Biology"
] | 1,247 | [
"Biogeography"
] |
55,849,011 | https://en.wikipedia.org/wiki/NGC%204519 | NGC 4519 is a barred spiral galaxy located about 70 million light-years away in the constellation Virgo. NGC 4519 was discovered by astronomer William Herschel on April 15, 1784. It has a companion galaxy known as PGC 41706 and is a member of the Virgo Cluster.
Physical characteristics
NGC 4519 has an asymmetric structure that contains a well-defined bar.
See also
List of NGC objects (4001–5000)
NGC 4498
References
External links
Virgo (constellation)
Barred spiral galaxies
4519
41719
7709
Astronomical objects discovered in 1784
Virgo Cluster
Discoveries by William Herschel | NGC 4519 | [
"Astronomy"
] | 127 | [
"Virgo (constellation)",
"Constellations"
] |
55,850,415 | https://en.wikipedia.org/wiki/Indigo%20Era | The Indigo Era (or Indigo economies) is a concept publicized by businessman Mikhail Fridman, describing what he views as an emerging new era of economies and economics based on ideas, innovation, and creativity, replacing those based on the possession of natural resources. Fridman is the co-founder of LetterOne, an international investment business, and first publicized the idea in early 2016. The word "indigo" was initially chosen based on the term indigo children, which has been used to describe people with unusual and innovative abilities.
Fridman describes the Indigo Era as a disruptive era driven by extraordinary levels of human creativity, where abnormally talented individuals and entities are able to realize new levels of human potential and economic achievement. It is "a new economic era where the main source of national wealth is no longer resource rent but the socio-economic infrastructure that allows every person to realise his or her intellectual or creative potential." But, according to Fridman – based on his observations of recent economic indicators, political and market volatility, and historical patterns – it is also an era that will generate winners and losers as lagging countries and groups fail to adapt quickly enough.
In late 2016 LetterOne's Global Perspectives journal published an Indigo Index, ranking 152 countries on their ability to compete and grow as economies move away from being powered by natural resources to being powered by ideas, creativity, and digital skills. In 2017 it launched the Indigo Prize, to award new concepts of economic measurement beyond mere GDP as countries in the 21st century transition into economies where innovation, creativity, and digital skills are economic drivers. The competition is intended to "stimulate debate about factors currently measured, given evolving economies, technology and skill bases, and what should now be taken into consideration in official economic statistics that measure the health, size and growth of a modern economy."
Origin
In an April 2016 article in RealClearPolitics, retitled and reprinted in May 2016 in the Jerusalem Post, and reprinted in November 2016 under the original title in LetterOne's Global Perspectives journal, Fridman wrote:
LetterOne's Global Perspectives website adds that the indigo symbolism "embodies a breaking of the norm, something that is highly reflective of the new era that we are entering into, one that lacks convention and is driven by innovation."
Analysis
In a series of articles published in 2016, Fridman cites the recent extreme volatility in markets, and worldwide political change and instability, as signs of an emerging global shift. He notes two frequently cited prominent indicators of an economic shift: the sharp decline in prices of natural resources including oil, and the slowing of China's economic growth despite this decline in the cost of natural resources.
He and other commentators also note the rise of populism and populist leaders and candidates, both right-wing and left-wing, as these changes occur. Meanwhile, companies like Apple and Google – digital and technological companies he calls "Indigo companies" – have replaced longterm traditional natural-resource or manufacturing companies such as Exxon as the world's largest companies.
Fridman observes that throughout history innovations, alternatives, and technologies have always overcome any perceived shortages of any natural resource. He therefore posits that the new "Indigo era", fueled by digital and technological resources, will be marked by a shift away from the struggle for natural resources and their perceived scarcity, to a reliance on ideas, innovation, and creativity and on supporting the intellectual and creative potential of each human being: "The world has entered a new era where the source of a nation's wealth is no longer natural resources. Intellectual capacity has now replaced land, raw materials and trade routes as the biggest source of wealth."
According to Fridman, three interconnected factors are needed for successful Indigo companies and an Indigo economy:
intuitive individual talent and a high level of education, plus the ability to form a team of equally educated and gifted people
a sophisticated and complex ecosystem, which includes not only legal systems that protect physical and intellectual property rights and protection from takeover by larger companies, but also thousands of suppliers and subcontractors to supply high-quality services which range from venture capital to marketing to web design and other technologies and services
a global digital infrastructure to distribute new products and accumulate customer data and customer-behavior insight
He notes that most emerging economies have focused on building physical structures (roads, buildings, cities, physical infrastructure) rather than the complex legal, political, and social systems, institutions, and changes that will support an effective free and innovative intellectual-resource economy. The freedoms, protections, and political and legal frameworks of developed Western countries rest on centuries-long histories, socio-political traditions, and mindsets, and therefore will be difficult to replicate quickly in emerging economies. Fridman singles out India as an emerging country that has adequate legal infrastructure and freedom to probably survive the Indigo shift.
Fridman considers the growth of Indigo economies to be a paradigm shift; he states that the pace at which technology is developing is creating worldwide tectonic shifts, and predicts huge global change over the next five to ten years. He and other analysts predict that the growing economic gap between free, creative economies and groups in contrast to repressive, authoritarian, totalitarian, or tradition-bound economies or groups will widen and create resentment and hostilities – whether this is between nations or within nations. Those left behind may be either emerging countries, or the average person – as opposed to the intellectual elites – within developed Western countries.
Authoritarian leaders and authoritarianism often rise during periods of uncertainty and insecurity and economic deprivation. Fridman maintains however that in this ever-changing new economic era, the main source of wealth in a country or region will no longer be a natural resource, but a social infrastructure that will allow everyone to realise their intellectual and creative potential. Therefore, he asserts that "The future Indigo economy is an economy of free people. And this means that the world will become more and more free."
In November 2016 LetterOne launched a journal, Global Perspectives, as a platform to explore "the new emerging economic era, the Indigo era, from different perspectives, including education, religion, politics, economics, history and business" and to examine "global issues through the eyes of leading commentators and business people around the world". The inaugural issue contained articles by Fridman, Dominic Barton, Michael Bloomberg, Stan Greenberg, Carl Bildt, Vince Cable, Ken Robinson, Brent Hoberman, Alex Klein, Deirdre McCloskey, Yuri Milner, Nick D'Aloisio, Lynda Gratton, Parag Khanna, Ian Goldin, George Freeman, Ian Bremmer, and others.
Indigo Index and Indigo Score
The November 2016 inaugural issue of Global Perspectives also published an Indigo Index, which rated 152 countries based on five key metrics for doing business as economies move away from being powered by natural resources to being powered by creativity and digital skills. The five metrics are: creativity and innovation, economic diversity, digital economy, freedom, and stability and legal frameworks, which were scored based on over 30 measures from published data sources such as the World Bank, UNESCO, the CIRI Human Rights Data Project, the Center for International Development at Harvard University, and the Global Education Monitoring Report. The index sought to measure a country's entrepreneurial ecosystem, and therefore its potential to adapt and develop.
Each country was given a combined overall Indigo Score, with 200 being the highest possible score. The 10 top-ranked countries were Sweden, Switzerland, Finland, Denmark, the UK, the Netherlands, Norway, Germany, Ireland, and Japan. The United States was 18th overall.
The report also included three key findings: Creativity and innovation was the biggest overall driver of high scores; this accentuated the importance of fostering entrepreneurialism and lifelong learning and of investing heavily in people. Nordic countries scored particularly high on the Indigo Index, with three Nordic countries in the top four and four Nordic countries in the top ten; this was attributed to their high rankings both in creativity and innovation and in freedom. And the lowest-scoring countries were beset with social and political problems, such as war, political turmoil, and corruption.
Indigo Prize
In July 2017, LetterOne's Global Perspectives journal announced the Indigo Prize, to stimulate discussion towards finding a new way of measuring the economy in the 21st century that moves beyond the limitations of mere GDP measurements.
Entrants were asked to submit an essay of up to 5,000 words answering the question:
Entries were due 15 September 2017, and were open worldwide to groups or individuals over the age of 16, with entries particularly encouraged from people at academic institutions, businesses, charities, think tanks, consultancies, or other organisations. The award amount was announced as £100,000, with second- and third-place winners to receive £25,000 and £10,000.
The judging panel included:
Dominic Barton, global managing partner at McKinsey & Company
Mervyn Davies, non-executive chairman of LetterOne, former Minister of State for Trade, Investment and Small Business, former chairman and CEO of Standard Chartered
Stephanie Flanders, chief market strategist at JP Morgan Asset Management and former BBC economics editor
Jim O'Neill, former Commercial Secretary to the Treasury, former Goldman Sachs chief economist
Lynda Gratton, author and professor of Management Practice at the London Business School
Brent Hoberman, founder of Lastminute.com and Founders Forum
Gus O'Donnell, chairman of Frontier Economics, former Cabinet Secretary
Ed Vaizey, Member of Parliament and former Minister for Culture, Communications and Creative Industries
Mikhail Fridman, co-founder of LetterOne
The winners of the inaugural Indigo Prize were announced on 25 October 2017. A joint first prize, £125,000 to be equally split, was awarded to two teams of writers: Diane Coyle and Benjamin Mitra-Kahn; and Jonathan Haskel, Carol Corrado, et al. A third place "Rising star" award of £10,000 was given to Alice Lassman.
Coyle was professor of economics at the University of Manchester, and Mitra-Kahn is chief economist at IP Australia. They proposed radically replacing GDP with a dashboard measuring six key assets: physical assets, natural capital, human capital, intellectual property, social and institutional capital, and net financial capital. Their essay stated that "GDP never pretended to be a measure of economic welfare", and proposed that the new measure should assess "the range of assets needed to maximise individuals' capabilities to lead the life they would like to lead"; this would include "financial and physical capital but also natural and intangible capital". They asserted that the new statistics should focus on measuring changes in the stock of important assets, rather than flows of income, expenditure, and output. Tracking the evolution of stocks of physical assets, financial assets and liabilities, natural capital, skill levels, and implicit state liabilities would better measure the sustainability of the economy. Coyle and Mitra-Kahn also proposed interim improvements to GDP measurements – such as better measurement of intangibles, adjusting for the distribution of income, and removing unproductive financial activity – before scrapping it entirely. Following her prize-winning essay, Coyle now leads the Six Capitals research project, funded by LetterOne, at the Bennett Institute for Public Policy at Cambridge University; the project was inaugurated in January 2019 and explores social and natural capital.
Haskel is professor of economics at Imperial College Business School. Rather than abandoning GDP, he proposed refining, updating, and extending the existing GDP measure. He proposed better measurement of services and intangibles, and direct measurement of the economic welfare being created by digital goods. His essay focused on the fact that economies have dramatically changed structure since GDP was originally developed, with more knowledge production, more digital goods, more free things and free information, and more intangible assets such as intellectual property. He also emphasized the importance of factoring in the environment, sustainability, and societal welfare, in addition to calculating the value of goods and services that are provided for free.
Lassman, the recipient of the "Rising star" award, was a 19-year-old geography student at Durham University. Her entry proposed a "Global Integration and Individual Potential" index, which measures each nation on two levels: its value relative to other nations, and the individuals and their contributions within each nation.
References
External links
The Indigo Era – Official website at Global Perspectives
Economic theories
International rankings
International macroeconomics
Economic forecasting
Economic indicators
Global economic indicators
Macroeconomic indicators
Comparative economic systems
Knowledge economy
Historical eras
2010s neologisms
Digital Revolution
Cultural trends
Information Age
21st century | Indigo Era | [
"Technology"
] | 2,574 | [
"Information Age",
"Computing and society",
"History of computing",
"Digital Revolution"
] |
62,787,541 | https://en.wikipedia.org/wiki/Marriage%20bed%20of%20Henry%20VII | The marriage bed of Henry VII (also known as the Paradise Bed, Bed of Roses) is a carved oak four-post bedstead bought in a dilapidated condition at an auction in Chester, England, in 2010. Since then the bed has been subjected to art historical investigation and advanced material analysis. It is suggested that the bed was made to celebrate the marriage of King Henry VII to Elizabeth of York on 18 January 1486.
Rediscovery
For over 15 years the bed was the prized feature of the honeymoon suite of the Redland House Hotel . It had been purchased from an Edinburgh-based antique dealer in the 1990s. In 2010 the Redland House Hotel underwent redevelopment and the bed was at risk of being thrown away but it was spotted by an antique dealer who suggested selling it at auction. The bed was dismantled and for a while lay in the hotel car park before being collected by the auctioneers. It was described in the auction catalogue as "profusely carved Victorian four poster bed with armorial shields". Ian Coulson, an antique bed restorer from Humshaugh, bought the bed at the on-line auction for £2,200.
Structural analysis
When Coulson finally got to examine the bed when it was delivered to his studio a few weeks after the sale he was immediately struck by the age of the wood, noting the degree of oxidisation, shrinkage, the damage caused by extensive woodworm infestation, that the carvings had been made by hand rather than by machine – all of which suggested to him that this was a bed of a considerable age. He began to suspect the bed was not a Victorian piece. A more detailed examination, which involved the opening of joints and the identification of later repairs, revealed evidence of medieval construction techniques, markings - clues to the original appearance and configuration of the bed. He noted that the main crest of the bed was missing. Coulson shared his findings with a range of other experts who agreed with his suggestion that this was a Tudor era bed. The distinctive Adam and Eve headboard and the carved detailing of the bedposts suggested that the bed was possibly the model for the forged beds and other pieces of furniture known to have been produced by George Shaw during the 1840s. Coulson has commissioned various forms of scientific analysis to determine the date and provenance of the bed.
DNA (deoxyribonucleic acid) analysis
The analysis of DNA extracted from several separate parts of the bed clearly demonstrated that the wood was a species of oak native to Continental Europe (probably the Baltic Region). DNA tests of the wood proved that it all came from a single continental European oak tree, of a type known to have been imported by the English royal family for their beds.
Dendrochronology (tree-ring dating)
Dendrochronology is a valuable tool for establishing the felling dates of trees, but the science is reliant on the existence of previously dated reference samples – or chronologies. In some cases, matches are simply not found, even for otherwise well documented timbers.
In 2014 a blind-testing dendrochronology investigation was commissioned. This analysis failed to make a match on European databases but did make an important new observation; cockchafer beetles had infested the wood used to make the bed on two separate occasions. This insect attack would have defoliated the infected trees limiting their growth and reducing the thickness of tree-rings during those two years. Dendrochronology protocols decree that trees exhibiting signs of this type of infestation and deformation are excluded from the tree-ring dating database.[i]
Another dendrochronology investigation carried out on the bed did not find a match in the European databases. it failed to note the repeated cockchafer infestation. A match was found with a tree felled sometime after 1756 in north-eastern US but as the DNA analysis had confirmed that it was made from European oak these findings are problematic.
Because of the cockchafer infestation dendrochronology on this bed will not provide a date for the age of its wood.
Paint research
Traces of paint were observed in the crevices of the mouldings suggesting that the bed had once been painted. Microscopic examination of these residues revealed that the bed had originally been painted. The presence of the pigment Coal Black, a pigment which dropped out of use by housepainters in the seventeenth-century, support the dating of the bed to the late medieval period. Traces of an early grained decoration and high quality finely ground yellow and red iron oxide pigments found on the bed provide evidence that the bed was originally painted – and at some expense.Fragments of paint containing large particles of the pigment natural ultramarine blue, a pigment more expensive that gold leaf during the Middle Ages, provided further evidence of the high status of the bed.
The findings of the paint research investigation were presented to leading paint research experts at an ICOM-CC held and published in post prints.
All of the research carried out on the bed was presented at a one-day symposium at the Victoria & Albert Museum in London, January 2019.
The bed itself featured in the "A Bed Of Roses" exhibition at Hever Castle. Foyle, who carried out research on the bed, described it as "one of the most significant artefacts of early Tudor history" and "the most important piece of furniture in England". The bed is held in Coulson's Langley Collection of historic beds and has been valued at £20 million.
Art historical investigation
It has been suggested that the bed was designed and built to celebrate the marriage of King Henry VII to Elizabeth of York on 18 January 1486. The marriage symbolised the end of the War of the Roses by joining Henry's House of Lancaster to Elizabeth's House of York and the bed's design reflected this featuring both the White Rose of York and the Red Rose of Lancaster. A carving on the headboard depicted the Royal couple as either Adam and Eve or Christ and the Virgin Mary defeating the animals that opposed Christ in Psalm 91 and bringing paradise to England (hence the bed's alternative name). The bed also includes the arms of France, reflecting Henry's possessions and ambitions there, as well as religious and fertility symbols.
The bed is thought the bed may have been gifted to Henry's stepfather, Thomas Stanley, in 1495 and remained in use by the Stanley family for a hundred years. It then vanishes from the historic record, surviving the wide-scale destruction of Royal furniture during the English Civil War, before reappearing in 1842 when copies of it were made but its true significance remained unknown. The bed, which is now valued at £20 million, is held in Coulson's Langley Collection and has featured in exhibitions.
Others, however, disagree with the hypothesis outlined above, which has been disseminated mainly by press coverage and has not been submitted to any peer reviewed journal. it is suggested that the bed was known to the 19th century antiquary, architect and faker George Shaw (1810–1876) who used it as the model for his own faked 'Tudor' beds.
Description
The bed measures in height, in width and in length. It is a four-poster bed with each of the posts topped with a carved lion (one of which has lost his tail) each of which holds a shield emblazoned with a rose. The posts support a tester which depicts the tree of life growing from Christ's cross. The frontispiece of the tester shows the arms of England and France quartered. The sides of the tester are formed in the shape of a knotted cord which may reflect Henry's links with the Order of Friars Minor (the Observant Friars). In the centre part of each side holds a shield showing the Cross of Saint George, a particular favourite of Henry's.
The headboard is in the form of a triptych with a central panel showing Henry and Elizabeth flanked by the arms of England and the arms of France, that was originally picked out in expensive ultramarine paint. The carving has been interpreted as an allegory of Adam and Eve and also of Christ and the Virgin Mary working to undo the sins of Adam and Eve and working to turn England into a paradise on Earth. The Royal couple are depicted overcoming the young lion, dragon and snake defeated by Christ in Psalm 91. Due to the connection to the story of Adam and Eve the bed is sometimes referred to as the "Paradise Bed". The inclusion of the arms of France reflect Henry's possessions on the continent and his future ambitions there.
The arms are repeated on the footboard. Other carvings include fertility symbols such as acorns, grapes and strawberries and symbols of the Tudor royalty such as stars, shields, lions and roses. These are similar to carvings in the Henry VII Chapel at Westminster Abbey, dating to 1503–1516. The rose carvings on the bed are a mixture of the Red Rose of Lancaster and the White Rose of York which dates it to the early period of Henry's monarchy, before the adoption of the combined Tudor rose.
History
The bed is believed to have been made for the wedding of Henry VII and Elizabeth of York, which took place on 18 January 1486. The marriage symbolised the end of the War of the Roses and joined Elizabeth's House of York with Henry's House of Lancaster, founding the House of Tudor. The marriage bed would have been used on the couple's marriage night and it is believed that their first son Arthur, Prince of Wales (born 19/20 September 1486) was conceived on this occasion. The couple married at Westminster Abbey and the bed stood in the Painted Chamber of the nearby Palace of Westminster, as confirmed by the bed's outline matching the shape of a mural known to have been painted there. The bed may have been influenced by Burgundian designs but archaeologist Jonathan Foyle believes it to have been made in Germany and that Henry himself may have helped to design the bed. The bed is decorated with symbology to support Henry's claim to the throne and the connection between the king and the state as well as religious and fertility symbols.
The bed may have accompanied Henry on his royal progresses about the kingdom. It is possible that the future Henry VIII and Henry and Elizabeth's six other children were conceived in the bed. The bed is believed to have been gifted to Thomas Stanley, Henry's stepfather in 1495, but this could have been another bed commissioned by Henry. It is possible that a similar bed was also commissioned by Henry for his hunting lodge at Knowsley Hall. The bed is thought to have remained in use at Stanley's seat at Lathom House for a century before vanishing from the historical record. During the English Civil War (1642–1651) many Royalist possessions were destroyed by Parliamentarian forces, including almost all Royal beds. Henry VII's bed is the only complete bed frame to survive this destruction, the only other Royal bed artefact being a fragment of the headboard of Henry VIII and Anne of Cleves. The bed reappeared in 1842, when it was apparently found by a George Shaw who seems to have sold copies of it without knowing its true significance. Shaw is believed to have kept the front crest of the bed but the remainder ended up at the Redland House Hotel in Chester.
References
External links
The Langley Collection
Beds
Individual pieces of furniture
English furniture
Early oak furniture
Henry VII of England
1486 works
Elizabeth of York
Culture of beds | Marriage bed of Henry VII | [
"Biology"
] | 2,350 | [
"Beds",
"Behavior",
"Sleep"
] |
62,787,574 | https://en.wikipedia.org/wiki/Kleptoprotein | A kleptoprotein is a protein which is not encoded in the genome of the organism which uses it, but instead is obtained through diet from a prey organism. Importantly, a kleptoprotein must maintain its function and be mostly or entirely undigested, drawing a distinction from proteins that are digested for nutrition, which become destroyed and non-functional in the process.
This phenomenon was first reported in the bioluminescent fish Parapriacanthus, which has specialized light organs adapted towards counter-illumination, but obtains the luciferase enzyme within these organs from bioluminescent ostracods, including Cypridina noctiluca or Vargula hilgendorfii.
See also
Kleptoplasty
References
Biological interactions
Proteins | Kleptoprotein | [
"Chemistry",
"Biology"
] | 162 | [
"Biomolecules by chemical classification",
"Behavior",
"Ethology",
"Biological interactions",
"nan",
"Molecular biology",
"Proteins"
] |
62,787,784 | https://en.wikipedia.org/wiki/3P%2BS | The 3P+S Input/Output Module was an S-100 expansion card introduced to the microcomputer market by Processor Technology. It supplied three parallel ports and one serial port, the latter of which conformed to the RS-232C standard. One of the three parallel ports was dedicated to interfacing with the host computer over the S-100 bus, while the other two were available for general use.
An Altair 8800 equipped with a 3P+S could use one of the parallel ports to accept input from a keyboard and another to output to a TV Typewriter, allowing the user to construct an all-in-one machine that did not need an external computer terminal to work. This also left the serial port free, which could be used to drive a teletype machine as a computer printer, or a punch tape system for storage.
Processor Technology later combined the 3P+S with the VDM-1 graphics card in a compact S-100 machine of their own to produce the Sol-20, the first all-in-one mass-produced personal computer.
References
S-100 machines | 3P+S | [
"Technology"
] | 229 | [
"Computing stubs",
"Computer hardware stubs"
] |
62,788,115 | https://en.wikipedia.org/wiki/Comet%20HLLV | The Comet HLLV was a proposed super heavy-lift launch vehicle designed for NASA's First Lunar Outpost program, which was in the design phase from 1992 to 1993 under the Space Exploration Initiative. It was a Saturn V-derived launch vehicle with modernized engines, stretched fuel tanks, and strap-on boosters. Its main goal was to support the First Lunar Outpost program and future human mission to Mars. It was designed to be inexpensive and simple while relying on existing technology to lower development costs.
Design
The Comet would have been capable of putting 254.4 tons into low Earth orbit and 97.6 tons to trans-lunar injection, roughly twice that of the Saturn V, making it one of the largest rockets ever designed in terms of payload. The vehicle resembled a Saturn V, but with stretched first and second stages, an increased-diameter third stage, and new side boosters. Additionally, the engines were updated to the F-1A and J-2S, and a sixth engine was added to the second stage. Each of the two side boosters had two F-1A engines. Development costs were expected to be modest due to reliance on Apollo-era technology.
A nuclear-powered variant of the third stage, with two 222.5-kN engines, was also considered. It would have reduced the rocket's size, but at a predicted development cost of $2 billion over a chemical-only design. The nuclear option was planned to be developed later to support crewed Mars missions. To this end, NASA's Lewis Research Center established a Nuclear Systems Office to develop and test a fully functional nuclear engine by 2005.
NLS derived launch vehicle
An alternate version of the launcher based on the then-in-development National Launch System was proposed. NASA's Marshall Spaceflight Center looked into the Comet rocket or a possible configuration with four F-1A boosters added to the basic 2-stage NLS vehicle. The main expected advantage was that the vehicle could rely on technology currently flying rather than having to resurrect 20 year old technology and manufacturing equipment.
References
NASA
Space launch vehicles of the United States
Spaceflight
Proposed space launch vehicles | Comet HLLV | [
"Astronomy"
] | 435 | [
"Spaceflight",
"Outer space"
] |
62,789,643 | https://en.wikipedia.org/wiki/Price-Whelan%201 | Price-Whelan 1 (PW 1) is a young stellar association or disrupting star cluster with low metallicity and extragalactic origin, more specifically the leading arm of the Magellanic gas stream originating in the Magellanic Clouds. Price-Whelan 1 was discovered by Adrian Price-Whelan using Gaia data and additional cluster members were identified using DECam data. The star cluster contains less than a thousand stars. The existence of Price-Whelan 1 suggests that the stream of gas extending from the Magellanic Clouds to our Milky Way is about half as far from the Milky Way as previously thought.
Structure
The star cluster has larger component 'a' and a smaller component 'b'. The component 'a' was later resolved in two components: an Eastern component 'aE' and a Western component 'aW'. The three components do not only differ in position, but also in stellar content.
The parent gas cloud of PW 1
Price-Whelan 1 is about ten degrees offset from the leading arm II. This difference is explained with the gas experiencing ram pressure as it passes through the hot gas of the Milky Way halo. The stars will not feel this force. Over time the gas and the stars will decouple, resulting in a different position and velocity for both components. Another possible origin of the star cluster could be the high-velocity cloud HVC 287.5+22.5+240, which has a similar metallicity compared with Price-Whelan 1. This cloud is part of the leading arm and displays a strong magnetic field, which could stabilize the cloud against the ram pressure. The cloud also shows traces of molecular hydrogen, which can also be found in star-forming regions.
References
External links
Discovery of a new star cluster: Price-Whelan 1 "Image of the Week" at ESA Gaia
The Milky Way’s Impending Galactic Collision Is Already Birthing New Stars Simons Foundation
Open clusters
Hydra (constellation) | Price-Whelan 1 | [
"Astronomy"
] | 406 | [
"Hydra (constellation)",
"Constellations"
] |
62,790,437 | https://en.wikipedia.org/wiki/A%20Lover%27s%20Oath | A Lover's Oath is a lost 1925 American silent fantasy film directed by Ferdinand P. Earle, jun. and featuring Ramon Novarro. The film is based upon the Rubaiyat of Omar Khayyam, as translated by Edward Fitzgerald, and included quotes of its text on intertitles. Actor Milton Sills was scenarist and editor for the film.
The film was shot in 1920–21 but not released in America until 1925. Actor Edwin Stevens died in 1923 before the film was released.
Published details of this film resemble those of Omar Khayyam, screened in Australia in 1923 to positive reviews.
Plot
As described in a film magazine review, the son of the chief of one desert tribe, who is betrothed to the daughter of the chief of a neighboring tribe, almost loses the young woman to a riotous rich man who attempts to abduct her. Desperate times for the young couple follow, but in the end they are united and made happy.
Cast
Production
For its 1925 release by Astor Pictures, a small distributor, Milton Sills edited the 1922 film to emphasize the role of Novarro, who by then was a rising star. Novarro refused to cooperate with a request for some new closeups, and reportedly some older clips of Novarro were edited into the film. Despite this, the film was not well received by critics or the public.
Preservation
A Lover's Oath is a lost save for a short segment (around thirty seconds) held by the Academy Film Archive; the archive preserved the film in 2009. An additional 135-foot section was apparently discovered and uploaded to YouTube.
References
External links
1925 films
1925 lost films
1920s American films
American black-and-white films
American silent feature films
Cultural depictions of Omar Khayyam
Lost American films
Lost fantasy films | A Lover's Oath | [
"Astronomy"
] | 361 | [
"Cultural depictions of Omar Khayyam",
"Cultural depictions of astronomers"
] |
62,791,988 | https://en.wikipedia.org/wiki/Joanna%20McKittrick | Joanna McKittrick (1954 – November 15, 2019) was an American engineer and college professor, the second woman to join the engineering faculty at the University of California, San Diego (UCSD).
Early life
Joanna McKittrick was born in New Jersey, the daughter of John R. McKittrick and Estella Ruth Pederson McKittrick. Her father was a doctor and her mother was a teacher. She earned a bachelor's degree in mechanical engineering from the University of Colorado, Boulder, and a master's degree from Northwestern University. She completed doctoral studies at the Massachusetts Institute of Technology, in the field of materials science and engineering.
Career
McKittrick was a professor in the engineering program at the University of California, San Diego, from 1988. She was the second woman to join the program (chemical engineer Jan B. Talbot was the first). McKittrick's research focused luminescent materials for medical, automotive, and aviation applications, and biomaterials such as chitin, keratin, and collagen. She worked with fellow UCSD professor Marc A. Meyers on biomaterials. "Mother Nature gives us templates," she explained in 2013, of her work studying spiders, seahorses, boxfish, and porcupines. "We are trying to understand them better so we can implement them in new materials."
McKittrick was a fellow of the American Ceramic Society. She wrote or co-wrote over a hundred academic journal articles, and was an editor of the Journal of the American Ceramic Society. McKittrick was a mentor for minority and women engineering students, including Lauren Rohwer and Olivia Graeve. She was faculty advisor of UCSD's student chapter of the National Society of Black Engineers.
Personal life
McKittrick died in November 2019, aged 65 years, at her home in La Jolla, California. Her colleagues held a memorial at the 8th International Conference on Mechanics of Biomaterials and Tissues a month later.
References
External links
Website of the McKittrick Group, showing her UCSD laboratory's ongoing work on luminescent materials and biological materials.
An interview with McKittrick at the blog Reigning It (January 19, 2016).
1954 births
2019 deaths
American materials scientists
University of Colorado Boulder alumni
Northwestern University alumni
MIT School of Engineering alumni
University of California, San Diego faculty
20th-century American women engineers
Women materials scientists and engineers
20th-century American engineers
21st-century American engineers
21st-century American women engineers
Fellows of the American Ceramic Society | Joanna McKittrick | [
"Materials_science",
"Technology"
] | 521 | [
"Women materials scientists and engineers",
"Materials scientists and engineers",
"Women in science and technology"
] |
62,792,531 | https://en.wikipedia.org/wiki/Oxyarsenides | Oxyarsenides or arsenide oxides are chemical compounds formally containing the group AsO, with one arsenic and one oxygen atom. The arsenic and oxygen are not bound together as in arsenates or arsenites, instead they make a separate presence bound to the cations (metals), and could be considered as a mixed arsenide-oxide compound. So a compound with OmAsn requires cations to balance a negative charge of 2m+3n. The cations will have charges of +2 or +3. The trications are often rare earth elements or actinides. They are in the category of oxypnictide compounds.
Some of these compounds are superconductors, but may require doping with fluoride or oxygen deficiency. Yet others undergo colossal magnetoresistance with a lowered electrical resistance in a magnetic field.
Many compounds are layered, containing two metals with the formula XZAsO, with an XAs layer alternating with a ZO layer.
Examples
Related
Related compounds include the oxynitrides, oxyphosphides, oxyantimonides and oxybismuthides.
References
Arsenides
Oxides | Oxyarsenides | [
"Chemistry"
] | 243 | [
"Oxides",
"Salts"
] |
62,793,183 | https://en.wikipedia.org/wiki/Oxybismuthides | Oxybismuthides or bismuthide oxides are chemical compounds formally containing the group BiO, with one bismuth and one oxygen atom. The bismuth and oxygen are not bound together as in bismuthates, instead they make a separate presence bound to the cations (metals), and could be considered as a mixed bismuthide-oxide compound. So a compound with OmBin requires cations to balance a negative charge of 2m+3n. The cations will have charges of +2 or +3. The trications are often rare earth elements or actinides. They are in the category of oxypnictide compounds.
Many of the bismuthide oxides have bismuth in an unusual -2 oxidation state. The ones with Ln2BiO2 have the anti-ThCr2Si2 structure. They include alternating layers of LnO (anti-fluorite-type) and LnBiO. The Eu4Bi2O has an anti-K2NiF4 structure, the same as for Na2Ti2As2O. Some other compounds contain calcium and a rare earth CaRE3BiO4, and Ca2RE8Bi3O10.
Some of these compounds are superconductors at very low temperatures and many are semiconductors at standard conditions.
Examples
References
Bismuth compounds
Bismuthides
Oxides
Mixed anion compounds | Oxybismuthides | [
"Physics",
"Chemistry"
] | 291 | [
"Matter",
"Mixed anion compounds",
"Oxides",
"Salts",
"Ions"
] |
62,793,986 | https://en.wikipedia.org/wiki/Non-fullerene%20acceptor | Non-fullerene acceptors (NFAs) are types of acceptors used in organic solar cells (OSCs). The name Fullerene comes from another type of acceptor-molecule which was used as the main acceptor material for bulk heterojunction Organic solar cells. Non-fullerene acceptors are thus defined as not being a part of this sort of acceptors.
Research in non-fullerene acceptors did not show promising results starting up when being compared to fullerene based organic solar cells. However, recent developments in this field launched a series of new opportunities for the NFA based OSCs. The most important breakthrough was the development of the small molecule acceptors (SMAs). These acceptors are showing promising results to be better alternatives for Fullerene acceptors because of their properties. The property that makes these SMAs such a big research topic is their tunability. SMAs can be modified to a much greater extent than Fullerene acceptors. There are, however, still many improvements to make on the design of the SMAs in order become profitable to use in OSCs.
Recent research on designing NFA-OSCs showed an efficiency of 15% with a so-called tandem solar cell which made use of Non-fullerene acceptors as well as fullerene acceptors. With a good chance that researchers will be able to boost this percentage up to 18%, it is clear that NFA-OSCs have a great potential in becoming a profitable photovoltaic in commercial application.
NFA Potential
Advantages
Fullerene acceptors (FAs) have been used extensively in OSCs. This is rationalized by several characteristics of fullerenes. The three-dimensional character causes them to be suitable materials for bulk heterojunction structures. Additionally, its electronic configuration (delocalized LUMOs) allows for efficient percolation and high electron mobility. Another consequence is that they are easily coupled to compatible donor polymers.
However, fullerene acceptor organic solar cells (FA-OSCs) encounter a limited efficiency. The energy levels in fullerene compounds are relatively constant and difficult to alter. Moreover, they employ weak absorption in the visible spectrum and the near-infrared spectrum and low thermal instability and photochemical instability. The acceptors need to be purified extensively, adding to the economical and temporal disadvantages of using FAs.
The organic NFAs, in the form of small molecular acceptors (SMAs), can be used to overcome these fullerene deficiencies. They have more structural degrees of freedom, allowing higher electron affinity tunability; they absorb incidental visible-NIR radiation more strongly; they are more stable; they are compatible with donor polymers and they are (in general) easier to synthesize. NF-OSCs with power conversion efficiencies (PCE) of over 13% have been reported, reaching a higher value than its FA-based counterpart.
Disadvantages
One of the downsides of using SMAs is the fact that, under atmospheric conditions, they tend to engage in disordered (anisotropic) states as a result of their planar structures. They are often planar as aromaticity is required for sufficient electron mobility. The lack of order may diminish electron transport and effective extraction routes that lead to induced current. Moreover, the corresponding lack of orientation affects donor-acceptor exciton formation. This makes them less compatible for bulk heterojunction blends than FAs.
Another downside to research on SMA usage is the profound scala of possibilities of donor-acceptor pairs that scientists are challenged to induce.
Physics
The mechanism of current induction in organic solar cells involves a charge transfer. After electromagnetic absorption and exciton formation in the electron donor polymer, the excited electron is moved towards the acceptor conduction band (LUMO) as a result of the lower energy value than the donor LUMO. This process is called a charge separation, and the corresponding energy value satisfies where CS denotes charge separation, A denotes the acceptor and D denotes the donor molecule. Along with the Coulombic potential that needs to be surpassed, the maximum energy obtained from the process is defined as the Charge Transfer energy, . The difference between the optical excitation energy (the optical band gap energy, ) and the charge transfer energy is the driving force of the system.
An advantage of NF-OSCs over current fullerene-based OSCs is that the SMAs used are relatively compatible with donors, as a result of their electronic affinity tunability. Their compatibility originates from their LUMO-energy value similarity. The driving force is minimized to solely Coulombic contributions (<0.3 eV) with negligible charge separation loss. This results in low potential spillage, , which depends explicitly on the value of the driving force, along with radiative and non-radiative losses during the current induction process. Thus, for NF-OSCs, , with q the electron's charge, is minimized, leading to a higher useful energy output. The result is a high open-circuit voltage of the solar cell compared to fullerene counterparts, with reports of values as high as 1.1V.
However, the diminished charge separation energy cost negatively influences the tendency of excited electrons in the donor conduction band to transport to the acceptor LUMO as it is less preferred energetically. This gives rise to the fact that electrons induced in the current are more energetic, but fewer electrons are induced. This means that the short-circuit current density and the fill factor (FF) are decreased.
In terms of the PCE, the higher open-circuit voltage is compensated by the lower short-circuit current density and fill factor. Researchers showed that ultrafast charge separation is possible with negligible driving force. In fact, the electrical external quantum efficiency is highest for donor-acceptor blends with lowest driving force.
Types
One of the main advantages of the non-fullerene acceptors is their ability to be tuned and customized by chemical modification. This in contrary to fullerene acceptors. It also immediately creates a bottleneck because of the huge amount of possibilities there are which could be applied as an SMA. A wide variety of SMAs are tested to be a successful acceptor, but two classes of SMAs have proven to give the best results concerning Power Conversion Efficiency (PCE) and have made the greatest attribute to the recent development in NFA-OSCs.
Rylene diimides
Rylene diimides are, as said, one of the two main subclasses which are a basis for acceptor-molecules in modern NFA-OSCs. Rylene diimides are industrial dyes and can be divided into, once again, two subclasses: Perylene Diimides (PDIs) and Naphthalene Diimides (NDIs). Rylene diimides consist of a planar rylene framework and numerous constructions can be made by attaching certain subgroups and by using more PDI molecules in one acceptor. The mono-PDI molecule is shown in the figure on the right.
Rylene diimides are considered good acceptors because of their favourable properties. Rylene diimides usually have high electron mobility values due to intermolecular π-stacking. These values are comparable to ones of fullerene acceptors. Furthermore, Rylene diimides also have a high absorbance spectrum in the visible area, high thermal and oxidative stability and their electric affinities can be tuned to a great extent by adding side groups and 3D-structure which leads to a significant higher open-circuit voltage ()
Challenges that must be faced by designing and improving Rylene diimides based OSCs are mainly concerned by synthesis of PDIs because the planar structure of the molecule makes that it tends to aggregate into a crystal structure. This greatly enhances the domain size, larger than the preferred 20 nm, in the bulk heterojunction which leads to a lower charge transport ability. Researchers have tried to reduce this aggregation by three structure adaptions, all focused on enhancing the mobility of Rylene diimides molecules. The first approach is to link two PDI molecules with a single carbon bond, to form a so-called twisted dimer. The second synthesis forms highly twisted 3D-structures of PDI molecules and the third approach forms a fused-ring structure. For all three possible ways, an example molecule is shown in the figure below. These derivatives are examples of acceptor-molecules which were tested and assessed in OSCs for their performance and PCE. Future research will focus on developing better PDIs resulting in higher PCE values for the OSC.
Fused-ring electron acceptors
Fused-ring electron acceptors (FREAs) are completely different from Rylene diimides. They consist of two electron withdrawing groups in between of a donor group. This donor group is a π-bridge of fused aromatic rings. FREAs have values for similar to those of fullerene acceptors and have a wide absorption range. Electron affinities can be tuned by substituting the side chains, the core and the end groups. Current research focusses on designing the best FREA with varying all these groups. Another development issue is the expensive synthesis of these molecules. Finding the most efficient synthetic route is therefore also an important subject concerning these acceptors
Future development
In current research, rylene diimides (for small band-gap energy donors) and FREAs (for large band-gap energy donors) have shown the most potential for becoming commercially viable solar cell materials for bulk heterojunction blend cells. Wide band gap donors are known to enhance voltage and diminish current density, but in combination with FREAs both values can be relatively high.
There are still a lot of improvements to be made before an NFA-OSC can be commercially profitable. First of all, the PCE should be increased to at least 15% since this is the minimal value for commercial application. As said, PCEs already have exceeded 13% so recent development is on the right track. PCEs can be increased by designing even better NFAs, for instance, on the level of electron mobility NFAs still can increase a lot compared to FAs ( for the best NFAs compared to for the best FAs). Improvements can also be made in the following aspects: better donor matching, tandem constructions, BHJ morphology and domain purity of the donor and acceptor.
Besides these theoretical research aspects, implementation in a life size commercial solar cell also brings a lot of challenges, such as easy and sustainable device fabrication methods and long-term stability of the organic compounds. Studies also show that with upscaling, the PCE in general drops. On all of these areas, NFA-OSCs show great potential but it will take a lot of research before a solid non-fullerene acceptor-organic solar cell can compete with inorganic solar cells.
See also
Rylene dye
References
Organic solar cells | Non-fullerene acceptor | [
"Chemistry",
"Materials_science"
] | 2,257 | [
"Organic solar cells",
"Polymer chemistry"
] |
62,795,655 | https://en.wikipedia.org/wiki/AE1/AE3 | AE1/AE3 is an antibody cocktail that is used in immunohistochemistry, being generally positive in the cytoplasm of carcinomas (cancers of epithelial origin).
Targets
The antibody cocktail binds to cytokeratin 1 - 8, 10, 14 - 16 and 19 (but not CK17 or CK18). It is therefore used as a marker of carcinomas, such as depth of invasion and metastases. For example, it is both relatively sensitive and specific for detection of breast cancer metastasis to sentinel lymph nodes.
It may cross-react with GFAP, leading to aberrant staining of glial tumors such as ependymoma, glioblastoma and schwannoma. It may also stain myofibroblasts and smooth muscle cells. Furthermore, it may stain nodal epithelial cells that has contaminated a tumor from recent biopsy.
See also
List of histologic stains that aid in diagnosis of cutaneous conditions
References
Antibodies
Pathology
Carcinoma | AE1/AE3 | [
"Biology"
] | 220 | [
"Pathology"
] |
62,798,551 | https://en.wikipedia.org/wiki/Elipovimab | Elipovimab (formerly known as GS-9722) is a first in class of effector-enhanced broadly neutralizing HIV-1 antibodies for the targeted elimination of HIV infected cells and is as of 2020 in phase 1b clinical testing, designed with the goal of reducing or eliminating the HIV reservoir in patients.
References
Antiviral drugs | Elipovimab | [
"Biology"
] | 71 | [
"Antiviral drugs",
"Biocides"
] |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.