id
int64
39
79M
url
stringlengths
31
227
text
stringlengths
6
334k
source
stringlengths
1
150
categories
listlengths
1
6
token_count
int64
3
71.8k
subcategories
listlengths
0
30
61,864,866
https://en.wikipedia.org/wiki/Electrification%20and%20controls%20technology
Electrification and controls technology are devices that control, service and enhance productivity of industrial handling. Controls interface with hardware such as receivers, cranes and hoists, through a network in order to ensure that equipment operates safely and effectively. Almost every business, including the food, chemical, and automobile industries, uses controls. Some examples of these gadgets are: Remote controls Festooning Drives Motors Conductor bars Anti-collision devices Weighing devices Brakes Resistors Cabling Industry definitions Conductor bar: Insulated energized rails that safely provide power, control and data to moving equipment from a fixed source, much like electric rails on a model train. Festoon system: A cable management system of rolling trolleys that properly support power, control and data cables to moving equipment from a fixed source. Cable reel: A cable management device designed to spool and store electrical power, control or data cable, as the equipment moves along its path of motion. Variable-frequency drive: A type of static controller that safely drives an electric motor by varying the frequency and voltage the motor is supplied. This device minimizes the wear and tear of the mechanical system while allowing precise control and maximizing operator safety. Radio remote control: Allows an operator to control different types of moving equipment and cranes, meanwhile, providing the operator the best vantage point to the load or operation and physical position for a safe working area. Load brake: A device used to safely stop linear or rotating motion of equipment through the use of power or friction. References Electrification Control engineering
Electrification and controls technology
[ "Engineering" ]
304
[ "Control engineering" ]
61,864,919
https://en.wikipedia.org/wiki/Vlad%20Voroninski
Vlad Y. Voroninski (born 21 March 1985) is a Russian-American mathematician and entrepreneur. Academic biography Voroninski received his B.S. and M.A degrees in Applied Mathematics from UCLA in 2008, summa cum laude. He earned his Ph.D. in mathematics from UC Berkeley in 2013, under the supervision of Emmanuel Candes and John Strain. He was on the faculty at the MIT Mathematics Department from 2013 to 2016. Research Voroninski's PhD thesis kicked off the study of phase retrieval in the applied mathematics community, by providing the PhaseLift algorithm along with the first mathematical recovery guarantees for phase retrieval. His research has also led to solutions to open problems in computer vision, quantum operator theory, optimization and the theory of deep learning and compressive sensing. More recently, Voroninski's research connected the fields of deep learning and inverse problems, resolving the sample complexity bottleneck for compressive phase retrieval. Awards and honors Voroninski was awarded the 2014 SIAM Outstanding Paper Prize, given to works that "exhibit originality, for example, papers that bring a fresh look at an existing field or that open up new areas of applied mathematics". His PhD thesis was awarded the university-wide Bernhard Friedman Memorial Prize from UC Berkeley. In addition he has received the SIAM Student Paper Prize and SIGEST Review Awards from SIAM. He received the George E.G. Sherwood Prize from the UCLA Mathematics Department in 2008, which is awarded to the top graduating senior, as well as the Computing Research Association Outstanding Undergraduate Award in 2007. Entrepreneurship From 2014 to 2016, Voroninski was the founding Chief Scientist at Sift Security, a cybersecurity machine learning startup which was acquired by Netskope in 2018. As of 2016, Voroninski is the CEO and co-founder of Helm.ai, a stealth mode AI software startup focusing on autonomous driving. References 21st-century American mathematicians Living people 21st-century Russian mathematicians Applied mathematicians 1985 births
Vlad Voroninski
[ "Mathematics" ]
400
[ "Applied mathematics", "Applied mathematicians" ]
61,865,163
https://en.wikipedia.org/wiki/Suicide%20bereavement
Suicide bereavement is the experience of those who are grieving the loss of someone to suicide. Over 800,000 individuals die by suicide every year. It was stated by Shneidman (1978) that every suicide leaves behind 6 "survivor-victims". However, new research shows that each suicide leaves behind approximately 135 who personally knew the decedent. The #not6 hashtag has been used by Cerel and colleagues to represent that suicide bereavement is many more than publicly portrayed. Individuals experiencing suicide bereavement experience different challenges than those otherwise bereaved. Across 41 studies that examined these differences, the suicide bereaved experienced higher levels of blaming, stigma, shame and rejection. Those who lose someone to suicide may experience delays in the healing process. Complicated grief Complicated grief is grief whose symptoms do not decline over time. 10% to 20% of individuals survivors develop complicated grief. Individuals who develop complicated grief are likely to experience physical impairments to their daily functioning, with accompanying suffering. These symptoms persist without proper treatment, which became Complicated Grief Treatment. Grief stages Elisabeth Kübler-Ross, a Swiss-American psychiatrist, developed a model that encapsulates five stages of grief that an individual is likely to experience after a loss:. However, there is no support for people going through these stages in order or experiencing all of them. Denial: Can help the individual minimize the overwhelming pain of loss. This stage allows the individual to slow the processing of painful imagery, potentially reducing the pain. Anger: The individual adjusts to their new reality while likely experiencing extreme emotional discomfort. Anger tends to be the first feeling the individual experiences when they realize that their emotions are related to their loss. Bargaining: In this stage, the individual attempts to negotiate with a higher power. Depression: During the grieving process, bargaining has failed and the individual faces the facts of their loss. They are more likely to internalize their feelings and experience extreme sadness. They may find it difficult to reach out to others and experience isolation. Acceptance: In this last stage, the individual no longer resists the reality of their situation, leaving mostly sadness and regret. Suicide risk Compared to other bereaved, the suicide bereaved are at a higher risk of experiencing suicidal ideation. According to a 2002 study, results showed that suicide bereaved individuals were 1.6 times more likely to experience thoughts of suicide, 2.9 times more likely to have a suicide plan, and 3.7 times more likely to have made a suicide attempt. These individuals may conclude that suicide is the only viable solution to their pain. Other suicide bereaved may want to feel closer to the person they lost by ending their life the same way their loved one did. Stigma Despite efforts to de-stigmatize suicide, unlike other types of death, suicide remains stigmatized. Individuals who have lost their loved ones to suicide have reported difficulty with talking to others about their loss due to feeling uncomfortable talking about the topic. Additionally, certain religions reject those who have committed suicide. The suicide bereaved individuals may feel shame, making them uncomfortable sharing their loss in their religious group. Insurance policies may exclude death benefits for suicides. Survivors of suicide often experience difficulty healing due to this stigma. Treatment Support groups Support groups can be beneficial for the suicide bereaved, because this is a nonthreatening place where they can experience catharsis among others in a similar situation. Homogeneous support groups are typically more helpful than broader bereavement support groups. Such groups offer reassurances that feelings are acceptable. Individuals also may receive coping recommendations for holidays, talking to others and other now-difficult situations. Support groups can be found by visiting the International Association for Suicide Bereavement (IASP) website, which is affiliated with the World Health Organization. The American Association of Suicidology (AAS) and the American Foundation for Suicide Prevention (AFSP) also has a list of 400+ support groups across the United States. Psychotherapy Support groups on their own may not help individuals who develop a mental illness, such as Major Depressive Disorder (MDD) or Posttraumatic Stress Disorder (PTSD). The suicide bereaved are encouraged to seek clinicians who specialize in psychotherapy and/or medication management. A combination of education, psychotherapy and medication have been shown to be the most effective treatments. Complicated grief treatment Complicated grief treatment (CGT) incorporates elements of cognitive behavioral therapy (CBT), exposure and motivational interviewing. This type of psychotherapy facilitates the essential process of accepting loss. The combination of exposure techniques with cognitive restructuring has been shown to help individuals experience life satisfaction and engage in meaningful relationships. References Suicide
Suicide bereavement
[ "Biology" ]
962
[ "Behavior", "Human behavior", "Suicide" ]
61,866,079
https://en.wikipedia.org/wiki/Jennifer%20Arcuri
Jennifer Marie Arcuri (born February 1985) is an American technology entrepreneur. She lived in London from 2011 to 2018, before moving back to California. Self-described as an "ethical hacker", she founded the white hat consultancy Hacker House in 2016 and organised the Innotech Network from 2012. Her connection to then Mayor of London Boris Johnson from 2012 came to national attention in the UK in September 2019 when he became Prime Minister, triggering investigations into alleged conflicts of interest. She said in 2021 that she had an affair with him from 2012 to 2016. Career Arcuri previously lived in California and New York. She studied politics at the University of Wisconsin-Madison, theatre at Pace University in New York, filmmaking at the University of Southern California, and film at the American University in Paris for a summer. She briefly modelled and acted, including in a short film titled Commute. In California, she worked in film, including digital distribution, and production, sending a short film titled La Valise to Cannes. Arcuri owned a video streaming site for filmmakers, Ubroadcast, until selling it to Diamond I in 2009. Technology In 2008, Arcuri first visited London to take part in a Bollywood film, Naughty @ 40. She moved to London in early 2011. Arcuri studied for an MBA at Hult International Business School, where she met Tom Hayes and founded software company Title X Technology with him in 2012, using developers in Bulgaria. She defended Hayes in the Wall Street Journal after he came under investigation for the Libor scandal in 2013, for which he was later sentenced to 14 years in jail. From 2012, Arcuri organised the Innotech Summit in London. She founded the Tech Hotel in Shoreditch and was involved with Founders for Schools. She also founded Playbox, a video social network for entrepreneurs. She received an entrepreneur's visa after three years in the UK, after raising over £200,000 in funding for Innotech. By 2016, Innotech Network was noted as a meeting place for the tech industry and policymakers such as Boris Johnson. In November 2016, she worked with Sky News on a report that showed that the UK NHS had spent nothing on cyber-security during 2015; she noted that security was generally lacking in NHS trusts. In 2016, she headed the Tech London Advocates working group on cyber tech. She also founded Hacker House in 2016, a consultancy that advises and trains on cybersecurity, coming out of a "Legislating LulzSec" event run by Innotech in October 2014. Hacker House received a grant from the UK Department for Digital, Culture, Media and Sport Cyber Skills Immediate Impact Fund in 2019, sharing £500,000 with three other projects. Arcuri has encouraged other women and girls to become involved in the sector and ran the PinkSheet, a list of UK women expert professionals. Recognition Arcuri was recognised several times by Computer Weekly: in 2016, she was named as one of five "rising stars" among women in UK IT; she was named 18th of 50 of the "most influential women in UK tech" in 2017; she was longlisted for the same award in 2018 and 2019. SC Magazine listed Arcuri among twenty "women to watch" in UK cybersecurity in 2017. Relationship to Boris Johnson Arcuri had a close friendship with then-Mayor of London Boris Johnson, with the Sunday Times describing him as a regular visitor to her flat, and implying they were in a sexual relationship. Innotech, her company, was awarded £10,000 from a mayoral fund in 2013, followed the next year by Arcuri being awarded £15,000 from a government programme. Johnson intervened to allow her onto three trade mission trips. the Sunday Times claimed in September 2019 that Johnson failed to declare his personal relationship as a conflict of interest. Later that month, the Greater London Authority referred Johnson and his actions in the matter to the Independent Office for Police Conduct (IOPC) "so it can assess whether or not it is necessary to investigate the former mayor of London for the criminal offence of misconduct in public office". The IOPC is involved because the Mayor is also London's police and crime commissioner. The London Assembly commenced its own investigation, but paused it at the IOPC's request in order to avoid overlap. On 9 November 2019, it was revealed that the IOPC, which had been due to publish a report on its investigation, had decided to do so after the general election of 12 December. On 22 May 2020, the IOPC announced that they would not proceed with a criminal investigation. The IOPC said in its statement "While there was no evidence that Mr Johnson influenced the payment of sponsorship monies or participation in trade missions, there was evidence to suggest that those officers making decisions about sponsorship monies and attendance on trade missions thought that there was a close relationship between Mr Johnson and Ms Arcuri, and this influenced their decision-making." On 17 October 2020, Arcuri said that her relationship with Johnson had been sexual. On 28 March 2021, she said their affair lasted from 2012 to 2016. Promotion of conspiracy theories In December 2021, Vice News reported that Arcuri was promoting QAnon and anti-vaccine conspiracy theories on her Twitter account and Telegram channel. Personal life Arcuri's grandfather was lawyer Richard Cates. His daughter Christine Jendrzejewski is Arcuri's mother. Arcuri is married to Matthew Hickey, the co-director of Hacker House, with whom she had a daughter in 2017. They moved to Orange County, California in June 2018. According to Andrew Neil, who is suing Arcuri for libel, since January 2021, Arcuri has lived in Panama City, Florida. References External links https://hacker.house Pink Sheet Database on the Internet Archive 21st-century American businesswomen 21st-century American businesspeople American expatriates in England Hackers 1985 births Date of birth missing (living people) People from Orange County, California American women company founders American company founders American Internet company founders Businesspeople from California Living people Hult International Business School alumni Place of birth missing (living people) University of Wisconsin–Madison College of Letters and Science alumni
Jennifer Arcuri
[ "Technology" ]
1,263
[ "Lists of people in STEM fields", "Hackers" ]
61,867,179
https://en.wikipedia.org/wiki/Auxarthron%20californiense
Auxarthron californiense is a fungus within the family Onygenaceae family and one of the type species of the genus Auxarthron . A. californiense is generally distributed around the world and it is frequently found on dung and in soil near the entrances of animal burrows. History and taxonomy As one of the first species selected in the genus Auxarthron, A. californiense was first isolated from pack rat dung in 1963, at California. The genus Auxarthron was erected by Orr and Kuehn to accommodate species with swollen septa of the peridial hyphae, formerly classified in genus of Gymnoascus. However, both Apinis(1964) did not accept this new genus, because they could also observe swollen septa in other species of Gymnoascus. Although this feature may not be significant for distinguishing new genera, the genus Auxarthron could also be identified on the basis of the presence of wall thickenings in the vicinity of septa which they called "knuckle joints". The genus Auxarthron is related to Amauroascus and Arachnotheca, but differs from these two genera in possessing dark ascomata with distinct appendages. In 1965, this taxon was examined on cellophane buried in calcareous fen soil and on bird dung. Growth and morphology Colonies grown at freezing agar are at first white, then yellow, tan to yellow-brown and granular. then the colonies turn powdery. Colonies grow slowly on YpSs agar and become umbonate in center. The color of the colonies are often brownish and the reverse is reddish-brown in center to yellowish at the margins. This restricted, granular appearance with a reddish-brown reverse colony makes it distinguished from other species in the genus. Cylindrical or oblong conidia are produced in asexual reproduction. Conidia are pale yellow-orange, smooth to slightly rugose. Both intercalary and apical conidia were observed. Appearance Ascomata of A. californiense are globose. Young ascomata are rosy, and turn orange-brown at maturity. Ascospores are oblate and they have a pale yellow-brown color. This punctate ascospores appears broadly around the colony, forming a reticulate structure. The peridial hyphae of A. californiense also have a pale orange-brown color. These septate peridial hyphae are asperulate and cuticularized. A loose network appears with all the peridial hyphae together. A. californiense have uncinate appendages. Different from other species in the genus, the appendages of A. californiense are arise at odd angles from the ascocarps. In other species, the appendages are projecting forth at right angles. A. californiense is similar to Uncinocarpus uncinatus, in having thick-walled appendages that are hooked at the tip. However, the crozier-shaped tips of A. californiense have an acute end while the tips of U. uncinatus have a blunt and parallel end. Habitat and ecology Auxarthron species, like most Onygenaceae, have usually been isolated from strongly anthropized soil, enriched with human or animal keratinaceous remnants. The genus Auxarthron is considered as keratinolytic on the basis of hair perforation studies. However, the keratin degradative ability was not found in the specie A. californiense, different from other species in the genus. By Phylogenetic studies, this absence of keratinolytic ability shows that during the course of evolution the keratinolytic ability of Auxarthron genus is decreasing. In 1998, A. californiense was isolated on the veil, tunic and wood fragments on the remains of a ninth century Longobard abbess at Pavia, Italy. The presence of A. californiense in this habitat indicates that A. californiense might be able to utilize some of the nutrients in human decomposing substances. References Onygenales Fungus species
Auxarthron californiense
[ "Biology" ]
887
[ "Fungi", "Fungus species" ]
61,867,655
https://en.wikipedia.org/wiki/Vivo%20NEX%203
The Vivo NEX 3 and Vivo NEX 3 5G are Android phablets developed by Vivo which were unveiled on 16 September 2019. Specifications Design The NEX 3 and NEX 3 5G both have an aluminum frame and curved Gorilla Glass on the back and Schott Xensation® Up glass on the front. The curved edges of the display do not allow for physical power and volume buttons; pressure-sensitive haptic keys are implemented instead. A small button at the top of the device is situated next to the pop-up camera mechanism and is used for hard reset and screen lock only. Hardware & Software Internally, the Snapdragon 855+ and Adreno 640 are used, with a 4500 mAh battery powering both. The 5G model also has a vapor chamber dedicated to the modem. The NEX 3 is available with 128 GB of non-expandable UFS 3.0 storage and 8 GB of RAM, while the NEX 3 5G is available with 256 GB of non-expandable UFS 3.0 storage and 8 or 12 GB of RAM. Both can quick charge at 44W via USB-C, but do not have wireless charging. Biometric options comprise an updated in-display optical fingerprint reader and facial recognition. The phones' rear cameras are housed in a centered circular module, consisting of a 64 MP main sensor, a 13 MP ultrawide sensor and a 13 telephoto sensor. A pop-up camera hides the 16 MP front sensor and a flash. The mechanism has been improved and is now faster and more durable, taking 0.65 seconds to operate. Both devices have one screen, lacking their predecessor's dual-screen functionality. The Vivo NEX 3 and NEX 3 5G have a 6.89" (175mm) FHD+ HDR10 display with a resolution of 1080 x 2256 pixels and a screen-body ratio of 93.6%. A Samsung-made Super AMOLED panel is used on both with a curved "Waterfall" design akin to the Mate 30 Pro's. Edge detection and palm rejection are implemented to prevent accidental touches. The phones' rear cameras are housed in a centered circular module, consisting of a 64 MP main sensor, a 13 MP ultrawide sensor and a 13 telephoto sensor. OIS is not present on any of the sensors. A pop-up camera hides the 16 MP front sensor and a flash. The mechanism has been improved and is now faster and more durable, taking 0.65 seconds to operate. Both phones run on Android 9.0 "Pie" with Vivo's Funtouch 9.1 skin. Successor Vivo announced the NEX 3S 5G on 10 March 2020 as a hardware revision of the NEX 3 5G. The device features an upgraded Snapdragon 865 SoC, Android 10 with Funtouch 10 and HDR10+ support for the display, and benefits from newer LPDDR5 RAM, UFS 3.1 and Bluetooth 5.1. The design remains the same, although there is now an orange color variant. References Android (operating system) devices Phablets Mobile phones introduced in 2019 Mobile phones with multiple rear cameras NEX 3 Mobile phones with 4K video recording Discontinued flagship smartphones
Vivo NEX 3
[ "Technology" ]
675
[ "Phablets", "Crossover devices", "Discontinued flagship smartphones", "Flagship smartphones" ]
61,869,688
https://en.wikipedia.org/wiki/Eva%20Regnier
Eva Dorothy Regnier (born 1971) is a decision scientist whose research concerns the interaction between human decision-making and environmental prediction. She is a professor of decision science in the Graduate School of Business and Public Policy of the Naval Postgraduate School. Education and career Regnier graduated from the Massachusetts Institute of Technology in 1992 with a bachelor's degree in environmental engineering science. After working from 1993 to 1996 in industry as an environmental engineer, she went to the Georgia Institute of Technology for a master's degree in operations research in 1999 and a Ph.D. in industrial engineering in 2001. Her dissertation, Discounted Cash Flow Methods and Environmental Decisions, was supervised by Craig Tovey. She joined the Defense Resources Management Institute of the Naval Postgraduate School in 2001, moved to the Graduate School of Business and Public Policy in 2017, and was promoted to full professor in 2019. Regnier was president of the INFORMS Forum on Women in Operations Research and Management Science for 2011. Contributions Regnier has published well-cited works on volatility in energy markets and on decision-making for evacuations based on hurricane predictions. Other topics in her research include correlations between pirate activity and predicted changes in climate and weather. Her work on hurricane evacuation was a finalist for the INFORMS Junior Faculty Forum award, and her work developing a tool to simulate the hurricane decision-making process was a finalist in the INFORMS MSOM Practice Based Research Competition. She received the INFORMS Decision Analysis Society Publication Award for her work on probability forecasting. Selected publications References External links Home page 1971 births Living people American industrial engineers American women engineers Operations researchers Environmental engineers MIT School of Engineering alumni Georgia Tech alumni Naval Postgraduate School faculty American women academics 21st-century American women
Eva Regnier
[ "Chemistry", "Engineering" ]
343
[ "Environmental engineers", "Environmental engineering" ]
61,869,767
https://en.wikipedia.org/wiki/Dirichlet%20series%20inversion
In analytic number theory, a Dirichlet series, or Dirichlet generating function (DGF), of a sequence is a common way of understanding and summing arithmetic functions in a meaningful way. A little known, or at least often forgotten about, way of expressing formulas for arithmetic functions and their summatory functions is to perform an integral transform that inverts the operation of forming the DGF of a sequence. This inversion is analogous to performing an inverse Z-transform to the generating function of a sequence to express formulas for the series coefficients of a given ordinary generating function. For now, we will use this page as a compendia of "oddities" and oft-forgotten facts about transforming and inverting Dirichlet series, DGFs, and relating the inversion of a DGF of a sequence to the sequence's summatory function. We also use the notation for coefficient extraction usually applied to formal generating functions in some complex variable, by denoting for any positive integer , whenever denotes the DGF (or Dirichlet series) of f which is taken to be absolutely convergent whenever the real part of s is greater than the abscissa of absolute convergence, . The relation of the Mellin transformation of the summatory function of a sequence to the DGF of a sequence provides us with a way of expressing arithmetic functions such that , and the corresponding Dirichlet inverse functions, , by inversion formulas involving the summatory function, defined by In particular, provided that the DGF of some arithmetic function f has an analytic continuation to , we can express the Mellin transform of the summatory function of f by the continued DGF formula as It is often also convenient to express formulas for the summatory functions over the Dirichlet inverse function of f using this construction of a Mellin inversion type problem. Preliminaries: Notation, conventions and known results on DGFs DGFs for Dirichlet inverse functions Recall that an arithmetic function is Dirichlet invertible, or has an inverse with respect to Dirichlet convolution such that , or equivalently , if and only if . It is not difficult to prove that is is the DGF of f and is absolutely convergent for all complex s satisfying , then the DGF of the Dirichlet inverse is given by and is also absolutely convergent for all . The positive real associated with each invertible arithmetic function f is called the abscissa of convergence. We also see the following identities related to the Dirichlet inverse of some function g that does not vanish at one: Summatory functions Using the same convention in expressing the result of Perron's formula, we assume that the summatory function of a (Dirichlet invertible) arithmetic function , is defined for all real according to the formula We know the following relation between the Mellin transform of the summatory function of f and the DGF of f whenever : Some examples of this relation include the following identities involving the Mertens function, or summatory function of the Moebius function, the prime zeta function and the prime-counting function, and the Riemann prime-counting function: Statements of the integral formula for Dirichlet inversion Classical integral formula For any s such that , we have that If we write the DGF of f according to the Mellin transform formula of the summatory function of f, then the stated integral formula simply corresponds to a special case of Perron's formula. Another variant of the previous formula stated in Apostol's book provides an integral formula for an alternate sum in the following form for and any real where we denote : Direct proof: from Apostol's book Define the Dirichlet series and its associated partial sums . Define the function . Use partial summation to write . Apply the Euler–Maclaurin summation formula to obtain an approximation for in terms of and its derivatives. Express the error term in the approximation as an integral of a certain function over the interval . Use Abel's summation formula to express as a sum of integrals involving and its derivatives. Express the integrals involving and its derivatives in terms of and its derivatives. Substitute the results from steps 4, 5, and 7 into the formula from step 3, and simplify to obtain the classical integral formula for Dirichlet inversion. This proof shows that the function can be recovered from its associated Dirichlet series by means of an integral, which is known as the classical integral formula for Dirichlet inversion. Special cases of the formula If we are interested in expressing formulas for the Dirichlet inverse of f, denoted by whenever , we write . Then we have by absolute convergence of the DGF for any that Now we can call on integration by parts to see that if we denote by denotes the antiderivative of F, for any fixed non-negative integers , we have Thus we obtain that We also can relate the iterated integrals for the antiderivatives of F by a finite sum of k single integrals of power-scaled versions of F: In light of this expansion, we can then write the partially limiting T-truncated Dirichlet series inversion integrals at hand in the form of Statements in the language of Mellin transformations The Dirichlet generating function of a sequence is the Mellin transform of the sequence, evaluated at : . The Dirichlet inverse of a sequence is related to the inverse Mellin transform of its generating function: , where is a real number greater than the abscissa of convergence of the Dirichlet series . The Mellin transform of a convolution of two sequences and is the product of their Mellin transforms: . If is a sequence and is a function such that the integral converges absolutely and uniformly for in some right half plane, then we can define a Dirichlet series by , and the Dirichlet series is the Mellin transform of . A formal generating-function-like convolution lemma Suppose that we wish to treat the integrand integral formula for Dirichlet coefficient inversion in powers of where , and then proceed as if we were evaluating a traditional integral on the real line. Then we have that We require the result given by the following formula, which is proved rigorously by an application of integration by parts, for any non-negative integer : So the respective real and imaginary parts of our arithmetic function coefficients f at positive integers x satisfy: The last identities suggest an application of the Hadamard product formula for generating functions. In particular, we can work out the following identities which express the real and imaginary parts of our function f at x in the following forms: Notice that in the special case where the arithmetic function f is strictly real-valued, we expect that the inner terms in the previous limit formula are always zero (i.e., for any T). See also Dirichlet series Dirichlet convolution Dirichlet inverse Arithmetic function Multiplicative function Dirichlet generating function (DGF) Notes References Number theory Analytic number theory Integer sequences
Dirichlet series inversion
[ "Mathematics" ]
1,451
[ "Sequences and series", "Analytic number theory", "Discrete mathematics", "Integer sequences", "Mathematical structures", "Recreational mathematics", "Mathematical objects", "Combinatorics", "Numbers", "Number theory" ]
64,210,333
https://en.wikipedia.org/wiki/2020%20Assam%20gas%20and%20oil%20leak
The 2020 Assam gas and oil leak, also referred as the Baghjan gas leak, was a blowout and methane leak that happened in Oil India Limited's Baghjan Oilfield in Tinsukia district, Assam, India on 27 May 2020. The blowout occurred at Well No. 5 in the Baghjan Oil Field, resulting in a leak of natural gas. The leaking well subsequently caught fire on 9 June 2020, and resulted in three deaths (Officially), large-scale local evacuations, and environmental damage to the nearby Dibru-Saikhowa National Park and Maguri-Motapung Wetland. The blowout was "killed" on 15 November 2020, 173 days after the blowout, with a technique called snubbing. Subsequently, Well No. 5 was abandoned on 3 December 2020. An investigation by a committee appointed by the National Green Tribunal revealed in November 2020 that Oil India Limited had failed to obtain legally required clearances to operate the oil field at Baghjan, and had failed to comply with provisions of environmental laws as well as with internal safety procedures in relation to drilling at the oil field. The findings reiterated a report by the Assam State Pollution Control Board, which had also reported that Oil India Limited did not obtain the necessary clearances before drilling. Background Location and environmental conditions The Baghjan Oil Field is located in Tinsukia district in the State of Assam, near Baghjan village, which has a resident population of 4,488 persons. The nearest towns are Doom Dooma and Tinsukia. Baghjan Oil Field is located near the Dibru-Saikhowa National Park in Assam, and is also in proximity to Maguri Motapung Beel, a natural wetland. Dibru-Saikhowa National Park is the only riverine island wildlife reserve, globally. It contains a variety of rare and endangered species, including the Gangetic dolphin, which is endangered, as well as herds of wild horses, tigers, capped langurs, hoolock gibbons, slow loris, and several rare butterflies and fish. The Dibru-Saikhowa National Park is also connected to Namdapha National Park via the Dehing Patkai Wildlife Sanctuary. These regions are part of the Indo-Burma Biodiversity Hotspot. Baghjan Well No. 5, from which the leak occurred, is located at a distance of 900 meters from the park and adjoins a buffer forested region surrounding the park. It is also close to the Indo-Burma Biodiversity Hotspot. On 18 January 2020, the Central Government exempted all oil and gas firms conducting exploratory drilling from legal requirements for environmental clearances. Prior to the leak, in 2020, the region witnessed protests after the Government of India announced that they had authorized Oil India Limited (OIL) to explore the area under Dibru-Saikhowa National Park for hydrocarbons. During May 2020, prior to and during the leak, Tinsukia district in Assam experienced heavy rains and flooding in two local rivers, Dibru and Lohit. The region was also placed in lockdown, following the COVID-19 pandemic in India. Environmental authorizations A subsequent investigation by Assam's State Pollution Control Board found that Oil India Limited had begun drilling at the Baghjan Oil Field without obtaining prior environmental clearances in accordance with legal requirements under the Water (Prevention of Control of Pollution) Act, 1974, Air (Prevention of Control of Pollution) Act 1981 and Environment Protection Act, 1986. Oil India Limited had also failed to conduct mandatory public hearings before beginning drilling. The Eco-Sensitive Zone around Dibru-Saikhowa National Park was reduced in size at the request of Oil India Limited in order to allow the well to be established. In 2013, the Baghjan Oil Field was inspected by a team from the National Board for Wildlife, which criticised the Government of India for ratifying the breach of wildlife norms after Oil India Limited had already begun operating in an eco-sensitive zone. Prior leaks In 2005, a blowout at an Oil India Limited well in Dikom in Assam resulted in the evacuation of 500 families. The leak subsequently caught fire, as well. On 3 February 2020 near Naharkatia in Assam, a stretch of the Burhi Dihing river, which is a tributary to the Brahmaputra river, caught fire after a punctured pipeline carrying crude oil from the Oil India Limited headquarters at Duliajan in Assam leaked, covering the water with a film of oil. The fire burned for a period of 48 hours before the leak was repaired and it could be extinguished. Oil India Limited officials stated that a technical error with their instruments had caused shut off valves in a storage tank to close, resulting in pressure building up in a pipeline carrying crude oil. This caused several leaks in the pipeline. Oil India Limited Officials blamed local residents for causing the fire, and stated that a number of leaks had previously occurred, as parts of attempts to steal quantities of crude oil from the delivery pipeline. The Central Pollution Control Board had directed Assam's State Pollution Control Board to investigate the damage to aquatic life and local environment caused by the leak. Drilling at Baghjan 5 Well The Baghjan Oil Field has 21 active wells, of which 4 produce natural gas, while the remaining produce oil. Baghjan Oil Field's Well No. 5 was established in 2006 and produces natural gas. It drills down to a depth of 3,870 meters, and was producing between 80,000 and 100,000 standard cubic meters of gas per day before the leak. Drilling at the Baghjan 5 well, where the leak occurred, was being outsourced to an Ahmedabad-based firm, John Energy. In May 2020, a new reservoir was being tapped for additional production at a depth of 3,729 meters, with workover operations taking place In addition, the infrastructure for the previous reservoir was being serviced, due to which production had been temporarily stopped. A device for pressure control, known as a blowout preventer, had been temporarily uninstalled while the extraction assembly was being serviced. In April 2020, following a lockdown caused by the COVID-19 pandemic in India, Oil India Limited suspended drilling at several of the wells in the Baghjan Oil Field, primarily because of the unavailability of employees. On 20 May 2020, before the leak, the Ministry of Environment, Forest and Climate Change announced that they would be authorizing exploration for hydrocarbons under the Dibru-Saikhowa National Park, at seven different locations. Oil India Limited stated that they would be using Extended Reach Drilling (ERD) to explore these wells without entering the National Park itself. Baghjan Oil Well 5, at which the leak occurred, will be one of the bases on which the ERD will extend under the National Park. Blowout, leak, and fire Initial occurrence On 27 May 2020, at 10:30 a.m., residents of Baghjan village in Assam reported hearing a loud sound from the nearby Baghjan Oil Field. Reports of the sound of the explosion were also confirmed from residents of Tinsukia town, which is located at a distance of 12 kilometres from the Oil Field. Local newspaper, the Sentinel, reported local accounts indicating that the gushing of natural gas from the well continued to produce an audible sound over the next few days. Although production had been temporarily stopped at Baghjan's Well No. 5 to explore a new sand, and to service the well, workers noticed that gas was escaping from the capped well, and had begun to evacuate the Oil Field before the incident. At 10:30 a.m, the blowout occurred, resulted in the leak of natural gas from the well, and causing a complete suspension in operations. Oil India Limited's official statement indicated that the well had "suddenly became very active" while workover operations were ongoing. OIL officials confirmed that the leak occurred from the existing sand which had been capped to allow exploration of the new sand. This resulted in a "fountain of crude oil" from the sand. The leaked gas consisted of a naturally occurring mix of propane, methane, propylene and other gases. Fires On 9 June 2020, Well No. 5 caught fire after efforts to cap the leak were unsuccessful. An official from Oil India Limited stated, "We cannot say how and why it happened," in regard to the fire, also noting that fires, in the case of a blowout, were not unexpected. The fire occurred at the plinth of the well while cleaning operations were under way. Following the fire, people who had not yet been evacuated from the local area left, as the fire spread to local grasslands. The fire quickly spread to a larger area and burned down nearby trees, crops and houses. 4 persons were injured, and 50 houses destroyed in the resulting fire. The Tinsukia District Collector stated that the condensate from the gas leak in surrounding areas had made the locale more vulnerable to fire and had allowed the fire to spread faster. On 10 June 2020, two firefighters, Durlov Gogoi and Tikheswar Gohain died fighting the fire. Their bodies were recovered from a water body near the site. Recently on 21 July 2020, an explosion occurred as the well number 5 and three foreign experts were injured in the incident. Evacuation Following the leak, 1610 local families, consisting of approximately 3000 persons were evacuated to relief camps. A safety zone consisting of a 1.5 km radius around the well was established. The National Disaster Relief Force was deployed to establish the relief camps. In addition to local residents, employees of Oil India Limited and their families were also evacuated from the area. The Indian Air Force and Indian Army also provided assistance. Environmental damage State Pollution Control Board officials stated that the leaked gas condensate had affected local agricultural crops and plants, including bamboo, tea, bananas and betel nuts. Wind conditions had carried the leaked gas towards the Dibru-Saikhowa National Park as well. On 29 May 2020, the carcass of a Gangetic dolphin covered in condensated oil was found in the Maguri Motapung Beel, a local wetland, and sent for a post-mortem by the Tinsukia Wildlife Division, to establish the cause of death. Wildlife Division officials noted that because of rain immediately following the leak, local waterbodies had been contaminated by condensate from the leak. On 31 May 2020, Assam Government officials confirmed that the State Pollution Control Board was investigating environmental damage as a result of the leak. Following the leak, the State Forest Department had also asked Oil India Limited to account for its actions after reports of dead fish in local waterbodies. On 5 June 2020, local residents protested near the Maguri Motapung Beel wetland, calling for the protection of the ecologically sensitive Dibru-Saikhowa National Park, to mark World Environment Day, and called for compensation for their loss of livelihoods as a result of the leak. Efforts to contain and stem the leak Initial response Operations at Baghjan Oil Field were suspended following the blowout and leakage of natural gas. OIL (Oil India Limited) requested assistance from the Crisis Management Team of the Oil and Natural Gas Corporation of India. IOC initially attempted to install a blowout preventer to bring the leak under control. OIL also sprayed the area with water. On 30 May 2020, four days after the initial leak, a spokesperson from Oil India Limited confirmed that "The well control operation is yet to start." Oil India Limited officials also stated that the process of controlling the well would take time, and that they were currently making arrangements to secure a water supply to help mitigate the leak. Attempts to cap the well Placing a blowout preventer In June 2020, Oil India Limited officials stated that they had constructed a temporary reservoir of water in the vicinity of the well, by placing pipes from the Dangori river. Water from the reservoir would be used to spray on to the well area to protect workers while they attempted to place a blowout preventer, using hydraulic devices, followed by injecting mud to completely shut down production of gas, or "kill" the well. Singapore-based Alert Disaster Control experts were called in to assist with the process. Initial efforts to control the leak were delayed because of severe flooding in Assam, which caused widespread damage and resulted in the deaths of 16 persons. Oil India Limited said that because of the floods, working conditions at the Oil Field were temporarily unsafe, as the Dangori river which was being used to pump water to the field, was overflowing. The nearby Doom Dooma-Baghjan bridge had collapsed due to the flooding, and consequently, access roads to the site were unavailable. There were two failed attempts to place the blowout preventer on Well No. 5. On 22 July 2020, during one such attempt, three Singaporean expert advisors were injured after a fire broke out. On 18 August 2020, a blowout preventer was successfully placed at the well head on a third attempt, but the process of 'killing the well' by injecting mud to completely stop production of natural gas from the well was not successful, and the well continued to leak. Oil India Limited officials stated that the reason for the failure was that a valve in the well casing collapsed. Snubbing and diversion On 3 September 2020, Assam's Commerce and Industry Minister, C. M. Patowary informed the Assam Legislative Assembly that experts from Canada had been brought in to attempt to cap the well again. Oil India Limited's officials stated that they were also considering an attempt to divert the gas produced at Well No. 5 into two separate channels. If successful, this will result the diversion of part of the gas being produced to the Baghjan Early Production System facility, while the remaining gas will be flared. Once the diversion has occurred, an attempt will be made to cap the well at a greater depth, in a process called snubbing. On 13 September 2020, Oil India Limited succeeded in diverting the gas leaking from Well No. 5 to nearby flare pits on a second attempt. The first attempt at such a diversion had been unsuccessful. The well has not as yet been controlled, and on 30 September 2020, Sushil Chandra Mishra, the managing director of OIL stated that it would take two more months to kill the well and stop the leak. On 3 September 2020, Assam's Commerce and Industry Minister, C. M. Patowary stated that it may take another two months to control the leak. On 13 September 2020, Oil India succeeded in diverting some of the leaking gas to flare pits, but has yet to completely control the leak and 'kill' the well. On 5 November 2020, a second attempt to snub the well was initiated. This included flying in 60 tonnes of the snubbing unit from Canada's Calgary by an Antonov An-24. In the final phase of snubbing, “kill fluid” or cement-laced chemical mud was injected to "kill" the blowout well. This was successfully completed on 15 November 2020. Subsequently, OIL initiated the process to abandon the Well No-5, which was completed on 3 December 2020. Consequences Fatalities and evacuations Two firefighters employed by Oil India Limited died on 10 June 2020 while fighting the fire resulting from the gas leak. Four other fire fighters suffered injuries. Local reports indicate that there have been more deaths in the area; an inquiry by a local magistrate is currently investigating these claims. On 10 September 2020, an engineer working with Oil India Limited, Arnab Kishore Bordoloi, died while working at the well site. The cause of death is yet to be established. As of September 2020, 500 persons are currently still residing in relief camps, with the surrounding area still being affected by audible sounds of the gas leak as well as the ongoing fire. Local reports indicate that the sound of the leak can be heard from a distance of 5 kilometres from the site of the leak. Litigation and compensation On 25 June 2020, the National Green Tribunal found a prima facie case had been made out against Oil India Limited for damage to the environment and local livelihoods because of the leak. The Tribunal ordered Oil India Limited to deposit a sum of 250 million with the District Magistrate of Tinsukia towards compensation, and constituted an expert committee to investigate the disaster and recommend how much compensation was payable to affected persons. The expert committee constituted by the National Green Tribunal recommended that completely damaged houses would result in compensation of 2,500,000, while those that were severely and moderately damaged would receive 1,000,000 and 250,000 each. From 24 August 2020 onwards, local residents staged a peaceful protest, establishing a camp in front of the office of the local Deputy Commissioner, to ensure that all persons whose homes were damaged in the fire and leak received more compensation. The Assam Government confirmed to the Assam Legislative Assembly that some families had been issued compensation. Oil India Limited had filed objections to these orders for compensation in the National Green Tribunal, but the Tribunal dismissed their objections on 8 August 2020. OIL announced that they will give ₹30,000 as compensation to all those families who are affected. On 1 October 2020, the Gauhati High Court agreed to hear a public interest litigation petition filed by two local residents, and ordered OIL as well as a number of government authorities, including the Assam Government, the National Board for Wildlife, the State Board for Wildlife, and the State and Central Boards for Pollution Control, to respond to concerns raised about drilling inside the Dibru-Saikhowa National Park. Internal investigation and criticism On 30 May 2020, an Oil India Limited spokesperson stated to press that a five-member investigative committee had been established to ascertain the causes of the blowout and leak. On 10 June 2020, Oil India Limited confirmed that two employees had been placed on suspension in connection with the leak. The Indian Oil Workers Union criticized the use of private contractors at Well No.5, calling on Oil India Limited to use only trained corporation employees for such work. Government investigations and environmental damage Assam Government The initial leak at the Baghjan Oil Field resulted in the contamination of water bodies in the nearby Maguri Motapung Beel wetland, as well as in the Dibru-Saikhowa National Park. Days after the leak, local forest officials reported seeing a layer of gas condensate on water bodies in the Dibru-Saikhowa National Park. An Environmental Impact Assessment Report is currently being prepared to assess the damage caused by the leak and fire. On 12 June 2020, the Assam state government ordered an investigation into the environmental impact of the leak, and asked for a report within a month. An expert committee was formed by Assam's Forest Department to investigate the environmental impact. On 21 June 2020, Assam's State Pollution Control Board (SPCB) ordered Oil India Limited to shut down all production at the Baghjan Oil Field, following an investigation that revealed that the company had begun drilling operations there without obtaining prior environmental clearances. The SPCB also found that the company had failed to comply with continuing reporting and certification requirements under environmental laws. Oil India Limited objected to this notice of closure, stating that it could potentially result in more blowouts at the wells. The closure notification was withdrawn three days after it was issued. Commenting on the government's measures following the disaster and intersecting it with the extractive industries in the district, sociologist Sanjay Barbora and geologist Sarat Phukan reflect,The government of Assam’s remedial measures like announcing the formation of Dehing Patkai as a national park, as well as the legal obligations pressed upon OIL to follow through compensatory payments to those affected by the Baghjan tragedy, raise hope and despair in equal measure. Hope, because of the enormous interest raised by ordinary citizens during a time when they were atomised by pandemic-induced lockdowns, where seams of the conflict and violent history were dug out as evidence of corporate greed and governmental collusion. Despair, because none of the contentious history and complicated contemporary realities seem to have found their way into the responses of the state (including the Gauhati High Court), the public sector (OIL, ONGC and CIL), or the tea companies. Union Government The Indian Government's Ministry of Petroleum and Natural Gas has also ordered an investigation into the gas leak. On 2 August 2020, the Wildlife Institute of India, which operates under the Ministry of Environment, Forest and Climate Change issued a report in which they stated that Oil India Limited had failed to address security concerns after two previous leaks in Dikom and Naharkatia in Assam. Their report indicated that between 60 and 70 hectares of land had been damaged by the leak, and noted particular concerns about the biodiversity in the Tinsukia district. The Wildlife Institute of India's report stated that the leak would have "....prolonged ill effects on all life forms, including humans," in the area. National Green Tribunal On 26 June 2020, the National Green Tribunal constituted an expert committee headed by a former judge, A.P. Katakey to investigate the reasons for, and the impact, of the gas blowout. The expert committee was constituted after a preliminary report requested by the Tribunal revealed that there were errors in internal safety procedures that resulted in the incident. The preliminary report indicated that compliance with these safety procedures could have prevented the blowout. The Katakey Committee reported in November 2020 that Oil India Limited had repeatedly violated provisions of the Air (Prevention and Control of Pollution) Act, 1981, the Water (Prevention and Control of Pollution) Act, 1974, and the Environment Protection Act 1986. The committee also found that Oil India Limited had failed to obtain the necessary clearances for the Baghjan Oil Field, and had not conducted Biodiversity Impact Assessment Study before beginning operations, which had been made mandatory by a Supreme Court order. In popular culture Discovery Channel made a two-part documentary on the disaster. It was aired in March 2021 and was also made available on YouTube. See also Visakhapatnam gas leak References 2020 disasters in India 2020 industrial disasters 2020s in Assam Environmental disasters in India Explosions in India Oil spills in India Petroleum Pollution events in 2020
2020 Assam gas and oil leak
[ "Chemistry" ]
4,574
[ "Petroleum", "Chemical mixtures" ]
64,210,761
https://en.wikipedia.org/wiki/Marasmius%20wynneae
Marasmius wynneae (sometimes incorrectly spelt Marasmius wynnei) is a species of gilled mushroom found in European woods. Naming This species was originally described in 1859 by the mycologists Miles Joseph Berkeley and Christopher Edmund Broome in the British "Annals and Magazine of Natural History". The original collection was done on the Coed Coch estate in Denbighshire, Wales, which then belonged to the Wynne family. The entry for the new species explicitly mentions Mrs. Wynne, and so it is clear that the mushroom was named for her and not her husband. A Latinized species name takes the genitive ending -i (or -ii) if it is dedicated to a man and it takes -ae (or -iae) if it is for a woman. Therefore, the spelling wynneae is correct here, even though the spelling in the original description of the species was wynnei. In English this fungus has the common name "Pearly Parachute". Description The same set of references applies to this section throughout. The hygrophanous cap is hemispherical and white to greyish or ochre when young. Later it expands to become flat and may turn grey or also take on violet tints. The diameter is from about 1 cm to 5 cm. The whitish thick and distant gills are almost free to emarginate. There is no ring or other veil remnant. The tough stem is up to about 7 cm long and up to 0.5 cm in diameter. Initially it is whitish and it becomes red brown from the base. The taste is mild and the smell is pleasant initially (of hay, melilot or bitter almonds) but after a moment it becomes unpleasant (of acid, or a drying facecloth). The spore powder is white and the spores are ellipsoid, sometimes elongated (with an aspect ratio up to about 2) and are around 6-8 μm by 3.5-4.5 μm. There are cheilocystidia which are club-shaped to spindle-shaped with lobes at one end. Distribution, habitat, ecology and human impact This saprobic mushroom grows in clusters on humus and litter in beech woods or other broad-leaved woods. The conspicuous white mycelium can bind leaf litter together in sheets. This mushroom is widespread from August to January in Europe, and has been reported from Algeria, Morocco, and Texas, but not from other regions. It is common in western temperate Europe, especially under beech. References Links wynneae Fungi described in 1859 Taxa named by Miles Joseph Berkeley Taxa named by Christopher Edmund Broome Fungus species
Marasmius wynneae
[ "Biology" ]
547
[ "Fungi", "Fungus species" ]
64,210,868
https://en.wikipedia.org/wiki/Pop-up%20bicycle%20lane
A pop-up bicycle lane (also known as a pop-up cycle path or corona cycle path) is a temporary bike lane that is used to test, pilot or trial new infrastructure to improve conditions for people riding bicycles. In the event that it is successful, interventions can be implemented permanently. During the COVID-19 pandemic in particular, many cities set up pop-up bike lanes to quickly provide more space and safety for cyclists in poor road traffic conditions. These were usually intended as temporary cycling infrastructure for the time of the Pandemic. The purpose was primarily to provide more capacity for the rapid increase in demand for cycling and provide a viable alternative to places in close proximity to other people such as public transport. The cycle paths, which are usually marked with yellow lines and construction site beacons, were usually established by redesignating the kerbside traffic lane or a previous parking lane as a cycle lane. In Berlin, the cost of one kilometre of pop-up cycle paths is around 9500 euros. History The term "pop-up bike lane" originated in North America, where, for example, the US city of New York City has launched a number of experiments with short-term cycling infrastructure. The COVID-19 pandemic led to the creation of more space for bicycle traffic in Colombia, initially in the capital city of Bogotá, over a total of more than one hundred kilometres of main roads. The measure was reported internationally. In Germany, pop-up cycle paths were initially set up in the Berlin district of Friedrichshain-Kreuzberg. The first pop-up cycle path in Berlin was created on 25 March 2020 at Hallesches Ufer. The pop-up cycle paths in Berlin were laid out for a limited period until 31 May 2020, with the prospect of a transfer to a permanent cycle infrastructure in accordance with the Berlin Mobility Act. However, the deadline was extended at the end of May until the end of the year. Mexico City announced a 54-kilometre pop up lane in Av Insurgentes and Eje 4 to create a mobility alternative to help decrease mass transit agglomeration in Metrobus lines. Its permanence will be evaluated according to use. Other Mexican cities that have created pop up bike lanes are Zapopan in Jalisco, San Pedro Garza García in Nuevo León and Puebla, Puebla. Concept The Berlin Friedrichshain-Kreuzberg district authority recommends a concept, after which pop-up cycle paths can be set up within ten days and in eleven steps, from the identification of the areas affected, consultation with the authorities to be involved, the ordering of measures and temporary signage, to completion. The following four basic principles are applied in the design process: traffic separation: physical separation of foot, bicycle and car traffic beyond markings, which, if possible, does not need to be crossed by car traffic for parking Forgiving infrastructure: infrastructure that minimises the risk of injury in the event of misbehaviour, for example by means of safety distances between the carriageway and the kerb or guide beacons or the use of separation elements made of yielding materials Predictability: easily understandable and less complex traffic routing Network approach: Establishment of a transport network to relieve individual road sections On roads with two lanes in both directions, the right-hand lane including a buffer zone for flowing motorised traffic is completely separated as a cycle lane and existing signs for stationary traffic are covered. Parking of motor vehicles is then no longer permitted and driving is possible on one strip for both cyclists and motorised traffic. On roads with two lanes and a parking strip in both directions, the right lane for cycle traffic, including a buffer zone for both stationary and moving motorised traffic, is separated and the parking strip is maintained. Motorised traffic can then drive on one lane in each direction. In this case, however, vehicles must cross the cycle lane when parking or entering, which is contrary to the above-mentioned principle of traffic separation. On roads with three lanes in each direction, where parking is allowed on the right lane, the right lane is separated as a cycle lane with a buffer zone to the middle lane and the middle lane is designated as a parking lane, so that flowing motorised traffic, stationary motorised traffic and cycle traffic are each provided with one lane. At crossroads with traffic lights, there are possible measures to protect straight ahead or right-hand cycle traffic from motorised traffic turning right. This includes the creation of a temporary protected intersection with temporary kerbstone extensions or alternatively traffic light phases with separate, exclusive green phases for pedestrian and bicycle traffic. If this is not possible, it is recommended to switch the green phase for cyclists before the green phase for motorised traffic. Political debate in Germany The establishment of pop-up cycle paths in Berlin by the red-red-green senate was praised by cycling associations such as the ADFC and the associations around the Initiative Volksentscheid Fahrrad and was mostly positively received in the social media. The Berlin-Brandenburg regional association of the ADAC criticised the measure and said that the senate would exploit an emergency situation to pursue particular interests. The CDU and FDP also accused the senate of instrumentalizing the pandemic to turn traffic around. The AfD spoke of "left-wing car-hating policies" and pointed to a decrease in the number of cyclists compared to last year. The ADFC, on the other hand, stated that the total number of distances travelled in the Corona crisis had decreased overall and evaluations by the traffic information centre and public transport showed "that this was far more drastically the case with car traffic, buses and trains than with cycling". As other cities in Germany initially did not want to set up temporary cycle paths, Deutsche Umwelthilfe sent applications to 204 city administrations, whereupon the cities of Cologne, Frankfurt am Main and Dresden, among others, wanted to consider the option. In several cities, including Stuttgart, cycling associations organised campaigns calling for the creation of pop-up cycle paths. After a female cyclist coming from a pop-up cycle path was killed by a truck driver turning right at the intersection of Petersburger Straße and Mühsamstraße in June 2020, Siegfried Brockmann, head of the accident research department of the insurers, criticised that pop-up cycle paths alone would not provide a safe solution for the intersection areas as the main danger spots and would thus make people think they were safe. To achieve sufficient safety, the intersections would have to be rebuilt and the traffic lights changed. Brockmann also criticised the short-term installation of the cycle paths without prior measurement of the respective traffic flows. The senate administration replied that the police were involved in every installation of a pop-up cycle path, "in order to consider safety aspects of the respective location together with the road traffic authority". The situation at crossroads and junctions would only change as a result of the provisional cycle lanes to the extent that the visibility conditions would improve significantly in each case. Scientific impact analysis A 2021 case-control study of cities found that redistributing street space for "pop-up bike lanes" during the COVID-19 pandemic leads to large additional increases in cycling. These may have substantial environmental and health benefits. See also Impact of the COVID-19 pandemic on the environment#Cycling Tactical urbanism References Cycleways Transport infrastructure
Pop-up bicycle lane
[ "Physics" ]
1,502
[ "Physical systems", "Transport", "Transport infrastructure" ]
64,212,921
https://en.wikipedia.org/wiki/Hyper-IL-6
Hyper-IL-6 is a designer cytokine, which was generated by the German biochemist Stefan Rose-John. Hyper-IL-6 is a fusion protein of the four-helical cytokine Interleukin-6 and the soluble Interleukin-6 receptor which are covalently linked by a flexible peptide linker. Interleukin-6 on target cells binds to a membrane bound Interleukin-6 receptor. The complex of Interleukin-6 and the Interleukin-6 receptor associate with a second receptor protein called gp130, which dimerises and initiates intracellular signal transduction. Gp130 is expressed on all cells of the human body whereas the Interleukin-6 receptor is only found on few cells such as hepatocytes and some leukocytes. Neither Interleukin-6 nor the Interleukin-6 receptor have a measurable affinity for gp130. Therefore, cells, which only express gp130 but no Interleukin-6 receptor are not responsive to Interleukin-6. It was found, however, that the membrane-bound Interleukin-6 receptor can be cleaved from the cell membrane generating a soluble Interleukin-6 receptor. The soluble Interleukin-6 receptor can bind the ligand Interleukin-6 with similar affinity as the membrane-bound Interleukin-6 receptor and the complex of Interleukin-6 and the soluble Interleukin-6 receptor can bind to gp130 on cells, which only express gp130 but no Interleukin-6 receptor. The mode of signaling via the soluble Interleukin-6 receptor has been named Interleukin-6 trans-signaling whereas Interleukin-6 signaling via the membrane-bound Interleukin-6 receptor is referred to as Interleukin-6 classic signaling. Therefore, the generation of the soluble Interleukin-6 receptor enables cells to respond to Interleukin-6, which in the absence of soluble Interleukin-6 receptor would be completely unresponsive to the cytokine. Molecular construction of Hyper-IL-6 In order to generate a molecular tool to discriminate between Interleukin-6 classic signaling and Interleukin-6 trans-signaling, a cDNA coding for human Interleukin-6 and a cDNA coding for the human soluble Interleukin-6 receptor were connected by a cDNA coding for a 13 amino acids long linker, which was long enough to bridge the 40 Å distance between the COOH terminus of the soluble Interleukin-6 receptor and the NH2 terminus of human Interleukin-6. The generated cDNA was expressed in yeast cells and in mammalian cells and it was shown that. Use of Hyper-IL-6 to analyse IL-6 signaling Hyper-IL-6 has been used to test which cells depend on Interleukin-6 trans-signaling in their response to the cytokine Interleukin-6. To this end, cells were treated with Interleukin-6 and alternatively with Hyper-IL-6. Cells, which respond to Interleukin-6 alone do express an Interleukin-6 receptor whereas cells, which only respond to Hyper-IL-6 but not to Interleukin-6 alone depend in their response to the cytokine on Interleukin-6 trans-signaling. It turned out that hematopoietic stem cells, neural cells, smooth muscle cells and endothelial cells are typical target cells of Interleukin-6 trans-signaling. The concept of Interleukin-6 trans-signaling The Hyper-IL-6 protein has also been used to explore the physiologic role of Interleukin-6 trans-signaling in vivo. It turned out that this signaling mode was involved in many types of inflammation and cancer. Hyper-IL-6 has helped to establish the concept of Interleukin-6 trans-signaling. Interleukin-6 trans-signaling mediates the pro-inflammatory activities of Interleukin-6 whereas Interleukin-6 classic signaling governs the protective and regenerative Interleukin-6 activities. Recently, in breast cancer patients, it was shown with the help of Hyper-IL-6 that IL-6 trans-signaling via phosphoinositid-3-kinase signaling activates disseminated cancer cells long before metastases are formed. In addition, it was demonstrated in mice that Hyper-IL-6 transneuronal delivery enabled functional recovery after severe spinal cord injury. References Cytokines Biochemistry Interleukins
Hyper-IL-6
[ "Chemistry", "Biology" ]
1,003
[ "Biochemistry", "Cytokines", "nan", "Signal transduction" ]
64,213,701
https://en.wikipedia.org/wiki/Reysa%20Bernson
Reysa Verhaeghe or Reysa Bernson (September 26, 1904 - 1944) was a French astronomer. She founded the Association Astronomique du Nord in 1923. She was particularly known for her work popularizing astronomy in France through public demonstrations and the development of planetaria. Early life Reysa Bernson was born in Lille September 28, 1904. Her parents, Désiré Verhaeghe and Dweira Bernson-Verhaeghe, were both doctors. In 1920, her parents separated, though they did not divorce. After the separation, Reysa frequently went by her mother's family name, Verhaeghe. Her father died when she was 26. Education Reysa Bernson attended high school at the Lycée Fénelon de Lille. In 1921, she obtained her baccalauréat Sciences et Langues and completed her first year of Russian studies at the University of Lille. In 1922, she obtained her baccalauréat from the Faculté des Sciences. In 1923, she finished her undergraduate degree in Russian studies from the University of Lille. In 1924, she finished a degree in science. In the following years, she would complete certifications in a number of fields: astronomy, chemistry, and radiotelegraphy. In 1927, she earned a teaching certification. In 1934, she earned a doctorate in astronomy from the University of Lille. Bernson was active in student life at the University of Lille. Bernson participated in restarting the student publication Annales de l'Association des Étudiates de Lille. In 1926, Bernson was a delegate to the 15th Congrès de l'Union Nationale des Étudiants de France in Poitiers. The following year, she became treasurer of the newly-formed Lille student union. On April 27 of 1927, she was elected vice-president of the National Student Union, the only woman elected to a cabinet position in the organization. In 1928 and from 1931 to 1934, along with her studies, Bernson worked as préparateur, or athletic trainer, at the University of Lille. Career Reysa Bernson was passionate about astronomy from a young age. On May 5, 1920, at the age of 16, she was admitted to the Société Astronomique de France. In 1923, Reysa Bernson founded the Association Astronomique du Nord (AAN), an organization with the aim of founding an observatory and a library, as well as gathering amateur astronomers in the region. Bernson also worked to generally introduce students throughout the region to astronomy. in 1932, she won the Henry Rey prize from the Société Astronomique de France in recognition of her work popularizing astronomy. In 1936, Reysa Bernson, along with Henri Lhote, created an Éclaireurs de France scouting group, the Groupe Camille Flammarion, to encourage youth participation in astronomy. The group arranged star-watching excursions and learned to read star maps. In 1937, Reysa Bernson was the secretary general of the Planetarium of the Exposition Internationale in Paris, which saw around 800,000 visitors during the six months of the exhibition. Two notable visitors to the Planetarium were astronomers Armand Delsemme and Gérard de Vaucouleurs. In 1938, Bernson was recognized for her work during the exhibition in encouraging more people to join the society by being awarded two prizes from the Société Astronomique de France: the Prix de l'Observatoire de la Guette and the Médaille commémorative. On June 2, 1940, twelve days before the Nazis invaded Paris, Reysa Bernson met with other members of the Société Astronomique at Camille Flammarion's grave at the Camille Fammarion Observatory near the city. Archival records indicate that Bernson remained professionally active in some capacity during the Occupation up to 1943. Death On February 23, 1944, Reysa Bernson, along with her mother, Dweira Bernson-Verhaeghe, were arrested in Dreux, to the west of Paris, for being Jewish. Despite the efforts of friends in Lille to save them, the women were sent shortly afterward to a concentration camp in Drancy, with a one-day stop in Chartres. On May 7, Reysa Bernson and her mother were sent to Auschwitz, where they were killed by the Nazis. Memorials There is an asteroid named after Bernson: 21114 Bernson. By decision of the municipal council in 2018, there is an allée Reysa Bernson in the Saint-Maurice Pellevoisin quarter of the city of Lille. References 1904 births 1944 deaths Scientists from Lille 20th-century French astronomers Jewish astronomers French women astronomers 20th-century French women scientists Drancy internment camp prisoners French people who died in Auschwitz concentration camp French Jews who died in the Holocaust
Reysa Bernson
[ "Astronomy" ]
994
[ "Astronomers", "Jewish astronomers" ]
64,215,003
https://en.wikipedia.org/wiki/Di%C3%B3si%E2%80%93Penrose%20model
The Diósi–Penrose model was introduced as a possible solution to the measurement problem, where the wave function collapse is related to gravity. The model was first suggested by Lajos Diósi when studying how possible gravitational fluctuations may affect the dynamics of quantum systems. Later, following a different line of reasoning, Roger Penrose arrived at an estimation for the collapse time of a superposition due to gravitational effects, which is the same (within an unimportant numerical factor) as that found by Diósi, hence the name Diósi–Penrose model. However, it should be pointed out that while Diósi gave a precise dynamical equation for the collapse, Penrose took a more conservative approach, estimating only the collapse time of a superposition. The Diósi model In the Diósi model, the wave-function collapse is induced by the interaction of the system with a classical noise field, where the spatial correlation function of this noise is related to the Newtonian potential. The evolution of the state vector deviates from the Schrödinger equation and has the typical structure of the collapse models equations: where is the mass density function, with , and respectively the mass, the position operator and the mass density function of the -th particle of the system. is a parameter introduced to smear the mass density function, required since taking a point-like mass distribution would lead to divergences in the predictions of the model, e.g. an infinite collapse rate or increase of energy. Typically, two different distributions for the mass density have been considered in the literature: a spherical or a Gaussian mass density profile, given respectively by and Choosing one or another distribution does not affect significantly the predictions of the model, as long as the same value for is considered. The noise field in Eq. () has zero average and correlation given by where “” denotes the average over the noise. One can then understand from Eq. () and () in which sense the model is gravity-related: the coupling constant between the system and the noise is proportional to the gravitational constant , and the spatial correlation of the noise field has the typical form of a Newtonian potential. Similarly to other collapse models, the Diósi–Penrose model shares the following two features: The model describes a collapse in position. There is an amplification mechanism, which guarantees that more massive objects localize more effectively. In order to show these features, it is convenient to write the master equation for the statistical operator corresponding to Eq. (): It is interesting to point out that this master equation has more recently been re-derived by L. Diósi using a hybrid approach where quantized massive particles interact with classical gravitational fields. If one considers the master equation in the position basis, introducing with , where is a position eigenstate of the -th particle, neglecting the free evolution, one finds with where is the mass density when the particles of the system are centered at the points , ..., . Eq. () can be solved exactly, and one gets where As expected, for the diagonal terms of the density matrix, when , one has , i.e. the time of decay goes to infinity, implying that states with well-localized position are not affected by the collapse. On the contrary, the off-diagonal terms , which are different from zero when a spatial superposition is involved, will decay with a time of decay given by Eq. (). To get an idea of the scale at which the gravitationally induced collapse becomes relevant, one can compute the time of decay in Eq. () for the case of a sphere with radius and mass in a spatial superposition at a distance . Then the time of decay can be computed) using Eq. () with where . To give some examples, if one considers a proton, for which  kg and  m, in a superposition with , one gets  years. On the contrary, for a dust grain with  kg and  m, one gets one gets  s. Therefore, contrary to what might be expected considering the weaknesses of gravitational force, the effects of the gravity-related collapse become relevant already at the mesoscopic scale. Recently, the model have been generalized by including dissipative and non-Markovian effects. Penrose's proposal It is well known that general relativity and quantum mechanics, our most fundamental theories for describing the universe, are not compatible, and the unification of the two is still missing. The standard approach to overcome this situation is to try to modify general relativity by quantizing gravity. Penrose suggests an opposite approach, what he calls “gravitization of quantum mechanics”, where quantum mechanics gets modified when gravitational effects become relevant. The reasoning underlying this approach is the following one: take a massive system of well-localized states in space. In this case, the state being well-localized, the induced space–time curvature is well defined. According to quantum mechanics, because of the superposition principle, the system can be placed (at least in principle) in a superposition of two well-localized states, which would lead to a superposition of two different space–times. The key idea is that since space–time metric should be well defined, nature “dislikes” these space–time superpositions and suppresses them by collapsing the wave function to one of the two localized states. To set these ideas on a more quantitative ground, Penrose suggested that a way for measuring the difference between two space–times, in the Newtonian limit, is where is the Newtonian gravitational acceleration at the point where the system is localized around . The acceleration can be written in terms of the corresponding gravitational potentials , i.e. . Using this relation in Eq. (), together with the Poisson equation , with giving the mass density when the state is localized around , and its solution, one arrives at The corresponding decay time can be obtained by the Heisenberg time–energy uncertainty: which, apart for a factor simply due to the use of different conventions, is exactly the same as the time decay derived by Diósi's model. This is the reason why the two proposals are named together as the Diósi–Penrose model. More recently, Penrose suggested a new and quite elegant way to justify the need for a gravity-induced collapse, based on avoiding tensions between the superposition principle and the equivalence principle, the cornerstones of quantum mechanics and general relativity. In order to explain it, let us start by comparing the evolution of a generic state in the presence of uniform gravitational acceleration . One way to perform the calculation, what Penrose calls “Newtonian perspective”, consists in working in an inertial frame, with space–time coordinates and solve the Schrödinger equation in presence of the potential (typically, one chooses the coordinates in such a way that the acceleration is directed along the axis, in which case ). Alternatively, because of the equivalence principle, one can choose to go in the free-fall reference frame, with coordinates related to by and , solve the free Schrödinger equation in that reference frame, and then write the results in terms of the inertial coordinates . This is what Penrose calls “Einsteinian perspective”. The solution obtained in the Einsteinian perspective and the one obtained in the Newtonian perspective are related to each other by Since the two wave functions are equivalent apart from an overall phase, they lead to the same physical predictions, which implies that there are no problems in this situation where the gravitational field always has a well-defined value. However, if the space–time metric is not well defined, then we will be in a situation where there is a superposition of a gravitational field corresponding to the acceleration and one corresponding to the acceleration . This does not create problems as long as one sticks to the Newtonian perspective. However, when using the Einstenian perspective, it will imply a phase difference between the two branches of the superposition given by . While the term in the exponent linear in the time does not lead to any conceptual difficulty, the first term, proportional to , is problematic, since it is a non-relativistic residue of the so-called Unruh effect: in other words, the two terms in the superposition belong to different Hilbert spaces and, strictly speaking, cannot be superposed. Here is where the gravity-induced collapse plays a role, collapsing the superposition when the first term of the phase becomes too large. Further information on Penrose's idea for the gravity-induced collapse can be also found in the Penrose interpretation. Experimental tests and theoretical bounds Since the Diósi–Penrose model predicts deviations from standard quantum mechanics, the model can be tested. The only free parameter of the model is the size of the mass density distribution, given by . All bounds present in the literature are based on an indirect effect of the gravitational-related collapse: a Brownian-like diffusion induced by the collapse on the motion of the particles. This Brownian-like diffusion is a common feature of all objective-collapse theories and, typically, allows to set the strongest bounds on the parameters of these models. The first bound on was set by Ghirardi et al., where it was shown that  m to avoid unrealistic heating due to this Brownian-like induced diffusion. Then the bound has been further restricted to  m by the analysis of the data from gravitational wave detectors. and later to  m by studying the heating of neutron stars. Regarding direct interferometric tests of the model, where a system is prepared in a spatial superposition, there are two proposals currently considered: an optomechanical setup with a mesoscopic mirror to be placed in a superposition by a laser, and experiments involving superpositions of Bose–Einstein condensates. See also Measurement problem Interpretation of quantum mechanics Penrose interpretation Gravitational decoherence Wave function collapse Objective-collapse theory Ghirardi–Rimini–Weber theory References Quantum measurement Interpretations of quantum mechanics
Diósi–Penrose model
[ "Physics" ]
2,046
[ "Interpretations of quantum mechanics", "Quantum measurement", "Quantum mechanics" ]
64,216,555
https://en.wikipedia.org/wiki/OpenVSP
OpenVSP (also Open Vehicle Sketch Pad) — is an open-source parametric aircraft geometry tool originally developed by NASA. It can be used to create 3D models of aircraft and to support engineering analysis of those models. History Predecessors to OpenVSP including VSP and Rapid Aircraft Modeler (RAM) were developed by J.R. Gloudemans and others for NASA beginning in the early 1990s. OpenVSP v2.0 was released as open source under the NOSA license in January 2012. Development has been led by Rob McDonald since around 2012 and has been supported by NASA and AFRL among other contributions. OpenVSP allows the user to quickly generate computer models from ideas, which can then be analyzed. As such, it is especially powerful in generating and evaluating unconventional design concepts. Features User interface OpenVSP displays a graphical user interface upon launch, built with FLTK. A workspace window and a "Geometry Browser" window open. The workspace is where the model is displayed while the Geometry Browser lists individual components in the workspace, such as fuselage and wings. These components can be selected, added or deleted, somewhat like a feature tree in CAD software such as Solidworks. When a component is selected in the Geometry Browser window, a component geometry window opens. This window is used to modify the component. OpenVSP also provides API capabilities which may be accessed using Matlab, Python or AngelScript. Geometry modelling OpenVSP offers a multitude of basic geometries, common to aircraft modelling, which users modify and assemble to create models. Wing, pod, fuselage, and propeller are a few available geometries. Advanced components like body of revolution, duct, conformal geometry and such are also available. Analysis tools Besides the geometry modeler, OpenVSP contains multiple tools that help with aerodynamic or structural analysis of models. The tools available are: CompGeom - mesh generation tool that can handle model intersection and trimming Mass Properties Analysis - to compute properties like centre of gravity and moment of inertia Projected Area Analysis - to compute project area CFD Mesh - to generate meshes that may be used in Computational fluid dynamics analysis software FEA Mesh - to generate meshes that may be used in FEA analysis software DegenGeom - to generate various simplified representations of geometry models like point, beam and camber surface models VSPAERO - for vortex lattice or panel method based aerodynamic and flight dynamic analysis Wave Drag Analysis - for estimating wave drag of geometries Parasite Drag Analysis - for estimating parasite drag of geometries based on parameters like wetted area and skin friction coefficient Surface fitting - for fitting a parametric surface to a point cloud Texture Manager - for applying image textures to geometry for aiding visualization FEA Structure - for creating internal structures such as ribs and spars Compatibility with other software OpenVSP permits import of multiple geometry formats like STL, CART3D (.tri) and PLOT3D. Point clouds may also be imported and used to fit a parametric surface. Geometry created in OpenVSP may be exported as STL, CART3D (.tri), PLOT3D, STEP and IGES, OBJ, SVG, DXF and X3D file formats. These file formats allow geometries to be used for mesh generation and in CFD or FEA software. Community repository OpenVSP Hangar OpenVSP Hangar (also VSP Hangar) provides users a place to upload models and promotes sharing of geometry created in OpenVSP. Each model is allowed revisions with accompanying details on source quality. Since end of 2023, OpenVSP Hangar has been closed and no backup downloads has been provided. OpenVSP Airshow On 22 August 2024, OpenVSP Airshow (also VSP Airshow), a successor to OpenVSP Hangar, has been launched. OpenVSP Workshop OpenVSP Workshop — is an offline event where developers and users meet to discuss progress and use of OpenVSP. The Workshop has been held annually since 2012 (except 2018). The 2020 and 2021 Workshops were held online due to the COVID-19 pandemic. The 2024 Workshop was held at the Museum of Flight in Seattle. Papers, slides and other workshops materials published on OpenVSP wiki site in a few days after workshops ends. OpenVSP Ground School OpenVSP Ground School is a set of comprehensive tutorials under development by Brandon Litherland at NASA. Ground school tutorials provide details on OpenVSP features and techniques, along with tutorials for beginner and advanced users, and are hosted on the Langley Research Center website. See also Comparison of computer-aided design software XFOIL References External links VSP Airshow Free and open-source software Computer-aided design software Aerospace companies Langley Research Center NASA Computational fluid dynamics Finite element software
OpenVSP
[ "Physics", "Chemistry" ]
987
[ "Computational fluid dynamics", "Fluid dynamics", "Computational physics" ]
64,217,569
https://en.wikipedia.org/wiki/Jacqueline%20Krim
Jacqueline Krim is an American condensed matter physicist specializing in nanotribology, the study of film growth, friction, and wetting of nanoscale surfaces. She is a Distinguished University Professor of Physics at North Carolina State University. Education and career Krim graduated from the University of Montana in 1978 and completed a Ph.D. in experimental condensed matter physics at the University of Washington in 1984. After postdoctoral research at Aix-Marseille University, she became a faculty member at Northeastern University, and moved to North Carolina State University in 1998. Recognition Krim is a fellow of the American Vacuum Society (1999) and the American Physical Society (2000). The Division of Materials Physics of the American Physical Society named her as their David Adler Lecturer for 2015. In 2019 she was elected as a fellow of the American Association for the Advancement of Science "for distinguished contributions to the understanding of atomic-scale friction, wetting and surface roughening and for exemplary efforts in scientific outreach and diversity". She received a National Science Foundation Presidential Young Investigator Award in 1986. References External links Krim Nanoscale Tribology Group Year of birth missing (living people) Living people American women physicists University of Montana alumni University of Washington alumni Northeastern University faculty North Carolina State University faculty Fellows of the American Physical Society Fellows of the American Association for the Advancement of Science 20th-century American physicists 20th-century American women scientists 21st-century American physicists 21st-century American women scientists Condensed matter physicists Tribologists American women academics
Jacqueline Krim
[ "Materials_science" ]
306
[ "Tribology", "Tribologists" ]
64,218,162
https://en.wikipedia.org/wiki/Iodine%20tribromide
Iodine tribromide is an interhalogen with chemical formula IBr3. Properties Iodine tribromide is a dark brown liquid that is miscible with ethanol and ethers. Uses Iodine tribromide can be used as a brominated flame retardant when producing semiconductors. It also can be used in dry etching. References Bromides Iodine compounds Interhalogen compounds
Iodine tribromide
[ "Chemistry" ]
84
[ "Bromides", "Interhalogen compounds", "Oxidizing agents", "Salts" ]
64,221,048
https://en.wikipedia.org/wiki/Glass%20cone
A glass cone is a glass production structure historically unique to the United Kingdom. A glass cone had a large central furnace, a circular platform where the glassblowers worked, and smaller furnaces around its wall to ensure the glass did not cool too quickly. There are four surviving glass cones: Catcliffe Glass Cone, South Yorkshire Lemington Glass Works, Newcastle upon Tyne Northern Glass Cone, Alloa Glass Works, a Scottish scheduled monument, Red House Cone in Wordsley, granted listed building status in 1966 See also Bottle oven, a bottle-shaped kiln typical of Stoke-on-Trent References Glass production Industrial processes
Glass cone
[ "Materials_science", "Engineering" ]
127
[ "Glass engineering and science", "Glass production" ]
64,221,230
https://en.wikipedia.org/wiki/List%20of%20Stadia%20games
This is a list of games that were available for purchase on the Stadia cloud gaming service from Google, which has now been discontinued. At the time of the service's shuttering in January 2023, there were titles on this list. Of these, five were Stadia exclusives and are marked in yellow and with (§). Some of the exclusive titles have since been released elsewhere, although Outcasters was already in maintenance mode at the time of shutdown and its publisher stated it would not be re-released. References Notes Citations Stadia Stadia games
List of Stadia games
[ "Technology" ]
117
[ "Computing-related lists", "Google lists" ]
64,221,731
https://en.wikipedia.org/wiki/Kirkpatrick%E2%80%93Reisch%20sort
Kirkpatrick–Reisch sorting is a fast sorting algorithm for items with limited-size integer keys. It is notable for having an asymptotic time complexity that is better than radix sort. References Sorting algorithms
Kirkpatrick–Reisch sort
[ "Mathematics", "Technology" ]
47
[ "Sorting algorithms", "Computer science stubs", "Computer science", "Computing stubs", "Order theory" ]
64,222,643
https://en.wikipedia.org/wiki/Dictionary%20of%20Minor%20Planet%20Names
The Dictionary of Minor Planet Names is a reference book containing information about the discovery and naming of 12,804 asteroids (March 2006). It is published by the International Astronomical Union. Editions (5,252 names, 7,041 numbered until June 1996) (+10,000 names) (+17,000 names) (22,000 names) Astronomy books
Dictionary of Minor Planet Names
[ "Astronomy" ]
77
[ "Astronomy books", "Astronomy book stubs", "Works about astronomy", "Astronomy stubs" ]
64,224,034
https://en.wikipedia.org/wiki/Togni%20reagent%20II
Togni reagent II (1-trifluoromethyl-1,2-benziodoxol-3(1H)-one) is a chemical compound used in organic synthesis for direct electrophilic trifluoromethylation. History Synthesis, properties, and reactivity of the compound were first described in 2006 by Antonio Togni and his coworkers at ETH Zurich. The article also contains information on Togni reagent I (1,3-dihydro-3,3-dimethyl-1-(trifluoromethyl)-1,2-benziodoxole). Preparation The synthesis consists of three steps. In the first step, 2-iodobenzoic acid is oxidized by sodium periodate and cyclized to 1-hydroxy-1,2-benziodoxol-3(1H)-one. The target compound can then be obtained by acylation with acetic anhydride and subsequent substitution reaction with trifluoromethyltrimethylsilane. Alternatively, trichloroisocyanuric acid can be used as oxidant in the place of sodium periodate for a newer one-pot synthesis method. Properties Physical properties The compound crystallized in a monoclinic crystal structure. The space group is P21/n with four molecules in the unit cell. From the crystallographic data, a density of 2.365 g·cm−3  was deduced. Chemical properties Pure Togni reagent II is metastable at room temperature. Heating it above the melting point will lead to strong exothermic decomposition, in which trifluoroiodomethane (CF3I) is released. The heat of composition at a temperature of 149 °C and higher has been determined to be 502 J·g−1. From recrystallization in acetonitrile, small amounts of trifluoromethyl-2-iodobenzoate and 2-iodobenzyl fluoride were observed as decomposition products. Togni reagent II reacts violently with strong bases and acids, as well as reductants. In tetrahydrofuran, the compound polymerizes. Uses Togni reagent II is used for trifluoromethylation of organic compounds. For phenolates, the substitution takes place preferably in the ortho position. It is possible to obtain a second substitution by using an excess of Togni reagent II. Reactions with alcohols yield the corresponding trifluoromethyl ethers. Trifluoromethylation of alkenes is possible under copper catalysis. References Trifluoromethyl compounds Iodanes Lactones Benzene derivatives Reagents for organic chemistry
Togni reagent II
[ "Chemistry" ]
590
[ "Iodanes", "Oxidizing agents", "Reagents for organic chemistry" ]
68,546,181
https://en.wikipedia.org/wiki/Manganese%28II%29%20chlorate
Manganese(II) chlorate is an unstable chemical compound with the formula Mn(ClO3)2. It is unstable even in dilute solution. As a hexahydrate, it is solid below −18°C. Above this it melts, to form an extremely explosive pink liquid. Preparation Manganese(II) chlorate was produced by the reaction of manganese(II) sulfate and barium chlorate. The water was removed by boiling in vacuum. Then the temperature was lowered to -80°C which resulted in a pink solid. Then it was cleaned with liquid nitrogen and potassium hydroxide to remove the decomposition products. Properties Manganese(II) chlorate forms the hexahydrate when solid, the water that cannot be removed. It decomposes above 6°C, to manganese(IV) oxide, chlorine dioxide, and water. In liquid form it is very viscous and extremely explosive. When heated to room temperature, it explodes with a sharp report. References Manganese(II) compounds Chlorates
Manganese(II) chlorate
[ "Chemistry" ]
223
[ "Chlorates", "Salts" ]
68,546,653
https://en.wikipedia.org/wiki/Lorenz%20Bruno%20Puntel
Lorenz (Lorencino) Bruno Puntel (; ; 22 September 1935 – 16 July 2024) was a German philosopher of Brazilian descent, who established the school of structural-systematic philosophy. Professor emeritus at the University of Munich, Puntel was named as one of the great contemporary philosophers, articulating his ideas from the most varied traditions. Background Puntel studied philosophy, theology, philology and psychology in Munich, Innsbruck, Vienna, Paris, and Rome. He studied philosophy in Munich (1968) and in Catholic theology (1969) in Innsbruck. He became a professor at the Institute of Philosophy at the University of Munich in 1975. He was a student of Karl Rahner and studied with Martin Heidegger, whose philosophy concerned him throughout his life. Puntel died in Augsburg, Bavaria on 16 July 2024, at the age of 88. Philosophical work Puntel's thought tries to reconstruct the systematics of philosophy from a very unique viewpoint, which involves the elaboration of a theoretical language, abandoning the idea of a language of predicates. Puntel drew on sources ranging from G. W. Leibniz, German idealism, Heidegger's phenomenology, and even analytic philosophy. Awards From 1983, Puntel was a visiting professor at Pittsburgh, Harvard and Princeton. Retired in 2001, in 2016, he received an honorary doctorate from the Munich School of Philosophy. He also received the Findlay Book Prize in 2011. Bibliography Analogy and historicity. Philosophical-historical-critical attempt at the basic problem of metaphysics. Herder Verlag, Freiburg 1969. Presentation, method and structure. Investigations in the Unity of Systematic Philosophy of G. W. F. Hegel. Bouvier Verlag, Bonn 1973. Theories of Truth in Modern Philosophy. A critical and systematic presentation. Wissenschaftliche Buchgesellschaft, Darmstadt 1978, . 3rd edition 1993. (Editor, Introduction) The concept of truth. New attempts at explanation. Scientific Book Society, Darmstadt 1987, . Basics of a theory of truth. W. de Gruyter, Berlin / New York 1990, . Structure and being. A theoretical framework for a systematic philosophy. Mohr Siebeck Verlag, Tübingen 2006, . Being and God. A systematic approach in dealing with M. Heidegger, E. Levinas and J.-L. Marion. Mohr Siebeck Verlag, Tübingen 2010, . (with Emmanuel Tourpe) Philosophy as a systematic discourse. Dialogues about the basics of a theory of beings. Karl Alber, Freiburg im Breisgau 2014. References External links Homepage at the LMU 1935 births 2024 deaths 20th-century anthropologists 20th-century Brazilian male writers 20th-century Brazilian philosophers 20th-century German educational theorists 20th-century German essayists 20th-century German male writers 20th-century German philosophers 21st-century anthropologists 21st-century Brazilian male writers 21st-century German educational theorists 21st-century essayists 21st-century German male writers 21st-century German philosophers Analytic philosophers Brazilian educational theorists Brazilian essayists Brazilian people of German descent Critical theorists Epistemologists German male essayists German male non-fiction writers Heidegger scholars Literacy and society theorists Ludwig Maximilian University of Munich alumni Academic staff of the Ludwig Maximilian University of Munich Metaphilosophers Metaphysics writers Ontologists Phenomenologists Philosophers of culture German philosophers of education Philosophers of history Philosophers of language Philosophers of mathematics Philosophers of mind Philosophers of religion Philosophers of science Philosophers of social science Philosophy academics Political philosophers Social anthropologists Social philosophers Sociologists of education Sociologists of religion Structuralists Theorists on Western civilization Writers about religion and science
Lorenz Bruno Puntel
[ "Mathematics" ]
738
[ "Philosophers of mathematics" ]
68,551,030
https://en.wikipedia.org/wiki/GRL-0617
GRL-0617 is a drug which is one of the first compounds discovered that acts as a selective small-molecule inhibitor of the protease enzyme papain-like protease (PLpro) found in some human pathogenic viruses, including the coronavirus SARS-CoV-2. It has been shown to inhibit viral replication in silico and in vitro. See also 3CLpro-1 Ebselen GC376 References Antiviral drugs 1-Naphthyl compounds Anilines Benzamides
GRL-0617
[ "Biology" ]
112
[ "Antiviral drugs", "Biocides" ]
68,551,049
https://en.wikipedia.org/wiki/Hemerochory
Hemerochory (Ancient Greek ἥμερος, hemeros: 'tame, ennobled, cultivated, cultivated' and Greek χωρίς choris: separate, isolated), or anthropochory, is the distribution of cultivated plants or their seeds and cuttings, consciously or unconsciously, by humans into an area that they could not colonize through their natural mechanisms of spread, but are able to maintain themselves without specific human help in their new habitat. Hemerochory is one of the main propagation mechanisms of a plant. Hemerochoric plants can both increase and decrease the biodiversity of a habitat. Categorisation Hemerochoric plants are classified according to the manner of introduction into, for example: Ethelochory: the conscious introduction by seed or young plants. Speirochory: the unintentional introduction by contaminated seed. Examples are the true chamomile and the cornflower. Agochory: the introduction by unintentional transport with, among other things, ships, trains and cars. These plants are common in port areas, roadsides, stations and railways. Division Chronologically the hemerochoric plants are divided in: Archaeophytes: plants that were introduced before the onset of world trade around the year 1500, or before the year 1492 (discovery of America). Neophytes: plants that were introduced later. Related terms Anthropochory is often used synonymously but does not mean exactly the same. Anthropochory is the spread by humans. The spread through domestic animals does not belong to the anthropochoric, but to the hemerochoric, because domestic animals belong to the human culture. Strictly speaking, anthropochoric means the spread through humans as a transport medium. These can also be native species that were either adapted from the outset to locations created by human cultural activity or have adapted to them afterwards; As a result, their area of distribution has often, but not always, increased. The term adventitious plants is sometimes used synonymously with hemerochory, but is often restricted to species that were intentionally brought into the area and then naturalized, sometimes also for species that have not (yet) firmly established themselves in their new habitat. Cosmochory is the spread by humans or robots of seeds external to earth. History Hemerochorous spread of plants through human cultural activity very likely already happened in the Stone Age, but demonstrably at the latest in antiquity, namely along old trade routes. Fruits such as apples and pears gradually made their way along the Silk Road from the area around the Altai Mountains to Greece and from there to the gardens of the Romans, who in turn brought these cultivated plants to Central Europe, and some of these plants were eventually able to survive outside the culture. Many useful plants, such as tomato, potato, pumpkin and French bean did not reach Central Europe until the 16th century, after the American continent was discovered, and are now grown worldwide. In the last 400 to 500 years the spread has expanded through trade and military campaigns, through explorers and missionaries. The latter brought countless plants with them from their travels both out of an interest in exotic plants, which were often included in the plant collections of princely courts, and for purely scientific purposes. In the context of botanical studies, the interest was often in the possible healing effects of these plants, but also in the expansion of botanical knowledge, or the plants were only used for collecting (herbaria). Some ornamental plants also came to Europe because they promised a lucrative business. This applies, for example, to the camellias, one of which is also grown as a tea plant in Japan and China. While this species turned out to be not cultivable in Central Europe, people very quickly discovered the aesthetic appeal of the other camellia species as an ornamental plant. Botanical gardens played a major role in the acclimatization of such plants from distant habitats. Forms Agochory Agochoric plants are those that are spread through accidental transport. Unlike speirochoric plants, they are usually not sown on human-prepared soil. On land, agochoric plants used to be common in harbors, at train stations or along railway lines. However, mainly aquatic plants are spread through agochory. Ballast water plays a major role in the agochoric spread of aquatic plants. Around the world, around ten billion tons of seawater and the organisms it contains are shipped in this way. Exporting countries in particular are affected by the spread of organisms through ballast water. The ships arrive at the ports with empty cargo hold, but fully pumped ballast tanks. In the draining of this ballast water, these ports receive thousands of cubic meters of seawater brimming with alien creatures now in a new environment. The seaweed Undaria pinnatifida, which is native to the Japanese coast, reached the Tasmanian coast via ballast water and has formed dense kelp forests along the coast since 1988, displacing the native flora and fauna. Caulerpa taxifolia is one of those plants that are often spread by ballast water. It is also spread by the fact that ships tear off parts of the algae with their anchors. Australia was the first country to introduce a ballast water policy back in 1990 and is now the most determined to address this problem. Ships were asked not to take in ballast water in shallow and polluted bays and not to refuel with ballast water during the night, since then many marine organisms that are otherwise on the seabed rise to the surface of the water. Ships should also exchange their ballast water 200 kilometers away from the coastal waters, so that on the one hand the offshore species are not introduced into the more sensitive coastal waters and, on the other hand, no inhabitants of the coastal zone are transported to other continents. Ethelochory Ethelochory is intentional transportation of plants or seeds to different regions for agricultural and gardening purposes. Numerous crops that are important for human nutrition have been willingly spread by humans. Wheat, barley, lentil, beans, flax and poppy seeds, for example, are not typical plants for Central Europe, although they are all archaeotypes. People brought them after the beginning of the Neolithic (about 6,500 years ago) gradually from the eastern Mediterranean to central Europe and the rest of the world through the upcoming centuries. In central Europe, it is especially Cyperus esculentus which has been classified since the 1980s among the invasive species, because their tubers have been spread en masse, by sticking to vehicles or machines. Many of the old cultivated plants have spread around the world, primarily through emigrants from Europe. Grown for at least 4,000 years, wheat was introduced to America in the 16th century and Australia in the 19th century. Orange, lemons, apricots and peaches were originally native to China. They probably came via the Silk Road as early as the 3rd century BC. In Asia Minor and from there through the Romans to the Mediterranean. European settlers, in turn, used these species to grow fruit in suitable regions of America. From the 16th century, ornamental plants were grown more and more. Species native to Europe were first introduced as garden plants. These include, for example, the gladioli, the ornamental onion, European bluebell, the snowdrop native to southeast Europe and the common clematis. Ornamental plants from more distant regions were added later. From East Asia in particular, a number of plants were introduced to Europe as exotic or for economic reasons. Speirochory Some plants were unintentionally introduced in this process; this unwanted hemerochory as a seed companion is called speirochory. Since every seed also contains seeds of the herbs of the field from which it comes, their competitors, the "weeds", were also sold through the trade in the seeds of the useful plant. The real chamomile is one of the plants that were unintentionally spread as a companion to seeds. Speirochoric plants are sown on human-prepared soil and are competitors of the crops. Plants that are considered to be archaeophytes, such as the poppy, native to the Mediterranean area, the real chamomile, the cornflower and field buttercup, spread through the seeds with the grain in Central Europe. In the meantime, the seeds are cleaned more thoroughly using modern methods and the cultivation is hardly contaminated by pesticides or other control techniques. In spite of this, Cuscuta campestris, which is classified as a problematic weed in Australia, was accidentally imported into the country together with basil seeds in 1981, 1988 and 1990. See also Assisted colonization Escaped plant Volunteer plant References Plant reproduction Invasive species Environmental conservation Environmental terminology Habitat Botany Introduced plants
Hemerochory
[ "Biology" ]
1,797
[ "Behavior", "Plant reproduction", "Plants", "Reproduction", "Invasive species", "Botany", "Pests (organism)" ]
68,552,209
https://en.wikipedia.org/wiki/Oxalate%20phosphite
The oxalate phosphites are chemical compounds containing oxalate and phosphite anions. They are also called oxalatophosphites or phosphite oxalates. Oxalate phosphates can form metal organic framework compounds. Related compounds include the nitrite oxalates, arsenite oxalates, phosphate oxalates and oxalatophosphonates. The oxalate ion is rectangular and planar. The phosphite ion is shaped as a triangular pyramid. Because of high charge and stiff shape they will bridge across more than one cation, in particular those hard cations with a higher charge such as +3. Hydrogen can convert some of the oxygen on the anions to OH and reduce the charge. Many oxalate phosphite compounds have microporous structures where amines direct the structure formation. List References Oxalates Phosphites Mixed anion compounds
Oxalate phosphite
[ "Physics", "Chemistry" ]
195
[ "Ions", "Matter", "Mixed anion compounds" ]
68,552,763
https://en.wikipedia.org/wiki/Zettascale%20computing
Zettascale computing refers to computing systems capable of calculating at least "1021 IEEE 754 Double Precision (64-bit) operations (multiplications and/or additions) per second (zettaFLOPS)". It is a measure of supercomputer performance, and is a hypothetical performance barrier. A zettascale computer system could generate more single floating point data in one second than was stored by the total digital means on Earth in the first quarter of 2011. Definitions Floating point operations per second (FLOPS) are one measure of computer performance. FLOPS can be recorded in different measures of precision, however the standard measure (used by the TOP500 supercomputer list) uses 64 bit (double-precision floating-point format) operations per second using the High Performance LINPACK (HPLinpack) benchmark. Forecasts In 2018, Chinese scientists predicted that the first zettascale system will be assembled in 2035. This forecast looks plausible from a historical point of view as it took some 12 years to progress from the terascale machines (1012) to petascale systems (1015) and then 14 more years to move to exascale computers (1018). Scientists forecast that the zettascale systems are likely to be data-centric; this proposition means that the system components will move to the data, not vice versa, as the data volumes in the future are anticipated to be so large that moving data will be too expensive. It is also forecasted that zettascale systems are expected to be decentralized—because such a model can be the shortest route to achieving zettascale performance, with millions of less powerful components linked and working together to form a collective hypercomputer that is more powerful than any single machine. Such decentralized systems may be designed to mimick complex biologic systems, and the next cybernetic paradigm may be based on liquid cybernetic systems with embodied intelligence solutions. Potential configuration China’s National University of Defense Technology propose the following metrics: Power consumption: 100 MW Power efficiency: 10 teraflops/watt Peak performance per node: 10 petaflops Communication bandwidth between nodes: 1.6 terabits/second I/O bandwidth: 10 to 100 petabytes/second Storage capacity: 1.0 zettabyte Floor space: 1000 square meters Problems As Moore's law nears its natural limits, supercomputing will face serious physical problems in moving from exascale to zettascale systems, making the decade after 2020 a vital period to develop key high-performance computing techniques. Many forecasters, including Gordon Moore himself, expect Moore's law to end by around 2025. Another challenge for reaching zettascale performance can be enormous energy consumption. Applications Zettascale computers will be able to accurately forecast global weather for 2 weeks in the future. Zettascale calculations will also be able to significantly reduce the time required for astrophysical simulations of such rare phenomena as black holes, neutron star mergers, and supernovae. For example, the calculating of a 3D model of shock wave instability from a collapsing supernova core, which takes 1 million hours on petaflops computers and 1000 hours on exaflops machines, can be done in just one hour on zettaflops systems. Zettascale or yottascale systems might be able to accurately model the whole human brain. See also Computer performance by orders of magnitude Exascale computing Petascale computing List of hypothetical technologies References External links Perspectives on High-Performance Computing in a Big Data World Towards Zettascale Computing on Exascale Platforms Supercomputing
Zettascale computing
[ "Technology" ]
765
[ "Supercomputing" ]
68,553,796
https://en.wikipedia.org/wiki/Genetic%20vaccine
A genetic vaccine (also gene-based vaccine) is a vaccine that contains nucleic acids such as DNA or RNA that lead to protein biosynthesis of antigens within a cell. Genetic vaccines thus include DNA vaccines, RNA vaccines and viral vector vaccines. Properties Most vaccines other than live attenuated vaccines and genetic vaccines are not taken up by MHC-I-presenting cells, but act outside of these cells, producing only a strong humoral immune response via antibodies. In the case of intracellular pathogens, an exclusive humoral immune response is ineffective. Genetic vaccines are based on the principle of uptake of a nucleic acid into cells, whereupon a protein is produced according to the nucleic acid template. This protein is usually the immunodominant antigen of the pathogen or a surface protein that enables the formation of neutralizing antibodies that inhibit the infection of cells. Subsequently, the protein is broken down at the proteasome into short fragments (peptides) that are imported into the endoplasmic reticulum via the transporter associated with antigen processing, allowing them to bind to MHCI-molecules that are subsequently secreted to the cell surface. The presentation of the peptides on MHC-I complexes on the cell surface is necessary for a cellular immune response. As a result, genetic vaccines and live vaccines generate cytotoxic T-cells in addition to antibodies in the vaccinated individual. In contrast to live vaccines, only parts of the pathogen are used, which means that a reversion to an infectious pathogen cannot occur as it happened during the polio vaccinations with the Sabin vaccine. Administration Genetic vaccines are most commonly administered by injection (intramuscular or subcutaneous) or infusion, and less commonly and for DNA, by gene gun or electroporation. While viral vectors have their own mechanisms to be taken up into cells, DNA and RNA must be introduced into cells via a method of transfection. In humans, the cationic lipids SM-102, ALC-0159 and ALC-0315 are used in conjunction with electrically neutral helper lipids. This allows the nucleic acid to be taken up by endocytosis and then released into the cytosol. Applications Examples of genetic vaccines approved for use in humans include the RNA vaccines tozinameran and mRNA-1273, the DNA vaccine ZyCoV-D as well as the viral vectors AZD1222, Ad26.COV2.S, Ad5-nCoV, and Sputnik V. In addition, genetic vaccines are being investigated against proteins of various infectious agents, protein-based toxins, as cancer vaccines, and as tolerogenic vaccines for hyposensitization of type I allergies. History The first use of a viral vector for vaccination – a Modified Vaccinia Ankara Virus expressing HBsAg – was published by Bernard Moss and colleagues. DNA was used as a vaccine by Jeffrey Ulmer and colleagues in 1993. The first use of RNA for vaccination purposes was described in 1993 by Frédéric Martinon, Pierre Meulien and colleagues and in 1994 by X. Zhou, Peter Liljeström, and colleagues in mice. Martinon demonstrated that a cellular immune response was induced by vaccination with an RNA vaccine. In 1995, Robert Conry and colleagues described that a humoral immune response was also elicited after vaccination with an RNA vaccine. While DNA vaccines were more frequently researched in the early years due to their ease of production, low cost, and high stability to degrading enzymes, but sometimes produced low vaccine responses despite containing immunostimulatory CpG sites, more research was later conducted on RNA vaccines, whose immunogenicity was often better due to inherent adjuvants and which, unlike DNA vaccines, cannot insert into the genome of the vaccinated. Accordingly, the first RNA- and DNA-based vaccines approved for humans were RNA and DNA vaccines used as COVID vaccines. Viral vectors had previously been approved as ebola vaccines. References Vaccines Nucleic acid vaccines Gene delivery
Genetic vaccine
[ "Chemistry", "Biology" ]
847
[ "Genetics techniques", "Molecular biology techniques", "Vaccination", "Vaccines", "Gene delivery" ]
68,553,867
https://en.wikipedia.org/wiki/Trichloronitrosomethane
Trichloronitrosomethane is a chlorinated nitrosoalkane. It is a deep blue liquid with powerful lachrymatory effects. Synthesis Trichloronitrosomethane can be produced with following methods: Oxidation of trichloromethylsulfinic acid with nitric acid. Reaction of sodium trichloromethylsulfinate with sodium nitrite and sodium nitrate or potassium nitrate in sulfuric acid. Pyrolysis of trichloroacethydroxamic acid. Chemistry Trichloronitrosomethane is an unstable substance. It slowly decomposes into nitrosyl chloride, nitrogen oxides, and chloropicrin over time. Trichloronitrosomethane can be reduced to phosgene oxime by hydrogen sulfide. See also Chloropicrin Trifluoronitrosomethane Phosgene oxime References Nitroso compounds Trichloromethyl compounds Lachrymatory agents Pulmonary agents
Trichloronitrosomethane
[ "Chemistry" ]
215
[ "Lachrymatory agents", "Pulmonary agents", "Chemical weapons" ]
68,554,761
https://en.wikipedia.org/wiki/Super-app
A super-app, also known as an everything-app, is a mobile or web application that can provide multiple services including payment and instant messaging services, effectively becoming an all-encompassing self-contained commerce and communication online platform that embraces many aspects of personal and commercial life. Notable examples of super-apps include Tencent's WeChat in China, Tata Neu in India, and Grab in Southeast Asia. For end users, a super-app is an application that provides a set of core features while also giving access to independently developed miniapps. For app developers, a superapp is an application integrated with the capabilities of platforms and ecosystems that allows third-parties to develop and publish miniapps. History The super-app term was first used to describe WeChat, which was coined the first super-app when it combined the instant messaging service with the digital wallet function. Recognition of WeChat as a super-app stems from its combination of messaging, payments, e-commerce, and much more within a single application, making it indispensable for many users. WeChat's establishment of the super-app model has led companies like Meta (formerly Facebook) to try to build similar applications outside of China. In India, Tata Group has announced that it is currently developing a super app named Tata Neu. Major Indian companies like Paytm, PhonePe, and ITC Maars also have apps in development that might constitute super-apps. In Southeast Asia, Grab and Gojek lay claim to the super-app classification despite lacking many of the features offered by WeChat. Accordingly, growth-stage companies like Shopee, Traveloka, and AirAsia have also expanded the range of services offered by their respective applications. Notable examples X Elon Musk in 2025 started hiring to build X into a superapp. Rubika Iranian applications that are funded by Iranian regime have millions of users are referred to as superapps than startups. Alipay Alipay is a third-party mobile and online payment platform established in Hangzhou, China in February 2004 by Alibaba Group and its founder Jack Ma. It operates in association with Ant Group, an affiliate company of the Chinese Alibaba Group. Gojek Gojek is an Indonesian on-demand multi-service digital platform and fintech payment super-app. Established in Jakarta in 2010, as a call center to connect consumers to courier delivery and two-wheeled ride-hailing services, it launched its mobile app in 2015 with four services: GoRide, GoSend, GoShop, and GoFood, which has since expanded to offer over 20 services. In 2021, it merged with another Indonesian unicorn, Tokopedia, forming the decacorn GoTo Gojek Tokopedia. Grab Grab is a Southeast Asian technology company headquartered in Singapore and Indonesia. Founded in 2012 as the MyTeksi app in Kuala Lumpur, Malaysia, it expanded the following year as GrabTaxi, before moving its headquarters to Singapore in 2014 and rebranding officially as Grab. In addition to ride-hailing and transportation services, the company's mobile app also offers food delivery and digital payment services. Tata Neu Tata Neu is a multi-purpose super-app, developed in India by the Tata Group. It is the country's first super-app. The app was launched to coincide with the start of a 2022 Indian Premier League cricket match. WeChat WeChat is a Chinese multi-purpose instant messaging, social media and mobile payment app. First released in 2011, it became the world's largest standalone mobile app in 2018, with over 1 billion monthly active users. WeChat provides text messaging, hold-to-talk voice messaging, broadcast (one-to-many) messaging, video conferencing, video games, the sharing of photographs and videos and location sharing. Criticism Although apps that fit the super-app classification can offer users a wider variety of services in comparison to single-purpose alternatives, internet regulators in regions such as the US and Europe have become more concerned about the overall power of the technology industry and have become more critical of companies developing such apps. In China, WeChat and other local firms have been ordered to open up their platforms to rivals by local regulators. There are also reports that suggest it might be difficult to replicate WeChat's super-app model. This stems partly from the peaking of smartphone penetration rates in many regions worldwide, which have led to overcrowded app stores and tighter restrictions on targeted advertising as regulators assert more control over the companies. From a technical viewpoint, single-purpose apps are comparatively faster, more responsive and easier to navigate than super-apps, which help improve the overall user experience. Super-apps are also likelier to store larger amounts of personal data to facilitate the delivery of their services, and users run a greater risk of becoming victims of severe data breaches. In 2020, this unfolded with Tokopedia, when the data of 91 million users was stolen and shared by hackers. It has also been noted that a user who loses access to their account or is banned from a super-app generally loses access to multiple real-life services and digital applications. See also Closed platform References E-commerce 2020s neologisms
Super-app
[ "Technology" ]
1,091
[ "Information technology", "E-commerce" ]
68,556,278
https://en.wikipedia.org/wiki/S416
S416 (GTPL-11164) is a drug which acts as a selective inhibitor of the enzyme dihydroorotate dehydrogenase (DHODH). This enzyme is involved in the synthesis of pyrimidine nucleosides in the body, which are required for the synthesis of DNA and RNA. This is an important rate-limiting step in the replication of viruses, and so DHODH inhibitors may have applications as broad-spectrum antiviral drugs. In tests in vitro, S416 was found to have antiviral activity against a range of pathogenic RNA viruses including influenza, Zika virus, Ebola virus and SARS-CoV-2. See also Brequinar Teriflunomide Leflunomide References Antiviral drugs Enzyme inhibitors 2-Chlorophenyl compounds Thiazoles Benzoic acids
S416
[ "Biology" ]
182
[ "Antiviral drugs", "Biocides" ]
52,970,612
https://en.wikipedia.org/wiki/Krishnan%20Raghavachari
Krishnan Raghavachari (born 3 April 1953, in Chennai, India) is a Professor of Chemistry at Indiana University Bloomington. Raghavachari began his education in his native India, completing his undergraduate degree in 1973 at Madras University and his masters from the Indian Institute of Technology in 1975. Following this, he moved to the United States to attend Carnegie-Mellon University for his doctorate under the tutelage of John Pople, completing it in 1981. Upon completing his degree, Raghavachari entered the private sector as a research scientist at Bell Labs. He served as a member of the technical staff until 1987 when he was named a distinguished member. In 2002, he joined the faculty at Indiana University. Raghavachari has been credited as one of the top quantum chemists in the United States and responsible for developing methods allowing for widespread use of computational chemistry. Among the methods he has developed over his career are CCSD(T), used to evaluate bond energies and the properties of molecules and the Gaussian-2, 3, and 4 methods. Over the course of his career, Raghavachari has given over 150 invited lectures, published over 320 scientific papers, and has been cited over 50,000 times by others in the field. He has also served as chair of the Theoretical Chemistry Subdivision of the American Chemical Society, on the editorial boards of the Journal of Physical Chemistry, Journal of Computational Chemistry, Theoretical Chemistry Accounts, and Journal of Materials Research. Honours and awards 2009 Davisson-Germer Prize in Surface Physics, American Physical Society 2008 Fellow of the Royal Society of Chemistry 2001 Fellow of the American Physical Society References 1953 births Living people People from Chennai IIT Madras alumni Indiana University Bloomington staff University of Madras alumni Carnegie Mellon University alumni 21st-century American chemists Theoretical chemists Members of the International Academy of Quantum Molecular Science Fellows of the American Physical Society Fellows of the Royal Society of Chemistry Indian emigrants to the United States American academics of Indian descent 21st-century Indian chemists
Krishnan Raghavachari
[ "Chemistry" ]
406
[ "Theoretical chemists", "American theoretical chemists" ]
52,971,323
https://en.wikipedia.org/wiki/Poseltinib
Poseltinib (HM71224, LY3337641) is an experimental Bruton's tyrosine kinase inhibitor for the treatment of rheumatoid arthritis. It was developed by Hanmi Pharmaceutical and licensed to Eli Lilly. Phase II clinical trials began in August 2016 in patients with rheumatoid arthritis. Additional phase II trials are planned for treatment of lupus, lupus nephritis, Sjögren's syndrome, and other immunological conditions. References Experimental drugs Acrylamides Phenylpiperazines Pyrimidines
Poseltinib
[ "Chemistry" ]
125
[ "Pharmacology", "Pharmacology stubs", "Medicinal chemistry stubs" ]
52,971,419
https://en.wikipedia.org/wiki/Align%20Biopharma
Align Biopharma is an initiative by six of the largest global pharmaceutical companies to create standards for physicians to access information on drugs. The companies participating include Allergan, AstraZeneca, Biogen, GlaxoSmithKline, Novartis and Pfizer, and the software company Veeva Systems. The group is led by vice presidents and technology officers from each participating company. It was created to address the expansion of personalized medicine and specialized drugs, which are not frequently prescribed and tailored prescribing information is needed for each patient. The stated goals of Align are to allow physicians to use a single login to access pharmaceutical information for all six companies, and create a standard for consenting to receive information from pharmaceutical companies. Currently, companies individually verify the health care credentials of each provider before allowing them access to information. References Pharmaceutical industry
Align Biopharma
[ "Chemistry", "Biology" ]
176
[ "Pharmacology", "Life sciences industry", "Pharmaceutical industry", "Medicinal chemistry stubs", "Pharmacology stubs" ]
52,971,422
https://en.wikipedia.org/wiki/Iota2%20Fornacis
{{DISPLAYTITLE:Iota2 Fornacis}} Iota2 Fornacis is the Bayer designation for a star in the southern constellation of Fornax. It is faintly visible to the naked eye on a dark night, having an apparent visual magnitude of 5.83. The distance to this star, based upon an annual parallax shift of 29.46 mas, is around 111 light-years. It is a member of the thin disk population of the Milky Way galaxy. This is an F-type main-sequence star with a stellar classification of . The suffix notation indicates that absorption lines of iron and the carbon–hydrogen G-band are abnormally weak. It has an estimated 1.42 times the mass of the Sun and 1.4 times the Sun's radius. The star is around 3.67 billion years old, and is spinning with a leisurely projected rotational velocity of 4.7 km/s. Iota2 Fornacis has a common proper motion companion, a magnitude 13.74 star at a position angle of 81.80 arcseconds along a position angle of 6°. This object has about 35% of the Sun's mass. References F-type main-sequence stars Fornax Fornacis, Zeta CD-30 973 016538 12288 777
Iota2 Fornacis
[ "Astronomy" ]
275
[ "Fornax", "Constellations" ]
52,971,772
https://en.wikipedia.org/wiki/Kin%20punishment
Kin punishment is the practice of punishing the family members of someone who is accused of committing a crime, either in place of or in addition to the perpetrator of the crime. It refers to the principle in which a family shares responsibility for a crime which is committed by one of its members, and it is a form of collective punishment. Kin punishment has been used as a form of extortion, harassment, and persecution by authoritarian and totalitarian states. Kin punishment has been practiced historically in Nazi Germany, China, Japan, and South Korea; and presently in North Korea. Traditional examples Europe Traditional Irish law required the payment of a tribute (Éraic) in reparation for murder or other major crimes. In the case of homicide, if the attacker fled, the fine had to be paid by the tribe to which he belonged. In medieval Welsh law, the kin of an offender was liable to make compensation for his wrongful act. This penalty (called Galanas) was generally limited to murder. The medieval Polish Główszczyzna fine functioned similarly to the Anglo-Saxon and Scandinavian weregild. Arabic states Traditional Arab society, which is clan-based, strongly adheres to the concept of collective responsibility. Bedouins recognize two main forms of penalty for a crime against a member. These are blood revenge, referred to as Qisas (قصا, "revenge") and blood money, Diyya (دية, "blood money"/"ransom"). In cases of severe crimes such as murder and rape, blood revenge is the prescribed punishment. If a murder occurs, clansmen of the victim have the right to kill the murderer or one of his male clansmen with impunity. Certain crimes justify multiple acts of revenge, for example, the murder of women and children is avenged fourfold. Crimes considered treacherous, such as the murder of a guest, are also avenged fourfold. Alternatively, a crime punishable by blood revenge can be commuted to a severe fine if the family of the offended party agrees to it. Blood money is paid jointly by the clan of the offending member to the clan of the victimized member. Bedouins differentiate between crimes in which the group must pay as a standing obligation without reimbursement from the perpetrator of the offense, and crimes where the latter must reimburse them. Crimes where the clan is obligated to pay a joint fee without any reimbursement are murder, violent assault, or insults and other offenses committed during a violent conflict. The collective payment of fines for such crimes is viewed as a justified contribution to the welfare of the injured party rather than a penalty to the perpetrator. Other offenses given a blood-price are crimes against property and crimes against honor. Concepts based on the Arabian laws of blood revenge and blood money are found in Islamic Sharia law, and are thus variously adhered to in Islamic states. After the Saudi Arabian trial and verdict of the assassination of Jamal Khashoggi, which the United Nations rapporteur on summary executions, Agnès Callamard, called "extrajudicial execution for which the state of Saudi Arabia is responsible,” Khashoggi's sons pardoned the five convicted officials on 22 May 2020, which means the officials would not be executed but their blood money will be paid to Khashoggi's family. China China historically adhered to the concept of liability among blood relatives. During the Qin and Han dynasties, families were subject to various punishments according to the punishment of the offending member. When the offense was punishable by death by severing the body at the waist, the offender's parents, siblings, spouse, and children were executed. When the offense was punishable by death and public display of the body, the offender's family was subject to imprisonment with hard labor. When the offender's sentence was exile, their kin was exiled along with them. The most severe punishment, given for capital offenses, was the nine familial exterminations (zú zhū (族誅), literally "family execution", and miè zú (灭族/滅族)), implemented by tyrannical rulers. This punishment entailed the execution of all the close and extended kin of the individual, categorized into nine groups: four generations of the paternal line, three from the maternal line, and two from the wife's. In the case of Confucian scholar Fang Xiaoru, his students and peers were uniquely included as a tenth group. During the Ming dynasty of China (1368–1644), 16 palace women attempted to assassinate the Jiajing Emperor. All were sentenced to death by slow slicing. Ten members of the women's families were also beheaded, while a further 20 were enslaved and gifted to ministers. Collective punishment was officially repealed by the government of the Qing dynasty (1644–1912) in 1905. Modern examples Nazi Germany In traditional Germanic law, the law of Germanic peoples (before the widespread adoption of Roman canon law) accepted that the clan of a criminal was liable for offenses committed by one of its members. In Nazi Germany, this concept was revived so that the relatives of persons accused of crimes against the state, including desertion, were held responsible for those crimes. North Korea Numerous testimonies of North Korean defectors confirm the practice of kin punishment (연좌제, yeonjwaje literally "association system") in North Korea, under which three to eight generations of a political offender's family can be summarily imprisoned or executed. Such punishment is based on internal Workers' Party protocols and lies outside the formal legal system. Relatives are not told why they fell under suspicion and the punishment extends to children born in prison. The association system was introduced with the North Korean state's founding in 1948, having previously existed under the Joseon kingdom. Israel In the State of Israel, there is a legal prohibition against punishing relatives for the actions of a family member. This principle is based on the idea that individuals should not be held accountable for the actions of others, especially their family members, who are presumed to be innocent unless proven otherwise. Despite this, pro-Palestinian groups accuse the demolition of the homes of terrorists, which directly led to deaths, as a form of kin punishment, since family members are also at risk of being harmed by this measure. The practice an appeals process against demolition was established in 1989, and consequently the number of demolitions declined. However, in an effort to prevent Palestinian suicide attacks, targeting Israeli civilians during the Second Intifada, the Supreme Court of Israel in July 2002 upheld the legality of expelling to Gaza family members that they had assisted in carrying out the attack, or had supported the terrorist's activities. It was limited the use of expulsion to cases where "that person, by his own deeds, constitutes a danger to security of the state." Expulsion to Gaza was discontinued after Israeli disengagement from the Gaza Strip., and this policy has been carried out only in cases where was involved murder. Venezuela The Independent International Fact-Finding Mission on Venezuela concluded in a September 2021 report that Venezuelan security and intelligence agents reportedly applied the principle of Sippenhaftung, using methods including and kidnapping and detention of relatives of critics, real or perceived, to accomplish arrests. See also Ancestral sin Blood vengeance Bloodline theory Cycle of violence Eye for an eye Family members of a traitor to the Motherland Guilt by association Nine familial exterminations References Collective punishment Determinism Punishment Political and cultural purges Victims of familial execution
Kin punishment
[ "Biology" ]
1,559
[ "Behavior", "Human behavior", "Kinship and descent" ]
52,972,490
https://en.wikipedia.org/wiki/European%20Forest%20Genetic%20Resources%20Programme
European Forest Genetic Resources Programme (EUFORGEN) is an international network that promotes the conservation and sustainable use of forest genetic resources in Europe as an integral part of sustainable forest management. It was established in 1994 as a result of a resolution adopted in 1990 by the first Ministerial Conference of the Forest Europe process. The programme's tasks include coordination and promotion of in situ and ex situ conservation of forest genetic resources, facilitation of the exchange of information, and increasing public awareness of the need to conserve forest genetic resources. EUFORGEN is funded by member countries and operates through working groups formed by experts from across Europe who meet to exchange knowledge, analyse policies and practice, and develop science-based strategies to improve the management of forest genetic resources. EUFORGEN was established in 1994. Its secretariat, hosted by the European Forest Institute, is located in Barcelona, Spain. Member countries , there are 28 member countries in Europe: Austria Belgium Croatia Czech Republic Denmark Estonia Finland France Germany Hungary Iceland Ireland Italy Lithuania Luxembourg Malta Netherlands Norway Poland Portugal Serbia Slovakia Slovenia Spain Sweden Switzerland Ukraine United Kingdom References Biology in Europe Forest conservation organizations Genetics databases Organisations based in Barcelona Plant breeding organizations Plant genetics Seed associations Trees of Europe
European Forest Genetic Resources Programme
[ "Biology" ]
242
[ "Plants", "Plant genetics" ]
52,973,193
https://en.wikipedia.org/wiki/Nano-FTIR
Nano-FTIR (nanoscale Fourier transform infrared spectroscopy) is a scanning probe technique that utilizes as a combination of two techniques: Fourier transform infrared spectroscopy (FTIR) and scattering-type scanning near-field optical microscopy (s-SNOM). As s-SNOM, nano-FTIR is based on atomic-force microscopy (AFM), where a sharp tip is illuminated by an external light source and the tip-scattered light (typically back-scattered) is detected as a function of tip position. A typical nano-FTIR setup thus consists of an atomic force microscope, a broadband infrared light source used for tip illumination, and a Michelson interferometer acting as Fourier-transform spectrometer. In nano-FTIR, the sample stage is placed in one of the interferometer arms, which allows for recording both amplitude and phase of the detected light (unlike conventional FTIR that normally does not yield phase information). Scanning the tip allows for performing hyperspectral imaging (i.e. complete spectrum at every pixel of the scanned area) with nanoscale spatial resolution determined by the tip apex size. The use of broadband infrared sources enables the acquisition of continuous spectra, which is a distinctive feature of nano-FTIR compared to s-SNOM. Nano-FTIR is capable of performing infrared (IR) spectroscopy of materials in ultrasmall quantities and with nanoscale spatial resolution. The detection of a single molecular complex and the sensitivity to a single monolayer has been shown. Recording infrared spectra as a function of position can be used for nanoscale mapping of the sample chemical composition, performing a local ultrafast IR spectroscopy and analyzing the nanoscale intermolecular coupling, among others. A spatial resolution of 10 nm to 20 nm is routinely achieved. For organic compounds, polymers, biological and other soft matter, nano-FTIR spectra can be directly compared to the standard FTIR databases, which allows for a straightforward chemical identification and characterization. Nano-FTIR does not require special sample preparation and is typically performed under ambient conditions. It uses an AFM operated in noncontact mode that is intrinsically nondestructive and sufficiently gentle to be suitable for soft-matter and biological sample investigations. Nano-FTIR can be utilized from THz to visible spectral range (and not only in infrared as its name suggests) depending on the application requirements and availability of broadband sources. Nano-FTIR is complementary to tip-enhanced Raman spectroscopy (TERS), SNOM, AFM-IR and other scanning probe methods that are capable of performing vibrational analysis. Basic principles Nano-FTIR is based on s-SNOM, where the infrared beam from a light source is focused onto a sharp, typically metalized AFM tip and the backscattering is detected. The tip greatly enhances the illuminating IR light in the nanoscopic volume around its apex, creating a strong near field. A sample, brought into this near field, interacts with the tip electromagnetically and modifies the tip (back)scattering in the process. Thus by detecting tip scattering, one can obtain information about the sample. Nano-FTIR detects the tip-scattered light interferometrically. The sample stage is placed into one arm of a conventional Michelson interferometer, while a mirror on a piezo stage is placed into another, reference arm. Recording the backscattered signal while translating the reference mirror yields an interferogram. The subsequent Fourier transform of this interferogram returns the near-field spectra of the sample. Placement of the sample stage into one of the interferometer's arms (instead of outside of the interferometer as typically implemented in conventional FTIR) is a key element of nano-FTIR. It boosts the weak near-field signal due to interference with the strong reference field, helps to eliminate the background caused by parasitic scattering off everything that falls into large diffraction-limited beam focus, and most importantly, allows for recording of both amplitude s and phase φ spectra of the tip-scattered radiation. With the detection of phase, nano-FTIR provides complete information about near fields, which is essential for quantitative studies and many other applications. For example, for soft matter samples (organics, polymers, biomaterials, etc.), φ directly relates to the absorption in the sample material. This permits a direct comparison of nano-FTIR spectra with conventional absorption spectra of the sample material, thus allowing for simple spectroscopic identification according to standard FTIR databases. History Nano-FTIR was first described in 2005 in a patent by Ocelic and Hillenbrand as Fourier-transform spectroscopy of tip-scattered light with an asymmetric spectrometer (i.e. the tip/sample placed inside one of the interferometer arms). The first realization of s-SNOM with FTIR was demonstrated in 2006 in the laboratory of F. Keilmann using a mid-infrared source based on a simple version of nonlinear difference-frequency generation (DFG). However, the mid-IR spectra in this realization were recorded using dual comb spectroscopy principles, yielding a discrete set of frequencies and thus demonstrating a multiheterodyne imaging technique rather than nano-FTIR. The first continuous spectra were recorded only in 2009 in the same laboratory using a supercontinuum IR beam also obtained by DFG in GaSe upon superimposing two pulsed trains emitted from Er-doped fiber laser. This source further allowed in 2011 for the first assessment of nanoscale-resolved spectra of SiC with excellent quality and spectral resolution. At the same time, Huth et al. in the laboratory of R. Hillenbrand used IR radiation from a simple glowbar source in combination with the principles of Fourier-transform spectroscopy, to record IR spectra of p-doped Si and its oxides in a semiconductor device. In the same work the term nano-FTIR was first introduced. However, an insufficient spectral irradiance of glowbar sources limited the applicability of the technique to the detection of strongly-resonant excitations such phonons; and the early supercontinuum IR laser sources, while providing more power, had very narrow bandwidth (<300 cm−1). Further attempt to improve the spectral power, while retaining the large bandwidth of a glowbar source was made by utilizing the IR radiation from a high temperature argon arc source (also known as plasma source). However, due to lack of commercial availability and rapid development of the IR supercontinium laser sources, plasma sources are not widely utilized in nano-FTIR. The breakthrough in nano-FTIR came upon the development of high-power broadband mid-IR laser sources, which provided large spectral irradiance in a sufficiently large bandwidth (mW-level power in ~1000 cm-1 bandwidth) and enabled truly broadband nanoscale-resolved material spectroscopy capable of detecting even the weakest vibrational resonances. Particularly, it has been shown that nano-FTIR is capable of measuring molecular fingerprints which match well with far-field FTIR spectra, owing to the asymmetry of the nano-FTIR spectrometer that provides phase and thus gives access to the molecular absorption. Recently, the first nanoscale-resolved infrared hyperspectral imaging of a co-polymer blend was demonstrated, which allowed for the application of statistical techniques such as multivariate analysis – a widely used tool for heterogeneous sample analysis. Additional boost to the development of nano-FTIR came from the utilization of the synchrotron radiation that provide extreme bandwidth, yet at the expense of weaker IR spectral irradiance compared to broadband laser sources. Commercialization The nano-FTIR technology has been commercialized by neaspec – a Germany-based spin-off company of the Max Planck Institute of Biochemistry founded by Ocelic, Hillenbrand and Keilmann in 2007 and based on the original patent by Ocelic and Hillenbrand. The detection module optimized for broadband illumination sources was first made available in 2010 as a part of the standard neaSNOM microscope system. At this time, broadband IR-lasers have not been yet commercially available, however experimental broadband IR-lasers prove that the technology works perfect and that it has a huge application potential across many disciplines. The first nano-FTIR was commercially available in 2012 (supplied with still experimental broadband IR-laser sources), becoming the first commercial system for broadband infrared nano-spectroscopy. In 2015 neaspec develops and introduces Ultrafast nano-FTIR, the commercial version of ultrafast nano-spectroscopy. Ultrafast nano-FTIR is a ready-to-use upgrade for nano-FTIR to enable pump-probe nano-spectroscopy at best-in-class spatial resolution. The same year the development of a cryo-neaSNOM – the first system of its kind to enable nanoscale near-field imaging & spectroscopy at cryogenic temperatures – was announced. Advanced capabilities Synchrotron beamlines integration Nano-FTIR systems can be easily integrated into synchrotron radiation beamlines. The use of synchrotron radiation allows for acquisition of an entire mid-infrared spectrum at once. Synchrotrons radiation has already been utilized in synchrotron infrared microscopectroscopy - the technique most widely used in biosciences, providing information on chemistry on microscales of virtually all biological specimens, like bone, plants, and other biological tissues. Nano-FTIR brings the spatial resolution to 10-20 nm scale (vs. ~2-5 μm in microspectroscopy), which has been utilized for broadband spatially-resolved spectroscopy of crystalline and phase-change materials, semiconductors, minerals, biominerals and proteins. Ultrafast spectroscopy Nano-FTIR is highly suitable for performing local ultrafast pump-probe spectroscopy due to intereferometric detection and an intrinsic ability to vary the probe delay time. It has been applied for studies of ultrafast nanoscale plasmonic phenomena in Graphene, for performing nanospectroscopy of InAs nanowires with subcycle resolution and for probing the coherent vibrational dynamics of nanoscopic ensembles. Quantitative studies The availability of both amplitude and phase of the scattered field and theoretically well understood signal formation in nano-FTIR allow for the recovery of both real and imaginary parts of the dielectric function, i.e. finding the index of refraction and the extinction coefficient of the sample. While such recovery for arbitrarily-shaped samples or samples exhibiting collective excitations, such as phonons, requires a resource-demanding numerical optimization, for soft matter samples (polymers, biological matter and other organic materials) the recovery of the dielectric function could often be performed in real time using fast semi-analytical approaches. One of such approaches is based on the Taylor expansion of the scattered field with respect to a small parameter that isolates the dielectric properties of the sample and allows for a polynomial representation of measured near-field contrast. With an adequate tip-sample interaction model and with known measurement parameters (e.g. tapping amplitude, demodulation order, reference material, etc.), the sample permittivity can be determined as a solution of a simple polynomial equation Subsurface analysis Near-field methods, including nano-FTIR, are typically viewed as a technique for surface studies due to short probing ranges of about couple tip radii (~20-50 nm). However it has been demonstrated that within such probing ranges, s-SNOM is capable of detecting subsurface features to some extents, which could be used for the investigations of samples capped by thin protective layers, or buried polymers, among others. As a direct consequence of being quantitative technique (i.e. capable of highly reproducible detection of both near-field amplitude & phase and well understood near-field interaction models), nano-FTIR also provides means for the quantitative studies of the sample interior (within the probing range of the tip near field, of course). This is often achieved by a simple method of utilizing signals recorded at multiple demodulation orders naturally returned by nano-FTIR in the process of background suppression. It has been shown that higher harmonics probe smaller volumes below the tip, thus encoding the volumetric structure of a sample. This way, nano-FTIR has a demonstrated capability for the recovery of thickness and permittivity of layered films and nanostructures, which has been utilized for the nanoscale depth profiling of multiphase materials and high-Tc cuprate nanoconstriction devices patterned by focused ion beams. In other words, nano-FTIR has a unique capability of recovering the same information about thin-film samples that is typically returned by ellipsometry or impedance spectroscopy, yet with nanoscale spatial resolution. This capability proved crucial for disentangling different surface states in topological insulators. Operation in liquid Nano-FTIR uses scattered IR light to obtain information about the sample and has the potential to investigate electrochemical interfaces in-situ/operando and biological (or other) samples in their natural environment, such as water. The feasibility of such investigations has already been demonstrated by acquisition of nano-FTIR spectra through a capping Graphene layer on top of a supported material or through Graphene suspended on a perforated silicon nitride membrane (using the same s-SNOM platform that nano-FTIR utilizes). Cryogenic environment Reveling the fundamentals of phase transitions in superconductors, correlated oxides, Bose-Einstein condensates of surface polaritons, etc. require spectroscopic studies at the characteristically nanometer length scales and in cryogenic environment. Nano-FTIR is compatible with cryogenic s-SNOM that has already been utilized for reveling a nanotextured coexistence of metal and correlated Mott insulator phases in Vanadium oxide near the metal-insulator transition. Special atmosphere environments Nano-FTIR can be operated in different atmospheric environments by enclosing the system into an isolated chamber or a glove box. Such operation has already been used for the investigation of highly reactive Lithium-ion battery components. Applications Nano-FTIR has a plenitude of applications, including polymers and polymer composites, organic films, semiconductors, biological research (cell membranes, proteins structure, studies of single viruses), chemistry and catalysis, photochemistry, minerals and biominerals, geochemistry, corrosion and materials sciences, low-dimensional materials, photonics, energy storage, cosmetics, pharmacology and environmental sciences. Materials and chemical sciences Nano-FTIR has been used for the nanoscale spectroscopic chemical identification of polymers and nanocomposites, for in situ investigation of structure and crystallinity of organic thin films, for strain characterization and relaxation in crystalline materials and for high-resolution spatial mapping of catalytic reactions, among others. Biological and pharmaceutical sciences Nano-FTIR has been used for investigation of protein secondary structure, bacterial membrane, detection and studies of single viruses and protein complexes. It has been applied to the detection of biominerals in bone tissue. Nano-FTIR, when coupled with THz light, can also be applied to cancer and burn imaging with high optical contrast. Semiconductor industry and research Nano-FTIR has been used for nanoscale free carrier profiling and quantification of free carrier concentration in semiconductor devices, for evaluation of ion beam damage in nanoconstriction devices, and general spectroscopic characterization of semiconductor materials. Theory High-harmonic demodulation for background suppression The nano-FTIR interferometrically detects light scattered from the tip-sample system, . The power at the detector can be written as where is the reference field. The scattered field can be written as and is dominated by parasitic background scattering, , from the tip shaft, cantilever sample roughness and everything else which falls into the diffraction-limited beam focus. To extract the near-field signal, , originating from the "hot-spot" below the tip apex (which carries the nanoscale-resolved information about the sample properties) a small harmonic modulation of the tip height H (i.e. oscillating the tip) with frequency Ω is provided and the detector signal is demodulated at higher harmonics of this frequency nΩ with n=1,2,3,4,... The background is nearly insensitive to small variations of the tip height and is nearly eliminated for sufficiently high demodulation orders (typically ). Mathematically this can be shown by expanding and into a Fourier series, which yields the following (approximated) expression for the demodulated detector signal: where is the complex-valued number that is obtained by combining the lock-in amplitude, , and phase, , signals, is the n-th Fourier coefficient of the near-field contribution and C. C. stands for the complex conjugate terms. is the zeroth-order Fourier coefficient of the background contribution and is often called the multiplicative background because it enters the detector signal as a product with . It cannot be removed by the high-harmonic demodulation alone. In nano-FTIR the multiplicative background is eliminated as described below. Asymmetric FTIR spectrometer To acquire a spectrum, the reference mirror is continuously translated while recording the demodulated detector signal as a function of the reference mirror position , yielding an interferogram . This way the phase of the reference field changes according to for each spectral component of the reference field and the detector signal can thus be written as where is the reference field at zero delay . To obtain the nano-FTIR spectrum, , the interferogram is Fourier-transformed with respect to . The second term in the above equation does not depend on the reference mirror position and after the Fourier transformation contributes only to the DC signal. Thus for only the near-field contribution multiplied by the reference field stays in the acquired spectrum: This way, besides providing the interferometric gain, the asymmetric interferometer utilized in nano-FTIR also eliminates the multiplicative background, which otherwise could be a source of various artifacts and is often overlooked in other s-SNOM based spectroscopies. Normalization Following the standard FTIR practice, spectra in nano-FTIR are normalized to those obtained on a known, preferably spectrally-flat reference material. This eliminates the generally unknown reference field and any instrumental functions, yielding spectra of the near-field contrast: Near-field contrast spectra are generally complex-valued, reflecting on the possible phase delay of the sample-scattered field with respect to the reference. Near-field contrast spectra depend nearly exclusively on the dielectric properties of sample material and can be used for its identification and characterization. Nano-FTIR absorption spectroscopy For the purpose of describing near-field contrasts for optically thin samples composed of polymers, organics, biological matter and other soft matter (so called weak oscillators), the near-field signal to a good approximation can be expressed as: , where is the surface response function that depends on the complex-valued dielectric function of the sample and can be also viewed as the reflection coefficient for evanescent waves that constitute the near field of the tip. That is, the spectral dependence of is determined exclusively by the sample reflection coefficient. The latter is purely real and acquires an imaginary part only in narrow spectral regions around the sample absorption lines. This means that the spectrum of an imaginary part of the near-field contrast resembles the conventional FTIR absorbance spectrum, , of the sample material:. It is therefore convenient to define the nano-FTIR absorption , which directly relates to the sample absorbance spectrum: It can be used for direct sample identification and characterization according to the standard FTIR databases without the need of modeling the tip-sample interaction. For phononic and plasmonic samples in the vicinity of the corresponding surface surface resonances, the relation might not hold. In such cases the simple relation between and can not be obtained, requiring modeling of the tip-sample interaction for spectroscopic identification of such samples. Analytical and numerical simulations Significant efforts have been put towards simulating nano-FTIR electric field and complex scattering signal through numerical methods (using commercial proprietary software such as CST Microwave Studio, Lumerical FDTD, and COMSOL Multiphysics) as well as through analytical models (such as through finite dipole and point dipole approximations). Analytical simulations tend to be more simplified and inaccurate, while numerical methods are more rigorous but computationally expensive. References External links Infrared Nanoscopy Laboratory of Fritz Keilman (Ludwigs-Maximilians-Universität) Nanooptics group of Rainer Hillenbrand (CIC nanoGUNE) Nano-Optics & Metamaterials group of Thomas Taubner (RWTH Aachen) Infrared Nano-Optics of Quantum Materials group of Dmitri Basov (UC San Diego) Scanning probe microscopy Infrared spectroscopy Scientific techniques
Nano-FTIR
[ "Physics", "Chemistry", "Materials_science" ]
4,360
[ "Spectrum (physical sciences)", "Infrared spectroscopy", "Scanning probe microscopy", "Microscopy", "Nanotechnology", "Spectroscopy" ]
52,973,276
https://en.wikipedia.org/wiki/Livestreamed%20crime
Livestreamed crime is a phenomenon in which criminal acts are publicly livestreamed on social media platforms such as Twitch or Facebook Live. Due to the fact that livestreams are accessible instantaneously, it is difficult to quickly detect and moderate violent content, and almost impossible to protect the privacy of victims or bystanders. Livestreaming crime allows anyone from the public to become a distant witness. "Trash streaming" is a "disturbing YouTube subculture" in which livestreamers solicit donations in exchange for carrying out dares on-stream, which are often abusive or criminal. History In April 2016, Marina Lonina, age 18; and Raymond Gates, age 29, were arrested in Ohio, US on charges that Gates raped an underage friend of Lonina's while Lonina live streamed the crime on Periscope. The prosecutor pointed out that Lonina, who was taken advantage of by a much older man, had become "caught up" in her excitement over the number of "likes" she was getting, and is shown on screen "laughing and giggling". Joss Wright of the Oxford Internet Institute pointed out that, given the "volume of content being created and uploaded every day, [there] is almost no practical way to prevent content like this being uploaded and shared". By May, The New York Times was including the Periscope rape as one of a series of recent cases in which crimes were live streamed. These included one in which a young woman in Égly, France, speaks via Periscope about her distress and suicidal thoughts and is apparently encouraged by viewers to kill herself, which she does by throwing herself under a train. Also included was the case of two teenagers who live stream themselves bragging and laughing as they beat up a drunken man in a bar in Bordeaux, France. Types Cybersex trafficking Cybersex trafficking, also referred to as live streaming sexual abuse, involves sex trafficking and the live streaming of coerced sexual abuse or rape. Victims are abducted, threatened, or deceived and transferred to "cybersex dens". The dens can be in any location where the cybersex traffickers have a computer, tablet, or phone with Internet connection. Perpetrators use social media networks, videoconferencing, pornographic video sharing websites, dating pages, chat rooms, apps, dark web sites, and other online platforms. Online payment systems and cryptocurrencies are often used for anonymity. Millions of reports of cybersex trafficking are sent to authorities annually. Local authorities in Cambodia have expressed concern that new laws and police procedures are necessary to combat this type of cybercrime. War crimes and cyberterrorism A war crime is the act of breaking the laws of war, and in recent years, there have been more instances of people or groups live-streaming such acts to instill fear into the public. Some groups create websites or use the dark web to host these live streams. Many times, the videos are of hostages or prisoners of war that are used as leverage or as a means of exerting fear and control. Instances 2008 8 June: A bystander livestreamed the Akihabara massacre on Ustream, attracting an audience of 2,000 viewers. Another user also used Ustream to livestream the massacre's aftermath, including police and public response. 2015 12 July: Two West Weber, Utah teenagers were arrested after filming themselves stealing ice cream from a truck while livestreaming on Periscope. 2016 14 June: 27-year-old Antonio Perkins was fatally shot while streaming on Facebook Live in Chicago's North Lawndale neighborhood. 1 August: Korryn Gaines of Randallstown, Maryland livestreamed her actions on Facebook and Instagram as she resisted arrest and started an armed standoff with the police, which ended with Gaines being fatally shot and her five-year-old son sustaining injuries to his face and arm after being hit by stray bullets. November: A Jackson, Michigan man livestreamed himself breaking into a home and shooting a woman on Facebook Live. 2017 1 January: A Reynoldsburg, Ohio woman livestreamed herself taping her toddler to a wall on Facebook Live. She later uploaded a video taunting viewers who had called child protective services on her. 3 January: A torture incident in Chicago, in which an 18-year-old mentally-disabled white male was filmed being physically and verbally abused by four black individuals (two men and two women), was livestreamed by one of the women on Facebook and sparked massive controversy. 21 January: In Uppsala, Sweden, two Afghan immigrants and one Swedish citizen livestreamed the gang rape of a woman on Facebook. 10 April: A Los Angeles man shot at passing cars and police officers while livestreaming on Facebook Live. 24 April: A Phuket man livestreamed himself killing his newborn daughter on Facebook. He soon after committed suicide. 2018 22 February: A Middletown, Connecticut man livestreamed himself on Facebook Live as he drove through the front doors of a hospital, before setting himself on fire. The man was treated for burn injuries and arrested. 4 April: A Detroit teenager accidentally shot his friend while playing with a gun on Instagram Live. 26 August: David Katz shot 12 people during a video game tournament, killing two. The shooting was livestreamed by the event's Twitch stream. 2019 15 March: Two mass shootings occurred at the Al Noor Mosque and the Linwood Islamic Centre in Christchurch, New Zealand, where 51 people were killed and 40 were injured by 28-year-old Brenton Tarrant from Australia. The first attack was livestreamed by the shooter on Facebook Live for 17 minutes; they were described by then-Prime Minister Jacinda Ardern as "one of the darkest days in New Zealand history". Following the shootings, Facebook announced restrictions on Facebook Live on those who posted violent extremist material. 22 March: Vlad Cristian Eremia, 26, stabbed a 77-year-old Catholic priest, Father Claude Grou, in Saint Joseph's Oratory in Montreal, Quebec, Canada; the attack was captured on a livestream by Salt + Light Television. 9 October: Stephan Balliet committed a shooting near a synagogue and kebab restaurant in Halle, Saxony-Anhalt, Germany, resulting in two dead and two others injured; the attack was livestreamed on Twitch. 29 December: Keith Thomas Kinnunen opened fire at the West Freeway Church of Christ in White Settlement, Texas, US, fatally shooting two people before he was shot and killed by an armed member of the congregation. The shooting was livestreamed on YouTube because the church livestreamed its services. 2020 8 February: Thai Army Sergeant Jakrapanth Thomma killed 29 and wounded 57 people in a mass shooting in Thailand. A portion of the shooting at the Terminal 21 Korat mall was livestreamed by the perpetrator on Facebook Live. 20 May: Armando "Junior" Hernandez, a 20-year-old livestreamed his attack at the Westgate Entertainment District in Glendale, Arizona, US where three people were wounded, on Snapchat. 26 May: A man in Stamford, Connecticut livestreamed himself on Instagram on a highway overpass saying he thought people were following him. He then began to fire on passing vehicles on the road below. July: A Tallahassee, Florida teen was murdered while livestreaming on Instagram in a church parking lot. 8 November: A Las Vegas man killed his girlfriend at her home while live on Facebook. Shortly before the livestream, the man also killed the woman's grandfather. 2021 6 January: Far-right personality Anthime Gionet, better known as "Baked Alaska", took part in the storming of the U.S. Capitol and livestreamed the event on DLive. 22 March: Ahmad Al Aliwi Al-Issa, age 21, shot and killed ten people in a mass shooting in Boulder, Colorado, US. A portion of the shooting at a King Soopers supermarket was livestreamed on YouTube by a bystander. 10 April: A Singapore student abroad in London was attacked with a knife during a robbery streamed by a bystander on YouTube. 29 April: A Prichard, Alabama man livestreamed himself abusing his wife on Facebook Live. 7 June: A Rochester, New York woman threatened to shoot a man and a woman over the course of four livestreams done on Facebook Live. 19 August: 15-year-old Hugo Jackson livestreamed himself via Twitch stabbing a faculty member at his secondary school in Eslöv, Sweden. The attacker used a head-mounted camera on a helmet to livestream the attack and wore body armour similar to Brenton Tarrant, whom he had quoted in his manifesto. 20 August: 19-year-old Aidan Ingalls shot a man to death and critically injured his wife on the South Haven Pier before turning the gun on himself halfway down. The entire shooting lasted around a minute and was captured on WWMT's beach camera, which was being livestreamed on YouTube. 2022 15 January: Malik Faisal Akram took multiple people hostage at Congregation Beth Israel, a Jewish synagogue in Colleyville, Texas, United States. A portion of the hostage-taking was livestreamed on the synagogue's Facebook account. 18 April: A Baton Rouge, Louisiana man fatally stabbed a woman while she was on Facebook Live. 14 May: Payton S. Gendron killed 10 people and injured three others while livestreaming the shooting on Twitch, inside and in the parking lot of the Tops Friendly Markets in Buffalo, New York, US. 11 of the 13 people shot were black. 15 June: A West Baton Rouge Parish, Louisiana reserve deputy filmed himself on Facebook firing at a fleeing vehicle as he and a full-time deputy investigated a vehicle theft. The reserve deputy was fired and later charged with illegal discharge of a weapon. 28 June: An Ypsilanti, Michigan man was shot and killed on Facebook Live due to a purported argument. 24 July: Brooklyn-based pastor Lamor Whitehead and his wife were robbed of more than $1,000,000 worth of jewelry during a livestreamed church service. 7 September: During a shooting spree in Memphis, Tennessee, Ezekiel Kelly, a 19-year-old man, livestreamed himself on Facebook Live entering an AutoZone store and critically wounding an employee. 22 December: A Handsworth, West Midlands man stabbed three people, killing one, in a "drink and drug-fuelled rampage" on Facebook Live. 2023 25 March: A Columbus, Mississippi woman fatally shot a man in a domestic violence incident on Facebook Live. 10 April: 25-year-old Connor James Sturgeon livestreamed on Instagram as he opened fire at a bank he previously worked at in Louisville, Kentucky, killing five and wounding eight before being fatally shot by police. 25 April: Two women in Holmes County, Mississippi were fatally shot on Facebook Live after a fight. 27 April: A Minneapolis man barricaded himself for several hours in a standoff with the FBI before being fatally shot, all while livestreaming on Facebook Live. 23 May: Two men were attacked with a knife while livestreaming on Douyin in Xiamen, China. 7 June: A Shreveport, Louisiana woman "fired wildly" during an argument on a Facebook Live broadcast. 16 June: A Doda, Jammu and Kashmir man murdered a relative with an axe during a Facebook Live broadcast. 23 July: After murdering his ex-girlfriend and her sister, a Tallahassee, Florida man streamed his suicide on Facebook Live. August: An Oklahoma teenager was accidentally shot in the face by another teenager. 11 August: A Gradačac man shot and killed his ex-wife and two other people live on Instagram. He also wounded a police officer and two others before killing himself. 7 October: During the 2023 Hamas-led attack on Israel, Hamas fighters livestreamed the attack through Facebook Live, sometimes from phones stolen from Israelis. 9 October: An 18-year-old Hebron man livestreamed on Facebook as he attempted to attack Israeli soldiers with a construction vehicle. 20 October: Two San Antonio teenagers accused of a fatal drive-by shooting livestreamed a car chase and their subsequent arrest on Instagram Live. 22 October: A San Antonio man shot at a dog during a livestream. 29 October: A shooting during Halloween festivities at Ybor City, a neighborhood in Tampa, which resulted in two dead was livestreamed on Instagram Live by victims. 5 November: A Calamba, Misamis Occidental radio host was shot and killed during a Facebook Live broadcast. 12 December: A group of friends in a Miami-Dade County rental were livestreaming when their house was shot at in a drive-by shooting, wounding a woman. 12 December: An Uber driver in Mexico was shot during a struggle while driving and streaming on TikTok. 15 December: 54-year-old Serhiy Batryn detonated three grenades during a meeting at Keretsky town hall, killing one and wounding 26. The meeting was being livestreamed at the time of the attack. 2024 2 January: A Waukesha, Wisconsin man engaged in a shootout with police on Facebook Live following a domestic violence call. 4 January: Student Dylan Butler livestreamed a mass shooting at his high school in Perry, Iowa, killing two people and injuring six others before committing suicide. 8 January: A San Rafael, California woman stabbed her mother to death on Facebook Live. 8 February: Indian politician Mauris Noronha shot and killed his rival Abhishek Ghosalkaron on Facebook Live before killing himself. 13 April: A man livestreamed himself as he fired several shots from a rifle while on the roof of a Marina del Rey, California apartment building. 9 May: A Korean YouTuber fatally stabbed another YouTuber on a livestream. 23 May: A Fort Wayne, Indiana man fired several shots at a random person inside a Kroger supermarket while on Facebook Live. 26 June: 19-year-old Jordan Patten livestreamed himself attempting to commit a mass stabbing in Newcastle but eventually gave up. No one was injured. 12 August: 18-year-old Arda Küçükyetim livestreamed himself stabbing and injuring five people outside a mosque in Eskişehir, Turkey. 30 November: A 17-year-old student livestreamed a hammer attack on two classmates at a school in Guadalajara, Jalisco, Mexico. 2025 22 January: A 17-year-old student at Antioch High School in Nashville, Tennessee live-streamed a shooting on Kick, killing one student and injuring another before killing himself. See also List of filmed mass shootings Beheading video Christchurch Call to Action Summit Hurtcore Shooting of Robert Godwin, a 2017 case in which a video of the crime was posted online (but not livestreamed) by the perpetrator Snuff film References External links Why rising numbers of criminals are using Facebook Live to film their acts Crimes Dark web Filmed killings Violence
Livestreamed crime
[ "Biology" ]
3,123
[ "Behavior", "Aggression", "Human behavior", "Violence" ]
52,973,407
https://en.wikipedia.org/wiki/S%20Eridani
The Bayer designation s Eridani and the variable star designation S Eridani are distinct. Due to technical limitations, both designations link here. For the star s Eridani, see HD 16754 S Eridani, see 64 Eridani HD 16754, s Eridanus (constellation)
S Eridani
[ "Astronomy" ]
64
[ "Eridanus (constellation)", "Constellations" ]
52,974,038
https://en.wikipedia.org/wiki/%CE%93-space
In mathematics, a -space (gamma space) is a topological space that satisfies a certain basic selection principle. An infinite cover of a topological space is an -cover if every finite subset of this space is contained in some member of the cover, and the whole space is not a member the cover. A cover of a topological space is a -cover if every point of this space belongs to all but finitely many members of this cover. A -space is a space in which every open -cover contains a -cover. History Gerlits and Nagy introduced the notion of γ-spaces. They listed some topological properties and enumerated them by Greek letters. The above property was the third one on this list, and therefore it is called the γ-property. Characterizations Combinatorial characterization Let be the set of all infinite subsets of the set of natural numbers. A set is centered if the intersection of finitely many elements of is infinite. Every set we identify with its increasing enumeration, and thus the set we can treat as a member of the Baire space . Therefore, is a topological space as a subspace of the Baire space . A zero-dimensional separable metric space is a γ-space if and only if every continuous image of that space into the space that is centered has a pseudointersection. Topological game characterization Let be a topological space. The -has a pseudo intersection if there is a set game played on is a game with two players Alice and Bob. 1st round: Alice chooses an open -cover of . Bob chooses a set . 2nd round: Alice chooses an open -cover of . Bob chooses a set . etc. If is a -cover of the space , then Bob wins the game. Otherwise, Alice wins. A player has a winning strategy if he knows how to play in order to win the game (formally, a winning strategy is a function). A topological space is a -space iff Alice has no winning strategy in the -game played on this space. Properties A topological space is a γ-space if and only if it satisfies selection principle. Every Lindelöf space of cardinality less than the pseudointersection number is a -space. Every -space is a Rothberger space, and thus it has strong measure zero. Let be a Tychonoff space, and be the space of continuous functions with pointwise convergence topology. The space is a -space if and only if is Fréchet–Urysohn if and only if is strong Fréchet–Urysohn. Let be a subset of the real line, and be a meager subset of the real line. Then the set is meager. References General topology
Γ-space
[ "Mathematics" ]
558
[ "General topology", "Mathematical structures", "Space (mathematics)", "Topological spaces", "Topology" ]
52,974,068
https://en.wikipedia.org/wiki/N%2CN%2CN%E2%80%B2%2CN%E2%80%B2-Tetramethylformamidinium%20chloride
N,N,N′,N′-Tetramethylformamidinium chloride is the simplest representative of quaternary formamidinium cations of the general formula [R2N−CH=NR2]+ with a chloride as a counterion in which all hydrogen atoms of the protonated formamidine [HC(=NH2)NH2]+ are replaced by methyl groups. Deprotonation results in the exceptionally basic bis(dimethylamino)carbene R2N−C̈−NR2. Preparation It is generated by protonation of (CH3)3COCH(N(CH3)2)2 (Bredereck's reagent). (CH3)3COCH(N(CH3)2)2 + H+ → (CH3)3COH + [CH(N(CH3)2)2]+ N,N,N′,N′-Tetramethylformamidinium chloride is also obtained in high yield (95%) in the reaction of dimethylformamide (DMF) with dimethylcarbamoyl chlorideThe conversion of DMF with thionyl chloride in a ratio of 3:1 obtains the product in a is significantly lower yield (72%) which appears, however, more realistic in view of the tricky handling of the chloride salt. Properties N,N,N′,N′-Tetramethylformamidinium chloride is a light yellow, strongly hygroscopic solid. For drying, the salt is dissolved in dichloromethane and the solution is treated with solid anhydrous sodium sulfate. After several dissolutions in dichloromethane and acetone, and precipitations with tetrahydrofuran, a colorless solid is obtained, which is stable under air and moisture sealing. The presumption of a mesomeric equilibrium between ionic formamidinium chloride and covalent bis(dimethylamino)chloromethane structure: could be decided by reaction with germanium dichloride or tin(II) chloride in favour of the presence of N,N,N′,N′-tetramethylformamidinium chloride. The hygroscopy of the chloride salt complicates the handling of the compound. Therefore, also syntheses of the much better processible salts N,N,N′,N′-tetramethylformamidinium methylsulfate (from the dimethylformamide–dimethylsulfate complex) and of N,N,N′,N′-tetramethylformamidinium p-toluenesulfonate (from dimethylformamide and p-toluenesulfonyl chloride) were investigated. Applications N,N,N′,N′-Tetramethylformamidinium chloride is useful as a reagent for aminomethylenation (that is, to introduce a =CH−NR1R2 function to CH-acidic compounds). For example, ethyl cyanoacetate reacts with the formamidinium salt in the presence of solid sodium hydroxide to give ethyl (dimethylaminomethylene)cyanoacetate in practically quantitative yields. The aminomethylenation provides intermediates for the synthesis of heterocycles such as indoles, pyrimidines, pyridines and quinolones. N,N,N′,N′-Tetramethylformamidinium chloride reacts with alkali metal dimethylamides (such as lithium dimethylamide or sodium dimethylamide) to tris(dimethylamino)methane in yields of 55% to 84%. The reaction product is suited as a reagent for formylation and aminomethylenation. From N,N,N′,N′-tetramethylformamidinium chloride and sodium ethoxide in ethanol, dimethylformamide diethyl acetal is formed in 68% yield. In aqueous sodium cyanide, N,N,N′,N′-tetramethylformamidinium reacts to bis(dimethylamino)acetonitrile. From N,N,N′,N′-tetramethylformamidinium and anhydrous hydrogen cyanide, dimethylaminomalonic acid dinitrile is obtained in 92% yield. N,N,N′,N′-Tetramethylformamidinium can be regaminated with cyclo-aliphatic amines to the corresponding heterocyclic formamidines. N,N,N′,N′-tetramethylformamidinium is a catalyst in the preparation of acyl chlorides from carboxylic acids and phosgene has been reported. Strong bases (such as phenyllithium) can abstract a proton from the formamidinium cation of N,N,N′,N′-tetramethylformamidinium forming bis(dimethylamino)carbene. See also Bis(dimethylamino)methane References Amidines Dimethylamino compounds
N,N,N′,N′-Tetramethylformamidinium chloride
[ "Chemistry" ]
1,115
[ "Bases (chemistry)", "Amidines", "Functional groups" ]
52,974,142
https://en.wikipedia.org/wiki/Creative%20Cities%20Network
The UNESCO Creative Cities Network (UCCN) is a flagship city programme of UNESCO launched in 2004 to promote cooperation among cities which have recognized culture and creativity as strategic drivers of sustainable urban development. , there are almost 300 cities from around 90 countries in the network. The network aims to foster mutual international cooperation with and between member cities committed to invest in creativity as a driver for sustainable urban development, social inclusion and cultural vibrancy. The Network recognizes the following creative fields: The overall situation and activities within the Network is reported in the UCCN Membership Monitoring Reports, each for a four-year period for a particular city. The Network recognizes the concept of creative tourism, defined as travel associated with creative experience and participation. Film Literature Music Crafts and Folk Arts Design Gastronomy Media Arts References External links UNESCO Organizations established in 2004 Creativity
Creative Cities Network
[ "Biology" ]
169
[ "Creativity", "Behavior", "Human behavior" ]
52,974,178
https://en.wikipedia.org/wiki/Ethyl%20cyanohydroxyiminoacetate
Ethyl cyanohydroxyiminoacetate (oxyma) is the oxime of ethyl cyanoacetate and finds use as an additive for carbodiimides, such as dicyclohexylcarbodiimide (DCC) in peptide synthesis. It acts as a neutralizing reagent for the basicity or nucleophilicity of the DCC due to its pronounced acidity (pKa 4.60) and suppresses base catalyzed side reactions, in particular racemization. Production Ethyl cyanohydroxyiminoacetate is obtained in the reaction of ethyl cyanoacetate and nitrous acid (from sodium nitrite and acetic acid) in 87% yield. Because of the rapid hydrolysis of the ester, the reaction should be carried out at pH 4.5, in buffered phosphoric acid the product can even be obtained in virtually quantitative yield. The compound can be purified by recrystallization from ethanol or ethyl acetate. Compared with the benzotriazole derivatives 1-hydroxybenzotriazole (HOBt) and 1-hydroxy-7-azabenzotriazole (HOAt) (which are widely used as peptide-linking reagents but are explosive), ethyl cyanohydroxyiminoacetate exhibits a markedly slowed thermal decomposition on heating. Properties Ethyl cyanohydroxyiminoacetate is a white solid which is soluble in many solvents common in the synthesis of peptides, such as dichloromethane or dimethylformamide (DMF). In crystalline form, the compound is present as an oxime, whereas it exists as a salt or in a strongly basic solution predominantly as a tautomeric nitrosoisomer in anionic form. Applications Owing to the simple preparative accessibility, the uncritical behavior at temperatures below 80 °C and in particular because of the high yields and the low racemization of the peptides obtained, ethyl cyanohydroxyiminoacetate has now become widely used as an additive in peptide syntheses. Ethyl cyanohydroxyiminoacetate can be used as a coupling additive in the conventional peptide linking in solution, as in automated Merrifield synthesis on a solid-phase peptide synthesis, together with coupling reagents such as carbodiimides (for example dicyclohexylcarbodiimide (DCC)), diisopropylcarbodiimide (DIC) or the water-soluble 1-ethyl-3-(3-dimethylaminopropyl)carbodiimide (EDCI)). For example, the stepwise liquid-phase synthesis of the dipeptide Z-L-Phg-L-Val-OMe yields the LL-product with 81-84% which is free from racemic DL dipeptide, using From N-protected Z-L-α-phenylglycine (with the benzyloxycarbonyl group, Z group) and L-valine methyl ester with the coupling reagent DIC and the additive ethyl cyanohydroxyiminoacetate. More recently, a variety of derivatives of ethyl cyanohydroxyiminoacetate (Oxyma) have been developed as acylation reagents, such as Fmoc-oxyma for the transfer of the fluorenylmethoxycarbonyl protective group or the coupling reagent COMU which is readily soluble as a dimethylmorpholine-uronium salt and which, like Oxyma, is superior to the standard additive HOBt for the suppression of racemization and acylation efficiency and is comparable to HOAt without presenting an explosion risk such as the benzotriazoles. With water-soluble derivatives of ethyl cyanohydroxyiminoacetate (glyceroacetonide-oxyma) as additive and DIC as coupling reagent even in weakly basic aqueous solutions the linking of protected amino acids to oligopeptides is possible with a yield of 95% and a diastereomeric excess of> 99% using the model substances Z-L-Phg-OH and L-H-Pro-NH2. In the coupling of amino acids, frequently occurring secondary reactions largely suppressed, which would be the formation of symmetrical acid anhydrides, racemization and epimerization and the cyclization to oxazolinones or - especially for dipeptides - to 2,5-diketopiperazines. References Oximes Cyanoacetate esters Reagents for biochemistry Reagents for organic chemistry Ethyl esters
Ethyl cyanohydroxyiminoacetate
[ "Chemistry", "Biology" ]
1,012
[ "Biochemistry methods", "Oximes", "Functional groups", "Reagents for organic chemistry", "Biochemistry", "Reagents for biochemistry" ]
52,974,247
https://en.wikipedia.org/wiki/Department%20of%20Pharmacology%2C%20University%20College%20London
The Department of Pharmacology at the University College London, the first of its kind in England, was founded in 1905 and remained in existence until 2007. Early history University College London (UCL) was founded in 1826. It was born in the ferment of radical London in the 1820s and 1830s and was heavily influenced by the Scottish and French Enlightenments. UCL was part of the radical opposition to the hegemony of Oxford and Cambridge. In medicine, UCL was a force in combatting the conservative and religious monopoly of the Royal Colleges of Physicians and Surgeons Although Edinburgh University was well ahead at the time, UCL had a professor of Materia Medica and Pharmacy, A.T.Thomson, from the start. Later this was renamed as the chair of Materia Medica and Therapeutics. Its best known holder was Sydney Ringer (1878–87), who worked on the isolated beating heart and is renowned for his eponymous salt solution which he designed to maximise the viability of isolated hearts. His textbook ‘Handbook of Therapeutics’ ran for 13 editions between 1869 and 1897. In 1905. Pharmacology was established as a distinct discipline within basic medical sciences at UCL. It was the first Department of Pharmacology in England. Most of the people involved in the development of quantitative analysis of drug-receptor interactions worked at some time in UCL's Departments of Pharmacology, or of Physiology or of Biophysics. The Department of Pharmacology Arthur Robertson Cushny FRS Arthur Cushny (1866–1926) was the first holder of the newly instituted Chair of Pharmacology, from 1905 until 1918. After graduating in medicine from Aberdeen, Cushny had studied in Berne, Würzburg, and Strasbourg, where he became Assistant to the famed Oswald Schmiedeberg. In 1893, at the age of 27, he was appointed Professor of Pharmacology at the University of Michigan, Ann Arbor. Eight years later Cushny came to the chair at UCL where he soon expanded the department from the single room he had been given. He raised the funds for the building which the remnants of the department still occupies. His main interests were in the heart and kidney. His work on the involvement of calcium in the action of digitalis was prescient. He was interested in optical isomers. Data from an early clinical trial using hyoscine isomers was used by William Sealy Gossett who, under the pseudonym "Student" published in 1908 the first small-sample test of significance, Student's t test. His use of these data has given rise to much discussion. Later reanalysis of the same data by a randomisation tests gave a similar result. Cushny published a textbook Textbook of Pharmacology and Therapeutics (eighth edition 1924). He introduced the Cushny myograph, an ingenious arrangement of counterbalanced levers that allowed the faithful recording of the rate and force of contraction of the rapidly beating animal heart. It was still in use in practical classes at UCL, and elsewhere, in the 1960s. Cushny left UCL in 1918, to become Professor of Materia Medica and Pharmacology at Edinburgh. he was succeeded by A.J. Clark/ Alfred Joseph Clark FRS A.J. Clark, FRS (1885–1941) held the established Chair of Pharmacology from 1918 to 1926. After qualifying in medicine, and serving as a field medical officer throughout the First World War, Clark had been appointed Professor of Pharmacology at the University of Cape Town where he remained until accepting the Chair of Pharmacology at UCL in 1920. His influence on the subject was profound. The distinguished physiologist and Nobel laureate A.V. Hill (Archibald Hill) had begun the quantitative study of the action of agonists on an isolated tissue (frog skeletal muscle) some years earlier. Clark took this much further and extended it to examine the actions of antagonists. The data he gathered on the exact relationship between agonist concentration and response, and on how this changed in the presence of a competitive antagonist, were published in two classic papers in the Journal of Physiology in 1926,. But he failed to work out a method for analysing properly the results of experiments with antagonists: that had to wait for his successor, Heinz Schild. Nevertheless, Clark was largely responsible for the transition of pharmacology from a descriptive subject to the quantitative science that it is today - this emphasis on quantitative approaches has remained strong throughout the subsequent history of the department. Clark's book The Mode of Action of Drugs on Cells (Williams & Wilkins, 1933) is a classic and the following quotation from it set the tone for the department for many years. "In the first place, there is no advantage in fitting curves by a formula unless this expresses some possible physico-chemical process, and it is undesirable to employ formulae that imply impossibilities. It is a question of finding a few systems so simple that it is possible to establish with reasonable probability the relation between the quantity of drug and the action produced." While at UCL Clark wrote the first edition of his textbook Applied Pharmacology in 1923, a book that was to be updated by two of his successors as Head of department, first by H.O. Schild and later by H.P. Rang, and is still extant in the form of the widely used textbook Rang & Dale's Pharmacology. In 1926 Clark followed his predecessor in moving to the University of Edinburgh. E.B.Verney FRS Ernest Basil Verney (1894–1967) succeeded Clark. He held the Chair of Pharmacology from 1926 to 1934. While at UCL Verney discovered the antidiuretic hormone and also the mechanism by which structures in the brain sense minute changes in blood osmotic pressure. Both findings were of profound importance for the understanding of water and electrolyte balance. Verney was also instrumental in arranging for Otto Krayer to come to the department, albeit for only a short period, following Krayer's exclusion from all academic positions in German universities because of his objection to the expulsion of Jewish scientists from their posts. Krayer was later to head the Department of Pharmacology at Harvard with the greatest distinction. In 1934 Verney moved to an academic post at the University of Cambridge where he later became the first Sheild Professor of Pharmacology J.H. Gaddum FRS John Gaddum (1900–1965) held the Chair of Pharmacology from 1935 to 1938. Like A.J. Clark, he had a profound interest in quantitative methods. He extended A.J. Clark's work on competitive antagonism, and applied the law of mass action to describe the relationship (the Gaddum equation) between receptor occupancy and the concentrations of an agonist and a competitive antagonist at equilibrium with the receptors in a tissue. The theory for two or more competing ligands had been known since Michaelis & Menten (1914), but Gaddum was the first to apply it in a pharmacological context). Like Clark before him, Gaddum failed to spot how to use the theory to estimate equilibrium constants. Gaddum was also a master of bioassay which was then the preferred, and usually the only, way to determine the concentrations of biologically active molecules such as labile neurotransmitters and the neuropeptides. F.R. Winton Frank Winton (1894–1985) held the Chair of Pharmacology from 1938 to 1961. His main scientific interest was in the control of blood flow to the kidney. Winton ran the department through the difficult war years when the Medical School was evacuated to Leatherhead, Surrey. He appointed the first two female academics in the department. Mary Lockett (1911–1982) was a lecturer in the department from 1945 - 1950. Hannah Steinberg arrived in the UK from Vienna on a Kindertransport train while still a schoolgirl, and she eventually became Professor of Psychopharmacology. Winton also worked hard and successfully to ensure that pharmacology had an appropriate place in the preclinical curriculum. He oversaw the extension of the department, including the Pharmacology Lecture Theatre (now the Schild Theatre). He was the author, with Leonard Bayliss, of a widely used textbook Human Physiology, first published in 1932. The 6th edition, 1968 was written by Olof J.C. Lippold and F.R Winton H.O. Schild FRS Heinz Otto Schild (1906–1984) held the Chair of Pharmacology from 1961 to 1973. He was born in Fiume (now Rijeka, Croatia), in 1908, when it was part of the Austro-Hungarian empire. He qualified in medicine in Munich and then worked with Straub, the leading German pharmacologist of the time. By good fortune, Schild had been accepted as a visiting worker by Sir Henry Dale and was in England when the National Socialists came to power in Germany. He decided to stay in Britain and became an assistant in the Department of Pharmacology in Edinburgh, then headed by A.J. Clark. When J. H. Gaddum was appointed to the chair at UCL, he invited Schild to join him as a Demonstrator. So began his long association with UCL, interrupted only by his bizarre internment on the Isle of Man as an ‘enemy alien’ at the outbreak of the Second World War. Following his release (greatly aided by F.R Winton's and Sir Henry Dale's appeals to the Home Office) he returned to his work in the department, then based in Leatherhead, and in 1961 became Winton's successor as Head of Department and Professor of Pharmacology. Schild made major contributions to receptor pharmacology, to the understanding of the mechanism of histamine release and to bioassay. Like Gaddum and Clark, he used quantitative approaches whenever possible. His name is immortalised by the Schild equation. He built on the work of Clark and Gaddum on competitive antagonism, by realising that the null method was key to extraction of physical equilibrium constants from simple functional experiments. Rather than looking at the depression by antagonist of the response to a fixed agonist concentration, he measured the dose-ratio, the factor by which the agonist concentration had to be increased in order to restore a given response in the presence of the antagonist. By measuring the dose-ratio as a function of antagonist, it was possible to estimate the dissociation equilibrium constant for the combination of the antagonist with its receptor. Crucially the estimate is not dependent on the nature of the agonist. Although Schild's derivation used the simplest possible model, it was subsequently shown that his equation is valid under much more general conditions. A.J. Clark's textbook was continued by Schild as Clark's Applied Pharmacology by Wilson & Schild. Heinz Schild was a generous and kindly Head of department. He appointed the third female member of academic staff, Dr M. Maureen Dale, a co-author of Rang & Dales Pharmacology. He oversaw the planning and introduction of a three-year B.Sc. course in Pharmacology which began in 1967 and continues to this day. Medical students were able to enter its final year and Schild, who never lost sight of the roots of the subject in medicine, was delighted that many took this opportunity. J.W.Black FRS Sir James Whyte Black (1924–2010) held the Chair of Pharmacology from 1973 to 1978. Jim Black and Heinz Schild knew each other well because Schild had acted as a consultant to the then Smith, Kline & French company during the time when Black was leading the team that developed the histamine receptor antagonists, H2 antagonists, which reduce secretion of gastric acid and which, at the time, transformed the treatment of gastric ulcers. Schild's methods for quantitative methods for analysis of drug antagonists were crucial for this work. Black introduced many changes to teaching in the department. One of the most important was the introduction of a BSc course in Medicinal Chemistry. His long experience in the pharmaceutical industry had convinced him that organic and physical chemists working on drug development with pharmacologists and biochemists would benefit greatly from a substantial knowledge of biology, certainly enough to allow them to understand and assess the kinds of measurements that their biological colleagues undertook. Though the students were based in the Department of Chemistry, they took also courses in physiology and pharmacology, particularly its molecular aspects. This BSc course, like that in Pharmacology, also flourished and continues today. Another important change was a sharp reduction in the number of experiments with animal tissues undertaken by medical students during their course in pharmacology. At the same time, the emphasis on the importance of observations on human subjects was increased. Black's appointment coincided with the onset of the straitened circumstances that all UK universities were to experience and that have continued in one form or another ever since. The changes he made helped the department to adjust to these harder times. To the regret of his Departmental staff, Black found that only the pharmaceutical industry could provide the facilities needed for the work he wished to pursue, and in 1978 he left to join the Wellcome Foundation. Black was knighted in 1981 and in 1988 he got the Nobel Prize in Physiology or Medicine along with Gertrude B. Elion and George H. Hitchings for their work on drug development. H.P.Rang FRS Humphrey Rang (born 1936) held the Chair of Pharmacology from 1979 to 1983. Rang qualified in medicine at UCL and had worked in H.O.Schild's laboratory while a medical student. He was the author of the first successful ligand-binding experiment of the modern era. This was based on his PhD work in Oxford, under William D.M. Paton. Rang had previously been the Professor of Pharmacology at Southampton and at St. George's Hospital Medical School. He brought with him David Colquhoun who was also returning to the department, having been appointed in 1964 as an assistant lecturer by H.O. Schild. These appointments greatly strengthened the interests and achievements of the department in fundamental aspects of pharmacology, particularly the study of ion channels and receptors. In collaboration with M. Maureen Dale (also appointed during Schild's Headship), Rang prepared the first edition of Pharmacology, the successor to Wilson & Schild's Applied Pharmacology. In 1983 Rang was offered and accepted the Directorship of the Sandoz Institute of Medical Research, a division of Sandoz, then an independent pharmaceutical company. The new Institute was located in UCL and developed a close relationship with the department, both in teaching, to which members of the Institute contributed, and in research. After 1983 The heads of department since 1983. After Rang's resignation, the Chair of Pharmacology became vacant. The Head of department from 1983 to 1987 was Donald.H Jenkinson He had done his PhD under Bernard Katz. in UCL's famous Department of Biophysics, and was yet another member of staff who had been invited to join the Department by Heinz Schild. During his tenure the Middlesex Hospital Medical School was merged with UCL's, including the two Departments of Pharmacology. During the 1980s the traditional role of Heads of department was replaced by rotating headships that were no longer associated necessarily with an established chair. Established chairs were, de facto, abolished as part of the move to corporatise universities. David Colquhoun FRS was appointed to the established chair in 1985. It was subsequently dubbed the A.J. Clark chair, in honour of Clark's role in the establishment of quantitative pharmacology. His work, with statistician Alan Hawkes and Bert Sakmann (Nobel prize 1991) established the department as the world leader in the theory and experiment of single ion channels. After retiring from the A.J.Clark chair in 2004, he worked on the misinterpretation of p values and its contribution to the irreproducibility that has come to light in some areas of science. D. A. Brown FRS (1936 - 2023) was appointed in 1987 as head of department (he had previously held the same position at the School of Pharmacy). In 1987, the merger with the Middlesex Hospital Medical School was completed and David Brown inherited the title Astor Chair of Pharmacology from Professor F Hobbiger who had held that title at the Middlesex. Brown's appointment was intended initially to be the start of a 5-year rotating headship, but when Colquhoun's turn became due, he decided that the job of Head of department would not allow enough time to do the algebra and program development with which he was involved. Donald Jenkinson likewise declined to take another turn. Luckily David Brown agreed to continue and he remained Head of department until 2002. His tenure saw a second merger, this time with the Department of Pharmacology at the Royal Free Medical School, headed by Professor Annette Dolphin, FRS. David Brown is renowned for his discovery of the acetylcholine (muscarinic)-sensitive potassium channel (M channel). The Wellcome Lab for Molecular Pharmacology The growing importance of molecular biology led Brown & Colquhoun to apply to the Wellcome Trust in 1990. They funded the building of the Lab for Molecular Pharmacology which Colquhoun directed until his retirement in 2004. Trevor G. Smart became Head of department in 2002, with the title of Schild Chair of Pharmacology. He also works in the ion channel field. After the demise of the department in 2007. Smart became head of the new Research Department of Neuroscience, Physiology and Pharmacology. Stephanie Schorge. In 2021, Professor Schorge succeeded Trevor Smart as head of the Research Department of Neuroscience, Physiology and Pharmacology. She is the first female head of pharmacology since 1905. Stuart Cull-Candy FRS. Stuart G. Cull-Candy works on glutamate-activated ion channels. He joined the Department from UCL's Department of Biophysics and holds the Gaddum Chair of Pharmacology. Lucia Sivilotti was appointed to the A.J. Clark chair in 2014. She has run the UCL Single Ion Channel group after Colquhoun's retirement in 2004. She continued and greatly extended the work in the field of single channel kinetics. She owns the web site OneMol where the group's analysis programs and publications can be downloaded. The association of the A.J. Clark chair with quantitative work on receptors has thus continued to the present day. The first of the nationwide Research Assessment Exercises took place in 1986. The UCL Department headed the list. It continued to be rated as the top Department of Pharmacology in each of the four research assessments that followed in 1989, 1992, 1996 and 2002. But this performance was not enough to save the department. The end of the Department of Pharmacology In 2004, Malcolm Grant became provost of UCL. He commissioned external reports on the reorganisation of the college. The distinguished vice-president of the University of Manchester, Richard Alan North FRS, was asked to assess several options for the reorganisation of the Faculty of Life Sciences. One was to create large Research Departments, including one of Neuroscience, Physiology and Pharmacology, from the existing academic Departments. Professor North's only comment on the options was that the proposed "research departments in Life Sciences were too big". Grant accepted the conclusions except for the part about the size of departments. On 24 May 2007 Grant persuaded the Academic Board to authorise him to act on its behalf and on 13 June 2007 the Department of Pharmacology was disestablished, after a century of distinction and innovation. The academic staff at the time had three main concerns about the proposals. (a) The separation of teaching from research is bad, especially for teaching: the fact that a degree is offered in, for example, Pharmacology without a Pharmacology department to support it, means that there is no guarantee that there will be staff qualified or fully motivated to teach it. Moreover, the collegiality that comes from designing and providing a first-rate degree course is lost. (b) The size of the merged department of Neurosciences, Physiology and Pharmacology means less interaction between staff, and less collegiate spirit. (c) The changes created two extra levels of administration, so that now five levels existed between academics and the provost. Staff were told at the time that the new organisation would be rolled out to other Faculties across UCL, though this has not happened. David Colquhoun has kept a personal diary of the process on his blog: In Memoriam Department of Pharmacology, UCL 1905 – 2007. On the positive side, UCL's current provost, Michael Arthur, has put much emphasis on the quality of teaching, and maintaining its connections with research. As of 2019, UCL still offers pharmacology degrees, though within the now merged Neuroscience, Physiology and Pharmacology department. A history of the combined department appears on UCL's web site. References History of University College London Pharmacology
Department of Pharmacology, University College London
[ "Chemistry" ]
4,484
[ "Pharmacology", "Medicinal chemistry" ]
52,974,301
https://en.wikipedia.org/wiki/Trans-Golgi%20network%20vesicle%20protein%2023%20A
Trans-Golgi network vesicle protein 23 A (TVP23A) is a protein coded for the TVP23A gene, formerly known as FAM18A. TVP23A is located on chromosome 16. It is known to have human paralogs, TVP23B and TVP23C, as well as orthologs in many different species, notably yeast, mice, and chickens. The general consensus on the TVP23A protein indicate that it has some function in the late Golgi apparatus and is involved in retrograde transport from endosomes back into the Golgi apparatus. The nature of this transport is still unknown. Gene Locus TVP23A is located at cytogenic band 16p13.13, on the negative strand of Chromosome 16. Alternative Names TVP23A stands for Trans-Golgi network Vesicle Protein 23A TVP23A, is the current name for the protein. Aliases of TVP23A include FAM18A, and rarely YDR084C. mRNA Isoforms There are two known isoforms of TVP23A, variant one and variant two, with variant one being the more common variant in humans. Protein Structure TVP23A is a member of the pfam superfamily containing the domain of unknown function 846 (DUF846). TVP23A has a predicted molecular weight of 24.1 kilodaltons, an isoelectric point of 6.5, and relatively high amounts of tryptophan and phenylalanine. The secondary structure of TVP23A consists primarily of alpha helices composing 4 transmembrane domains. There is not much information on the tertiary structure of TVP23A or its homologs. iTASSER was used to generate a prediction for the folding pattern of TVP23A, which supports the presence of multiple helix structures. Expression TVP23A is ubiquitously expressed in all human tissues. There is evidence of higher expression in the brain tissue of mice. The promoter for TVP23A is GXP_91266, spanning 1403 base pairs located on the negative strand of chromosome 16. Function The hypothesized function of TVP23A is a transmembrane protein involved in retrograde transport of vesicles from early endosomes into the late Golgi apparatus. TVP23A interactions with SNARE TVI1 were found to be required for retrograde transport. Interacting Proteins TVP23A has been found to interact with four different proteins via Yeast two hybrid arrays. Two of these proteins, YIPF1 and YIPF2, are believed to be Golgi transport proteins. Homology Homologous Domains TVP23A is a DUF846 containing protein, which is homologous throughout TVP-type proteins. This domain contains the 4 transmembrane domains of TVP23A Paralogs TVP23A has two paralogs, TVP23B and TVP23C. TVP23B and TVP23C are 96% similar to each other, and both are located on chromosome 17. Due to the locations of these three genes, and their identities to each other, it is probable that ancestral TVP23 underwent duplication and translocation, giving rise to TVP23A on chromosome 16 and TVP23B/C on chromosome 17, which then underwent a second duplication to form TVP23B and TVP23C. Orthologs TVP23A has been found in all multicellular eukaryotes, including fungi. This gene has not been found in bacteria. References Genes Proteins
Trans-Golgi network vesicle protein 23 A
[ "Chemistry" ]
765
[ "Biomolecules by chemical classification", "Proteins", "Molecular biology" ]
52,974,387
https://en.wikipedia.org/wiki/Mu%20Fornacis
μ Fornacis (Latinised as Mu Fornacis) is the Bayer designation for a star in the southern constellation of Fornax. With an apparent visual magnitude of 5.27, it is visible to the naked eye. The distance to this star, as determined by its annual parallax shift of 10.18 mas, is around 320 light years. This is an A-type main sequence star with a stellar classification of A0 V. It is a Be star that displays "central quasi emission" (CQE) bumps in its spectrum due to a surrounding shell of material. The star has an estimated 3.1 times the mass of the Sun and 2.5 times the Sun's radius. Mu Fornacis is spinning rapidly with a projected rotational velocity of 320 km/s and is around 188 million years old. It radiates 68.7 times the solar luminosity from its outer atmosphere at an effective temperature of 11,745 K. Mu Fornacis appears to be emitting an infrared excess at a wavelength of 22 μm, which could be due to an orbiting debris disk. References A-type main-sequence stars Be stars Circumstellar disks Fornax Fornacis, Lambda CD-31 882 013709 10320 652 0089
Mu Fornacis
[ "Astronomy" ]
271
[ "Fornax", "Constellations" ]
52,975,690
https://en.wikipedia.org/wiki/Robert%20Kezaala
Dr. Robert Kezaala is a medical doctor, epidemiologist, scholar and public health leader in the field of immunization and health emergencies. Currently he is serving as a Senior Health Advisor and team lead for Accelerated Immunization Initiatives: measles, rubella, epidemic meningitis and yellow fever control and Immunization in Emergencies at the United Nations Children’s Fund. Education Kezaala received his medical degree from Makerere University in Kampala, Uganda. He also holds an MPH from the Royal Tropical Institute (KIT) in Amsterdam, Netherlands majoring in epidemiology and health planning. Career Kezaala has over 30 years of professional experience in public health including 24 years at international level. In the late 1980s, Dr. Kezaala practiced as Medical Officer in Karamoja province in the northeast of Uganda with recognized work in immunization and Tuberculosis control. From 1992 to 1993, Dr Kezaala worked with UNDP in multi-sectoral HIV/AIDS control where he managed the collaborative program that supported Uganda government efforts to address the AIDS epidemic. Thereafter, until 1998, he worked with the International Federation of Red Cross and Red Crescent Societies (IFRC) as Regional Health Delegate for Eastern and Southern Africa, managing a variety of health interventions including HIV/AIDS control, community water and sanitation and refugee health and emergency response when he led IFRC's initial health response in Goma during the 1994 Rwanda crisis. Subsequently, he joined the World Health Organization (WHO), where he worked for 14 years, first as epidemiologist and Team Lead for WHO-EPI in Ethiopia. From 2001 to 2005, Kezaala headed Measles Control for the Africa Region of WHO, when the Africa region registered a reduction in measles mortality by 70%. He spent the next seven years serving as a medical officer with the Polio Eradication Initiative at the WHO headquarters in Geneva, Switzerland. While here, he worked in country support across the globe, including Chad, Pakistan, the Horn of Africa and served as the outbreak response manager for the 2010 Polio outbreak that affected Tajikistan, Kazakhstan and Russia. During the stint in GPEI, Dr Kezaala developed the Short Interval Additional Dose (SIAD) tactical approach that has since become a standard for Polio outbreak response. In 2012, Dr Kezaala served as WHO liaison officer to the US Centers for Disease Control (CDC) in setting up the Emergency Operations Centre (EOC) for CDC's global Polio Eradication Initiative. Since June 2012, Dr Kezaala has served as Senior Health Advisor at the UNICEF headquarters in New York in charge of the Accelerated Immunization Initiatives - responsible for Measles and Rubella control, Yellow Fever, epidemic Meningitis and immunization in emergency settings. In 2016, he was instrumental as liaison officer to WHO in the response to the central Africa Yellow Fever outbreak that affected Angola and the Democratic Republic of Congo. Numerous news outlets and reports such as U.S. News, CNN, TV2Africa, allAfrica have quoted Dr. Kezaala. He is also a thought leader in topics related to public health, vaccines, health diplomacy and on Uganda. References Ugandan public health doctors UNICEF people Living people Ugandan officials of the United Nations Year of birth missing (living people) Vaccination advocates
Robert Kezaala
[ "Biology" ]
685
[ "Vaccination", "Vaccination advocates" ]
52,975,808
https://en.wikipedia.org/wiki/Slosh%20baffle
A slosh baffle is a device used to dampen the adverse effects of liquid slosh in a tank. Slosh baffles have been implemented in a variety of applications including tanker trucks, and liquid rockets, although any moving tank containing liquid may employ them. Baffle rings Baffle rings are rigid rings placed within the inside of a tank to retard the flow of liquid between sections. The location and orifice size of the rings yield varying performance for a given application. See also Baffle blocks References Fluid dynamics
Slosh baffle
[ "Chemistry", "Engineering" ]
108
[ "Piping", "Chemical engineering", "Fluid dynamics stubs", "Fluid dynamics" ]
52,977,719
https://en.wikipedia.org/wiki/Low-code%20development%20platform
A low-code development platform (LCDP) provides a development environment used to create application software, generally through a graphical user interface (as opposed to only writing code, though some coding is possible and may be required). A low-coded platform may produce entirely operational applications, or require additional coding for specific situations. Low-code development platforms are typically on a high abstraction level, and can reduce the amount of traditional time spent, enabling accelerated delivery of business applications. A common benefit is that a wider range of people can contribute to the application's development, not only those with coding skills, but good governance is needed to be able to adhere to common rules and regulations. LCDPs can also lower the initial cost of setup, training, deployment, and maintenance. Low-code development platforms trace their roots back to fourth-generation programming language and the rapid application development tools of the 1990s and early 2000s. Similar to these predecessor development environments, LCDPs are based on the principles of model-driven architecture, automatic code generation, and visual programming. The concept of end-user development also existed previously, although LCDPs brought some new ways of approaching this development. The low-code development platform market traces its origins back to 2011. The specific name "low-code" was not put forward until 9 June, 2014, when it was used by the industry analyst Forrester Research. Along with no-code development platforms, low-code was described as "extraordinarily disruptive" in Forbes magazine in 2017. Use As a result of the microcomputer revolution, businesses have deployed computers widely across their employee bases, enabling widespread automation of business processes using software. The need for software automation and new applications for business processes places demands on software developers to create custom applications in volume, tailoring them to organizations' unique needs. Low-code development platforms have been developed as a means to allow for quick creation and use of working applications that can address the specific process and data needs of the organization. Reception Research firm Forrester estimated in 2016 that the total market for low-code development platforms would grow to $15.5 billion by 2020. Segments in the market include database, request handling, mobile, process, and general purpose low-code platforms. Low-code development's market growth can be attributed to its flexibility and ease. Low-code development platforms are shifting their focus toward general purpose of applications, with the ability to add in custom code when needed or desired. Mobile accessibility is one of the driving factors of using low-code development platforms. Instead of developers having to spend time creating multi-device software, low-code packages typically come with that feature as standard. Because they require less coding knowledge, nearly anyone in a software development environment can learn to use a low-code development platform. Features like drag and drop interfaces help users visualize and build the application Security and compliance concerns Concerns over low-code development platform security and compliance are growing, especially for apps that use consumer data. There can be concerns over the security of apps built so quickly and possible lack of due governance leading to compliance issues. However, low-code apps do also fuel security innovations. With continuous app development in mind, it becomes easier to create secure data workflows. Still the fact remains that low-code development platforms that do not apply and strictly adhere to normalized systems theory do not solve the challenge of increasing complexity due to changes. Criticisms Some IT professionals question whether low-code development platforms are suitable for large-scale and mission-critical enterprise applications. Others have questioned whether these platforms actually make development cheaper or easier. Additionally, some CIOs have expressed concern that adopting low-code development platforms internally could lead to an increase in unsupported applications built by shadow IT. See also DRAKON End-user computing End-user development Flow-based programming List of online database creator apps List of low-code development platforms Visual programming language Backend as a service References Enterprise architecture Software development
Low-code development platform
[ "Technology", "Engineering" ]
804
[ "Software engineering", "Computer occupations", "Software development" ]
52,979,724
https://en.wikipedia.org/wiki/Alan%20Mackay-Sim
Alan Mackay-Sim (16 May 1951 – 4 January 2023) was an Australian biomedical scientist specialising in adult stem cell research, and winner of the 2017 Australian of the Year. Some of his research focused on olfactory ensheathing cells, which are cells in the human nose that interact with the nervous system to cause a sense of smell. His research into stem cells contributed to the treatment of spinal cord injuries. The restoration of mobility to Darek Fidyka, a Polish paraplegic man, used research by Mackay-Sim. Academic and professional life Mackay-Sim received his PhD from Macquarie University in 1980, then subsequently studied at the University of Pennsylvania and the University of Wyoming. Mackay-Sim was the director of the National Centre for Adult Stem Cell Research at Griffith University before his retirement in 2015. Personal life Mackay-Sim was born on 16 May 1951. He grew up in Roseville, New South Wales, the third of four brothers. He attended North Sydney Boys High School. Mackay-Sim was married, with two children. In 2015, he was diagnosed with multiple myeloma, and was treated with stem cell transplant. Mackay-Sim retired in 2015, and lived in Currimundi, Queensland. He died on 4 January 2023, at the age of 71. Awards Member of the Order of Australia, for "significant service to tertiary education, and to biomedical science", in the 2021 Queen's Birthday Honours. Australian of the Year, 2017 Queenslander of the Year, 2003 Neil Hamilton Fairley Medal, 2018 Australasian Neuroscience Society Distinguished Achievement Award, 2017 Eureka Science Prize, 2011 References External links Professor Alan Mackay-Sim at Griffith University 1951 births 2023 deaths People with multiple myeloma Stem cell researchers Australian medical researchers Australian of the Year Award winners Academic staff of Griffith University Macquarie University alumni Members of the Order of Australia University of Pennsylvania alumni University of Wyoming alumni People educated at North Sydney Boys High School Australian republicans
Alan Mackay-Sim
[ "Biology" ]
395
[ "Stem cell researchers", "Stem cell research" ]
52,980,074
https://en.wikipedia.org/wiki/Five%20safes
The Five Safes is a framework for helping make decisions about making effective use of data which is confidential or sensitive. It is mainly used to describe or design research access to statistical data held by government and health agencies, and by data archives such as the UK Data Service. Two of the Five Safes refer to statistical disclosure control, and so the Five Safes is usually used to contrast statistical and non-statistical controls when comparing data management options. Concept The Five Safes proposes that data management decisions be considered as solving problems in five 'dimensions': projects, people, settings, data and outputs. The combination of the controls leads to 'safe use'. These are most commonly expressed as questions, for example: These dimensions are scales, not limits. That is, solutions can have a mix of more or fewer controls in each dimension, but the overall aim of 'safe use' independent of the particular mix. For example, a public use file available for open download cannot control who uses it, where or for what purpose, and so all the control (protection) must be in the data itself. In contrast, a file which is only accessed through a secure environment with certified users can contain very sensitive information: the non-statistical controls allow the data to be 'unsafe'. One academic likened the process to a graphic equalizer, where bass and treble can be combined independently to produce a sound the listener likes. There is no 'order' to the Five Safes, in that one is necessarily more important than the others. However, Ritchie argued that the 'managerial' controls (projects, people, setting) should be addressed before the 'statistical' controls (data, output). The Five Safes concept is associated with other topics which developed from the same programme at ONS, although these are not necessarily implemented. Safe people is associated with 'active researcher management', while safe outputs is linked with principles-based output statistical disclosure control. The Five Safes is a positive framework, describing what is and is not. The EDRU ('evidence-based, default-open, risk-managed, user-centred') attitudinal model is sometimes used to give a normative context The 'data access spectrum' From 2003 the Five Safes was also represented in a simpler form as a 'Data Access Spectrum'.The non-data controls (project, people, setting, outputs) tend to work together, in that organisations often see these as a complementary set of restrictions on access. These can then be contrasted with choices about data anonymisation to present a linear representation of data access options. This presentation is consistent with the idea of 'data as a residual', as well as data protection laws of the time which often characterised data simply as anonymous or not anonymous. A similar idea had already been developed independently in 2001 by Chuck Humphrey of the Canadian RDC network, the 'continuum of access'. More recently, The Open Data Institute has developed a 'Data Spectrum toolkit' which includes industry-specific examples. History and terminology The Five Safes was devised in the winter of 2002/2003 by Felix Ritchie at the UK Office for National Statistics (ONS) to describe its secure remote-access Virtual Microdata Laboratory (VML). It was described at this time as the 'VML Security Model'. This was adopted by the NORC data enclave, and more widely in the US, as the 'portfolio model' (although this is now also used to refer to a slightly different legal/statistical/educational breakdown). In 2012 the framework as was still being referred to as the 'VML security model', but its increasing use among non-UK organisations led to the adoption of the more general and informative phrase 'Five Safes'. The original framework only had four safes (projects, people, settings and outputs): the framework was used to describe highly detailed data access through a secure environment, and so the 'data' dimension was irrelevant. From 2007 onwards, 'safe data' was included as the framework was used to a describe a wider range of ONS activities. As the US version was based upon the 2005 specification, some US iterations uses have the original four dimensions (eg). Some discussions, such as the OECD, use the term 'secure' instead 'safe'. However, the use of both these terms can cause presentational problems: less control in a particular dimension could be seen to imply 'unsafe users' or 'insecure settings', for example, which distracts from the main message. Hence, the Australian government uses the term "five data sharing principles". The 'Anonymisation Decision-Making Framework' uses a framework based on the Five Safes but relabelling "projects", "people", and "settings" as "governance", "agency" and "infrastructure", respectively; "Output" is omitted, and "safe use" becomes "functional anonymisation". There is no reference to the Five Safes or any associated literature. The Australian version was required to include references to the Five Safes, and presented it as an alternative without comment. Application The framework has had three uses: pedagogical, descriptive, and design. Since 2016, it has also been used, directly and indirectly in legislation. See for more detailed examples. Pedagogy The first significant use of the framework, other than internal administrative use, was to structure researcher training courses at the UK Office for National Statistics from 2003. UK Data Archive, Administrative Data Research Network, Eurostat, Statistics New Zealand, the Mexican National Institute of Statistics and Geography, NORC, Statistics Canada and the Australian Bureau of Statistics, amongst others, have also used this framework. Most of these courses are for researchers using restricted-access facilities; the Eurostat courses are unusual in that they are designed for all users of sensitive data. Description The framework is often used to describe existing data access solutions (e.g. UK HMRC Data Lab, UK Data Service, Statistics New Zealand) or planned/conceptualised ones (e.g. Eurostat in 2011). An early use was to help identify areas where ONS' still had 'irreducible risks' in its provision of secure remote access. The framework is mostly used for confidential social science data. To date it appears to have made little impact on medical research planning, although it is now included in the revised guidelines on implementing HIPAA regulations in the US, and by Cancer Research UK and the Health Foundation in the UK. It has also been used to describe a security model for the Scottish Health Informatics Programme. Design In general the Five Safes has been used to describe solutions post-factum, and to explain/justify choices made, but an increasing number of organisations have used the framework to design data access solutions. For example, the Hellenic Statistical Agency developed a data strategy built around the Five Safes in 2016; the UK Health Foundation used the Five Safes to design its data management and training programmes. Use in the private sector is less common but some organisations have incorporated the Five Safes into consulting services. In 2015 the UK Data Service organized a workshop to encourage data users from the academic and private sectors to think about how to manage confidential research data, using the Five Safes to demonstrate alternative options and best practice. Early adopters for strategic design use were in Australia: both the Australian Bureau of Statistics and the Australian Department of Social Service used the Five Safes as an ex ante design tool. In 2017 the Australian Productivity Commission recommended adopting a version of the framework to support cross-government data sharing and re-use. This underwent extensive consultation and culminated in the DAT Act 2022. Since 2020 the Five Safes has been the overriding framework for the design of new secure facilities and data sharing arrangements in the UK for public health and social sciences. This has been promoted by the Office for Statistics Regulation, the UK Statistics Authority, NHS DIgital, and the research funding bodies Administrative Data Research UK and DARE UK. Regulation and legislation Three laws have incorporated the Fives Safes. They are explicit in the South Australian Public Sector (Data Sharing) Act 2016, and implicit in the research provisions of the UK Digital Economy Act 2017. The Australian Data Availability and Transparency Act 2022 renames the Five Safes as the Five Data Sharing Principles. Public engagement The UK Data Service has produced a blog and video for the general public about the use of Five Safes in re-using administrative data. Statistics New Zealand produced a non-technical description, as did ONS for Data Privacy Day 2017. The Australian Federal Government has produced several videos on data sharing, including the Data Sharing Principles. Criticism In the 2020 paper, "Not fit for Purpose: A critical analysis of the ‘Five Safes’", the authors argue that Five Safes is fundamentally flawed due to its disconnection from existing legal protections, its appropriation of safety notions without strong technical measures, and its static view of disclosure risk. Others have argued that the Five Safes has too little content to be useful, or is a box-ticking exercise, or that more 'safes' are needed. Green and Ritchie (2023) provide an extensive review of these critiques and proposals. References External links fivesafes.org- original concept website aihw.gov.au/.../the-five-safes-framework - example as implemented by Australian Institute of Health and Welfare ukdataservice.ac.uk/.../what-is-the-five-safes-framework - example as implemented by UK Data Service Confidentiality Data management Disclosure Human subject research National statistical services Privacy
Five safes
[ "Technology" ]
1,978
[ "Data management", "Data" ]
52,981,048
https://en.wikipedia.org/wiki/Species%20pool
The ecological and biogeographical concept of the species pool describes all species available that could potentially colonize and inhabit a focal habitat area. The concept lays emphasis on the fact that "local communities aren't closed systems, and that the species occupying any local site typically came from somewhere else", however, the species pool concept may suffer from the logical fallacy of composition. Most local communities, however, have just a fraction of its species pool present. It is derived from MacArthur and Wilson's Island Biogeography Theory that examines the factors that affect the species richness of isolated natural communities. It helps to understand the composition and richness of local communities and how they are influenced by biogeographic and evolutionary processes acting at large spatial and temporal scales. The absent portion of species pool—dark diversity—has been used to understand processes influencing local communities. Methods to estimate potential but absent species are developing. It has been hypothesized that there might be a direct correlation between species richness and the size of the species pool for plant communities. Elsewhere, it was reported that "trade-offs and species pool structure (size and trait distribution) determines the shape of the plant productivity-diversity relationship. References Biogeography Measurement of biodiversity
Species pool
[ "Biology" ]
253
[ "Biogeography", "Biodiversity", "Measurement of biodiversity" ]
52,981,231
https://en.wikipedia.org/wiki/Cognition%20enhanced%20Natural%20language%20Information%20Analysis%20Method
Cognition enhanced Natural language Information Analysis Method (CogNIAM) is a conceptual fact-based modelling method, that aims to integrate the different dimensions of knowledge: data, rules, processes and semantics. To represent these dimensions world standards SBVR, BPMN and DMN from the Object Management Group (OMG) are used. CogNIAM, a successor of NIAM, is based on the work of knowledge scientist Sjir Nijssen. CogNIAM structures knowledge, gathered from people, documentation and software, by classifying it. For this purpose CogNIAM uses the so-called ‘Knowledge Triangle’. The outcome of CogNIAM is independent of the person applying it. The resulting model allows the knowledge to be expressed in diagrammatic form as well as in controlled natural language. The different dimensions of knowledge CogNIAM recognises 4 different dimensions of knowledge: Data: What are the facts? Process: How are facts generated/deleted/altered? Semantics: What do the facts mean? Rules: What conditions apply on the facts? These dimensions influence each other heavily. Rules restrict data, Semantics describe the concepts and terms used in processes etc., therefore The aim of CogNIAM is to integrate these different dimensions. Structuring knowledge As mentioned earlier, CogNIAM classifies knowledge using the knowledge triangle . The knowledge that can be mapped to the knowledge triangle is structurally relevant and can be verbalised. Knowledge that cannot be verbalised, for example the ‘Mona Lisa’, is not included. Also the knowledge must be structurally relevant. Not structurally relevant is for example motivation (the why?). It is important information, but it is not an added value to the model. The remaining knowledge can be mapped to the knowledge triangle. The knowledge triangle consists of three levels Level 1 – The level of facts The majority of knowledge consists of concrete facts. Facts describe possible current, past or future states. In CogNIAM a fact is defined as “a proposition taken to be true by a relevant community”. An example of a level 1 fact is: “The capital of Italy is Rome.” Level 2 – The domain-specific level In this level the rules that govern the facts of level 1 are specified. For the example above a rule governing the level 1 facts could be “a country has exactly one capital”. This is a rule that ensures no untrue states or disallowed transitions between different states can occur at level 1. Besides rules level 2 contains six more knowledge categories, which are discussed in the next chapter. Level 3 – The generic level This level is not associated to any specific domain, it says nothing about capitals or countries. As level 2 governs the facts on level 1, the generic level governs the knowledge categories of level 2. It consists of the same knowledge categories, but here they are applied to the content of level 2. In other words, level 3 contains the rules that determine the rules. The generic level can also be seen as a domain-specific level with the domain being ‘domain-specific knowledge’. As a result, level 3 also governs itself. Knowledge categories Level 2 and 3 of the knowledge triangle consist of seven knowledge categories: Concept definitions describe the meaning of every term or group of terms at the fact level. A large part of the semantics dimension can be found here. Fact types provide the functionality to define which kinds of facts are considered to be within the scope of the domain of interest. Communication patterns: Fact communication patterns act as a communication mechanism to be used as a template to communicate facts using terms the subject matter expert is familiar with Rule communication patterns act as communication mechanism for the rules (see below) of the conceptual schema. Rules, distinguishing between: Integrity or validation rules, also known as constraints, restrict the set of facts and the transitions between the permitted sets of facts to those that are considered useful. In terms of data quality, integrity rules are used to guarantee the quality of the facts. Derivation rules are used to derive or calculate new information (facts) based on existing information. Exchange rules transfer facts into the administration of that domain or remove facts from the administration. In other words, they specify how facts are added and/or removed from the fact base so that the system stays in sync with the communication about the outside world. Event rules specify when to update the set of ground facts by a derivation rule or exchange rule in the context of a process description. Process descriptions specify the fact consuming and/or fact generating activities (the exchange and/or derivation rules) to be performed by the different actors for that process, as well as the event rules invoking the execution of those exchange and derivation rules in an ordered manner. Actors, identifying the involved participants and their responsibilities in the processes (in terms of the exchange and derivation rules they need to execute). Services, identifying the realisations of the process descriptions in terms of information products to be delivered or consulted References Data modeling Data modeling languages
Cognition enhanced Natural language Information Analysis Method
[ "Engineering" ]
1,008
[ "Data modeling", "Data engineering" ]
52,981,810
https://en.wikipedia.org/wiki/International%20Society%20of%20Explosives%20Engineers
The International Society of Explosives Engineers (ISEE) is a tax-exempt professional body founded in 1974 to "advance the science and art of explosives engineering." Headquartered in Cleveland, Ohio, it is the primary international organization for explosives engineers. Current president of ISEE is Janeen T. Smith. History ISEE was founded in Pittsburgh in 1974, when a small group of explosives engineers came together to discuss how to advance the science and art of explosives engineering. Since then the ISEE has grown to over 4000 members with 45 local chapters. Publications The ISEE has 2 primary peer-reviewed publications, and an industry handbook. Journals Journal of Explosives Engineering Blasting and Fragmentation Journal Books Blasters' Handbook, 18th Edition Conferences and education Each year more than 1500 blasters, manufacturers, government officials and industry leaders, come together for an annual conference. The Blasters Weekend is a preconference event targeted at field personnel focusing on more practical knowledge and skills. The main conference is 3 days of technical papers and presentations. Both conference events qualify for continuing education units (CEUs) and/or professional development hours (PDHs) for most states licensing programs. ISEE also founded ISEE children's funds which is now contributing in educational expenses of dependents of a person who was killed in an explosive accident or suffered a related incident. Awards and designations Industry awards and designations presented at the annual national conference. Blaster's Leadership Award Driller's Leadership Award Distinguished Service Award Paper of the Year President's Award References External links Industrial organization Explosives engineering 1974 establishments in Pennsylvania Organizations established in 1974
International Society of Explosives Engineers
[ "Engineering" ]
320
[ "Explosives engineering" ]
52,982,183
https://en.wikipedia.org/wiki/Toxoplasma%20lactate%20dehydrogenase%201%20regulatory%20UTR
In Toxoplasma gondii the translational repression of lactate dehydrogenase 1 (LDH1) was discovered to be mediated through its 5′UTR. A small nucleotide regulatory RNA hairpin was shown to be essential for the repression. It is possible that this hairpin may act as the nucleation site for the binding of a trans-acting factor(s) that allow for the translational repression. References RNA Non-coding RNA Cis-regulatory RNA elements
Toxoplasma lactate dehydrogenase 1 regulatory UTR
[ "Chemistry" ]
104
[ "Biochemistry stubs", "Molecular and cellular biology stubs" ]
52,982,637
https://en.wikipedia.org/wiki/Applied%20Imagination
Applied Imagination is an influential 1953 book on creative ideation by Alex Faickney Osborn, in which he introduces the technique of brainstorming. Chapters The all-importance of imagination Indispensability of creativity in science Careers depend largely on creativity Creativity in leadership and professions Imagination can improve personal relations Universality of imaginative talent Ways by which creativity can be developed Our new environment - its effect on creativity Other factors that tend to cramp creativity Creative and non-creative forms of imagination The process of ideation vary widely Orientation calls for setting our sights Preparation and analysis go hand in hand Editions Osborn, Alex F. (1953). Applied Imagination: Principles and Procedures of Creative Problem Solving. New York: Charles Scribner's Sons, 1953. Revised edition, New York, Scribner, 1957 3rd ed. New York C. Scribner 1963 French translation by Georges Rona and Pierre Dupont, L'Imagination constructive. Principes et processus de la Pensée créative et du Brainstorming, Paris, Dunod, 1959. Chinese translation by Ikkō Shō, 応用想像力 Taipei : Kyōshi Kōgyō Sōsho Shuppan Kofun Yūgen Kōshi, 1965 References Creativity Group problem solving methods 1953 non-fiction books
Applied Imagination
[ "Biology" ]
257
[ "Creativity", "Behavior", "Human behavior" ]
52,983,512
https://en.wikipedia.org/wiki/Pi%20Fornacis
π Fornacis (Latinised as Pi Fornacis) is the Bayer designation for a binary star system in the southern constellation of Fornax. It has an apparent visual magnitude of 5.360, which is bright enough to be seen with the naked eye on a dark night. With an annual parallax shift of 11.08 mas, it is estimated to lie around 294 light years from the Sun. At that distance, the visual magnitude is diminished by an interstellar absorption factor of 0.10 due to dust. This system is a member of the thin disk population of the Milky Way galaxy. The primary, component A, is an evolved G-type giant star with a stellar classification of G8 III. It has an estimated mass slightly higher than the Sun, but has expanded to more than nine times the Sun's radius. The star is roughly five billion years old and is spinning slowly with a projected rotational velocity of 0.9 km/s. Pi Fornacis A radiates 57.5 times the solar luminosity from its outer atmosphere at an effective temperature of 5,048 K. A companion, component B, was discovered in 2008 using the AMBER instrument of the Very Large Telescope facility. At the time of discovery, this star lay at an estimated angular separation of from the primary along a position angle of . The preliminary orbital period for the pair is 11.4 years, and the semimajor axis is at least 70 mas. The orbit is highly inclined to the line of sight from the Earth. References G-type giants Binary stars Fornax Fornacis, Pi CD-30 703 012438 09440 594
Pi Fornacis
[ "Astronomy" ]
343
[ "Fornax", "Constellations" ]
51,433,179
https://en.wikipedia.org/wiki/Cosmon
In physical cosmology, a cosmon or cosmonium is a hypothetical form of matter. The idea was originally proposed by Georges Lemaître, who suggested the concept of a 'primeval atom’ (L'Hypothèse de l'Atome Primitif) 1946, leading up to the theory of the Big Bang. He illustrated the idea by imagining an object 30 times larger than the volume of the sun containing all the matter of the Universe. Its density would be around . In his view, this occurred somewhere between 20 and 60 billion years ago. The idea of a primeval “super-atom” lived on and was developed forward by Maurice Goldhaber in 1956. In his proposal there would have been a point, which had been called a universon, that would have collapsed into a cosmon and an anti-cosmon pair. Goldhaber was questioned why is there any matter if equal amounts of matter and antimatter were formed in the Big Bang. One explanation for this is the asymmetry of matter meaning that there could have been slightly more matter than antimatter, for instance 1001 matter particles to every 1000 antimatter. In Goldhaber's model, cosmon and anticosmon would have flown apart and therefore explaining issue without asymmetry. In 1989, Hans Dehmelt attempted to modernize the idea of the primeval atom. In this hypothesis, The cosmonium would have been the heaviest form of matter at the beginning of the Big Bang. References Physical cosmology Hypothetical elementary particles
Cosmon
[ "Physics", "Astronomy" ]
325
[ "Astronomical sub-disciplines", "Theoretical physics", "Unsolved problems in physics", "Astrophysics", "Hypothetical elementary particles", "Physics beyond the Standard Model", "Physical cosmology" ]
51,433,654
https://en.wikipedia.org/wiki/Sempati%20Air%20Flight%20304
Sempati Air Flight 304 (SSR304) was a scheduled domestic passenger flight, flying from Husein Sastranegara International Airport in Bandung, West Java to Halim Perdanakusuma Airport in Jakarta. On 17 July 1997, the aircraft operating the flight, a Fokker F27 Friendship 600 leased from Trigana Air Service, registered PK-YPM, crashed into a densely populated neighbourhood of Margahayu in Bandung after suffering an engine failure shortly after take-off killing 28 of the occupants and injuring several on the ground. Subsequent investigation by the Indonesian National Transportation Safety Committee found out that the aircraft had suffered an engine failure on its left engine; the pilot failed to execute a proper go-around procedure and crashed into the densely populated residential area of Margahayu. Accident The aircraft took off from Husein Sastranegara Airport at 11:46 WIB (04:46 UTC), with Bambang Rudy Santoso as PIC. Approximately 3 minutes after take off, the left engine of the Fokker F27 failed and began spewing thick smoke. The pilot contacted Husein Sastranegara ATC, choosing to land at Sulaiman Airbase runway 13. During the emergency landing procedure, the aircraft clipped houses in Margahayu, Bandung and crashed into a muddy field killing 28 people and severely injuring several people on the ground. Most of the victims were families travelling for school holidays. Aircraft The aircraft involved in the crash was a Fokker F27 Friendship 600 built in 1969 with a serial number of 10415, delivered to Merpati Nusantara Airlines as PK-MHF in 1974. Later, it was sold to Trigana Air Service in 1993 as PK-YPM, Sempati Air in January 1995 and returned to Trigana Air Service in August 1995. References Airliner accidents and incidents caused by mechanical failure Airliner accidents and incidents caused by engine failure Aviation accidents and incidents in 1997 Aviation accidents and incidents in Indonesia Accidents and incidents involving the Fokker F27 1997 disasters in Indonesia July 1997 events in Asia
Sempati Air Flight 304
[ "Materials_science" ]
435
[ "Airliner accidents and incidents caused by mechanical failure", "Mechanical failure" ]
51,434,698
https://en.wikipedia.org/wiki/Infinitesimal%20cohomology
In mathematics, infinitesimal cohomology is a cohomology theory for algebraic varieties introduced by . In characteristic 0 it is essentially the same as crystalline cohomology. In nonzero characteristic p showed that it is closely related to etale cohomology with mod p coefficients, a theory known to have undesirable properties. References . . Algebraic geometry Cohomology theories
Infinitesimal cohomology
[ "Mathematics" ]
80
[ "Fields of abstract algebra", "Algebraic geometry" ]
51,434,916
https://en.wikipedia.org/wiki/Operational%20continuity
Operational continuity refers to the ability of a system to continue working despite damages, losses, or critical events. In the Human Resources and Organizational domain, including IT, it implies the need to determine the level of resilience of the system, its ability to recover after an event, and build a system that resists to external and internal events or is able to recover after an event without losing its external performance management capability. Organizational Continuity is achieved only with specific corporate planning. In the material domain, it determines the need to adopt redundant systems, performance monitoring systems, and can even imply the practice to cannibalize or to remove serviceable assemblies, sub-assemblies or components from a repairable or serviceable item of equipment to install them on another, in order to keep the external system performance active. Operational continuity may be referred to single systems, single individuals, up to teams or entire complex systems such as IT infrastructures, implying the ability of an organization or system to continue to provide its mission. See also Business continuity Disaster response Emergency management References External links National Institute of Science and Technology (NIST) Special Publication 800-34: Contingency Planning Guide for Information Technology Systems The Professional Practices for Business Continuity Management, Disaster Recovery Institute International (DRI), 2017. Business continuity Information technology Systems theory
Operational continuity
[ "Technology" ]
264
[ "Information and communications technology", "Information technology" ]
51,437,144
https://en.wikipedia.org/wiki/Well%20of%20Dina%20Nath
The Well of Dina Nath () was intended to be a water well in the Wazir Khan Chowk in Lahore, Pakistan. The well's construction by in the 19th century by a Sikh nobleman sparked controversy, given its location in the immediate vicinity of the Wazir Khan Mosque. History The well was commissioned by Raja Dina Nath in the mid 19th century under the reign of Ranjit Singh. The well was not built as an open well, but is instead enclosed within a walled structure. Legend It is said that Nath wished to build his well near the site of a well dug Sufi saint Said Soaf despite strong objections from local Muslim leaders who viewed construction of a second well to be antagonistic to the saint's memory. Disregarding their warnings and objections, Dina Nath ordered construction to begin on the site. After 200 metres of digging, labourers could not tap a water source, and refused to dig any further, much to the embarrassment of Dina Nath. The well has remained dry ever since, and remains a local monument. Restoration The well fell into disrepair, and was eventually surrounded by illegally constructed shops which had encroached upon the Wazir Khan Chowk. In 2012, the Aga Khan Trust for Culture and the Government of Punjab launched restoration efforts which have since removed the illegal shops, restoring public access to the well. See also Haveli Dina Nath References Buildings and structures in Lahore Walled City of Lahore Sikh architecture Water wells
Well of Dina Nath
[ "Chemistry", "Engineering", "Environmental_science" ]
299
[ "Hydrology", "Water wells", "Environmental engineering" ]
51,437,973
https://en.wikipedia.org/wiki/Panting%20%28ship%20construction%29
Panting refers to the tendency of steel hull plating to flex in and out like an oil can being squeezed when a ship is pitching. This occurs when a ship is making headway in waves. Panting creates significant stress on a ship's hull. It is potentially dangerous and can result in flooding and the separation of the hull and deck. The British battleship HMS Rodney suffered significant leaking from panting. Addressing panting is an essential component of ship design. It is typically countered by reinforcing the bow and the stern with beams and stringers. References Naval architecture Shipbuilding
Panting (ship construction)
[ "Engineering" ]
114
[ "Naval architecture", "Shipbuilding", "Marine engineering" ]
51,443,362
https://en.wikipedia.org/wiki/Data%20augmentation
Data augmentation is a statistical technique which allows maximum likelihood estimation from incomplete data. Data augmentation has important applications in Bayesian analysis, and the technique is widely used in machine learning to reduce overfitting when training machine learning models, achieved by training models on several slightly-modified copies of existing data. Synthetic oversampling techniques for traditional machine learning Synthetic Minority Over-sampling Technique (SMOTE) is a method used to address imbalanced datasets in machine learning. In such datasets, the number of samples in different classes varies significantly, leading to biased model performance. For example, in a medical diagnosis dataset with 90 samples representing healthy individuals and only 10 samples representing individuals with a particular disease, traditional algorithms may struggle to accurately classify the minority class. SMOTE rebalances the dataset by generating synthetic samples for the minority class. For instance, if there are 100 samples in the majority class and 10 in the minority class, SMOTE can create synthetic samples by randomly selecting a minority class sample and its nearest neighbors, then generating new samples along the line segments joining these neighbors. This process helps increase the representation of the minority class, improving model performance. Data augmentation for image classification When convolutional neural networks grew larger in mid-1990s, there was a lack of data to use, especially considering that some part of the overall dataset should be spared for later testing. It was proposed to perturb existing data with affine transformations to create new examples with the same labels, which were complemented by so-called elastic distortions in 2003, and the technique was widely used as of 2010s. Data augmentation can enhance CNN performance and acts as a countermeasure against CNN profiling attacks. Data augmentation has become fundamental in image classification, enriching training dataset diversity to improve model generalization and performance. The evolution of this practice has introduced a broad spectrum of techniques, including geometric transformations, color space adjustments, and noise injection. Geometric Transformations Geometric transformations alter the spatial properties of images to simulate different perspectives, orientations, and scales. Common techniques include: Rotation: Rotating images by a specified degree to help models recognize objects at various angles. Flipping: Reflecting images horizontally or vertically to introduce variability in orientation. Cropping: Removing sections of the image to focus on particular features or simulate closer views. Translation: Shifting images in different directions to teach models positional invariance. Color Space Transformations Color space transformations modify the color properties of images, addressing variations in lighting, color saturation, and contrast. Techniques include: Brightness Adjustment: Varying the image's brightness to simulate different lighting conditions. Contrast Adjustment: Changing the contrast to help models recognize objects under various clarity levels. Saturation Adjustment: Altering saturation to prepare models for images with diverse color intensities. Color Jittering: Randomly adjusting brightness, contrast, saturation, and hue to introduce color variability. Noise Injection Injecting noise into images simulates real-world imperfections, teaching models to ignore irrelevant variations. Techniques involve: Gaussian Noise: Adding Gaussian noise mimics sensor noise or graininess. Salt and Pepper Noise: Introducing black or white pixels at random simulates sensor dust or dead pixels. Data augmentation for signal processing Residual or block bootstrap can be used for time series augmentation. Biological signals Synthetic data augmentation is of paramount importance for machine learning classification, particularly for biological data, which tend to be high dimensional and scarce. The applications of robotic control and augmentation in disabled and able-bodied subjects still rely mainly on subject-specific analyses. Data scarcity is notable in signal processing problems such as for Parkinson's Disease Electromyography signals, which are difficult to source - Zanini, et al. noted that it is possible to use a generative adversarial network (in particular, a DCGAN) to perform style transfer in order to generate synthetic electromyographic signals that corresponded to those exhibited by sufferers of Parkinson's Disease. The approaches are also important in electroencephalography (brainwaves). Wang, et al. explored the idea of using deep convolutional neural networks for EEG-Based Emotion Recognition, results show that emotion recognition was improved when data augmentation was used. A common approach is to generate synthetic signals by re-arranging components of real data. Lotte proposed a method of "Artificial Trial Generation Based on Analogy" where three data examples provide examples and an artificial is formed which is to what is to . A transformation is applied to to make it more similar to , the same transformation is then applied to which generates . This approach was shown to improve performance of a Linear Discriminant Analysis classifier on three different datasets. Current research shows great impact can be derived from relatively simple techniques. For example, Freer observed that introducing noise into gathered data to form additional data points improved the learning ability of several models which otherwise performed relatively poorly. Tsinganos et al. studied the approaches of magnitude warping, wavelet decomposition, and synthetic surface EMG models (generative approaches) for hand gesture recognition, finding classification performance increases of up to +16% when augmented data was introduced during training. More recently, data augmentation studies have begun to focus on the field of deep learning, more specifically on the ability of generative models to create artificial data which is then introduced during the classification model training process. In 2018, Luo et al. observed that useful EEG signal data could be generated by Conditional Wasserstein Generative Adversarial Networks (GANs) which was then introduced to the training set in a classical train-test learning framework. The authors found classification performance was improved when such techniques were introduced. Mechanical signals The prediction of mechanical signals based on data augmentation brings a new generation of technological innovations, such as new energy dispatch, 5G communication field, and robotics control engineering. In 2022, Yang et al. integrate constraints, optimization and control into a deep network framework based on data augmentation and data pruning with spatio-temporal data correlation, and improve the interpretability, safety and controllability of deep learning in real industrial projects through explicit mathematical programming equations and analytical solutions. See also Oversampling and undersampling in data analysis Surrogate data Generative adversarial network Variational autoencoder Data pre-processing Convolutional neural network Regularization (mathematics) Data preparation Data fusion References Machine learning
Data augmentation
[ "Engineering" ]
1,328
[ "Artificial intelligence engineering", "Machine learning" ]
41,387,300
https://en.wikipedia.org/wiki/Empty%20forest
Empty forest is a term coined by Kent H. Redford's article "The Empty Forest" (1992), which was published in BioScience. An "empty forest" refers to an ecosystem that is void of large mammals. Empty forests are characterized by an otherwise excellent habitat, and often have large, fully grown trees, although they lack large mammals as a result of human impact. Empty forests show that human impact can destroy an ecosystem from within as well as from without. Many of the large mammals that are disappearing, such as deers and tapirs, are important for seed dispersion. Many tree species that are very localized in their dispersion rely on mammals rather than the wind to disperse their seeds. Furthermore, when seed predation is down, trees with large seeds begin to completely dominate those with small seeds, changing the balance of plant life in an area. Predatory large mammals are important for increasing overall biodiversity by making sure that smaller predators and herbivores do not become overabundant and dominate. An absence of large predators seems to result in uneven densities of prey species. Even though certain animals may not have become completely extinct, they may have lowered in numbers to the point that they have suffered an ecological extinction. The animals that have most likely suffered an ecological extinction in neotropical forests are the ones who are the most important predators, large seed dispersers, and seed predators. The defaunation of large mammals can be done by direct or indirect means. Any type of human activity not aimed at the animals in question that results in the defaunation of those animals is indirect. The most common means of indirect defaunation is habitat destruction. However, other examples of indirect means of defaunation of large mammals would be the over-collection of fruits and nuts or over-hunting of prey that large mammals need for food. Another example of an indirect means of the defaunation of large mammals is through the by-products of modern human activities such as mercury and smoke, or even noise pollution. There are two categories of direct defaunation. They include subsistence hunting and commercial hunting. The most common species of animals hunted are typically the largest species in their area. The large mammals in an area are often represented by only a few species, but make up a major part of the overall biomass. In areas with only moderate hunting, the biomass of mammalian game species decreases by 80.7%. In areas with heavy hunting, the biomass of mammalian game species can decrease by 93.7%. See also Silent Spring Pollinator decline References Ecology
Empty forest
[ "Biology" ]
521
[ "Ecology" ]
41,387,674
https://en.wikipedia.org/wiki/Dorymyrmex%20bicolor
Dorymyrmex bicolor is a species of ant in the Dolichoderinae subfamily, known as bicolored pyramid ant due to the two-toned body color and the shape of their mounds. Dorymyrmex bicolor was recently known as Conomyrma bicolor but has been renamed to Dorymyrmex bicolor. It has a single petiole and a slit-like orifice which releases chemical compounds. This ant does not have the capability to sting. It is primarily found in arid desert regions in Central and South America and the southwestern United States. Tool use Dorymyrmex bicolor regularly interacts with three different species of Myrmecocystus, another genus of ants. Between the two genera of ants, in the southwestern region, there is much overlap of food sources and space between Dorymyrmex bicolor and Myrmecocystus. The Myrmecocystus ants secrete a substance from their poison gland onto a food source they find in order to repel other ants. Dorymyrmex bicolor exhibits a different type of interference behavior. The workers of Dorymyrmex bicolor will surround the entrance to a nest of Myrmecocystus and will drop stones and other objects down the entrance, in an attempt to block the entrance. As many as 10-30 workers of Dorymyrmex bicolor have been observed to drop stones in an opposing nest, but only 5 workers are required to drop stones and other small objects at an efficient rate that will affect the Myrmecocystus nest. The number of Dorymyrmex bicolor workers in an area will have a reducing effect on the number of Myrmecocystus workers in an area, sometimes to drastic effects. Distribution The known range of Dorymyrmex bicolor stretches from the Southwestern United States (California, Texas, Arizona, New Mexico, and Oklahoma), northern and southern Mexico (including Baja California), El Salvador, Guatemala, Belize, Peru, Honduras, and a few Caribbean island nations (i.e. Jamaica). Behavior Dorymyrmex bicolor has a primarily foraging focused behavior, because of this they are known to influence rates of seed germination and distribution in plants. They move quickly and are active from early to late afternoon. They build crater-shaped nests made of fine sand, as they primarily live in desert-like areas though they can live in areas with higher humidity levels. References External links Dorymyrmex Hymenoptera of North America Fauna of California Insects of the United States Insects described in 1906 Tool-using animals
Dorymyrmex bicolor
[ "Biology" ]
545
[ "Ethology", "Behavior", "Tool-using animals" ]
41,388,326
https://en.wikipedia.org/wiki/PLISSIT%20model
The PLISSIT model, also known as the PLISSIT model of sex therapy, is a modeling system used in the field of sexology to determine the different levels of intervention for individual clients. The model was created in 1976 by Jack S. Annon. The letters of the name refer to the four different levels of intervention that a sexologist can apply: permission (P), limited information (LI), specific suggestions (SS), and intensive therapy (IT). The model is also used outside the field of sexology, especially in fields involving extensive or life-threatening surgery. Structure The model created by Annon has four levels of increasing intervention and interaction related to what kind of and how much help is given to a client. The varying levels largely revolve around what the client is looking for and how comfortable they are in discussing sexuality and sexual health. The first level is permission, which involves the sexologist giving the client permission to feel comfortable about a topic or permission to change their lifestyle or to get medical assistance. This level was created because many clients only require the permission to speak and voice their concerns about sexual issues in order to understand and move past them, often without needing the other levels of the model. The sexologist, in acting as a receptive, nonjudgmental listening partner, allows the client to discuss matters that would otherwise be too embarrassing for the individual to discuss. The second level is limited information, wherein the client is supplied with limited and specific information on the topics of discussion. Because there is a significant amount of information available, sexologists must learn what sexual topics the client wishes to discuss, so that information, organizations, and support groups for those specific subjects can be provided. The third level is specific suggestions, where the sexologist gives the client suggestions related to the specific situations and assignments to do in order to help the client fix the mental or health problem. This can include suggestions on how to deal with sex related diseases or information on how to better achieve sexual satisfaction by the client changing their sexual behavior. The suggestions may be as simple as recommending exercise or can involve specific regimens of activity or medications. The fourth and final level is intensive therapy, which has the sexologist refer the client to other mental and medical health professionals that can help the client deal with the deeper, underlying issues and concerns being expressed. This level, with the onset of the internet age, may also refer to a sexologist suggesting professional online resources for the client to browse about their specific issue in a more private setting. EX-PLISSIT model The PLISSIT model was extended in 2006 by Sally Davis and Bridget Taylor because of concerns that practitioners often bypass permission-giving and go straight to providing information (sometimes merely in the form of a leaflet), without giving patients the opportunity to express any concerns they might have. The extended model, named the EX-PLISSIT model, places permission-giving at the core. By giving people explicit permission to discuss any concerns they have about their sexuality, the healthcare professional affirms the individual as a sexual being. Any information or suggestions that follow, are then specific to the needs of that person. The EX-PLISSIT model also requires further permission-giving in the form of 'review', whereby the healthcare professional asks the patient to review the interaction and is given the opportunity to express any further worries or concerns. In addition, this model requires the professional to reflect on their interactions, challenging assumptions and extending their knowledge. References Further reading Annon, Jack S. (1975) The Behavioral Treatment of Sexual Problems. Vol 1: Brief Therapy. Harper & Row Annon, Jack S. (1976) The Behavioral Treatment of Sexual Problems Vol. 2: Intensive Therapy. Harper & Row Counseling Human sexuality Sex therapy
PLISSIT model
[ "Biology" ]
768
[ "Human sexuality", "Behavior", "Human behavior", "Sexuality" ]
41,390,065
https://en.wikipedia.org/wiki/Loewy%20decomposition
In the study of differential equations, the Loewy decomposition breaks every linear ordinary differential equation (ODE) into what are called largest completely reducible components. It was introduced by Alfred Loewy. Solving differential equations is one of the most important subfields in mathematics. Of particular interest are solutions in closed form. Breaking ODEs into largest irreducible components, reduces the process of solving the original equation to solving irreducible equations of lowest possible order. This procedure is algorithmic, so that the best possible answer for solving a reducible equation is guaranteed. A detailed discussion may be found in. Loewy's results have been extended to linear partial differential equations (PDEs) in two independent variables. In this way, algorithmic methods for solving large classes of linear PDEs have become available. Decomposing linear ordinary differential equations Let denote the derivative with respect to the variable . A differential operator of order is a polynomial of the form where the coefficients , are from some function field, the base field of . Usually it is the field of rational functions in the variable , i.e. . If is an indeterminate with , becomes a differential polynomial, and is the differential equation corresponding to . An operator of order is called reducible if it may be represented as the product of two operators and , both of order lower than . Then one writes , i.e. juxtaposition means the operator product, it is defined by the rule ; is called a left factor of , a right factor. By default, the coefficient domain of the factors is assumed to be the base field of , possibly extended by some algebraic numbers, i.e. is allowed. If an operator does not allow any right factor it is called irreducible. For any two operators and the least common left multiple is the operator of lowest order such that both and divide it from the right. The greatest common right divisior is the operator of highest order that divides both and from the right. If an operator may be represented as of irreducible operators it is called completely reducible. By definition, an irreducible operator is called completely reducible. If an operator is not completely reducible, the of its irreducible right factors is divided out and the same procedure is repeated with the quotient. Due to the lowering of order in each step, this proceeding terminates after a finite number of iterations and the desired decomposition is obtained. Based on these considerations, Loewy obtained the following fundamental result. The decomposition determined in this theorem is called the Loewy decomposition of . It provides a detailed description of the function space containing the solution of a reducible linear differential equation . For operators of fixed order the possible Loewy decompositions, differing by the number and the order of factors, may be listed explicitly; some of the factors may contain parameters. Each alternative is called a type of Loewy decomposition. The complete answer for is detailed in the following corollary to the above theorem. Corollary 1 Let be a second-order operator. Its possible Loewy decompositions are denoted by , they may be described as follows; and are irreducible operators of order ; is a constant. The decomposition type of an operator is the decomposition with the highest value of . An irreducible second-order operator is defined to have decomposition type . The decompositions , and are completely reducible. If a decomposition of type , or has been obtained for a second-order equation , a fundamental system may be given explicitly. Corollary 2 Let be a second-order differential operator, , a differential indeterminate, and . Define for and , is a parameter; the barred quantities and are arbitrary numbers, . For the three nontrivial decompositions of Corollary 1 the following elements and of a fundamental system are obtained. is not equivalent to . Here two rational functions are called equivalent if there exists another rational function such that There remains the question how to obtain a factorization for a given equation or operator. It turns out that for linear ode's finding the factors comes down to determining rational solutions of Riccati equations or linear ode's; both may be determined algorithmically. The two examples below show how the above corollary is applied. Example 1 Equation 2.201 from Kamke's collection. has the decomposition The coefficients and are rational solutions of the Riccati equation , they yield the fundamental system Example 2 An equation with a type decomposition is The coefficient of the first-order factor is the rational solution of . Upon integration the fundamental system and for and respectively is obtained. These results show that factorization provides an algorithmic scheme for solving reducible linear ode's. Whenever an equation of order 2 factorizes according to one of the types defined above the elements of a fundamental system are explicitly known, i.e. factorization is equivalent to solving it. A similar scheme may be set up for linear ode's of any order, although the number of alternatives grows considerably with the order; for order the answer is given in full detail in. If an equation is irreducible it may occur that its Galois group is nontrivial, then algebraic solutions may exist. If the Galois group is trivial it may be possible to express the solutions in terms of special function like e.g. Bessel or Legendre functions, see or. Basic facts from differential algebra In order to generalize Loewy's result to linear PDEs it is necessary to apply the more general setting of differential algebra. Therefore, a few basic concepts that are required for this purpose are given next. A field is called a differential field if it is equipped with a derivation operator. An operator on a field is called a derivation operator if and for all elements . A field with a single derivation operator is called an ordinary differential field; if there is a finite set containing several commuting derivation operators the field is called a partial differential field. Here differential operators with derivatives and with coefficients from some differential field are considered. Its elements have the form ; almost all coefficients are zero. The coefficient field is called the base field. If constructive and algorithmic methods are the main issue it is . The respective ring of differential operators is denoted by or . The ring is non-commutative, and similarly for the other variables; is from the base field. For an operator of order the symbol of L is the homogeneous algebraic polynomial where and algebraic indeterminates. Let be a left ideal which is generated by , . Then one writes . Because right ideals are not considered here, sometimes is simply called an ideal. The relation between left ideals in and systems of linear PDEs is established as follows. The elements are applied to a single differential indeterminate . In this way the ideal corresponds to the system of PDEs , for the single function . The generators of an ideal are highly non-unique; its members may be transformed in infinitely many ways by taking linear combinations of them or its derivatives without changing the ideal. Therefore, M. Janet introduced a normal form for systems of linear PDEs (see Janet basis). They are the differential analog to Gröbner bases of commutative algebra (which were originally introduced by Bruno Buchberger); therefore they are also sometimes called differential Gröbner basis. In order to generate a Janet basis, a ranking of derivatives must be defined. It is a total ordering such that for any derivatives , and , and any derivation operator the relations , and are valid. Here graded lexicographic term orderings are applied. For partial derivatives of a single function their definition is analogous to the monomial orderings in commutative algebra. The S-pairs in commutative algebra correspond to the integrability conditions. If it is assured that the generators of an ideal form a Janet basis the notation is applied. Example 3 Consider the ideal in term order with . Its generators are autoreduced. If the integrability condition is reduced with respect to , the new generator is obtained. Adding it to the generators and performing all possible reductions, the given ideal is represented as . Its generators are autoreduced and the single integrability condition is satisfied, i.e. they form a Janet basis. Given any ideal it may occur that it is properly contained in some larger ideal with coefficients in the base field of ; then is called a divisor of . In general, a divisor in a ring of partial differential operators need not be principal. The greatest common right divisor (Gcrd) or sum of two ideals and is the smallest ideal with the property that both and are contained in it. If they have the representation and , for all and , the sum is generated by the union of the generators of and . The solution space of the equations corresponding to is the intersection of the solution spaces of its arguments. The least common left multiple (Lclm) or left intersection of two ideals and is the largest ideal with the property that it is contained both in and . The solution space of is the smallest space containing the solution spaces of its arguments. A special kind of divisor is the so-called Laplace divisor of a given operator , page 34. It is defined as follows. Definition Let be a partial differential operator in the plane; define and be ordinary differential operators with respect to or ; for all i; and are natural numbers not less than 2. Assume the coefficients , are such that and form a Janet basis. If is the smallest integer with this property then is called a Laplace divisor of . Similarly, if , are such that and form a Janet basis and is minimal, then is also called a Laplace divisor of . In order for a Laplace divisor to exist the coeffients of an operator must obey certain constraints. An algorithm for determining an upper bound for a Laplace divisor is not known at present, therefore in general the existence of a Laplace divisor may be undecidable. Decomposing second-order linear partial differential equations in the plane Applying the above concepts Loewy's theory may be generalized to linear PDEs. Here it is applied to individual linear PDEs of second order in the plane with coordinates and , and the principal ideals generated by the corresponding operators. Second-order equations have been considered extensively in the literature of the 19th century,. Usually equations with leading derivatives or are distinguished. Their general solutions contain not only constants but undetermined functions of varying numbers of arguments; determining them is part of the solution procedure. For equations with leading derivative Loewy's results may be generalized as follows. Theorem 2 Let the differential operator be defined by where for all . Let for and , and be first-order operators with ; is an undetermined function of a single argument. Then has a Loewy decomposition according to one of the following types. The decomposition type of an operator is the decomposition with the highest value of . If does not have any first-order factor in the base field, its decomposition type is defined to be . Decompositions , and are completely reducible. In order to apply this result for solving any given differential equation involving the operator the question arises whether its first-order factors may be determined algorithmically. The subsequent corollary provides the answer for factors with coefficients either in the base field or a universal field extension. Corollary 3 In general, first-order right factors of a linear pde in the base field cannot be determined algorithmically. If the symbol polynomial is separable any factor may be determined. If it has a double root in general it is not possible to determine the right factors in the base field. The existence of factors in a universal field, i.e. absolute irreducibility, may always be decided. The above theorem may be applied for solving reducible equations in closed form. Because there are only principal divisors involved the answer is similar as for ordinary second-order equations. Proposition 1 Let a reducible second-order equation where . Define , for ; is a rational first integral of ; and the inverse ; both and are assumed to exist. Furthermore, define for . A differential fundamental system has the following structure for the various decompositions into first-order components. The are undetermined functions of a single argument; , and are rational in all arguments; is assumed to exist. In general , they are determined by the coefficients , and of the given equation. A typical example of a linear pde where factorization applies is an equation that has been discussed by Forsyth, vol. VI, page 16, Example 5 (Forsyth 1906) Consider the differential equation . Upon factorization the representation is obtained. There follows Consequently, a differential fundamental system is and are undetermined functions. If the only second-order derivative of an operator is , its possible decompositions involving only principal divisors may be described as follows. Theorem 3 Let the differential operator be defined by where for all . Let and are first-order operators. has Loewy decompositions involving first-order principal divisors of the following form. The decomposition type of an operator is the decomposition with highest value of . The decomposition of type is completely reducible In addition there are five more possible decomposition types involving non-principal Laplace divisors as shown next. Theorem 4 Let the differential operator be defined by where for all . and as well as and are defined above; furthermore , , . has Loewy decompositions involving Laplace divisors according to one of the following types; and obey . If does not have a first order right factor and it may be shown that a Laplace divisor does not exist its decomposition type is defined to be . The decompositions , , and are completely reducible. An equation that does not allow a decomposition involving principal divisors but is completely reducible with respect to non-principal Laplace divisors of type has been considered by Forsyth. Example 6 (Forsyth 1906) Define generating the principal ideal . A first-order factor does not exist. However, there are Laplace divisors and The ideal generated by has the representation , i.e. it is completely reducible; its decomposition type is . Therefore, the equation has the differential fundamental system and Decomposing linear PDEs of order higher than 2 It turns out that operators of higher order have more complicated decompositions and there are more alternatives, many of them in terms of non-principal divisors. The solutions of the corresponding equations get more complex. For equations of order three in the plane a fairly complete answer may be found in. A typical example of a third-order equation that is also of historical interest is due to Blumberg. Example 7 (Blumberg 1912) In his dissertation Blumberg considered the third order operator It allows the two first-order factors and . Their intersection is not principal; defining it may be written as . Consequently, the Loewy decomposition of Blumbergs's operator is It yields the following differential fundamental system for the differential equation . , , and are an undetermined functions. Factorizations and Loewy decompositions turned out to be an extremely useful method for determining solutions of linear differential equations in closed form, both for ordinary and partial equations. It should be possible to generalize these methods to equations of higher order, equations in more variables and system of differential equations. References Differential equations
Loewy decomposition
[ "Mathematics" ]
3,142
[ "Mathematical objects", "Differential equations", "Equations" ]
41,390,796
https://en.wikipedia.org/wiki/Galileo%20%281968%20film%29
Galileo (also known as Galileo Galilei) is a 1968 Italian–Bulgarian biographical drama film directed by Liliana Cavani. It depicts the life of Galileo Galilei and particularly his conflicts with the Catholic Church over his scientific theories. Plot Astronomer Galileo Galilei teaches at the University of Padua. While he questions the ideas of Ptolemy and Aristotle, the official scientific dogmas imposed by the Catholic Church, he remains secretive about his doubts. His more candid friend, philosopher Giordano Bruno, is reported to the Inquisition for his revolutionary ideas and later executed as a heretic. Still, Galileo continues his studies with a telescope constructed by Dutch technicians and perfected by him, and comes to the conclusion that Copernico's heliocentric system is valid. He publishes his discoveries in a book, which leads to a series of interrogations by the Inquisition. Facing a possible death sentence, Galileo publicly recants his theories. Cast Cyril Cusack as Galileo Galilei Georgi Kaloyanchev as Giordano Bruno Nevena Kokanova as Marina Nicolai Doicev as Cardinal Bellarmino Georgi Cherkelov as Paolo Sarpi Piero Vida as Pope Urban VIII Gigi Ballista as Dominican Commissioner Paolo Graziosi as Gian Lorenzo Bernini Maia Dragomanska as Galilei's daughter Lou Castel as Father Charles Giulio Brogi as Sagredo Production and release Originally intended as a miniseries co-produced by Italian and Bulgarian film companies, radio and television company RAI refused to broadcast the finished film and sold the distribution rights to Cineriz, who trimmed the originally 105 minutes long film to 92 minutes running time. Galileo was shown in competition at the 1968 Venice International Film Festival. Home media Galileo was released in 2010 as a Region 2 DVD. See also Life of Galileo, a play by Bertolt Brecht Galileo, a 1975 film adaptation of Brecht's play References External links Article by Cristina Olivotto and Antonella Testa comparing the 1968 and 1975 films 1968 drama films 1960s biographical drama films 1960s historical drama films Bulgarian biographical drama films 1960s Italian-language films Bulgarian historical drama films Films scored by Ennio Morricone Films directed by Liliana Cavani Italian biographical drama films Italian historical drama films Films shot in Bulgaria Films shot at Cinecittà Studios Films set in the 1590s Films set in Rome Cultural depictions of Galileo Galilei Cultural depictions of Giordano Bruno 1960s Italian films
Galileo (1968 film)
[ "Astronomy" ]
498
[ "Cultural depictions of astronomers", "Cultural depictions of Galileo Galilei" ]
41,391,212
https://en.wikipedia.org/wiki/Timeline%20of%20women%20in%20mathematics%20in%20the%20United%20States
There is a long history of women in mathematics in the United States. All women mentioned here are American unless otherwise noted. Timeline 19th Century 1829: The first public examination of an American girl in geometry was held. 1886: Winifred Edgerton Merrill became the first American woman to earn a PhD in mathematics, which she earned from Columbia University. 1891: Charlotte Angas Scott of Britain became the first woman to join the American Mathematical Society, then called the New York Mathematical Society. 1894: Charlotte Angas Scott of Britain became the first woman on the first Council of the American Mathematical Society. 20th Century 1913: Mildred Sanderson earned her PhD for a thesis that included an important theorem about modular invariants. 1927: Anna Pell-Wheeler became the first woman to present a lecture at the American Mathematical Society Colloquium. 1943: Euphemia Haynes became the first African-American woman to earn a Ph.D. in mathematics, which she earned from Catholic University of America. 1949: Gertrude Mary Cox became the first woman elected into the International Statistical Institute. 1956: Gladys West began collecting data from satellites at the Naval Surface Warfare Center Dahlgren Division. Her calculations directly impacted the development of accurate GPS systems. 1962: Mina Rees became the first person to receive the Award for Distinguished Service to Mathematics from the Mathematical Association of America. 1966: Mary L. Boas published Mathematical Methods in the Physical Sciences, which was still widely used in college classrooms as of 1999. 1970s 1970: Mina Rees became the first female president of the American Association for the Advancement of Science. 1971: Mary Ellen Rudin constructed the first Dowker space. The Association for Women in Mathematics (AWM) was founded. It is a professional society whose mission is to encourage women and girls to study and to have active careers in the mathematical sciences, and to promote equal opportunity for and the equal treatment of women and girls in the mathematical sciences. It is incorporated in the state of Massachusetts. The American Mathematical Society established its Joint Committee on Women in the Mathematical Sciences (JCW), which later became a joint committee of multiple scholarly societies. 1973: Jean Taylor published her dissertation on "Regularity of the Singular Set of Two-Dimensional Area-Minimizing Flat Chains Modulo 3 in R3" which solved a long-standing problem about length and smoothness of soap-film triple function curves. 1974: Joan Birman published the book Braids, Links, and Mapping Class Groups. It has become a standard introduction, with many of today's researchers having learned the subject through it. 1975: Julia Robinson became the first female mathematician elected to the National Academy of Sciences. 1976-1977: Marjorie Rice, an amateur mathematician, discovered four new types of tessellating pentagons in 1976 and 1977. 1979: Dorothy Lewis Bernstein became the first female president of the Mathematical Association of America. Mary Ellen Rudin became the first woman to present the Mathematical Association of America’s Earle Raymond Hedrick Lectures, intended to showcase skilled expositors and enrich the understanding of instructors of college-level mathematics. 1980s 1981: Canadian-American mathematician Cathleen Morawetz became the first woman to give the Gibbs Lecture of the American Mathematical Society. 1981: Doris Schattschneider became the first female editor of Mathematics Magazine, a refereed bimonthly publication of the Mathematical Association of America. 1983: Julia Robinson became the first female president of the American Mathematical Society, and the first female mathematician to be awarded a MacArthur Fellowship. 1987: Eileen Poiani became the first female president of Pi Mu Epsilon. 1988: Doris Schattschneider became the first woman to present the Mathematical Association of America’s J. Sutherland Frame Lectures. 1990s 1992: Gloria Gilmer became the first woman to deliver a major National Association of Mathematicians lecture (it was the Cox-Talbot address). 1995: Margaret Wright became the first female president of the Society for Industrial and Applied Mathematics. 1996: Joan Birman became the first woman to receive the Mathematical Association of America’s Chauvenet Prize, an annual award for expository articles. 1998: Melanie Wood became the first female American to make the U.S. International Math Olympiad Team. She won silver medals in the 1998 and 1999 International Mathematical Olympiads. 21st Century 2002: Melanie Wood became the first American woman and second woman overall to be named a Putnam Fellow in 2002. Putnam Fellows are the top five (or six, in case of a tie) scorers on William Lowell Putnam Mathematical Competition. 2004: Melanie Wood became the first woman to win the Frank and Brennie Morgan Prize for Outstanding Research in Mathematics by an Undergraduate Student. It is an annual award given to an undergraduate student in the US, Canada, or Mexico who demonstrates superior mathematics research. Alison Miller became the first female gold medal winner on the U.S. International Mathematical Olympiad Team. 2006: Stefanie Petermichl, a German mathematical analyst then at the University of Texas at Austin, became the first woman to win the Salem Prize, an annual award given to young mathematicians who have worked in Raphael Salem's field of interest, chiefly topics in analysis related to Fourier series. She shared the prize with Artur Avila. 2019: Karen Uhlenbeck became the first woman to win the Abel Prize, with the award committee citing "the fundamental impact of her work on analysis, geometry and mathematical physics." Marissa Kawehi Loving became the first Native Hawaiian woman to earn a PhD in mathematics when she graduated from the University of Illinois Urbana-Champaign in 2019. In addition to being Native Hawaiian, she is also black, Japanese, and Puerto Rican. 2020: Lisa Piccirillo published a mathematical proof in the journal Annals of Mathematics determining that the Conway knot is not a smoothly slice knot, answering an unsolved problem in knot theory first proposed over fifty years prior by English mathematician John Horton Conway. See also Timeline of women in mathematics References Further reading A Brief History of the Association for Women in Mathematics: The Presidents' Perspectives, by Lenore Blum (1991) History of women in the United States Women in mathematics in America mathematics usa Women women in mathematics in America
Timeline of women in mathematics in the United States
[ "Technology" ]
1,245
[ "Women in science and technology", "Women in mathematics" ]
41,392,084
https://en.wikipedia.org/wiki/Cyclic%20subspace
In mathematics, in linear algebra and functional analysis, a cyclic subspace is a certain special subspace of a vector space associated with a vector in the vector space and a linear transformation of the vector space. The cyclic subspace associated with a vector v in a vector space V and a linear transformation T of V is called the T-cyclic subspace generated by v. The concept of a cyclic subspace is a basic component in the formulation of the cyclic decomposition theorem in linear algebra. Definition Let be a linear transformation of a vector space and let be a vector in . The -cyclic subspace of generated by , denoted , is the subspace of generated by the set of vectors . In the case when is a topological vector space, is called a cyclic vector for if is dense in . For the particular case of finite-dimensional spaces, this is equivalent to saying that is the whole space . There is another equivalent definition of cyclic spaces. Let be a linear transformation of a topological vector space over a field and be a vector in . The set of all vectors of the form , where is a polynomial in the ring of all polynomials in over , is the -cyclic subspace generated by . The subspace is an invariant subspace for , in the sense that . Examples For any vector space and any linear operator on , the -cyclic subspace generated by the zero vector is the zero-subspace of . If is the identity operator then every -cyclic subspace is one-dimensional. is one-dimensional if and only if is a characteristic vector (eigenvector) of . Let be the two-dimensional vector space and let be the linear operator on represented by the matrix relative to the standard ordered basis of . Let . Then . Therefore and so . Thus is a cyclic vector for . Companion matrix Let be a linear transformation of a -dimensional vector space over a field and be a cyclic vector for . Then the vectors form an ordered basis for . Let the characteristic polynomial for be . Then Therefore, relative to the ordered basis , the operator is represented by the matrix This matrix is called the companion matrix of the polynomial . See also Companion matrix Krylov subspace External links PlanetMath: cyclic subspace References Linear algebra
Cyclic subspace
[ "Mathematics" ]
449
[ "Linear algebra", "Algebra" ]
41,392,177
https://en.wikipedia.org/wiki/Rui%20L.%20Reis
Rui Luís Reis (born 19 April 1967) is a Portuguese scientist known for his research in tissue engineering, regenerative medicine, biomaterials, biomimetics, stem cells, and biodegradable polymers. Reis is a professor of at the University of Minho in Braga and Guimarães. He is the Founding Director of the 3B's Research Group, part of the Research Institute on Biomaterials, Biodegradables and Biomimetics (I3Bs) of UMinho (www.i3bs.uminho.pt), a group that specializes in the areas of Regenerative Medicine, Tissue Engineering, Stem Cells and Biomaterials. He is also the Director of the ICVS/3B's Associate Laboratory of UMinho. He is the CEO of the European Institute of Excellence on Tissue Engineering and Regenerative Medicine. Rui L. Reis was, from 2013 to 2017, the Vice-Rector (vice-president) for research and innovation of UMinho. From 2007 to 2021 Reis was the editor-in-chief of the Journal of Tissue Engineering and Regenerative Medicine. From 2016 to 2018, he was president of the Tissue Engineering and Regenerative Medicine International Society (TERMIS). Reis is in the board of several scientific societies, companies and associations. From 2017 to 2019, he was the President of TECMINHO - the technology transfer office of the University of Minho. Reis is the CEO of the European Institute of Excellence on Tissue Engineering and Regenerative Medicine in Avepark, Guimarães. He co-founded different start up companies originating from the research and activities of 3B's research group, such as Stemmatters and HydruStent/HydruMedical. Reis is the current president of the I3B's research institute, and one of the most cited Portuguese researchers in science. Biography Reis was born and has always lived in Porto, being one of three children of a chemical engineering professor and a domestic. Reis spent a small part of his childhood in Metangula, Mozambique, a small town near Lake Niassa, while his father was engaged in military service during the Portuguese Colonial War. He is married with Olga Paiva and has one son, Bernardo Reis (born in 2001). He is a strong supporter of FC Porto. Reis graduated in Metallurgical Engineering, University of Porto, Portugal, in 1990. He then completed a master's degree at the University of Porto, Portugal, in 1994. Reis did his PhD on Polymer Engineering – Biomaterials, Regenerative Medicine & Tissue Engineering, in the University of Minho, Portugal and Brunel University London, in 1999. He also completed a Doctor of Science (D.Sc.) degree on Biomedical Engineering - Biomaterials & Tissue Engineering, by University of Minho, Portugal, in 2007. Reis has also received two Honoris Causa degrees: A first in Medicine from University of Granada, Spain, in 2010 and a second in Engineering from University Polytechnica of Bucharest, Romania, in 2018. Career Reis is a researcher who has been involved in the field of biomaterials since 1990. He has worked with several universities and companies abroad. Some of Reis' research has been on liver and neurological tissues regeneration, new strategies for antimicrobial materials, innovative high-throughput approaches for studying cell/materials interactions, as well as on TE approaches for developing different 3D disease models, including different cancer models, and therapies for treatment of diabetes and Alzheimers. Reis has also been responsible for several cooperation programs with universities and companies worldwide. He has coordinated four major EU research projects, including the STREP "HIPPOCRATES". Under HORIZON 2020, Reis was the coordinator of the ERA Chairs FoReCast grant for 3B's-UMinho. He has coordinated two TWINNING projects Gene2Skin and Chem2Nature, and is currently coordinating another TWINNING project. Until 2021, he was the coordinator of the 15 MEuros EC funded TEAMING proposal, "The Discoveries Centre for Regenerative and Precision Medicine" with UCL - University College London, UPorto, UAveiro, ULisboa, and UNova Lisboa. He is also the PI of a major project of the Portuguese roadmap for strategic infrastructures, TERM Research Hub. Honours and awards 2002: Jean LeRay Award by the European Society for Biomaterials for outstanding contributions to the biomaterials field as a young scientist 2007: Pfizer Award for Clinical Research 2007: START Innovation Award 2011: George Winter Award by the European Society for Biomaterials 2011: Gold Medal of Scientific Merit from the City of Guimarães 2014: Clemson Award for Contributions to the Literature by the Society for Biomaterials (SFB, USA) 2014: Nomination as a Commander (Comendador, a kind of knighthood) of the Military Order of Saint James of the Sword by the Portuguese President of the Republic 2015: International Fellow of Tissue Engineering of Regenerative Medicine (FTERM), Boston 2016: Induction as a foreigner member of the National Academy of Engineering (NAE) of the USA 2018: IET A F Harvey Prize – Institute of Engineering and Technology 2018: Induction as Fellow of the European Alliance for Medical and Biological Engineering and Science (EAMBES) 2018: UNESCO-Equatorial Guinea International Prize for Research in the Life Sciences 2018: Induction as Fellow of the American Institute for Medical and Biological Engineering (AIMBE) 2018: Honoris Causa degree awarded by the University Politechnica of Bucharest (UPB) 2019: Career Achievement Award of the Tissue Engineering and Regenerative Medicine International Society, TERMIS-EU 2020: Gold Medal of the City of Braga 2022: Klaas de Groot Award by the European Society for Biomaterials References External links University of Minho, 3B's Research Group in Biomaterials, Biodegradables and Biomimetics. Rui L. Reis. 1967 births Living people Portuguese medical researchers People from Porto Biomedical engineers Biomaterials Tissue engineering University of Porto alumni University of Minho alumni
Rui L. Reis
[ "Physics", "Chemistry", "Engineering", "Biology" ]
1,293
[ "Biomaterials", "Biological engineering", "Cloning", "Chemical engineering", "Materials", "Tissue engineering", "Matter", "Medical technology" ]
41,392,354
https://en.wikipedia.org/wiki/Zden%C4%9Bk%20%C5%A0vestka
Zdeněk Švestka (30 September 1925 – 2 March 2013) was a Czech astronomer. For several decades he was the world's leading expert on solar flares. He studied mathematics and physics at Charles University, Prague, until graduating in 1948. Together with Cornelis de Jager, he was the co-founder and editor of the journal Solar Physics. For 38 years, from the establishment of the journal in 1967 until his retirement in 2005, he handled all papers on solar flares, while De Jager took care of everything else. The minor planet 17805 Švestka was named after him. References Czech astronomers 2013 deaths 1925 births Scientists from Prague Charles University alumni
Zdeněk Švestka
[ "Astronomy" ]
139
[ "Astronomers", "Astronomer stubs", "Astronomy stubs" ]
41,392,746
https://en.wikipedia.org/wiki/Melvin%20Fitting
Melvin Fitting (born January 24, 1942) is a logician with special interests in philosophical logic and tableau proof systems. He was a professor at Lehman College and the Graduate Center of the City University of New York from 1968 to 2013. At the Graduate Center he was in the departments of Computer Science, Philosophy, and Mathematics, and at Lehman College he was in the department of Mathematics and Computer Science. He is now Professor emeritus. Fitting was born in Troy, New York. His undergraduate degree is from Rensselaer Polytechnic Institute, and his doctorate is from Yeshiva University, both in mathematics. His thesis advisor was Raymond Smullyan. In June 2012 Melvin Fitting was given the Herbrand Award by the Conference on Automated Deduction, for distinguished contributions to automated deduction. A loose motivation for much of Melvin Fitting's work can be formulated succinctly as follows. There are many logics. Our principles of reasoning vary with context and subject matter. Multiplicity is one of the glories of modern formal logic. The common thread tying logics together is a concern for what can be said (syntax), what that means (semantics), and relationships between the two. A philosophical position that can be embodied in a formal logic has been shown to be coherent, not correct. Logic is a tool, not a master, but it is an enjoyable tool to use. Notes References External links Melvin Fitting, official homepage The CUNY Graduate Center, faculty page Mathematical Genealogy Project 1942 births Living people 20th-century American philosophers 21st-century American philosophers American logicians CUNY Graduate Center faculty Lehman College faculty Computability theorists Rensselaer Polytechnic Institute alumni People from Troy, New York Proof theorists Yeshiva University alumni Mathematicians from New York (state)
Melvin Fitting
[ "Mathematics" ]
354
[ "Proof theorists", "Proof theory" ]
41,393,355
https://en.wikipedia.org/wiki/Restoration%20economy
The restoration economy is the economic activity associated with regenerative land use, such as ecological restoration activities. It stands in contrast to economic activity premised on sprawl, or on the extraction or depletion of natural resources. The term is meant to convey that activities meant to repair past damage to natural and human communities are often economically beneficial at local, regional, and national scales. Overview The "restoration economy" refers to an economic model based on repurposing, renewing and reconnecting the natural, built and socioeconomic environments. The phrase gained popularity with the publication in 2002 of The Restoration Economy by Storm Cunningham. That book created an eight-sector taxonomy for "restorative development", with a chapter on each. Four professional / technical / scientific sectors primarily focused on the built environment: brownfield remediation, infrastructure renewal, heritage restoration and catastrophe reconstruction. The other four sectors primarily focused on the natural environment: ecological restoration, watershed restoration, fishery restoration and regenerative agriculture. All eight sectors help to revitalize the socioeconomic environment. On the natural resources side of the equation, the "restoration economy" refers to the employment, capital, resources, and economic activity that emerge from investments in ecological restoration, or "the process of assisting the recovery of an ecosystem that has been degraded, damaged, or destroyed." Restoration projects can include habitat enhancement, water quality improvement, invasive species removal, forest thinning for canopy diversification, or any other activity that aims to improve the natural function of an ecosystem. While investments in restoration benefit the environment, restoration projects also require workers, materials, and services to implement. The marketplace for these goods and services can create employment, spur business and workforce development, and increase activity in local economies. Activities that use byproducts of restoration work are also sometimes considered as part of the restoration economy; for example, the use of small trees and/or shrubs from forest diversification or thinning projects as biomass to produce heat or energy. History The emergence of the restoration economy as a concept followed closely from historic shifts in natural resource policy in the mid-1990s, when the northern spotted owl and several species of salmon with habitat in the Pacific Northwest of North America were listed as endangered. The listings created a significant shift in forest management policies across the region, leading to drastic decreases in logging and other natural resource extraction activities that would further destroy habitats for these species. The listings also marked a shift toward more sustainable land management through forest and watershed restoration, as policymakers began to realize that environmental restoration work could provide social and economic, in addition to ecological, benefits. Federal and state agencies sought to replace some of the economic activity that was lost as extraction activities declined with economic activity from restoration work. The Northwest Forest Plan (1994) initiated several programs intended to transition traditional logging communities and workforces toward livelihoods in restoration work. The Collaborative Forest Landscape Restoration Act (2009), the National Forest Economic Action Plan, and stewardship contracting opportunities also focus on marrying ecological with socioeconomic benefits from national forest management. In Oregon, decreased timber yields and the potential for additional salmon listings that would further affect traditional land management activities led to the creation of community-based watershed councils under the Oregon Watershed Health Program (1993), and later the Oregon Plan for Salmon and Watersheds (1997). With dedicated state lottery funding, these watershed councils continue to promote voluntary restoration actions that improve salmon habitat by encouraging collaboration at the local level and promoting both socioeconomic and ecological benefits through restoration. Businesses, employees, and working conditions Agencies and nongovernmental organizations that implement restoration work create employment by hiring staff and contracting with businesses that hire employees to perform the actual restoration work. In Oregon, contracting businesses involved in restoration projects tend to be small businesses (as defined by the Small Business Administration), family-owned, and subject to large seasonal fluctuations in available work. Differences among businesses in the state appear to be related to customer base—businesses that work for non-federal customers such as watershed councils tend to perform more equipment-based work close to their home, versus businesses working for federal agencies who tend to travel more often and perform more labor-intensive work. These findings have important implications, as restoration investments are frequently intended to increase economic activity in rural communities affected by decreased logging activity, and these benefits are lost to local economies if contractors from outside of the area capture the contract. In addition, labor-intensive activities (e.g. planting trees, thinning) tend to pay less than traditional logging jobs, be less reliable, and be dominated by immigrant workers. In equipment-based and technically intensive restoration activities, working conditions are generally thought to be better. Measuring economic activity United States Several studies have examined the economic impacts and activity created by restoration work in numbers. These have typically been for individual projects, e.g. a large hazardous fuels reduction contract in Arizona, or a $113 million mine reclamation project in Montana that was estimated to generate a total of 3,563 full-time equivalent (FTE) positions. A study looking at the economic out of a restoration program across an entire county (Humboldt County, California) determined that the $12.5 million invested in the program in 2002 supported 300 direct jobs that year. A study that looked at investments in restoration projects through both federal agencies and nongovernmental organizations across the state of Oregon found that restoration investments have similar economic and employment impacts as other public infrastructure projects such as road building. Per million invested, 15-24 jobs were created, and each invested dollar generated an additional 1.4 to 2.4 dollars in economic activity as it cycled through the economy, depending on the specific type of project. Workers, or restoration practitioners, in the restoration economy are also vulnerable to policy changes, and the "volatility of restoration funding." In Humboldt County, California, private sector businesses and non-governmental organizations are often compromising with bureaucratic hindrances and drastic changes in funding. South Africa Restoration economy has been implemented in South Africa as well. For example, the ARISE Project in the Giyani district of the Limpopo province, was a poverty-alleviation program contracted to a private company in 1999 by the Department of Environmental Affairs and Tourism, with the intention of raising the overall quality of life in that region. The program created jobs in ecological restoration of the surrounding landscape and helped to economically empower rural Africans in the short run. Restoration activities included creating fences around off-limit areas, planting approximately 8000 trees and creating a fifty-meter buffer zone on both sides of the Klein Letaba River. Over 323 jobs were created, and employees were earning nearly twice the average daily income ($5.30 per household from $2.20). In South Africa, while the ARISE project was able to positively affect the Giyani District by providing jobs for a short period to low-income earners, it did not increase the likelihood of long employment later in the commercial sector. Namibia In its efforts to control woody plant encroachment, the Namibian government deliberately supports the commercial utilisation of excess wood as an eocomic incentive and pull factor for restoration. This has resulted in the emergence of a so-called bush-biomass sector, boasting various different value chains. At the core of Namibia´s restoration economy approach is the sustainable harvesting of shrubs. Such harvesting is meant to reduce the density woody plants and allow for a better between grass and bushes in the affected savanna rangelands. A total of over 12,000 workers find employment in harvesting and processing of excess wood. It was estimated that the bush control and utilisaiton could lead to an aggregated net benefit of around USD 3.8 billion over a time period of 25 years, taking into consideration . Ongoing benefits Although the potential exists for a win-win situation pairing environmental betterment with socioeconomic benefits, the dual-benefits potential from this kind of work is often overlooked or under-emphasized. Like its emergence from policy with aims to provide socioeconomic benefits in concert with natural resource work, a robust and enduring restoration economy depends on policy that continues to prioritize multiple benefits for both human and natural systems, in addition to initiating investments in restoration activities. Indices of performance and knowledge of the traits associated with successful establishment and persistence in restored vegetation are potentially of great benefit to practitioners and policymakers involved in restoration increased land productivity, enhanced groundwater recharge, woody utilisation and carbon offsets. See also Restoration ecology Northwest Forest Plan Ecosystem services Green economy References External links Institute for Resilient Organizations, Communities, and Environments, University of Oregon US Forest Service, Collaborative Forest Landscape Restoration Program Ecotrust Whole Watershed Restoration Initiative Society for Ecological Restoration Environmental economics Economy by field Ecological restoration
Restoration economy
[ "Chemistry", "Engineering", "Environmental_science" ]
1,771
[ "Environmental economics", "Environmental social science", "Ecological restoration", "Environmental engineering" ]
41,393,993
https://en.wikipedia.org/wiki/OMPL
OMPL (Open Motion Planning Library) is a software package for computing motion plans using sampling-based algorithms. The content of the library is limited to motion planning algorithms, which means there is no environment specification, no collision detection or visualization. This is intentional as the library is designed to be easily integrated into systems that already provide the additional needed components. For example, OMPL is integrated with ROS and MoveIt!. In 2012 OMPL won the Grand Prize at the Open Source Software World Challenge. Implementation OMPL is written in C++ but also offers Python bindings. The library includes implementations for a large number of planning algorithms, all of these being implemented on top of the same base functionality. The base functionality OMPL provides for planners is thread safe. Adding new motion planning algorithms to OMPL is easy, thus facilitating comparisons between existing algorithms and evaluations of new ideas. Target audience Teaching One of the design goals for OMPL is clarity of concepts used. This equates to having C++ classes that correspond to concepts found in the literature. Such a design facilitates using OMPL for education. Furthermore, the authors provide free course materials and assignments for use in conjunction with OMPL. Industrial use The first use for OMPL was actually at Willow Garage, where the library was started, to do motion planning for the PR2 arms. As such, the library was hardened to run reliably and efficiently. Afterwards OMPL started to be used (via ROS and MoveIt!) for hundreds of different types of robots . Research OMPL includes tools for benchmarking, providing a common platform for easily developing and testing new algorithms. References External links Computer libraries Robot kinematics Software using the BSD license Robotics software Free software programmed in C++ Cross-platform free software Free software for Linux Free software for Windows Free software for macOS
OMPL
[ "Technology", "Engineering" ]
378
[ "IT infrastructure", "Robotics engineering", "Computer libraries", "Robotics software", "Robot kinematics" ]
41,394,066
https://en.wikipedia.org/wiki/Spherical%20roller%20thrust%20bearing
A spherical roller thrust bearing is a rolling-element bearing of thrust type that permits rotation with low friction, and permits angular misalignment. The bearing is designed to take radial loads, and heavy axial loads in one direction. Typically these bearings support a rotating shaft in the bore of the shaft washer that may be misaligned in respect to the housing washer. The misalignment is possible due to the spherical internal shape of the house washer. Construction Spherical roller thrust bearings consist of a shaft washer (for radial bearings often called "inner ring"), a housing washer (for radial bearings often called "outer ring"), asymmetrical rollers and a cage. There are also bearing units available that can take axial loads in two directions. History The spherical roller thrust bearing was introduced by SKF in 1939. The design of the early bearings is similar to the design that is still in use in modern machines. Designs The internal design of the bearing is not standardised by ISO, so it varies between different manufacturers and different series. Some of the design parameters are: Roller shape and dimensions Flange design Non-rotational notches in house washer The spherical roller thrust bearings have the highest load rating density of all thrust bearings. Dimensions External dimensions of spherical roller bearings are standardised by ISO in the standard ISO 104:2015. Some common series of spherical roller bearings are: 292 293 294 Materials Bearing rings and rolling elements can be made of a number of different materials, but the most common is "chrome steel", a material with approximately 1.5% chrome content. Such "chrome steel" has been standardized by a number of authorities, and there are therefore a number of similar materials, such as: AISI 52100 (USA), 100CR6 (Germany), SUJ2 (Japan) and GCR15 (China). Some common materials for bearing cages: Sheet steel (stamped or laser-cut) Brass (stamped or machined) Steel (machined) The choice of material is mainly done by the manufacturing volume and method. For large-volume bearings, cages are often of stamped sheet-metal, whereas low volume series often have cages of machined brass or machined steel. Manufacturers Some manufacturers of spherical roller bearings are SKF, Schaeffler, Timken Company and NSK Ltd. Applications Spherical roller thrust bearings are used in industrial applications, where there are heavy axial loads, moderate speeds and possibly misalignment. Some common application areas are: Gearboxes Pulp and paper processing equipment (notably refiners) Marine propulsion and offshore drilling Cranes and swing bridges Water turbines Extruders for injection molding See also Bearing (mechanical) Rolling-element bearing Self-aligning ball bearing Spherical plain bearing Spherical roller bearing Tapered roller bearing Thrust bearing References Bearings (mechanical) Rolling-element bearings Mechanical engineering Swedish inventions
Spherical roller thrust bearing
[ "Physics", "Engineering" ]
585
[ "Applied and interdisciplinary physics", "Mechanical engineering" ]
41,394,765
https://en.wikipedia.org/wiki/York%20Leeman%20Road%20depot
The York Leeman Road railway depot, located in York, England, is a passenger multiple unit depot opened in May 2007 by Siemens. It services TransPennine Express Class 185s and locomotives. The facility's shed code is YK. History Before the 1870s, the site area was known as Bishop Fields; it was undeveloped and in agricultural use. In 1877, the new Holgate railway station (see York railway station), and its associated loop line, opened. The loop line passed through Bishop Fields and through the 20th century surrounding land north of Leeman Road. It was extensively developed, much for railway use including a large engine shed to the east, with sidings and a large carriage shed to the west. In the latter part of the 20th century, there was some contraction; the carriage shed was removed and the engine shed ceased to have an operational role until it became part of the National Railway Museum in 1975. In 2004, the site of the depot was occupied by a mixture of disused and in-use railway sidings. TransPennine Express depot In 2003, the new TransPennine Express franchise was awarded to First TransPennine Express (FTPE). As part of the franchise agreement, FTPE was to introduce a new fleet of 100 mph trains, together with new maintenance facilities for the fleet; the main depot was to be in Manchester, with a secondary depot in York. An order for 51 Class 185 diesel multiple units, and the associated maintenance facilities, was placed with Siemens in 2003. The facilities required by the franchise agreement included: a one road three-car length shed, with sidings for 8 three-car trains; siding facilities for controlled emission toilet servicing and fuelling; train electric supply (125 A three phase); offices and stores; and a capacity overhead hoist. In 2004, Siemens submitted a planning application for a depot on Leeman Road to York City Council, the plans were approved in May 2005. A ground breaking ceremony took place in December 2005. The main contractors for the £10m works were the Spencer Group. The depot was opened in May 2007, by Prince Andrew, Duke of York. Since April 2016, the Class 185s have been operated by TransPennine Express (TPE). In 2018, work began on an £11 million upgrade to the depot, which included signalling upgrades and road lengthening. The project was undertaken to allow the depot to accommodate Class 68 with Mark 5A carriages and Class 802 units. Allocation and stabling As of 2018, the depot's allocation consists of TransPennine Express locomotives and Desiros. Notes References Sources External links Railway depots in Yorkshire Rail transport in York Siemens Mobility projects
York Leeman Road depot
[ "Technology", "Engineering" ]
544
[ "Siemens Mobility projects", "Transport systems" ]
41,395,058
https://en.wikipedia.org/wiki/Journal%20of%20Cleaner%20Production
The Journal of Cleaner Production is a peer-reviewed academic journal covering transdisciplinary research on cleaner production. It is published by Elsevier. The job of editor-in-chief is shared jointly by Cecília Maria Villas Bôas de Almeida (Paulista University), and Yutao Wang (Fudan University). The former and founding editor-in-chief was Donald Huisingh (University of Tennessee). The Journal of Cleaner Production serves as a transdisciplinary, international forum for the exchange of information and research concepts, policies, and technologies designed to help ensure progress towards making societies and regions more sustainable. It aims to encourage innovation and creativity, new and improved products, and the implementation of new, cleaner structures, systems, processes, products and services. It is also designed to stimulate the development and implementation of prevention oriented governmental policies and educational programmes. References External links Elsevier academic journals Academic journals established in 1993 English-language journals Environmental science journals Journals published between 27 and 51 times per year
Journal of Cleaner Production
[ "Environmental_science" ]
206
[ "Environmental science journals", "Environmental science journal stubs" ]
41,395,916
https://en.wikipedia.org/wiki/Observation%20arc
In observational astronomy, the observation arc (or arc length) of a Solar System body is the time period between its earliest and latest observations, used for tracing the body's path. It is usually given in days or years. The term is mostly used in the discovery and tracking of asteroids and comets. Arc length has the greatest influence on the accuracy of an orbit. The number, spacing of intermediate observations, and timestamps have a lesser effect. Short arcs A very short arc leaves a high uncertainty parameter. The object might be in one of many different orbits, at many distances from Earth. In some cases, the initial arc was too short to determine if the object was in orbit around the Earth, or orbiting out in the asteroid belt. With a 1-day observation arc, was thought to be a trans-Neptunian dwarf planet, but is now known to be a 1 km main-belt asteroid. With an observation arc of 3 days, was thought to be a Mars-crossing asteroid that could be a threat to Earth, but was later found to be another main-belt asteroid. A relatively modest observation arc may allow finding an older "precovery" photo, providing a much longer arc and a more precise orbit. An observation arc less than 30 days can make it difficult to recover an Inner Solar System object more than a year after the last observation, and may result in a lost minor planet. Due to their greater distance from the Sun and slow movement across the sky, trans-Neptunian objects with observation arcs less than several years often have poorly constrained orbits. As a general rule objects discovered when they are currently farther from the Sun will have greater uncertainties in their initial orbits if the observation arcs are short. which was discovered when 100+ AU from the Sun and has only been observed 9 times over 2 years will require an observation arc of several years to refine the uncertainties in the orbital period and aphelion (farthest distance from the Sun). with only 4 observations over 1 day has uncertainties so large that the error bars are not really meaningful and just show that the uncertainties are very large. On its discovery date is estimated to have been from Earth. Oort cloud comet C/2017 K2 was announced when it had a short 2.6 day observation arc, was estimated to be from the Sun, and was estimated to come to perihelion around 10 AU from the Sun in 2027. But it is now known that C/2017 K2 was discovered when it was 16 AU from the Sun and will come to perihelion 1.8 AU from the Sun on 19 December 2022. It took an observation arc of about 200 days to rule out a Mars impact by Oort cloud comet C/2013 A1 (Siding Spring). Interstellar objects Interstellar objects generally require an observation arc of 2–3 weeks using hundreds of observations to confirm that an interloper has a hyperbolic excess velocity (interstellar speed) of more than a few km/s. Comet C/2008 J4 (McNaught) was only observed 22 times over an observation arc of 15 days, and due to an insufficient number of observations generates a low inbound interstellar speed of 3.9 km/s, but the uncertainties in the eccentricity easily produce a closed orbit with . Comet C/1999 U2 (SOHO) with an almost meaningless observation arc of 1 day shows a very dubious interstellar speed of 17 km/s, but could easily have a closed orbit with an eccentricity as low as 0.7. Earth approaches With an observation arc of 257 years, the uncertainty in Comet Swift–Tuttle's closest approach to Earth on 5 August 2126 is about ±10 thousand km. With an observation arc of ~1 year, the uncertainty in 's closest approach to Earth on 23 March 2147 is about ±2 million km. Even though C/1991 L3 (Levy) has a longer observation arc than C/2001 OG108, it has significantly fewer observations which generates a greater uncertainty. In contrast, comet C/2022 A1 (Sarneczky) was discovered on 2 January 2022 when it was 1.3 AU from the Sun, and announced on 7 January 2022 with only a 5-day observation arc. It made its closest Earth approach the next day with a 3-sigma uncertainty region of ±1 million km. The large uncertainty was a result of the short arc and discovery distance. See also Precovery Uncertainty parameter U References External links How to determine the orbit of a comet? (ESA 7 March 2014) Asteroids Observational astronomy Orbits
Observation arc
[ "Astronomy" ]
946
[ "Observational astronomy", "Astronomical sub-disciplines" ]
41,396,440
https://en.wikipedia.org/wiki/Frank%E2%80%93Kasper%20phases
Topologically close pack (TCP) phases, also known as Frank-Kasper (FK) phases, are one of the largest groups of intermetallic compounds, known for their complex crystallographic structure and physical properties. Owing to their combination of periodic and aperiodic structure, some TCP phases belong to the class of quasicrystals. Applications of TCP phases as high-temperature structural and superconducting materials have been highlighted; however, they have not yet been sufficiently investigated for details of their physical properties. Also, their complex and often non-stoichiometric structure makes them good subjects for theoretical calculations. History In 1958, Frederick C. Frank and John S. Kasper, in their original work investigating many complex alloy structures, showed that non-icosahedral environments form an open-end network which they called the major skeleton, and is now identified as the declination locus. They came up with the methodology to pack asymmetric icosahedra into crystals using other polyhedra with larger coordination numbers. These coordination polyhedra were constructed to maintain topological close packing (TCP). Unit-cell geometries classification Based on the tetrahedral units, FK crystallographic structures are classified into low and high polyhedral groups denoted by their coordination numbers (CN) referring to the number of atom centering the polyhedron. Some atoms have an icosahedral structure with low coordination, labeled CN12. Some others have higher coordination numbers of 14, 15 and 16, labeled CN14, CN15, and CN16, respectively. These atoms with higher coordination numbers form uninterrupted networks connected along the directions where the five-fold icosahedral symmetry is replaced by six-fold local symmetry. The sites of 12-coordination are called minor sites and those with more than 12-fold coordination are major sites. Classic FK phases The most common members of a FK-phases family are: A15, Laves phases, σ, μ, M, P, and R. A15 phases A15 phases are intermetallic alloys with an average coordination number (ACN) of 13.5 and eight A3B stoichiometry atoms per unit cell where two B atoms are surrounded by CN12 polyhedral (icosahedra), and six A atoms are surrounded by CN14 polyhedral. Nb3Ge is a superconductor with A15 structure. Laves phases The three Laves phases are intermetallic compounds composed of CN12 and CN16 polyhedra with AB2 stoichiometry, commonly seen in binary metal systems like MgZn2. Due to the small solubility of AB2 structures, Laves phases are almost line compounds, though sometimes they can have a wide homogeneity region. σ, μ, M, P, and R phases The sigma (σ) phase is an intermetallic compound known as the one without definite stoichiometric composition and formed at the electron/atom ratio range of 6.2 to 7. It has a primitive tetragonal unit cell with 30 atoms. CrFe is a typical alloy crystallizing in the σ phase at the equiatomic composition. With physical properties adjustable based on its structural components, or its chemical composition provided a given structure. The μ phase has an ideal A6B7 stoichiometry, with its prototype W6Fe7, containing rhombohedral cell with 13 atoms. While many other Frank-Kasper alloy types have been identified, more continue to be found. The alloy Nb10Ni9Al3 is the prototype for the M phase. It has orthorhombic space group with 52 atoms per unit cell. The alloy Cr9Mo21Ni20 is the prototype for the P-phase. It has a primitive orthorhombic cell with 56 atoms. The alloy Co5Cr2Mo3 is the prototype for the R-phase which belongs to the rhombohedral space group with 53 atoms per cell. Applications FK phase materials have been pointed out for their high-temperature structure and as superconducting materials. Their complex and often non-stoichiometric structure makes them good subjects for theoretical calculations. A15, Laves and σ are the most applicable FK structures with interesting fundamental properties. The A15 compounds include important intermetallic superconductors such as: Nb3Sn, Nb3Al, and V3Ga with applications including wires for high-field superconducting magnets. Nb3Sn is also being investigated as a potential material for fabricating superconducting radio frequency cavities. Small extents of σ phase considerably decreases the flexibility and impairment in erosion resistance. While addition of refractory elements like W, Mo, or Re to FK phases helps to enhance the thermal properties in such alloys as steels or nickel-based superalloys, it increases the risk of unwanted precipitation in intermetallic compounds. See also Complex metallic alloys References Crystal structure types
Frank–Kasper phases
[ "Physics", "Chemistry", "Materials_science" ]
1,036
[ "Inorganic compounds", "Metallurgy", "Crystal structure types", "Crystallography", "Alloys", "Intermetallics", "Condensed matter physics" ]
41,396,808
https://en.wikipedia.org/wiki/Ardwick%20train%20depot
The Ardwick train depot is a passenger multiple unit maintenance depot in Ardwick, Greater Manchester. The depot was opened in 2006 for the servicing of Siemens-built Class 185 DMUs, which are used on the TransPennine Express franchise. It was electrified in 2012–13 to allow the servicing of Siemens Class 350/4 EMUs. Location Ardwick depot is located approximately east of Ardwick railway station, beside the Hope Valley line and south of the A635 road. The depot code is AK. History In the 1840s, the Sheffield, Ashton-under-Lyne and Manchester Railway was constructed between Manchester Store Street station (now known as Manchester Piccadilly) and Sheffield Victoria railway station. Initially, the site of the depot was just outside the urban spread of Manchester; the area was used for brick manufacture. Subsequently, the area east of Ardwick station and north of the line was used for railway sidings, including goods sheds; the wider area became known as Ardwick West Goods Depot and Mineral Yard, later Ardwick West Freight Terminal. The site remained in railway use through most of the 20th century, but became disused by 1990. Siemens train maintenance depot In 2003, Siemens was announced as the preferred bidder to supply and maintain a fleet of diesel multiple units for use on the TransPennine Express franchise. The £260 million contract included the provision and supply of maintenance facilities, with the primary facility in Manchester, along with a secondary depot in York. In 2004, Siemens applied to construct sidings, train facilities and offices on the Ardwick site. The groundbreaking ceremony took place in March 2005. The depot was constructed on the site over 13 months, being completed on 3 April 2006, at a cost of about £30m. The designers were Burks Green; Taylor Woodrow were the main contractors; and NG Bailey and Atkins Rail were subcontractors. The main shed was a building capable of holding eight trains – a four-road shed with two three-car trains per road – with facilities including wheel lathe, train jacks, and bogie and engine drops. The depot was officially opened on 16 May 2006 by Transport Secretary Douglas Alexander. In February 2012, Siemens obtained a contract to build and maintain twenty Class 350 Desiro 25 kV AC electric multiple units, including ten Class 350/4 units for First TransPennine Express. The contract required the overhead electrification of the depot. Work to modify the depot began in September 2012, with Spencer Rail (Spencer Group) as the main contractor. The depot was extended at the western end by to allow four -long car units to be accommodated, with one shed road and all of the outside track electrified. The estimated cost of the project was £5 million. The depot's electrification was ceremonially opened by local MP, Lucy Powell, on 27 September 2013. The first of the Class 350/4 trains arrived at the depot on 28 November 2013. References Notes Sources Railway depots in England Rail transport in Greater Manchester Siemens Mobility projects
Ardwick train depot
[ "Technology", "Engineering" ]
613
[ "Siemens Mobility projects", "Transport systems" ]
41,396,994
https://en.wikipedia.org/wiki/Illegitimate%20receiver
An illegitimate receiver is an organism that intercepts another organism's signal, despite not being the signaler's intended target. In animal communication, a signal is any transfer of information from one organism to another, including visual, olfactory (e.g. pheromones), and auditory signals. If the illegitimate receiver's interception of the signal is a means of finding prey, the interception is typically a fitness detriment (meaning that it reduces survival or reproductive ability) to either the signaler or the organism meant to legitimately receive the signal, but it is a fitness advantage to the illegitimate receiver because it provides energy in the form of food. Illegitimate receivers can have important effects on the evolution of communication behaviors. Fitness Benefits and Costs Fitness Benefits Illegitimate receivers can benefit by intercepting signals to locate prey, or, if they are parasites or parasitoids, by intercepting signals to locate host organisms. In addition to locating prey by intercepting signals given by the prey organism, some animals use the signals of other predators to find carcasses that they can scavenge off of. Other organisms benefit by illegitimately receiving the signals of rivals and using this information to improve their own chances of winning in competition for resources, including mates. Fitness Costs Illegitimate receivers can experience fitness costs if they respond to signals given off by illegitimate signalers, which are organisms that utilize deceptive signals to reduce receiver fitness, typically by preying on or parasitizing the organism that responds. Illegitimate receivers may also experience fitness costs if intercepting signals not intended for them reduces their likelihood of receiving signals that are directed at them, such as the mating calls of members of their own species or the warning calls of rivals. Examples Redeye bass (Micropterus coosae) and midland water snakes (Nerodia sipedon pleuralis) respond to acoustic and visual signals in male tricolor shiners (Cyprinella trichroistia) when detecting prey. Male Great Bowerbirds sometimes steal nest decorations, which are intended to attract mates, from their rivals and use these decorations in their own nests. Male túngara frogs (Physalaemus pustulosus) give off mating calls consisting of both "whines" and "chucks," with songs that contain chucks favored by females over those containing only whines. However, the fringe-lipped bat (Trachops cirrhosus), a natural predator of the túngara frog, is an illegitimate receiver of these songs and uses them to locate its prey. These bats are especially attracted to frog songs containing the chuck element, and so túngara frogs rarely incorporate chucks into their calls. In fact, the frogs have been shown to typically only incorporate the chuck element into their songs when they are congregated in large groups, as this reduces the chance of being eaten via the dilution effect. On the island of Kauai, females of a species of parasitoid fly, Ormia ochracea, respond to the stridulation mating calls of male field crickets (Teleogryllus oceanicus) by locating the crickets and then laying their lethal larvae on them. In response to this, male field crickets have evolved via a "flatwing" mutation to no longer produce mating songs. Another example of evolution in response to illegitimate receivers is that of the Great tits. These European songbirds have evolved to use "seet" calls in order to avoid having their signals illegitimately received by hawks or owls. Great tits use two different calls to warn one another of nearby predators: When the predators are flying nearby, the great tits use a "seet" alarm call; however, when the predators are perched nearby, the great tits use a mobbing call. The mobbing call is at a much higher frequency than the seet call, allowing for the great tits to recruit nearby individuals of their species when mobbing perched predators in an attempt to chase them out of the area. Meanwhile, the lower frequency of the seet call allows the great tits to warn one another of the danger without attracting the unwanted attention of the mobile hawk or owl. Louder calls are also more frequently exhibited in birds inhabiting more protected habitats, while softer seet calls are more common in unprotected, open areas. See also Illegitimate Signalers Illegitimate signalers utilize deceptive signals to reduce the receiver's fitness while increasing their own. Examples include the case of the Photinus and Photuris fireflies, as well as aggressive mimicry. Honest Signals Honest signals are signals used by one organism to convey true information to another individual. An example is the begging calls of bird chicks. Animal Communication Animal communication includes any transfer of information between individuals, including illegitimate receiving and signaling. References Animal communication Evolutionary biology
Illegitimate receiver
[ "Biology" ]
977
[ "Evolutionary biology" ]
41,397,181
https://en.wikipedia.org/wiki/Environmental%20hermeneutics
Environmental hermeneutics is a term for a wide range of scholarship that applies the techniques and resources of the philosophical field of hermeneutics to environmental issues. That is to say it addresses issues of interpretation as they relate to nature and environmental issues broadly conceived to include wilderness, ecosystems, landscapes, ecology, the built environment (architecture), life, embodiment, and more. Work in environmental philosophy, ecocriticism, environmental theology, ecotheology, and similar disciplines may overlap the field of environmental hermeneutics. In the public sphere, much of the focus on “the environment” is concerned with discovering scientific facts and then reporting how policy can act on these facts. On its face, philosophical hermeneutics might appear to be an unrelated enterprise. But... even the facts of the sciences are given meaning by how humans interpret them. Of course this does not mean that there are no facts, or that all facts must come from scientific discourse. Rather... [it calls] for mediation—the mediation that grounds the interpretive task of connecting fact and meaning through a number of different structures and forms. (Clingerman, et al. 2013, emphasis added) See also Ecosemiotics References Notes Bibliography Abram, D. (1996). The Spell of the Sensuous, New York: Vintage. Brown, C. S. and T. Toadvine (2003). Eco-Phenomenology: Back to the Earth Itself, New York: SUNY Press. Clingerman, F. and B. Treanor, M. Drenthen, D. Utsler (2013). Interpreting Nature: The Emerging Field of Environmental Hermeneutics, New York: Fordham University Press. Clingerman, F. and M. Dixon (2011). Placing Nature on the Borders of Religion, Philosophy and Ethics, London: Ashgate. Cronon, William (1992). “A Place for Stories: Nature, History, and Narrative.” Journal of American History 78 (1992): 1347–76. Drenthen, M. and J. Keulartz (2014). Environmental Aesthetics: Crossing Divides and Breaking Ground, New York: Fordham University Press Gare, Arran (1998). “MacIntyre, Narratives, and Environmental Ethics.” Environmental Ethics 20 (1998): 3-21. Keller, K. (2003). The Face of the Deep, London: Routledge Kohák, E. (1984). The Embers and the Stars: A Philosophical Inquiry into the Moral Sense of Nature, Chicago: University of Chicago Press. Mugerauer, R. (1995). Interpreting Environments, Austin, TX: University of Texas Press Plumwood, Val (1993). Feminism and the Mastery of Nature, London: Routledge Treanor, Brian (2014). Emplotting Virtue: A Narrative Approach to Environmental Virtue Ethics. New York: SUNY Press. Van Buren, John (1995). “Critical Environmental Hermeneutics.” Environmental Ethics 17 (1995): 259–275. Wood, D. (2006). “On the Way to Econstruction” in Environmental Philosophy, vol. 3, issue 1 (Spring 2006) Environmental ethics Environmental philosophy Hermeneutics
Environmental hermeneutics
[ "Environmental_science" ]
676
[ "Environmental philosophy", "Environmental social science", "Environmental ethics" ]
41,397,239
https://en.wikipedia.org/wiki/Kepler-24
Kepler-24 is a star in the northern constellation of Lyra. It is located at the celestial coordinates: Right Ascension , Declination . With an apparent visual magnitude of 15.5, this star is too faint to be seen with the naked eye. Planetary system Two planetary candidates b and c were discovered in 2011, and were confirmed in 2012 together with d and e. References Lyra G-type main-sequence stars 1102 Planetary transit variables Planetary systems with four confirmed planets J19213918+3820375
Kepler-24
[ "Astronomy" ]
109
[ "Lyra", "Constellations" ]
41,397,398
https://en.wikipedia.org/wiki/Kepler-25
Kepler-25 is a star in the northern constellation of Lyra. It is slightly larger and more massive than the Sun, with a luminosity 2 times that of the Sun. With an apparent visual magnitude of 10.6, this star is too faint to be seen with the naked eye. Planetary system In 2011, two candidate planets were found transiting this star by the Kepler space telescope. These planets are very close to yet not lie in the 1:2 orbital resonance to each other, indicating the absence of other planetary objects in the inner part of the planetary systems. These planets were confirmed through transit-timing variation method. A third planet was discovered through follow-up radial velocity measurements and was confirmed in January 2014. The plane of planetary orbits is well aligned with the equatorial plane of the star, misalignment angle equal to 7° References Lyra F-type main-sequence stars 244 Planetary transit variables Planetary systems with three confirmed planets J19063321+3929164
Kepler-25
[ "Astronomy" ]
204
[ "Lyra", "Constellations" ]
41,397,440
https://en.wikipedia.org/wiki/Kepler-26
Kepler-26 is a star in the northern constellation of Lyra. It is located at the celestial coordinates: Right Ascension Declination . With an apparent visual magnitude of 15.5, this star is too faint to be seen with the naked eye. Planetary system The two planets, Kepler-26b and Kepler-26c, were discovered by transit method in late 2011, and classified as small (sub-Neptune) gas giants in 2016. In 2012, the planetary candidate Kepler-26d was also detected, and confirmed in 2014. The planet Kepler-26e was discovered on a much wider orbit in 2014. References Lyra M-type main-sequence stars 250 Planetary transit variables Planetary systems with four confirmed planets J18594583+4633595
Kepler-26
[ "Astronomy" ]
158
[ "Lyra", "Constellations" ]
47,352,090
https://en.wikipedia.org/wiki/Education%20for%20Nature%20Vietnam
Education for Nature Vietnam (ENV) was set up in 2000 and according to their website is Vietnam's "first local non-governmental organization to focus on wildlife protection." They have offices in Hanoi. There are three main planks to the work of ENV: Persuading the Vietnamese public of the need to protect nature and wildlife Convincing Vietnamese society that using animal products is hastening the extinction of endangered species Working with the Vietnamese authorities to strengthen wildlife protection laws and enforce the current legislation to its full extent to stop illegal wildlife trade Wildlife trade in Vietnam As well as being a source of endangered wildlife and a consumer of protected animal products, Vietnam is also a major international transit hub for illegal wildlife goods from other countries in South East Asia and as far afield as South Africa. The main destination for much of the illegal wildlife trade is China. Due to the burgeoning economies of both China and Vietnam in recent years, the expanding middle class, with larger disposable incomes, created a surge in demand for wildlife trade products such as rhino horn and tiger bone paste. Two species of pangolin, believed to be the most trafficked mammal in the world and native to Vietnamese forests, are particularly under threat. Many thousands are caught and traded between China and Vietnam every year. Pangolin scales are in demand for use in traditional medicine, and the meat is served in restaurants as a high-end delicacy. Rhinos, annihilated in Vietnam in 2010, are also threatened with extinction as a result of wildlife trade: in 2007, 13 were killed for their horns. By 2014, 1215 rhinos were poached and killed in South Africa. Approach ENV aims to reduce illegal wildlife trade in three ways by: Mobilizing the public to support wildlife protection Strengthening wildlife crime law enforcement by direct support and assistance, and mobilizing public opinion and participation Working with high-level decision-makers and government agencies to formulate and improve wildlife-related policy and legislation. Reducing demand ENV strives to curb demand for wildlife products in Vietnam via Public Service Announcements (PSAs) on TV and radio. They concentrate their efforts on tigers, rhinos, bears and pangolins. The emphasis of these short infomercials is that wildlife trade products, such as pangolin scales, rhino horn and tiger bones, lack medicinal value. A recent 2016 campaign, for instance, pointed out that rhino horn is made of keratin, the same substance as human hair and nails. One would, therefore, be as well eating one's own finger nails and hair. It is well documented that there is now a growing social stigma attached to the use of wildlife products as vulgar status symbols. Money does not buy taste or a social conscience it would appear. But just as important is the fact that being a consumer of illegal wildlife products helps to line the pockets of organized crime and encourages cruelty to animals. The PSAs are aired across Vietnam on up to 80 TV channels, reaching millions of viewers. ENV also produce radio adverts that are regularly broadcast on Voice of Vietnam radio. Since 2008, ENV has partnered with Voice of Vietnam to produce a radio show about wildlife protection topics every month. ENV also works with well-known Vietnamese celebrities to spread the wildlife protection message to their fans and the general public. Further efforts to cut consumer demand have included the establishment of Wildlife Safety Zones in conjunction with ministries, government offices, corporate partners and markets across Vietnam. Among the partners signed up are the US Embassy, Mercedes-Benz, BMW, and the Daewoo Hotel. Outreach events are also run at universities, parks and shopping malls to promote public awareness and involvement in their campaigns. For example, ENV, in partnership with the South African organization Endangered Wildlife Trust, launched a targeted campaign to convince shoppers to ‘Say no to rhino horn’ through awareness events and viral media activities. Strengthening enforcement In addition to reducing demand, ENV are heavily involved in ensuring law enforcement agencies prosecute wildlife crime offenders to the full extent of the law. In recent years there has been a perception that the authorities are not taking wildlife crime as seriously as they should and are reluctant to prosecute anyone other than low level players involved in wildlife trafficking. In 2005 ENV established a Wildlife Crime Unit (WCU) to encourage the Vietnamese public to report wildlife offences. The WCU operates the Wildlife Crime Hotline 1-800-1522, a national toll-free hotline that the public can use to report wildlife crimes throughout the country. Crimes can also be reported via an email hotline and a smart phone application. Since its establishment, ENV's national crime database has recorded over 10,000 reports of wildlife crimes. Getting the public involved is a major concern for ENV. It has established wildlife protection volunteer clubs in over 15 major cities across Vietnam. These clubs carry out awareness events, monitor businesses and report wildlife crimes. They also have promoted greater public involvement in wildlife protection. Since 2013 ENV has cracked down on consumer demand by targeting major cities including Hanoi, Huế, Dong Ha, Ho Chi Mi, Vinh and Da Nang. Their campaign involves surveying restaurants, hotels, bars, traditional medicine shops, pet shops and markets across selected districts. Any violations discovered are reported to the district People's Committee, along with a request that they work with local authorities to tackle each violation. Two months later, a follow-up team from ENV inspects the establishments that had previously been reported and tracks any changes. Report cards are then sent to the People's Committees, summarizing their effectiveness compared to colleagues in different districts. In areas where the process has been completed, wildlife crime has fallen by between 39% and 77%. ENV also tackles illegal online wildlife trade through its internet crime campaign. Thousands of links reported as selling wildlife have been removed, and numerous websites and forums have joined ENV's wildlife safety zone by banning all wildlife advertising. Accompanying this outreach, the wildlife protection organization undertakes wildlife crime investigations. Current investigations focus on crime syndicates that support the illegal transnational trade of endangered species, such as a major criminal network that smuggled frozen tigers from Laos into Vietnam. An investigation into the marine turtle trade in Vietnam resulted in a seizure of 10 tons of marine turtles, and an ongoing criminal investigation by the police. ENV produced a film about the case, and sent it to hundreds of legislators to encourage them to prosecute the kingpin of the marine turtle trade. In May 2020, ENV and Four Paws rescued a bear cub from illegal wildlife trade and brought to Bear Sanctuary Ninh Binh. Improving policy and legislation The third leg of the ENV strategy stool is to work with the highest levels of government to bring about change on a national level by improving legislation and ensuring sound policy in support of wildlife protection. Ultimately their goal is a legal framework in Vietnam that effectively protects endangered species. But much of their day-to-day work revolves around helping to address conflicts and loopholes in existing legislation. . Among ENV's most notable cases recently include: Successfully campaigning to prevent authorities from auctioning off tiger products seized from poachers, arguing that it would increase demand and undermine wildlife protection efforts. ENV's intervention helped to stop this practice. Working with authorities to close down bear bile tourism in Halong Bay. From 2007, hundreds of Korean tourists arrived each day to watch a live bear bile extraction and to buy the bile to take back to their country. After an intensive joint enforcement campaign by Quang Ninh People's Committee, relevant government agencies, and ENV, bear bile tourism in Ha Long was shut down in May 2014, putting the bear farms out of business. Since 2005, there has been a 72% decline in the number of bears caged on farms and exploited for their bile in Vietnam. ENV also fought for any new bears being kept illegally to be confiscated by the government, not left in their owner's hands. Since September 2011, there hasn't been a single case where an illegal bear was discovered and not confiscated. ENV is currently opposing proposals to legalize the farming of endangered species in Vietnam, as they risk increasing demand for wildlife trade products, as well as complicating enforcement efforts. ENV also focuses on investigating, prosecuting and punishing major wildlife crime figures. By working to improve the penal code and fighting for the implementation of Decree 160 at provincial level, ENV aims to make it easier to tackle the kingpins of the illegal wildlife trade. ENV also works to improve awareness of current wildlife protection law in all provinces of Vietnam, enabling more effective law enforcement throughout the country. ENV encourages local authorities to comply with Decree 160 by not auctioning off endangered wildlife such as pangolins after confiscation. References Environmental organizations based in Vietnam Wildlife conservation
Education for Nature Vietnam
[ "Biology" ]
1,793
[ "Wildlife conservation", "Biodiversity" ]
47,353,676
https://en.wikipedia.org/wiki/Kethoxal
Kethoxal (3-ethoxy-1,1-dihydroxy-2-butanone) is an organic compound that has antiviral and anaplasmosis properties. It also forms a stable covalent adduct with guanine, which makes it useful for nucleic acid structure determination. Nucleic acid binding Kethoxal, as with other 1,2-dicarbonyl compounds, reacts with nucleic acids. It has high specificity for guanine over other ribonucleotides. In whole RNA, it reacts preferentially with guanine residues that are not involved in hydrogen-bonding. It can thus be used to probe the interactions involved with the secondary structure and other binding interactions of RNA and help with nucleic acid sequence analysis. The binding is reversible, which allows the kethoxal to be removed and the original RNA recovered. References Diols Ketones Ethoxy compounds Antiviral drugs
Kethoxal
[ "Chemistry", "Biology" ]
204
[ "Ketones", "Antiviral drugs", "Biocides", "Functional groups" ]
47,353,702
https://en.wikipedia.org/wiki/Research%20Council%20of%20Norway%27s%20Award%20for%20Excellence%20in%20Communication%20of%20Science
The Research Council of Norway's Award for Excellence in Communication of Science () is awarded annually by the Research Council of Norway, a Norwegian government body. According to its bylaws, the prize is to be given in order to "reward and stimulate the dissemination of research to a broad audience. The dissemination must be of high quality in both form and content." The price is worth . Award winners 2020 Anne Sverdrup-Thygeson 2019 Audun Rikardsen University of Tromsø 2018 Bjørn H. Samset, Centre for International Climate and Environmental Research 2017 Henrik H. Svensen, University of Oslo 2016 Ingun Grimstad Klepp, Oslo Metropolitan University 2015 Anine Kierulf, University of Oslo 2014: Frank Aarebrot, Department of Comparative Politics, University of Bergen 2013: Jorunn Sundgot-Borgen, Sports Medicine Section, Norwegian School of Sport Sciences 2012: Nils Christian Stenseth, Department of Biology and the Centre for Ecological and Evolutionary Synthesis, University of Oslo 2011: Kalle Moene, Head of the Centre of Equality, Social Organization and Performance (ESOP), University of Oslo 2010: The website, Forskning.no 2009: Jørn H. Hurum, Natural History Museum at the University of Oslo 2008: Ole Didrik Lærum, Gades Institute, University of Bergen. 2007: Egil Lillestøl, Department of Physics and Technology, University of Bergen 2006: Wenche Blomberg, Institute of Criminology and Legal Sociology, University of Oslo 2005: Terje Tvedt, Head of Research at the Center for Development Studies, University of Bergen 2004: Odd Aksel Bergstad, MAR-ECO Project 2003: Bente Træen, Department of Psychology, University of Tromsø 2002: Thomas Hylland Eriksen, Department of Social Anthropology, University of Oslo 2001: Michael Baziljevich, Department of Physics, University of Oslo 2000: NRK P2s pop-science radio program Verdt å vite 1999: Henry Valen, Norwegian Institute of Social Research 1998: Dag O. Hessen, University of Oslo 1997: Asgeir Brekke, University of Tromsø, and Professor Alv Egeland, University of Oslo 1996: Lisbet Holtedahl, University of Tromsø References Awards established in 1996 1996 establishments in Norway 1996 in science Science communication awards Norwegian science and technology awards
Research Council of Norway's Award for Excellence in Communication of Science
[ "Technology" ]
508
[ "Science and technology awards", "Science communication awards" ]
47,355,235
https://en.wikipedia.org/wiki/David%20Prior%20%28entrepreneur%29
David Prior is an Australian businessman who became notable in the yoghurt business. He built up the yoghurt brand five:am from scratch in 2009, and sold it in 2014 to UK-based consumer products company PZ Cussons for $80 million. In 2013 the company employed 65 staff, supplied over a thousand outlets around Australia, delivered 250% increase in revenue over the previous year and won the 2013 "Company to Watch" award. Before that, he partnered with his father in 1997 to create the sustainable packaging company Baroda, running it successfully for ten years before selling it to packaging giant Visy. In January 2015 Prior bought the mothballed Scotch distillery Bladnoch and said he plans to resume production from the 200-year-old establishment. Background and education David Prior was born and raised in Melbourne, Victoria. David attended Haileybury and went on to complete a Bachelor of Business degree from Monash University. David completed his MBA from Melbourne Business School in 1999. Early career – Baroda Packaging In 1997, Prior and his father, Malcolm, built the sustainable packaging company, Baroda. The company grew to large success over the next 10 years and was listed as one of BRW’s Fastest Growing Companies. In 2007 Baroda was sold to packaging giant, Visy. five:AM With some money from the Baroda sale and his own personal savings, Prior founded organic yoghurt brand five:AM in 2009. Prior is quoted as saying "There was no way that the banks were going to get involved unless they could see that the founder had put everything on the line." Prior stated that the inspiration for the name "five:AM" was his morning routine, "You know, you get up early, you win the morning. You get up, you do your thing. And, in my case, I meditate, I do yoga, swim or go for a surf or whatever. And by the time you hit your office desk you, you're charged. You're energised." five:AM became a market leader in the Australian organic food category since its launch in 2011. Products were stocked in Australian supermarket giants, Woolworths and Coles, as well as internationally to South-East Asia, the Middle East and China. In 2013, five:AM won the 2013 ‘Company to Watch’ award. In 2014 Prior sold five:AM to UK company, PZ Cussons for $80m. Bladnoch Distillery Soon after the sale of five:AM, Prior purchased Bladnoch Distillery, a 200-year-old Scotch Whisky distillery located in the Scottish Lowlands. Bladnoch Distillery is the fourth oldest distillery in Scotland and the oldest independently owned distillery in Scotland. Bladnoch Distillery is located in the small town of Bladnoch, roughly 140 km south-west of Glasgow. Prior is considered to be the first Australian to own a Scotch Whisky distillery. After purchasing the distillery in 2015, Prior began the extensive £5m refurbishment, installing brand new distilling equipment and two additional stills. Prior also hired awarded Master Distiller, Ian MacMillan, and the distillery began producing liquid again in 2017. Prior has launched two brands under Bladnoch Distillery. Bladnoch , the namesake brand, is a line of rare Single Malt Scotch Whiskies, and Pure Scot, a premium Blended Scotch Whisky brand that is highly awarded by international spirits competitions. Philanthropy Prior founded not-for-profit organisation, Culture Is Life, in 2015 which aims to prevent youth suicide in Indigenous communities by strengthening Aboriginal culture. Prior also founded the Prior Family Foundation with his wife, Sallie, which supports projects that work to preserve culture, wildlife, the environment and assist in disaster relief. Awards and recognition Distinguished Alumni Award 2017 – Melbourne Business School Commitment to Conservation Award 2017 - Australia Zoo Wildlife Warriors References Living people Businesspeople from Melbourne Year of birth missing (living people) Whisky distillers People educated at Haileybury (Melbourne) Drink distillers
David Prior (entrepreneur)
[ "Chemistry" ]
825
[ "Distillation", "Drink distillers" ]
47,356,548
https://en.wikipedia.org/wiki/World%20Wide%20What%3F
World Wide What? is a 2015 British film written, directed and filmed by Adam Townsend and Andy Trace of Cavalier, and produced by Poppy Gaye of Founders Forum. The film shows a parallel universe in which Tim Berners-Lee failed to invent the World Wide Web and the subsequent impact that would have on the lives of high-profile tech entrepreneurs. The film is narrated by Stephen Fry, and includes cameos from Jimmy Wales, Arianna Huffington, Sean Parker, Tim Berners-Lee, Reid Hoffman, Michael Bloomberg, Michael Acton Smith, Martha Lane Fox, Baroness Lane-Fox of Soho, Brent Hoberman, and Steve Case. Plot The film begins in 1989 at the offices of the European Organization for Nuclear Research, also known as CERN, where Tim Berners Lee is working on his aspirations for the World Wide Web. He reflects to himself in private that this project could go on to promote "world peace" and also help people "share pictures of kittens". Upon leaving his office, Stephen Fry describes how a short-sighted assistant could have changed the course of humanity if she had accidentally put Tim Berners Lee's research in a waste paper basket. The film then goes on to depict a succession of sketches played out by global tech entrepreneurs, in which they are forced to carry out analogue versions of their digital jobs. In order of appearance: Arianna Huffington is a newspaper girl, throwing paper copies of the Huffington Post onto doorsteps Sean Parker is unable to keep himself occupied with Napster and creates his own selfie empire of polaroids instead Steve Case is a postman delivering huge quantities of junk mail to people whilst repeatedly stating You've got mail, a phrase synonymous with AOL Venture Capitalist Danny Rimer is left creating his own infomercials instead of investing in startup company Dropbox Bebo founder Michael Birch desperately tries to release funds from mafia bank owner Max Levchin, famous for founding PayPal Skype founder Niklas Zennström, as an expert in communications, is asked to crack the Enigma Machine Martha Lane Fox, Baroness Lane-Fox of Soho visits the man who helped her float lastminute.com to try and book a holiday Brent Hoberman is a holiday sales agent, desperately trying to sell a holiday Reid Hoffman is a secretary using a Rolodex to categorise the entire world Michael Acton Smith is a street vendor selling Moshi Monsters toys Fon founder Martín Varsavsky is a nuisance caller, asking people if they want a new deal on their 'fon' Jonathan Goodwin (entrepreneur) is a hunter looking for a Unicorn (finance) Candy Crush and King founder Ricardo Zucconi is selling sweets in a sweet shop Michael Bloomberg is still working at Bloomberg, but as a journalist working on the latest edition of his paper Jimmy Wales stays true to the principles of Wikipedia whilst working as a librarian, encouraging people to make edits to books The film then returns to CERN to find Tim Berners Lee desperately looking for his research for creating the World Wide Web. As he becomes more and more distressed, his assistant walks through the door holding the papers, stating that she had read them overnight. When Tim Berners Lee asks her if she thinks the research stands a chance of working, she ponders "50/50", much to his disappointment. Production The film was funded by YouTube and King, with Jimmy Wales as executive producer. The script was written by filmmaker Adam Townsend of Cavalier, at the time with Freud Communications after the initial concept was developed by Jimmy Wales and Founders Forum. The scenes were filmed in London, Los Angeles, New York, Washington and Suffolk. The scenes that took place in America were directed by Declan Masterson. References 2015 films Films shot in London Films shot in Los Angeles Films shot in New York (state) Films shot in Washington, D.C. Films shot in Suffolk British alternative history films Mass media technology English inventions Digital technology History of Europe 2010s English-language films 2010s British films
World Wide What?
[ "Technology" ]
803
[ "Information and communications technology", "Mass media technology", "Digital technology" ]
47,357,235
https://en.wikipedia.org/wiki/Neural%20efficiency%20hypothesis
The neural efficiency hypothesis proposes that while performing a cognitive task, individuals with higher intelligence levels exhibit lower brain activation in comparison to individuals with lower intelligence levels. This hypothesis suggests that individual differences in cognitive abilities are due to differences in the efficiency of neural processing. Essentially, individuals with higher cognitive abilities utilize fewer neural resources to perform a given task than those with lower cognitive abilities. History Since the late 19th century, there has been a growing interest among psychologists to understand the influence of individual differences in intelligence and the underlying neural mechanisms of intelligence. The Neural efficiency hypothesis was first introduced by Haier et al. in 1988 through a Position Emission Tomography (PET) study aimed at investigating the relationship between intelligence and brain activation. PET is a type of nuclear medicine procedure that measures the metabolic activity of the cells of body tissues. During the study, participants underwent PET of the head while completing different cognitive tasks such as Raven's Advanced Progressive Matrices (RAPM) and Continuous Performance Tests (CPT). The PET Scans showed that task performance activated specific regions of the participant's brain. Also, a negative correlation was found between brain glucose metabolism levels and intelligence test scores. The results of the study indicated that individuals with higher intelligence levels exhibited lower levels of brain glucose metabolism while solving cognitive tasks. A few years later, Haier confirmed the results of the study by replicating it while considering learning as a factor. Research The early studies mainly focused on certain cognitive tasks such as intelligence tests to test the hypothesis, potentially confounding efficiency during the intelligence-test performance with neural efficiency in general. To overcome this limitation recent studies have refined and expanded the hypothesis by applying and testing it in various domains. In one study, researchers used a personal decision-making task to test the NEH which included questions about preferences like, “which profession do you prefer?”. Subjective preferences were used to force participants to make decisions, and preference ratings were used to manipulate the level of decisional conflict. The study found that individuals with higher intelligence test scores displayed less brain activity during simple tasks and greater brain activity during complex tasks, compared to individuals with lower intelligence test scores. This suggested that smarter people can use their brains more effectively by turning on only the areas that are required for the activity at hand. Also, more intelligent people displayed quicker reaction times during challenging tasks. These findings offered fresh evidence in support of the NEH and indicated that the neural efficiency of highly intelligent people can be applied to tasks that are different from typical intelligence tests. Another study focused on understanding the effect of long-term specialized training on an athlete's neural efficiency, using functional neuroimaging while performing a sport-specific task. The results of this study showed that athletes with prolonged experience or “experts” in their domains performed better than novices in terms of speed, accuracy, and efficiency, with lower activity levels in the sensory and motor cortex and less energy expenditure. These findings supported the Neural Efficiency Hypothesis (NEH) and proved that individuals who are highly skilled and experienced have more efficient brain functioning. Limitations Recent studies on the Neural Efficiency Hypothesis have identified several limitations in the former research. They have also found several moderating variables, such as task complexity, sex and task type. Task complexity The difficulty level of the task is one of the key moderating variables that influence the neural efficiency hypothesis. In a study, it was found that the hypothesis only holds for easy tasks. For difficult tasks, intelligent individuals may show increased brain activation. The study revealed that participants with high IQ showed weaker activation during easy tasks but had a significant increase from easy to difficult tasks. This pattern was not observed in the average IQ group. The study suggests that the relationship between intelligence and brain activation depends on the difficulty of the task. Sex and task type Former studies have primarily used uniform tasks and have mainly focused on male participants. One study found that neural efficiency was influenced by sex and task content. The study tried to examine possible sex differences in human brain functioning. It aimed at investigating the relationship between intelligence and cortical activation during the cognitive performance in various versions of a task, using brain imaging techniques. The results of the study suggested that, In the verbal task, the females were more likely to produce cortical activation patterns consistent with the NEH. Whereas, in the figural task, the expected neural activation was primarily in the males in comparison to the female participants. This suggested the role of sex and task type as moderating variables. References Intelligence Cognitive tests Biological hypotheses
Neural efficiency hypothesis
[ "Biology" ]
909
[ "Biological hypotheses" ]
47,357,437
https://en.wikipedia.org/wiki/FG-15
The FG-15 (AKA DFH-2 AKM and SpaB-170) was a Chinese spin stabilized apogee kick motor burning HTPB-based composite propellant. It was developed by China Hexi Chemical and Machinery Corporation (also known as the 6th Academy of CASIC) for use in the Dong Fang Hong 2 satellite bus for insertion into GSO orbit. It has a total nominal mass of , of which is propellant load and its burn out mass is . It has an average thrust of with a specific impulse of 289 seconds burning for 35 seconds, with a total impulse of . This motor introduced a series of innovations for the Chinese solid motor industry: first use of glass fibre wound cases, carbon/carbon nozzle throat insert material, contoured divergent nozzle, and non-destructive test of the motor grain. The initial version had the manufacturing code FG-15 was associated to the DFH-2 bus and flew twice. The FG-15B was used by the DFH-2A bus and flew five times. See also China Hexi Chemical and Machinery Corporation Dong Fang Hong 2 Long March 3 References Rocket engines of China Solid-fuel rockets
FG-15
[ "Astronomy" ]
246
[ "Rocketry stubs", "Astronomy stubs" ]
47,357,663
https://en.wikipedia.org/wiki/List%20of%20United%20States%20Army%20tactical%20truck%20engines
In the late 1930s the US Army began setting requirements for custom built tactical trucks, winning designs would be built in quantity. As demand increased during WWII some standardized designs were built by other manufactures. Most trucks had gasoline (G) engines until the early 1960s, when multifuel (M) and diesel (D) engines were introduced. Since then diesel fuel has increasingly been used, the last gasoline engine vehicles were built in 1985. Most engines have been water-cooled with inline (I) cylinders, but V types (V) and opposed (O) engines have also been used. Three air-cooled engines were used in two very light trucks. Gasoline engines up to WWII were often valve in block design (L-head), during the war more overhead valve (ohv) engines were used, and after the war all new engines (except 1 F-head and 1 Overhead camshaft (ohc)) have been ohv. All diesel engines have ohv, they can be naturally aspired, supercharged (SC), or turbocharged (TC). The same engines have been used in different trucks, and larger trucks often have had different engines during their service life. Because of application and evolution, the same engine often has different power ratings. Ratings are in SAE gross horsepower. The front of an engine is the fan end, the rear is the flywheel end, right and left are as viewed from the rear, regardless of how the engine is mounted in the vehicle. Engines in the tables are water-cooled and naturally aspirated unless noted. AM General/Kaiser Jeep/Willys Caterpillar Chevrolet – GMC Continental Cummins Detroit Diesel Dodge-Fargo Ford Hall-Scott Hercules International Harvester LeRoi Mack REO Notes References Works cited Military trucks of the United States United States Army tactical truck engines Tactical truck engines United States Army tactical truck engines
List of United States Army tactical truck engines
[ "Technology", "Engineering" ]
380
[ "Internal combustion engine", "Combustion engineering", "Engines" ]
47,358,172
https://en.wikipedia.org/wiki/Dehydroamino%20acid
In biochemistry, a dehydroamino acid or α,β-dehydroamino acid is an amino acids, usually with a C=C double bond in its side chain. Dehydroamino acids are not coded by DNA, but arise via post-translational modification. Examples A common dehydroamino acid is dehydroalanine, which otherwise exists only as a residue in proteins and peptides. The dehydroalanine residue is obtained dehydration of serine-containing protein/peptide (alternatively, removal of H2S from cysteine). Another example is dehydrobutyrine, derived from dehydration of threonine. Generally, amino acid residues are unreactive toward nucleophiles, but the dehydroamino acids are exceptions to this pattern. For example, dehydroalanine adds cysteine and lysine to form covalent crosslinks. An unusual dehydroamino acid is dehydroglycine (DHG) because it does not contain a carbon-carbon double bond. Instead it is the imino acid of glyoxalic acid. It arises by the radical-induced degradation of tyrosine. N-Acyl derivatives Dehydroamino acids do not feature amino-alkene groups, but the corresponding N-acylated derivatives are known. These derivatives, also known as N-acylamino acrylates, are prochiral substrates for asymmetric hydrogenation. The 2001 Nobel Prize in Chemistry was awarded to William S. Knowles for his synthesis of L-DOPA from the N-acylacrylate. References Amino acids
Dehydroamino acid
[ "Chemistry" ]
354
[ "Amino acids", "Biomolecules by chemical classification" ]
47,359,227
https://en.wikipedia.org/wiki/M59-UCD3
M59-UCD3 is an ultra-compact dwarf galaxy located near the Messier 59 galaxy. , it is the second-densest galaxy currently observed, second to M85-HCC1. References See also M60-UCD1 (densest galaxy known, as of 2013) M85-HCC1 (densest galaxy known, as of 2015) Virgo (constellation) Dwarf galaxies 20150727
M59-UCD3
[ "Astronomy" ]
90
[ "Virgo (constellation)", "Galaxy stubs", "Astronomy stubs", "Constellations" ]
47,360,092
https://en.wikipedia.org/wiki/Lasiodiplodia%20citricola
Lasiodiplodia citricola is an endophytic fungus. It was first isolated in northern Iran, and is named after its first known host, citrus plants. It has since been isolated in other plants in other continents, and is considered a plant pathogen. L. citricola is phylogenetically related to L. parva, but conidia of the former are longer and wider. Description Its conidiomata are stromatic and pycnidial; mycelium is uniloculate, up to in diameter, non-papillate and with a central ostiole. Its paraphyses are hyaline, cylindrical and thin-walled. Conidiophores are absent in this species. Conidiogenous cells are holoblastic and also hyaline. Conidia are aseptate, ellipsoid to ovoid and with longitudinal striations. References Further reading Chen, S. F., et al. "First report of Lasiodiplodia citricola and Neoscytalidium dimidiatum causing death of graft union of English walnut in California." Fungal Diversity 67 (2014): 157–179. Chen, S. F., et al. "First report of Lasiodiplodia citricola associated with stem canker of peach in California, USA." Journal of Plant Pathology 95.3 (2013). Van der Linde, Johannes Alwyn, et al. "Lasiodiplodia species associated with dying Euphorbia ingens in South Africa." Southern Forests: a Journal of Forest Science 73.3-4 (2011): 165–173. Marques, Marília W., et al. "Species of Lasiodiplodia associated with mango in Brazil." Fungal Diversity 61.1 (2013): 181–193. External links MycoBank Botryosphaeriaceae Fungal citrus diseases Fungi described in 2010 Fungus species
Lasiodiplodia citricola
[ "Biology" ]
413
[ "Fungi", "Fungus species" ]