source
stringlengths
31
203
text
stringlengths
28
2k
https://en.wikipedia.org/wiki/Strand%20sort
Strand sort is a recursive sorting algorithm that sorts items of a list into increasing order. It has O(n2) worst time complexity which occurs when the input list is reverse sorted. It has a best case time complexity of O(n) which occurs when the input is a list that is already sorted. The algorithm first moves the first element of a list into a sub-list. It then compares the last element in the sub-list to each subsequent element in the original list. Once there is an element in the original list that is greater than the last element in the sub-list, the element is removed from the original list and added to the sub-list. This process continues until the last element in the sub-list is compared to the remaining elements in the original list. The sub-list is then merged into a new list. Repeat this process and merge all sub-lists until all elements are sorted. This algorithm is called strand sort because there are strands of sorted elements within the unsorted elements that are removed one at a time. This algorithm is also used in J Sort for fewer than 40 elements. Example This example is based on the description of the algorithm provided in the book, IT Enabled Practices and Emerging Management Paradigms. Step 1: Start with a list of numbers: {5, 1, 4, 2, 0, 9, 6, 3, 8, 7 } Step 2: Next move the first element of the list into a new sub-list:  sub-list contains {5} Step 3: Then iterate through the original list and compare each number to 5 until there is a number greater than 5. 1 < 5 so 1 is not added to the sub-list. 4 < 5 so 4 is not added to the sub-list. 2 < 5 so 2 is not added to the sub-list. 0 < 5 so 0  is not added to the sub-list. 9 > 5 so 9 is added to the sub-list and removed from the original list. Step 4: Now compare 9 with the remaining elements in the original list until there is a number greater than 9.   6 < 9 so 6 is not added to the sub-list. 3 < 9 so 3 is not added to the sub-list. 8 < 9 so 8 is not added to the sub-list. 7 < 9
https://en.wikipedia.org/wiki/Neon%20Museum%2C%20Warsaw
Neon Museum, also the Museum of Neon () is a museum located in Warsaw's Praga-Południe. The institution documents and protects Polish and Eastern Bloc light advertisements created after World War II. It is the first in Poland and one of the few museums of neon signs in the world. The museum is located at ul. Mińska 25, on the premises of Soho Factory. It was established in 2012. History The history of the museum began in 2005 when Ilona Karwińska saved the Berlin neon sign from Marszałkowska Street in Warsaw. The collection of the museum features about 100 neon lights from all over Poland  Most of the neons come from the 1960s and 1970s. The nine largest neon signs including GŁÓWNA KSIĘGARNIA TECHNICZNA (MAIN TECHNICAL BOOKSTORE), Jubiler, dworzec kolejowy CHODZIEŻ (CHODZIEŻ railway station), KINO PRAHA (PRAHA cinema) and WARSZAWA WSCHODNIA are located on different Soho Factory buildings. The museum also looks after some neon signs in Warsaw, including Mermaid on Grójecka Street. The museum has a large archive containing blueprints, photographs and original plans. In 2013, the museum together with RWE organized Neon for Warsaw competition. Gallery References External links Official website of the Museum Museums in Warsaw Praga-Północ Museums established in 2012 Neon lighting Signage 2012 establishments in Poland
https://en.wikipedia.org/wiki/Small%20RNA%20sequencing
Small RNA sequencing (Small RNA-Seq) is a type of RNA sequencing based on the use of NGS technologies that allows to isolate and get information about noncoding RNA molecules in order to evaluate and discover new forms of small RNA and to predict their possible functions. By using this technique, it is possible to discriminate small RNAs from the larger RNA family to better understand their functions in the cell and in gene expression. Small RNA-Seq can analyze thousands of small RNA molecules with a high throughput and specificity. The greatest advantage of using RNA-seq is represented by the possibility of generating libraries of RNA fragments starting from the whole RNA content of a cell. Introduction Small RNAs are noncoding RNA molecules between 20 and 200 nucleotide in length. The item "small RNA" is a rather arbitrary term, which is vaguely defined based on its length comparing with regular RNA such as messenger RNA (mRNA). Previously bacterial short regulatory RNAs have been referred to as small RNAs, but they are not related to eukaryotic small RNAs. Small RNAs include several different classes of noncoding RNAs, depending on their sizes and functions: snRNA, snoRNA, scRNA, piRNA, miRNA, YRNA, tsRNA, rsRNA, and siRNA. Their functions go from RNAi (specific for endogenously expressed miRNA and exogenously derived siRNA), RNA processing and modification, gene silencing (i.g. X chromosome inactivation by Xist RNA), epigenetics modifications, protein stability and transport. Small RNA sequencing Purification This step is very critical and important for any molecular-based technique since it ensures that the small RNA fragments found in the samples to be analyzed are characterized by a good level of purity and quality. There are different purification methods that can be used, based on the purposes of the experiment: acid guanidinium thiocyanate-phenol-chloroform extraction: it is based on the use of a guanidinium-thiocyanate solution combined with acid
https://en.wikipedia.org/wiki/Oceanography%20and%20Marine%20Biology%3A%20An%20Annual%20Review
Oceanography and Marine Biology: An Annual Review is an annual review of oceanography and marine biology that has been published since 1963. It was originally edited by Harold Barnes. It was originally published by Aberdeen University Press and Allen & Unwin but is now published by CRC Press, part of Taylor & Francis. The 55th volume was published in 2017. References University of Aberdeen Monographic series Marine biology Oceanography Academic journals established in 1963 Biology journals Taylor & Francis academic journals CRC Press books
https://en.wikipedia.org/wiki/Wasabi%20Technologies
Wasabi Technologies, Inc. is an American object storage service provider based in Boston, Massachusetts that sells one product, an object storage service called Wasabi Hot Cloud Storage. The company was co-founded in September 2015 by David Friend and Jeff Flowers and launched its cloud storage product in May 2017. History Friend and Flowers were previously co-founders of Carbonite, an online backup service, among other companies. Friend also previously launched ARP Instruments, Computer Pictures, Pilot Software, and Faxnet. The company was initially called "BlueArchive" at its founding, but was later renamed to "Wasabi Technologies, Inc." after hot Japanese horseradish. Wasabi Technologies, Inc. was launched with a single data center location in Ashburn, Virginia, a limitation that generated concern in the online tech forum, Hacker News. On January 30, 2018, Frost & Sullivan gave Wasabi Technologies, Inc, the 2017 North American Technology Innovation Award. On March 20, 2018, the company announced a satellite-based cloud storage system, collaborating with SpaceBelt to use satellites as data centers. On June 28, 2018, Chris Fenton (former President and General Manager of DMG Entertainment) joined the advisory board. Wasabi also launched its new data center in Hillsboro, Oregon. In October 2018, the company announced plans to open its first European data center in the Netherlands. On March 5, 2019, the company announced that its third data center, the first one located in Europe, opened in Amsterdam, Netherlands. On April 2, 2019, Wasabi launched its partner network, a volume-based incentive program. In September 2019, Wasabi Technologies announced it had received an investment of an undisclosed amount from NTT DOCOMO Ventures. The companies also made a joint announcement of the availability of Wasabi hot cloud storage as part of its Enterprise Cloud service, which became available in the US, EMEA and APAC regions. In October 2019, the company announce
https://en.wikipedia.org/wiki/Dirichlet%20average
Dirichlet averages are averages of functions under the Dirichlet distribution. An important one are dirichlet averages that have a certain argument structure, namely where and is the Dirichlet measure with dimension N. They were introduced by the mathematician Bille C. Carlson in the '70s who noticed that the simple notion of this type of averaging generalizes and unifies many special functions, among them generalized hypergeometric functions or various orthogonal polynomials:. They also play an important role for the solution of elliptic integrals (see Carlson symmetric form) and are connected to statistical applications in various ways, for example in Bayesian analysis. Notable Dirichlet averages Some Dirichlet averages are so fundamental that they are named. A few are listed below. R-function The (Carlson) R-function is the Dirichlet average of , with . Sometimes is also denoted by . Exact solutions: For it is possible to write an exact solution in the form of an iterative sum where , is the dimension of or and . S-function The (Carlson) S-function is the Dirichlet average of , References Calculus
https://en.wikipedia.org/wiki/Larisa%20Maksimova
Larisa Lvovna Maksimova (; born 1943) is a Russian mathematical logician known for her research in non-classical logic. Education and career Maksimova was born on November 5, 1943, in Kochenyovo, the daughter of two biologists who had temporarily moved there from Tomsk State University to escape the war. She grew up in Novosibirsk, where her parents became geographers at the Novosibirsk Pedagogical Institute. She studied mechanics and mathematics at Novosibirsk State University, publishing her first paper on Wilhelm Ackermann's axioms for strict implication in relevance logic in 1964 and graduating in 1965. Meanwhile, in 1964, she joined the Sobolev Institute of Mathematics, and has remained there for the rest of her career. She defended her doctorate at Novosibirsk State University in 1968, a year after the death of her primary mentor at the university, Anatoly Maltsev. She completed a habilitation at the Sobolev Institute in 1986, and was promoted to full professor in 1993. Recognition Maksimova won the Maltsev Prize of the Russian Academy of Sciences in 2009, for her papers on definability and interpolation in non-classical logic. With several others from the Sobolev Institute, she won the Russian Federation Government Prize in Education in 2010. She is the subject of a festschrift, Larisa Maksimova on Implication, Interpolation, and Definability (Sergei Odintsov, ed., Springer, 2018). Books Maksimova's books include Problems in Set Theory, Mathematical Logic and the Theory of Algorithms (with Igor Lavrov, Izdat Nauka, 1975, 1984, and 1995; translated into English by Valentin Shehtman, Kluwer, 2003) Interpolation and Definability: Modal and Intuitionistic Logics (with Dov Gabbay, Clarendon Press, 2005) References 1943 births Living people Russian mathematicians Russian women mathematicians Women mathematicians Mathematical logicians Women logicians Novosibirsk State University alumni
https://en.wikipedia.org/wiki/Anne%20Schilling
Anne Schilling is an American mathematician specializing in algebraic combinatorics, representation theory, and mathematical physics. She is a professor of mathematics at the University of California, Davis. Education Schilling completed her Ph.D. in 1997 at Stony Brook University. Her dissertation, Bose-Fermi Identities and Bailey Flows in Statistical Mechanics and Conformal Field Theory, was supervised by Barry M. McCoy. From 1997 until 1999, she was a postdoctoral fellow at the Institute for Theoretical Physics at Amsterdam University and from 1999 until 2001, she was a C.L.E. Moore Instructor at the Mathematics Department at M.I.T.. After that she joined the faculty at the Department of Mathematics at UC Davis. Books With Thomas Lam, Luc Lapointe, Jennifer Morse, Mark Shimozono, and Mike Zabrocki, Schilling is the author of the research monograph -Schur Functions and Affine Schubert Calculus (Fields Institute Monographs 33, Springer, 2014). With Isaiah Lankham and Bruno Nachtergaele, Schilling is the author of the textbook on linear algebra, Linear Algebra as an Introduction to Abstract Mathematics (World Scientific, 2016). With Daniel Bump, she is the author of a more advanced book on crystal bases in representation theory, Crystal Bases: Representations and Combinatorics (World Scientific, 2017). Recognition Schilling was included in the 2019 class of fellows of the American Mathematical Society "for contributions to algebraic combinatorics, combinatorial representation theory, and mathematical physics and for service to the profession". For the academic year 2012–2013 she was awarded a Simons Fellowship. In 2002 she received a Humboldt Research Fellowship and was a Fulbright Scholar from 1992-1993 as a doctoral student. Schilling was selected to be the 43rd Emmy Noether Lecturer at the Joint Mathematics Meetings to be held in San Francisco on January 3 – 6, 2024. References External links Home page Year of birth missing (living people) Living people 2
https://en.wikipedia.org/wiki/BlackArch
BlackArch is a penetration testing distribution based on Arch Linux that provides a large number of security tools. It is an open-source distro created specially for penetration testers and security researchers. The repository contains more than 2800 tools that can be installed individually or in groups. BlackArch Linux is compatible with existing Arch Linux installations. Overview BlackArch is similar in usage to both Parrot OS and Kali Linux when fully installed, with a major difference being BlackArch is based on Arch Linux instead of Debian. BlackArch only provides the Xfce desktop environment in the "Slim ISO" but provides multiple preconfigured Window Managers in the "Full ISO". Similar to Kali Linux and Parrot OS, BlackArch can be burned to an ISO image and run as a live system. BlackArch can also be installed as an unofficial user repository on any current Arch Linux installation. Packages BlackArch currently contains 2859 packages and tools, along with their dependencies. BlackArch is developed by a small number of cyber security specialists and researchers that add the packages as well as dependencies needed to run these tools. Tools categories within the BlackArch distribution (Counting date: 9 June 2022): blackarch-anti-forensic: 2 tools blackarch-automation: 106 tools blackarch-automobile: 3 tools blackarch-backdoor: 47 tools blackarch-binary: 68 tools blackarch-bluetooth: 25 tools blackarch-code-audit: 29 tools blackarch-cracker: 166 tools blackarch-crypto: 74 tools blackarch-database: 6 tools blackarch-debugger: 16 tools blackarch-decompiler: 17 tools blackarch-defensive: 45 tools blackarch-disassembler: 19 tools blackarch-dos: 30 tools blackarch-drone: 4 tools blackarch-exploitation: 180 tools blackarch-fingerprint: 30 tools blackarch-firmware: 4 tools blackarch-forensic: 124 tools blackarch-fuzzer: 84 tools blackarch-hardware: 6 tools blackarch-honeypot: 16 tools blackarch-ids: 1 tool blackarch-keylogger: 3 tools
https://en.wikipedia.org/wiki/Trajectory%20inference
Trajectory inference or pseudotemporal ordering is a computational technique used in single-cell transcriptomics to determine the pattern of a dynamic process experienced by cells and then arrange cells based on their progression through the process. Single-cell protocols have much higher levels of noise than bulk RNA-seq, so a common step in a single-cell transcriptomics workflow is the clustering of cells into subgroups. Clustering can contend with this inherent variation by combining the signal from many cells, while allowing for the identification of cell types. However, some differences in gene expression between cells are the result of dynamic processes such as the cell cycle, cell differentiation, or response to an external stimuli. Trajectory inference seeks to characterize such differences by placing cells along a continuous path that represents the evolution of the process rather than dividing cells into discrete clusters. In some methods this is done by projecting cells onto an axis called pseudotime which represents the progression through the process. Methods Since 2015, more than 50 algorithms for trajectory inference have been created. Although the approaches taken are diverse there are some commonalities to the methods. Typically, the steps in the algorithm consist of dimensionality reduction to reduce the complexity of the data, trajectory building to determine the structure of the dynamic process, and projection of the data onto the trajectory so that cells are positioned by their evolution through the process and cells with similar expression profiles are situated near each other. Trajectory inference algorithms differ in the specific procedure used for dimensionality reduction, the kinds of structures that can be used to represent the dynamic process, and the prior information that is required or can be provided. Dimensionality reduction The data produced by single-cell RNA-seq can consist of thousands of cells each with expression levels r
https://en.wikipedia.org/wiki/Systems%20Engineering%20Body%20of%20Knowledge
The Systems Engineering Body of Knowledge (SEBoK), formally known as Guide to the Systems Engineering Body of Knowledge, is a wiki-based collection of key knowledge sources and references for systems engineering. The SEBoK is a curated wiki meaning that the content is managed by an editorial board, and updated on a regular basis. This wiki is a collaboration of three organizations: 1) International Council on Systems Engineering (INCOSE), 2) IEEE Systems Council, and 3) Stevens Institute of Technology. The most recent version (v.2.5) was released on October 15, 2021. History The Guide was developed over three years, from 2009 to 2012, through the contributions of 70 authors worldwide. During this period, three prototype versions were created. The first prototype (v.0.25) was a document that was released for review in September 2010. However, the final versions were all published online as agreed by the authors in January 2011. This switch to a wiki-based SEBoK began with v.0.50. The first version of the SEBoK for public use was published online in September 2012. The initial release was named 2012 product of the year by the International Council on Systems Engineering. Since then, the guide had several revisions and minor updates leading to the 19th release, as of November 2018. Version 1.7, released on October 27, 2016, added a new Healthcare Systems Engineering knowledge area. Knowledge areas According to the site, the guide has a total of 26 knowledge areas distributed among the different parts. However, the majority of these knowledge areas can be grouped to form nine general knowledge areas. The general and specific knowledge areas are: Science & Technology Knowledge Introduction to Life Cycle Processes Life Cycle Models Concept Definition Domain Technology Knowledge System Definition System Realization System Deployment and Use Operational Environment Knowledge Engineering Discipline/ Specialty Knowledge Systems Engineering and Software Engine
https://en.wikipedia.org/wiki/Search%20engine%20privacy
Search engine privacy is a subset of internet privacy that deals with user data being collected by search engines. Both types of privacy fall under the umbrella of information privacy. Privacy concerns regarding search engines can take many forms, such as the ability for search engines to log individual search queries, browsing history, IP addresses, and cookies of users, and conducting user profiling in general. The collection of personally identifiable information (PII) of users by search engines is referred to as tracking. This is controversial because search engines often claim to collect a user's data in order to better tailor results to that specific user and to provide the user with a better searching experience. However, search engines can also abuse and compromise its users' privacy by selling their data to advertisers for profit. In the absence of regulations, users must decide what is more important to their search engine experience: relevance and speed of results or their privacy, and choose a search engine accordingly. The legal framework for protecting user privacy is not very solid. The most popular search engines collect personal information, but other search engines that are focused on privacy have cropped up recently. There have been several well publicized breaches of search engine user privacy that occurred with companies like AOL and Yahoo. For individuals interested in preserving their privacy, there are options available to them, such as using software like Tor which makes the user's location and personal information anonymous or using a privacy focused search engine. Privacy policies Search engines generally publish privacy policies to inform users about what data of theirs may be collected and what purposes it may be used for. While these policies may be an attempt at transparency by search engines, many people never read them and are therefore unaware of how much of their private information, like passwords and saved files, are collecte
https://en.wikipedia.org/wiki/Wiley%20Interdisciplinary%20Reviews%3A%20Systems%20Biology%20and%20Medicine
Wiley Interdisciplinary Reviews: Systems Biology and Medicine (abbreviated WIREs Systems Biology and Medicine) is a bimonthly peer-reviewed interdisciplinary scientific review journal covering systems biology and medicine. It was established in 2009 and is published by John Wiley & Sons as part of its Wiley Interdisciplinary Reviews journal series. The editors-in-chief are Joseph H. Nadeau (Pacific Northwest Research Institute) and Shankar Subramaniam (University of California, San Diego). According to the Journal Citation Reports, the journal has a 2020 impact factor of 5.000, ranking it 48th out of 140 journals in the category "Medicine, Research & Experimental". References External links Systems biology Biology journals General medical journals Academic journals established in 2009 Bimonthly journals Wiley (publisher) academic journals English-language journals
https://en.wikipedia.org/wiki/One%20UI
One UI is a user interface (UI) developed by Samsung Electronics for its Android devices running Android 9 "Pie" and later. Succeeding Samsung Experience and TouchWiz, it is designed to make using larger smartphones easier and be more visually appealing. To provide more clarity, some elements of the UI are tweaked to match colors that are based on the color of the user's phone. It was announced at Samsung Developer Conference in 2018, and was unveiled in Galaxy Unpacked in February 2019 alongside the Galaxy S10 series, Galaxy Buds and the Galaxy Fold. It is also the software layer for their smartwatch Tizen and Wear OS platform, which Samsung co-developed with Google. As of 2021, it is also the software layer for the Microsoft Windows platform on Galaxy Book devices. Features One UI was designed as part of a goal to make Samsung's hardware and software "work together in perfect harmony" and provide a more "natural" experience on large-screen smartphones. One UI displays most of the features that were in the Samsung Experience UX. A prominent design pattern in many of Samsung's system applications is to intentionally place common features and user interface elements along the middle of the screen rather than near the top. This makes them easier to reach with a user's thumb when using the device one-handed. For similar reasons, apps utilise large headers to push their main content towards the vertical centre of the screen. The navigation bar supports the use of gestures and the usual 3-button system, while a system-wide "night mode" was also added (which gives UI elements and supported applications a darkened color scheme). As with Android Pie upstream, the Overview screen of recent apps uses a horizontal layout, as opposed to the vertical layout of previous versions. One UI Core One UI Core is a slimmed down version of the original One UI feature set aimed towards the budget-friendly A, F, and M series low and mid-range devices. These variants of the system usu
https://en.wikipedia.org/wiki/Leontovich%20boundary%20condition
The Leontovich boundary condition is a boundary condition in classical electrodynamics that relates to the tangential components of the electric Et and magnetic Ht fields on the surface of well-conducting bodies. Definition As originally formulated by Russian physicist Mikhail Leontovich, the boundary condition is given as where and represent the tangential components of the electric and magnetic fields, is the effective surface impedance, and is a unit normal pointing into the conducting material. This condition is accurate when the conductivity of the conductor is large, which is the case for most metals. More generally, for cases when the radii of curvature of the conducting surface is large with respect to the skin depth, the resulting fields on the interior can be well approximated by plane waves, thus giving rise to the Leontovitch condition. Applications The Leontovich boundary condition is useful in many scattering problems where one material is a metal with large (but finite) conductivity. As the condition provides a relationship between the electric and magnetic fields at the surface of the conductor, without knowledge of the fields within, the task of finding the total fields is considerably simplified. References Boundary conditions Electrodynamics
https://en.wikipedia.org/wiki/Voice%20computing
Voice computing is the discipline that develops hardware or software to process voice inputs. It spans many other fields including human-computer interaction, conversational computing, linguistics, natural language processing, automatic speech recognition, speech synthesis, audio engineering, digital signal processing, cloud computing, data science, ethics, law, and information security. Voice computing has become increasingly significant in modern times, especially with the advent of smart speakers like the Amazon Echo and Google Assistant, a shift towards serverless computing, and improved accuracy of speech recognition and text-to-speech models. History Voice computing has a rich history. First, scientists like Wolfgang Kempelen started to build speech machines to produce the earliest synthetic speech sounds. This led to further work by Thomas Edison to record audio with dictation machines and play it back in corporate settings. In the 1950s-1960s there were primitive attempts to build automated speech recognition systems by Bell Labs, IBM, and others. However, it was not until the 1980s that Hidden Markov Models were used to recognize up to 1,000 words that speech recognition systems became relevant. Around 2011, Siri emerged on Apple iPhones as the first voice assistant accessible to consumers. This innovation led to a dramatic shift to building voice-first computing architectures. PS4 was released by Sony in North America in 2013 (70+ million devices), Amazon released the Amazon Echo in 2014 (30+ million devices), Microsoft released Cortana (2015 - 400 million Windows 10 users), Google released Google Assistant (2016 - 2 billion active monthly users on Android phones), and Apple released HomePod (2018 - 500,000 devices sold and 1 billion devices active with iOS/Siri). These shifts, along with advancements in cloud infrastructure (e.g. Amazon Web Services) and codecs, have solidified the voice computing field and made it widely relevant to the public at lar
https://en.wikipedia.org/wiki/Video%20manipulation
Video manipulation is a type of media manipulation that targets digital video using video processing and video editing techniques. The applications of these methods range from educational videos to videos aimed at (mass) manipulation and propaganda, a straightforward extension of the long-standing possibilities of photo manipulation. This form of computer-generated misinformation has contributed to fake news, and there have been instances when this technology was used during political campaigns. Other uses are less sinister; entertainment purposes and harmless pranks provide users with movie-quality artistic possibilities. History The concept of manipulating video can be traced back as far as the 1950s when the 2-inch Quadruplex tape used in videotape recorders would be manually cut and spliced. After being coated with ferrofluid, the two ends of tape that were to be joined were painted with a mixture of iron filings and carbon tetrachloride, a toxic and carcinogenic compound to make the tracks in the tape visible when viewed through a microscope so that they could be aligned in a splicer designed for this task As the video cassette recorder developed in the 1960s, 1970s, 1980s, and 1990s, the ability to record over an existing magnetic tape became possible. This led to the concept of overlaying specific parts of film to give the illusion of one consistently recorded video, which is the first identifiable instance of video manipulation. In 1985, Quantel released The Harry, the first all-digital video editing and effects compositing system. It recorded and applied effects to a maximum of 80 seconds of 8-bit uncompressed digital video. A few years later, in 1991, Adobe released its first version of Premiere for the Mac, a program that has since become an industry standard for editing and is now commonly used for video manipulation. In 1999, Apple released Final Cut Pro, which competed with Adobe Premiere and was used in the production of major films such as The R
https://en.wikipedia.org/wiki/Quantum%20speed%20limit
In quantum mechanics, a quantum speed limit (QSL) is a limitation on the minimum time for a quantum system to evolve between two distinguishable states. QSL are closely related to time-energy uncertainty relations. In 1945, Leonid Mandelstam and Igor Tamm derived a time-energy uncertainty relation that bounds the speed of evolution in terms of the energy dispersion. Over half a century later, Norman Margolus and Lev Levitin showed that the speed of evolution cannot exceed the mean energy, a result known as the Margolus–Levitin theorem. Realistic physical systems in contact with an environment are known as open quantum systems and their evolution is also subject to QSL. Quite remarkably it was shown that environmental effects, such as non-Markovian dynamics can speed up quantum processes, which was verified in a cavity QED experiment. QSL have been used to explore the limits of computation and complexity. In 2017, QSLs were studied in a quantum oscillator at high temperature. In 2018, it was shown that QSL are not restricted to the quantum domain and that similar bounds hold in classical systems. In 2021, both the Mandelstam-Tamm and the Margolus-Levitin QSL bounds were concurrently tested in a single experiment which indicated there are "two different regimes: one where the Mandelstam-Tamm limit constrains the evolution at all times, and a second where a crossover to the Margolus-Levitin limit occurs at longer times." Mandelstam-Tamm limit Let be the Bures metric, defined byIf a quantum system is evolving under a time-dependent Hamiltonian , then its velocity according to Bures metric is upper bounded bywhere is the uncertainty in energy at time . Two corollaries: The time taken to evolve from to is , where is the time-averaged uncertainty in energy. The time taken to evolve from one pure state to another pure state orthogonal to it is . Applications Computation machinery is constructed out of physical matter that follows quantum mechanics, and each
https://en.wikipedia.org/wiki/Chiplet
A chiplet is a tiny integrated circuit (IC) that contains a well-defined subset of functionality. It is designed to be combined with other chiplets on an interposer in a single package. A set of chiplets can be implemented in a mix-and-match "Lego-like" assembly. This provides several advantages over a traditional system on chip: Reusable IP (intellectual property): the same chiplet can be used in many different devices Heterogeneous integration: chiplets can be fabricated with different processes, materials, and nodes, each optimized for its particular function Known good die: chiplets can be tested before assembly, improving the yield of the final device Multiple chiplets working together in a single integrated circuit may be called a multi-chip module, hybrid IC, 2.5D IC, or an advanced package. Chiplets may be connected with standards such as UCIe, bunch of wires (BoW), OpenHBI, and OIF XSR. The term was coined by University of California, Berkeley professor John Wawrzynek as a component of the RAMP Project (research accelerator for multiple processors) in 2006 extension for the Department of Energy, as was RISC-V architecture. References Further reading Integrated circuits Semiconductor devices
https://en.wikipedia.org/wiki/NetQoS
NetQoS, which sells network performance management software and services, was co-founded by Joel Trammell in 1999 and acquired by CA Technologies in 2009. The company's name refers to Network Quality of Service. Their ReportAnalyzer provides "real-time visibility into network traffic" and seeks to improve network performance. Offerings introduced shortly before the company was acquired by CA Technologies include: Performance Center Anomaly Detection Network latency calculator Earlier offerings include: SuperAgent a software/hardware package to identify applications that use excessive bandwidth Notability NetQos' products were cited by over 100 articles regarding NetQos patents and prior art. References History of software American companies established in 1999
https://en.wikipedia.org/wiki/Healthcare%20Environmental%20Services
Healthcare Environment Services (HES) Limited (company number SC173861) was a company based in Shotts. It claimed to be the largest independent medical waste management solutions company in the UK. On 30 April, HES was placed into liquidation History HES was a private limited company, incorporated on 26 March 1997. Its registered office address was Hassockrigg Ecopark, Shotts Road, Shotts, Lanarkshire, ML7 5TQ. Companies House lists two stated "Persons with significant control", Mr Garry Pettigrew (Director and Secretary) and Mrs Alison Pettigrew (Director). As of the 1st if May, 2019 the companies status on Companies House states that HES are "In Liquidation". What relationship, if any, exists between HES Ltd, One Waste Solution Limited, HEG Sustainable Solutions Limited, Healthcare Sharp Systems oLimited, Healthcare Environmental (Group) Limited, and Healthcare Washroom Services Limited is unclear, other than the fact that Mr Garry Pettigrew is a named Officer in each. HES Ltd acquired GW Butler Ltd in 2014 which led to the Competition and Markets Authority (CMA) issuing a formal enforcement order in November 2014. On 18 March 2018 the CMA cleared the completed acquisition. All the 400 staff at HES were given redundancy notices on 27 December 2018 as a result of the scandal outlined below and the company went into liquidation in April 2019. Operational Sites England The company operated waste management facilities in England: HES Ltd, North Tyne, Newcastle HES Ltd, Normanton Scotland The company operated two [three?] waste management facilities in Scotland: HES Ltd, Gourdie Industrial Estate, Dundee HES Ltd, Hassockrigg, Shotts HES Ltd, Aberdeen (?) Contracts In 2009 the HES Ltd won a 10-year contract to dispose of NHS Scotland's clinical waste, including waste from all the Scottish hospitals, GP's surgeries, dental practices and pharmacies, from the incumbent service provider Stericycle. Bidding for the next 10 year NHS Scotland contract (NP8
https://en.wikipedia.org/wiki/Bitwarden
Bitwarden is a freemium open-source password management service that stores sensitive information such as website credentials in an encrypted vault. The platform offers a variety of client applications including a web interface, desktop applications, browser extensions, mobile apps, and a command-line interface. Bitwarden offers a free US or European cloud-hosted service as well as the ability to self-host. Desktop applications are available for Windows, MacOS, and Linux. Browser extensions include Chrome, Firefox, Safari, Edge, Opera, Vivaldi, Arc, Brave and Tor. Mobile apps for Android, iPhone, and iPad are available. Client functionalities include 2FA login, passwordless login, biometric unlock, random password generator, password strength testing tool, login/form/app autofill, syncing across unlimited platforms and devices, storing unlimited number of items, sharing credentials, and storing a variety of information including credit cards. Features Overall security Open-source codebase Zero-knowledge encryption, i.e., the company can't see the vault contents End-to-end encryption of the stored vault data Uses AES-CBC 256-bit to encrypt vault data, and PBKDF2 SHA-256 / Argon2id to derive user's encryption key from the entered password. Third-party independent application/code-library/network-infrastructure audits and bug bounty program Vault storage Cloud synchronization (Microsoft Azure)with free version being able to sync across unlimited platforms and devices Can choose either US cloud or European cloud, and migrate between them Can self-host the Bitwarden server on-premises, or with services such as DigitalOcean Availability Variety of client applications including a web interface, desktop applications, browser extensions, mobile apps, and a command-line interface. Desktop applications are available for Windows, MacOS, and Linux Browser extensions are available for Chrome, Firefox, Safari, Edge, Opera, Vivaldi, Brave and Tor. Mobile apps a
https://en.wikipedia.org/wiki/Combinatorial%20matrix%20theory
Combinatorial matrix theory is a branch of linear algebra and combinatorics that studies matrices in terms of the patterns of nonzeros and of positive and negative values in their coefficients. Concepts and topics studied within combinatorial matrix theory include: (0,1)-matrix, a matrix whose coefficients are all 0 or 1 Permutation matrix, a (0,1)-matrix with exactly one nonzero in each row and each column The Gale–Ryser theorem, on the existence of (0,1)-matrices with given row and column sums Hadamard matrix, a square matrix of 1 and –1 coefficients with each pair of rows having matching coefficients in exactly half of their columns Alternating sign matrix, a matrix of 0, 1, and –1 coefficients with the nonzeros in each row or column alternating between 1 and –1 and summing to 1 Sparse matrix, a matrix with few nonzero elements, and sparse matrices of special form such as diagonal matrices and band matrices Sylvester's law of inertia, on the invariance of the number of negative diagonal elements of a matrix under changes of basis Researchers in combinatorial matrix theory include Richard A. Brualdi and Pauline van den Driessche. References Linear algebra Combinatorics
https://en.wikipedia.org/wiki/SCMOS
sCMOS (scientific Complementary Metal–Oxide–Semiconductor) are a type of CMOS image sensor (CIS). These sensors are commonly used as components in specific observational scientific instruments, such as microscopes and telescopes. sCMOS image sensors offer extremely low noise, rapid frame rates, wide dynamic range, high quantum efficiency, high resolution, and a large field of view simultaneously in one image. The sCMOS technology was launched in 2009 during the Laser World of Photonics fair in Munich. The companies Andor Technology, Fairchild Imaging and PCO Imaging developed the technology for image sensors as a joint venture. Technical details Prior to the introduction of the technology, scientists were limited to using either CCD or EMCCD cameras, both of which had their own set of technical limitations. While back-illuminated electron-multiplying CCD (EMCCD) cameras are optimal for purposes requiring the lowest noise and dark currents, sCMOS technology's higher pixel count and lower cost result in its use in a wide range of precision applications. sCMOS devices can capture data in a global-shutter “snapshot” mode over all the pixels or rectangular subsets of pixels, and can also operate in a rolling-shutter mode. The cameras are available with a monochrome sCMOS image sensors or with RGB sCMOS image sensors. With sCMOS, digital information for each frame is generated rapidly and with an improved low-light image quality. The sCMOS sensor's low read noise and larger area provides a low-noise, large field-of-view (FOV) image that enables researchers to scan across a sample and capture high-quality images. Some disadvantages at this time, (2023), with sCMOS cameras versus related technologies are: sCMOS sensors tend be more expensive than traditional CMOS sensors. sCMOS sensors have a limited resolution compared to other types of sensors like CCD. In practice The New York University School of Medicine uses sCMOS cameras for their research. They were used t
https://en.wikipedia.org/wiki/Perlmutter%20%28supercomputer%29
Perlmutter (also known as NERSC-9) is a supercomputer delivered to the National Energy Research Scientific Computing Center of the United States Department of Energy as the successor to Cori. It is being built by Cray and is based on their Shasta architecture which utilizes Zen 3 based AMD Epyc CPUs ("Milan") and next-generation Nvidia Tesla GPUs. Its intended use-cases are nuclear fusion simulations, climate projections and material and biological research. Phase 1, completed 27 May 2022, reached 70.9 PFLOPS of processing power. It is named in honour of Nobel prize winner Saul Perlmutter. References Cray products GPGPU supercomputers Nvidia United States Department of Energy facilities
https://en.wikipedia.org/wiki/Interference%20channel
In information theory, the interference channel is the basic model used to analyze the effect of interference in communication channels. The model consists of two pairs of users communicating through a shared channel. The problem of interference between two mobile users in close proximity or crosstalk between two parallel landlines are two examples where this model is applicable. Unlike in the point-to-point channel, where the amount of information that can be sent through the channel is limited by the noise that distorts the transmitted signal, in the interference channel, it is mainly the signal from the other user that hinders the communication. However, the transmitted signals are not purely random (otherwise they would not be decodable), and, therefore, the users can reduce the effect of the interference by decoding the undesired signal. Discrete memoryless interference channel The mathematical model for this channel is the following: where, for : is the message to be transmitted by user ; is the channel input symbol ( is a sequence of symbols) of user ; is the channel output symbol ( is a sequence of symbols) of user ; is the estimate of the transmitted message by user ; and is the noisy memoryless channel, which is modeled by a conditional probability distribution. The capacity of this channel model is not known in general; only for special cases of the capacity has been calculated, e.g., in the case of strong interference or deterministic channels. References Further references R. Ahlswede, “The Capacity Region of a Channel with Two Senders and Two Receivers,” The Annals of Probability, vol. 2, No. 5, pp. 805–814, Oct. 1974. A. B. Carleial, "Interference Channels", IEEE Transactions on Information Theory, vol. 24, no. 1, pp. 60-70, Jan. 1978. T. S. Han and K. Kobayashi, “A New Achievable Rate Region for the Interference Channel,” IEEE Trans. Inf. Theory, vol. 27, No. 1, pp. 49–60, Jan. 1981. R. H. Etkin, D. Tse, and H. Wang, “Gaussi
https://en.wikipedia.org/wiki/Ethiopian%20Biotechnology%20Institute
The Bio and Emerging Technology Institute (BETin), formerly known as Ethiopian Biotechnology Institute (EBTi), is a research institute owned by the Government of Ethiopia. It was established in 2016 in Addis Ababa, Ethiopia. History The Bio and Emerging Technology Institute was the first research institute to prepare and publish a national biotechnology policy to the Ministry of Science and Technology (MoST) for review and endorsement of its approval by the Council of Ministers (CoM). References External links Bio and Emerging Technology Institute Science and technology in Ethiopia Government agencies of Ethiopia Innovation ministries Biotechnology organizations
https://en.wikipedia.org/wiki/Peg%20loom
A peg loom is a simple weaving loom. Handheld weaving sticks use the same principle. A peg loom is a board, usually wooden, with one or more rows of holes, and a set of wooden or nylon pegs which fit into these holes. Each peg is a dowel with a hole drilled along its diameter near one end. Handheld weaving sticks are similar to the pegs, but tapered at the hole end and pointed at the other end. Plastic looms are also made for the educational market. Double-length warp threads are threaded through the hole in each peg or stick, and the loose ends knotted. The pegs are inserted into the loom, or the sticks are handheld, and the weft thread is woven around the pegs or sticks. As the work progresses, it is slid off the pegs or sticks onto the trailing warp threads. Different yarns can be used to create patterns, and when using sticks a curved piece of work can be created by weaving less often round certain sticks. Looms are made in a range of sizes. As an example, one English company makes looms from , with 9-63 holes. Most of their looms have three rows of holes, with 6mm pegs at spacings of 12.5mm and 9mm holes at spacings of 18.5 and 25.4mm. A larger scale peg loom has been used to create sleeping mats for homeless people from recycled plastic carrier bags. It has been said that stick weaving was used by people of the Great Lakes region of North America in the 1500s when French trappers first encountered them, and that in the 18th century it was taught to children to develop their manual dexterity before they entered the weaving trade. References Weaving equipment
https://en.wikipedia.org/wiki/Font%20Bomb
Font Bomb is a JavaScript bookmarklet to "Blow Up" web pages. When the script is loaded, clicking on a web page starts a countdown. When the countdown reaches zero, it uses Cascading Style Sheets to scatter nearby text across the page. The script wraps all affected letters in a tag, so that they can be moved individually. References JavaScript Web development External links Main site
https://en.wikipedia.org/wiki/MLOps
MLOps or ML Ops is a paradigm that aims to deploy and maintain machine learning models in production reliably and efficiently. The word is a compound of "machine learning" and the continuous development practice of DevOps in the software field. Machine learning models are tested and developed in isolated experimental systems. When an algorithm is ready to be launched, MLOps is practiced between Data Scientists, DevOps, and Machine Learning engineers to transition the algorithm to production systems. Similar to DevOps or DataOps approaches, MLOps seeks to increase automation and improve the quality of production models, while also focusing on business and regulatory requirements. While MLOps started as a set of best practices, it is slowly evolving into an independent approach to ML lifecycle management. MLOps applies to the entire lifecycle - from integrating with model generation (software development lifecycle, continuous integration/continuous delivery), orchestration, and deployment, to health, diagnostics, governance, and business metrics. According to Gartner, MLOps is a subset of ModelOps. MLOps is focused on the operationalization of ML models, while ModelOps covers the operationalization of all types of AI models. Definition MLOps is a paradigm, including aspects like best practices, sets of concepts, as well as a development culture when it comes to the end-to-end conceptualization, implementation, monitoring, deployment, and scalability of machine learning products. Most of all, it is an engineering practice that leverages three contributing disciplines: machine learning, software engineering (especially DevOps), and data engineering. MLOps is aimed at productionizing machine learning systems by bridging the gap between development (Dev) and operations (Ops). Essentially, MLOps aims to facilitate the creation of machine learning products by leveraging these principles: CI/CD automation, workflow orchestration, reproducibility; versioning of data, model,
https://en.wikipedia.org/wiki/The%20Creepy%20Line
The Creepy Line is a 2018 American documentary exploring the influence Google and Facebook have on public opinion, and the power the companies have that is not regulated or controlled by national government legislation. The title is taken from a quote by Eric Schmidt, who when describing Google's use of personal information stated that the company did not cross the line that an ordinary user would find unacceptable. It was released in October 2018. Synopsis The film contrasts the notion of fake news which is visible, with the invisible ranking or masking of information by Google and Facebook, and features headshot interviews with Robert Epstein, Jaron Lanier, Jordan Peterson and Peter Schweizer. Reception The Verge reviewed The Creepy Line, writing that "Despite its dark name, The Creepy Line appeals to the comforting logic of conspiracy: when something goes wrong in life, it’s because an all-powerful entity thinks you’re important enough to attack." References External links American documentary films Films about social media Criticisms of software and websites Documentary films about Google 2010s English-language films 2010s American films
https://en.wikipedia.org/wiki/Offshore%20embedded%20anchors
Offshore embedded anchors are anchors intended for offshore use that derive their holding capacity from the frictional, or bearing, resistance of the surrounding soil, as opposed to gravity anchors, which derive their holding capacity largely from their weight. As offshore developments move into deeper waters, gravity-based structures become less economical due to the large size needed and the consequent cost of transportation. Each of several embedded-anchor types presents its own advantages for anchoring offshore structures. The choice of anchoring solution depends on multiple factors, such as the type of offshore facility that requires mooring, its location, economic viability, the lifetime of its use, soil conditions, and resources available. Examples of facilities that may need mooring offshore are floating production storage and offloading (FPSO) units, mobile offshore drilling units, offshore oil production platforms, wave power and other renewable energy converters, and floating liquefied natural gas facilities. Drag-embedment anchors Drag-embedment anchors (DEA) derive their holding capacity from being buried, or embedded, deep within the seabed with their anchoring capacity being directly related to embedment depth. DEAs are installed by means of dragging, using a mooring chain or wire, this relatively simple means of installation making the DEA a cost-effective option for anchoring offshore structures. DEAs are commonly used for temporary moorings of offshore oil and gas structures, e.g. mobile offshore drilling units. Their use in only temporary mooring situations may be largely attributed to uncertainty involving the anchor's embedding trajectory and placement in the soil, which results in uncertainty with regard to the anchor's holding capacity. Under ideal conditions, DEAs are one of the most efficient types of anchors, with holding capacities ranging from 33 to greater than 50 times their weight; and such efficiency gives DEAs an inherent advan
https://en.wikipedia.org/wiki/Gianluca%20Tognon
Gianluca Tognon is an Italian food scientist, researcher, public health expert, author and speaker. In 2017 he has founded the consulting company The Food Scientist, which is based in Sweden. Education Gianluca Tognon has obtained a degree in Biological Sciences from the University of Pavia in 2000. The title of his dissertation was Effects of the antitumor drug Paclitaxel on human colon cancer. In 2003 he has obtained a specialist degree in Pharmacological research from Regione Lombardia (Italy). In 2005 he has completed a Medical specialty degree in Food science and human nutrition at the University of Pavia. Tognon has also been rewarded with the titles of Advanced speaker and Advanced communicator by the American association Toastmasters Internationals. Career Gianluca Tognon started his career in Italy, after the degree in biological sciences at the State University of Pavia. In 2000 he has been granted a fellowship in pharmacology research to work, from September 2000 to September 2003 at the Oncology Department of the “Mario Negri” pharmacological institute in Milan, where he contributed to the testing of two new anticancer drugs (Aplidine and ET-743) developed by the Spanish multinational company Pharma Mar. During 2004 and 2005 he has worked at the Department of Food Safety of the University of Milan, where he developed a new technique for mycotoxin identification based on the electronic nose. In November 2005 he joined the Cancer Prevention and Research Institute and worked as researcher in the Epidemiology unit where he worked within two EU projects: the Diogenes (Diet, obesity and genes) and the Idames (Innovative dietary assessment methods for epidemiological studies and public health). From April 2010 to March 2017 he worked at the Department of Public Health and Community Medicine at the University of Gothenburg (Sweden). He has published several scientific manuscripts on the association between adherence to the Mediterranean diet and longevity
https://en.wikipedia.org/wiki/Kiwi%20drive
A Kiwi drive is a holonomic drive system of three omni-directional wheels (such as omni wheels or Mecanum wheels), 120 degrees from each other, that enables movement in any direction using only three motors. This is in contrast with non-holonomic systems such as traditionally wheeled or tracked vehicles which cannot move sideways without turning first. This drive system is similar to the Killough platform which achieves omni-directional travel using traditional non-omni-directional wheels in a three wheel configuration. Named after the Flightless national bird of New Zealand The Kiwi References Robotics
https://en.wikipedia.org/wiki/Killough%20platform
A Killough platform is a three-wheel drive system that uses traditional wheels to achieve omni-directional movement without the use of omni-directional wheels (such as omni wheels/Mecanum wheels). Designed by Stephen Killough, after which the platform is named, with help from Francois Pin, wanted to achieve omni-directional movement without using the complicated six motor arrangement required to achieve a controllable three caster wheel system (one motor to control wheel rotation and one motor to control pivoting of the wheel). He first looked into solutions by other inventors that used rollers on the rims larger wheels but considered them flawed in some critical way. This led to the Killough system: With Francois Pin, who helped with the computer control and choreography aspects of the design, Killough and Pin readied a public demonstration in 1994. This led to a partnership with Cybertrax Innovative Technologies in 1996 which was developing a motorized wheelchair. By combining two the motion of two-wheel the vehicle can move in the direction of the perpendicular wheel or by rotating all the wheels in the same direction the vehicle can rotate in place. By using the resultant motion of the vector addition of the wheels a Killough platform is able to achieve omni-directional motion. References Robotics
https://en.wikipedia.org/wiki/Solid%20cell%20nests
Solid cell nests, often abbreviated as SCN, also known as solid cell rests, are specific groups of cells found in the thyroid gland of babies. Typically they are a fraction of a millimeter in size but can rarely become larger. They are considered to be the remains of the ultimobranchial body that exists in early development. Discovery Solid cell nests were discovered in 1907 by pathologist Sophia Getzowa, as documented in her paper titled "Über die Glandula parathyreoidea, intrathyreoideale Zellhaufen derselben und Reste des postbranchialen Körpers". References Pathology
https://en.wikipedia.org/wiki/Resident%20engineer
In general, a resident engineer is a person who works at or from the clients' side of a project. He or she possess a high degree of technical and social skills. One of the main goals of the designated role is to foster knowledge transfer. The role exists across various industries with historic roots in the construction business. Construction A resident engineer is a specific construction occupation. It often describes an engineer employed to work from site for the client or the design engineer. The duties include supervision of and issuing of instructions to the contractor and to report regularly to the designer and/or client. The role was common historically and was also defined in the Institution of Civil Engineers Conditions of Contract and FIDIC contracts. It is not a defined role in the commonly used and more modern NEC Engineering and Construction Contract in which similar roles are named as the client's "project manager" and "supervisor". Rail According to Caltrain Standard Manuals "Resident Engineer's Manual", the resident is "the individual designated to administer construction management matters on specific contracts, with authority delegated by the Agency’s Project Manager." See also Construction management Virtual team Distributed development References Engineering occupations
https://en.wikipedia.org/wiki/Fair%20division%20experiments
Various experiments have been made to evaluate various procedures for fair division, the problem of dividing resources among several people. These include case studies, computerized simulations, and lab experiments. Case studies Allocating indivisible heirlooms 1. Flood describes a division of a gift containing 5 parcels: whiskey, prunes, eggs, suitcase, etc. The division was done using the Knaster auction. The resulting division was fair, but in retrospect it was found that coalitions could gain from manipulation. 2. When Mary Anna Lee Paine Winsor died at the age of 93, her estate included two trunks of silver, that had to be divided among her 8 grandchildren. It was divided using a decentralized, fair and efficient allocation procedure, which combined market equilibrium and a Vickrey auction. Although most participants did not fully understand the algorithm or the preference information desired, it handled the major considerations well and was regarded as equitable. Allocating unused classrooms In California, the law says that public school classrooms should be shared fairly among all public school pupils, including those in charter schools. Schools have dichotomous preferences: each school demands a certain number of classes, it is happy if it got all of them and unhappy otherwise. A new algorithm allocates classrooms to schools using a non-trivial implementation of the randomized leximin mechanism. Unfortunately it was not deployed in practice, but it was tested using computer simulations based on real school data. While the problem is computationally-hard, simulations show that the implementation scales gracefully in terms of running time: even when there are 300 charter schools, it terminates in a few minutes on average. Moreover, while theoretically the algorithm guarantees only 1/4 of the maximum number of allocated classrooms, in the simulations it satisfies on average at least 98% of the maximum number of charter schools that can possibly be satisf
https://en.wikipedia.org/wiki/Tannery%27s%20theorem
In mathematical analysis, Tannery's theorem gives sufficient conditions for the interchanging of the limit and infinite summation operations. It is named after Jules Tannery. Statement Let and suppose that . If and , then . Proofs Tannery's theorem follows directly from Lebesgue's dominated convergence theorem applied to the sequence space . An elementary proof can also be given. Example Tannery's theorem can be used to prove that the binomial limit and the infinite series characterizations of the exponential are equivalent. Note that Define . We have that and that , so Tannery's theorem can be applied and References Mathematical analysis Limits (mathematics)
https://en.wikipedia.org/wiki/Mathematical%20modelling%20competition
Mathematical modelling competitions are team competitions for students that aim to promote mathematical modelling to solve problems of real-world importance. Several types of math contests exist. Contests are held at all levels, from grade school to undergraduate college students. See also Mathematical Contest in Modeling MathWorks Math Modeling Challenge International Mathematical Olympiad List of mathematics competitions References Mathematics competitions Mathematics education
https://en.wikipedia.org/wiki/Beilinson%E2%80%93Bernstein%20localization
The Beilinson–Bernstein localization theorem is a foundational result of geometric representation theory, a part of mathematics studying the representation theory of e.g. Lie algebras using geometry. Statement Let G be a reductive group over the complex numbers, and B a Borel subgroup. Then there is an equivalence of categories On the left is the category of D-modules on G/B. On the right χ is a homomorphism χ : Z(U(g)) → C from the centre of the universal enveloping algebra, corresponding to the weight -ρ ∈ t* given by minus half the sum over the positive roots of g. The above action of W on t* = Spec Sym(t) is shifted so as to fix -ρ. Twisted version There is an equivalence of categories for any λ ∈ t* such that λ-ρ does not pair with any positive root α to give a nonpositive integer (it is "regular dominant"): Here χ is the central character corresponding to λ-ρ, and Dλ is the sheaf of rings on G/B formed by taking the *-pushforward of DG/U along the T-bundle G/U → G/B, a sheaf of rings whose center is the constant sheaf of algebras U(t), and taking the quotient by the central character determined by λ (not λ-ρ). Example: SL2 The Lie algebra of vector fields on the projective line P1 is identified with sl2, and via It can be checked linear combinations of three vector fields C ⊂ P1 are the only vector fields extending to ∞ ∈ P1. Here, is sent to zero. The only finite dimensional sl2 representation on which Ω acts by zero is the trivial representation k, which is sent to the constant sheaf, i.e. the ring of functions O ∈ D-Mod. The Verma module of weight 0 is sent to the D-Module δ supported at 0 ∈ P1. Each finite dimensional representation corresponds to a different twist. History The theorem was introduced by . Extensions of this theorem include the case of partial flag varieties G/P, where P is a parabolic subgroup in and a theorem relating D-modules on the affine Grassmannian to representations of the Kac–Moody algebra in . Refere
https://en.wikipedia.org/wiki/White%20matter%20dissection
White matter dissection refers to a special anatomical technique able to reveal the subcortical organization of white matter fibers in the human or animal cadaver brain. The first studies of cerebral white matter (WM) were described by Galen and by the subsequent efforts of Vesalius on human cadaver specimens. The interest for the deep anatomy of the brain pushed anatomist during centuries to create and develop different techniques for specimen preparation and dissection in order to better reveal the complex white matter architectural organization. However, the biggest impact on the dissection of white matter anatomy was made by Joseph Klingler who developed a new method for specimens preparation and dissection. This technique became more feasible and widely used due to an increased quality of dissection and surprising quality of anatomical details. Klingler developed a new method of brain fixation, by freezing already formalin-fixed brains before dissection. First, the water crystallization induced by freezing disrupts the structure of the grey matter (which has a high water content). This process made possible to peel off the cortex from the brain surface without damaging the subcortical white matter organization underneath. Second, the freezing process along the WM fibers, induced a clear separation between them facilitating the dissection by progressive peeling of the fibers. White matter fibre dissection is nowadays considered as a valuable tool to enhance our knowledge about brain connectivity, and has been used to validate tractographic results and vice versa with good consistency between the two techniques, but also for neurosurgical training and neuroanatomical teaching. References Neuroscience Neurosurgery
https://en.wikipedia.org/wiki/Logeion
Logeion is an open-access database of Latin and Ancient Greek dictionaries. Developed by Josh Goldenberg and Matt Shanahan in 2011, it is hosted by the University of Chicago. Apart from simultaneous search capabilities across different dictionaries and reference works, Logeion offers access to frequency and collocation data from the Perseus Project. Features Having started out as an aggregator for Latin and Ancient Greek dictionaries, Logeion has implemented multiple new features in its development. These include: the integration of reference works on antiquity; frequency and collocation data from the Perseus Project; corpus examples, equally retrieved from the Perseus Project; references to relevant chapters in a number of (English-language) textbooks. Furthermore, an iOS app was developed by Joshua Day in 2013. The app's second version, launched in 2018, is also available for Android devices. Dictionaries As of November 2018, Logeion contains the following dictionaries. Dictionaries with full-text search Ancient Greek dictionaries Autenrieth, G. (1891). A Homeric Dictionary for Schools and Colleges. New York: Harper and Brothers. Liddell, H. G., & Scott, R. (1889). An Intermediate Greek Lexicon. Oxford: Clarendon Press. Liddell, H. G., Scott, R., Jones, H. S., & McKenzie, R. (1940). A Greek-English Lexicon. Oxford: Clarendon Press. Slater, W. L. (1969). Lexicon to Pindar. Berlin: De Gruyter. Latin dictionaries Lewis, Ch. T., & Short, Ch. (1879). A Latin Dictionary. Oxford: Clarendon Press. Lewis, Ch. T. (1890). An Elementary Latin Dictionary. New York: American Book Company. Dictionaries without full-text search Ancient Greek dictionaries Adrados, F. R., & Somolinos, J. R. (Eds.). Diccionario Griego-Español. Madrid: CSIC. Muñoz Delgado, L. (2001). Léxico de magia y religión en los papiros mágicos griegos. Madrid: CSIC. Latin dictionaries Babeliowsky, J. K. L., den Hengst, D., Holtland, W., van Lakwijk, W., Marcelis, J. Th. K., Pinkster, H., Smolenaa
https://en.wikipedia.org/wiki/Library%20of%20Latin%20Texts
The Library of Latin Texts (LLT) is a subscription-based database of Latin texts, from antiquity up to the present day. Started in 1991 as the Cetedoc Library of Christian Latin Texts (CLCLT), it continues to be developed by the Centre ‘Traditio Litterarum Occidentalium’ and is hosted by Brepols Publishers. History In 1991, development of the Cetedoc Library of Christian Latin Texts (CLCLT) started, with the aim of encompassing the entirety of Christian Latin literature. This digital database, initially released as a CD-ROM, was produced by the Cetedoc, led by prof. Paul Tombeur at the Université catholique de Louvain. Since 2001, the activities of Cetedoc have been continued in Turnhout, Belgium by the Centre ‘Traditio Litterarum Occidentalium’ (CTLO), still led by Paul Tombeur. In 2002, it was decided to expand the database's chronological scope beyond medieval and patristic times, its name was changed to Library of Latin Texts. In 2009, a B series (LLT-B) was added to the original LLT (hence known as LLT-A). The LLT-B's scope was to accelerate the growth of the database by directly adopting the text of existing editions, without the intensive research work that is applied to the texts of the LLT-A. This work includes verifying facts related to the text and correcting errors in the printed edition. Difference between the LLT and similar databases Unlike similar initiatives, like Corpus Corporum and The Latin Library, the LLT is not an open-access database. This allows for the adoption of copyrighted editions. In fact, while open-access initiatives have to rely on out-of-copyright, possibly outdated editions, the LLT's policy is to select texts that have "been edited according to best contemporary scholarly practice". The texts edited in Brepols' Corpus Christianorum series form the core of the LLT, even though numerically, they are outnumbered by texts edited in other publishers' series and, if no modern edition is available, by out-of-copyright editions. Nev
https://en.wikipedia.org/wiki/GuySpy
GuySpy is a location based online social network, gay dating app. It is available on Android, iOS and via web browser. The app uses location based services like GPS to connect gay, bi and curious men. Through their platform users are able to create profiles with detailed descriptions and with multiple photos. They are able to chat with guys around them and send additional photos, have video messaging and provide their exact location. There is a voice option within the app as well. The app is similar to Grindr, GuySpy also has a map feature that allows the users to search for other guys in other towns and cities. The app was officially launched in 2011 for GayWhistler's WinterPride festival. In 2013 GuySpy launched The PEARL Pride Party app, which was helping users to keep up to date, build and maintain a personal schedule, share information via social networks related to global LGBT pride events and circuit parties. The app was founded in 2010 originally by Stark Mobile and was purchased by Pink Triangle Press in 2017. Controversies GuySpy run into difficulties with Facebook's Community Standards a couple times. In 2012 their Facebook page was blocked for 30 days and later they posted an image of a male and it was taken down within few hours. GuySpy has challenged Facebook's community standards homophobic as they allowed similar images that showed females by other pages. The app's 2013 billboard campaign in Vancouver received complaints. The complainant alleged that the advertisement offended standards of public decency. The ad was showing two men, one sitting on the other's back with the phrase "Taste the difference" According to the council's opinion the words crossed the line of the acceptability of Clause 14. See also List of LGBT social networking services Homosocialization Timeline of online dating services Tinder References LGBT social networking services LGBT online dating services Geosocial networking Social networking services Mobile social
https://en.wikipedia.org/wiki/Carla%20Cotwright-Williams
Carla Denise Cotwright-Williams (born November 6) is an American mathematician who works as a Technical Director and Data Scientist for the United States Department of Defense. She was the second African-American woman to earn a doctorate in mathematics at the University of Mississippi. Early life and education She is the daughter of a police officer and grew up in South Central Los Angeles. Moving to a better neighborhood in Los Angeles as a teenager. She went to Westchester High School and attended summer enrichment programs for underrepresented students there that included courses at the University of California, Los Angeles, and a field trip to see the Space Shuttle at NASA's Armstrong Flight Research Center on Edwards Air Force Base. She graduated in 1991. As an undergraduate at California State University, Long Beach, Cotwright-Williams started in engineering. Then, as a math major, she struggled initially and earned low enough grades to be academically disqualified from the university, but worked hard to return as a student in good standing, eventually earning a bachelor's degree in mathematics in 2000. She then earned a master's degree in mathematics from Southern University in Baton Rouge, Louisiana, in 2002. Initially intending to follow a science & math Ph.D. track, she was persuaded to shift to pure mathematics under the mentorship of an African-American professor, Stella R. Ashford, who became the supervisor for her master's thesis in number theory, Unique Factorization in Bi-Quadratic Number Fields. She went on to doctoral studies at the University of Mississippi, where she became president of the Graduate Student Council and earned a second master's degree there along the way in 2004. She completed her Ph.D. at the University of Mississippi in 2006. Her dissertation was supervised by T. James Reid and concerned matroid theory. She was the second African-American woman to earn a doctorate in mathematics at the university, and was part of a group of
https://en.wikipedia.org/wiki/Swelling%20index
Swelling index may refer to the following material parameters that quantify volume change: Crucible swelling index, also known as free swelling index, in coal assay Swelling capacity, the amount of a liquid that can be absorbed by a polymer Shrink–swell capacity in soil mechanics Unload-reload constant (κ) in critical state soil mechanics Mechanics Materials science
https://en.wikipedia.org/wiki/BIM%20Collaboration%20Format
The BIM Collaboration Format (BCF) is a structured file format suited to issue tracking with a building information model. The BCF is designed primarily for defining views of a building model and associated information on collisions and errors connected with specific objects in the view. The BCF allows users of different BIM software, and/or different disciplines to collaborate on issues with the project. The use of the BCF to coordinate changes to a BIM is an important aspect of OpenBIM. The format was developed by Tekla and Solibri and later adopted as a standard by buildingSMART. Most major BIM modelling software platforms support some integration with BCF, typically through plug-ins provided by the BCF server vendor. Although the BCF was originally conceived as a file base there are now many implementations using the cloud-based collaborative workflow described in the BCF API, including Open Source implementation as part of the Open Source BIM collective. Research work has been done in Denmark looking into using the BCF for a broader range of information management and exchange in the architecture, engineering and construction (AEC) sector. Supporting software There are two main categories of support for the BCF: authoring software and coordination software. Authoring software can generate and share BCF issues. Coordination software is most powerful at coordinating issues and presenting a user interface for the management and tracking of issues. Coordination software is typically a web-based service which allows for real-time coordination across multiple authoring software platforms and geographies. Most BIM software has a mix of these functions. The BCF is supported natively by authoring software such as Vectorworks, ArchiCAD, Tekla Structures, Quadri, DDS CAD, BIMcollab ZOOM, BIMsight, Solibri, Navisworks, and Simplebim. Standalone BCF plugins include BCF Manager, and BCFier. Coordination software as cloud services offering BCF based issue tracking includ
https://en.wikipedia.org/wiki/He%20Jiankui%20affair
The He Jiankui affair is a scientific and bioethical controversy concerning the use of genome editing following its first use on humans by Chinese scientist He Jiankui, who edited the genomes of human embryos in 2018. He became widely known on 26 November 2018 after he announced that he had created the first human genetically edited babies. He was listed in the Time's 100 most influential people of 2019. The affair led to ethical and legal controversies, resulting in the indictment of He and two of his collaborators, Zhang Renli and Qin Jinzhou. He eventually received widespread international condemnation. He Jiankui, working at the Southern University of Science and Technology (SUSTech) in Shenzhen, China, started a project to help people with HIV-related fertility problems, specifically involving HIV-positive fathers and HIV-negative mothers. The subjects were offered standard in vitro fertilisation services and in addition, use of CRISPR gene editing (CRISPR/Cas9), a technology for modifying DNA. The embryos' genomes were edited to remove the CCR5 gene in an attempt to confer genetic resistance to HIV. The clinical project was conducted secretly until 25 November 2018, when MIT Technology Review broke the story of the human experiment based on information from the Chinese clinical trials registry. Compelled by the situation, he immediately announced the birth of genome-edited babies in a series of five YouTube videos the same day. The first babies, known by their pseudonyms Lulu () and Nana (), are twin girls born in October 2018, and the second birth or the third baby born was in 2019, named Amy. He reported that the babies were born healthy. His actions received widespread criticism, and included concern for the girls' well-being. After his presentation on the research at the Second International Summit on Human Genome Editing at the University of Hong Kong on 28 November 2018, Chinese authorities suspended his research activities the following day. On 30 Dec
https://en.wikipedia.org/wiki/Cryptographic%20agility
In cryptographic protocol design, cryptographic agility or crypto-agility is the ability to switch between multiple cryptographic primitives. Then the systems implementing a particular standard can choose which combination of primitives they want to use. The primary goal of cryptographic agility was to enable rapid adaptations of new cryptographic primitives and algorithms without making disruptive changes to the systems' infrastructure. Cryptographic agility acts as a safety measure or an incident response mechanism when a cryptographic primitive of a system is discovered to be vulnerable. A security system is considered crypto agile if its cryptographic algorithms or parameters can be replaced with ease and is at least partly automated. The impending arrival of a quantum computer that can break existing asymmetric cryptography is raising awareness of the importance of cryptographic agility. Example The X.509 public key certificate illustrates crypto-agility. A public key certificate has cryptographic parameters including key type, key length, and a hash algorithm. X.509 version v.3, with key type RSA, a 1024-bit key length, and the SHA-1 hash algorithm were found by NIST to have a key length that made it vulnerable to attacks, thus prompting the transition to SHA-2. Importance With the rise of secure transport layer communication in the end of the 1990s, cryptographic primitives and algorithms have been increasingly more popular; as an example, by 2019 more than 80% of all website employed some form of security measures. Furthermore, cryptographic techniques are widely incorporated to protect applications and business transactions. However, as cryptographic algorithms are deployed, researching their security intensifies, and new attacks against cryptographic primitives (old and new alike) are discovered in short intervals. Crypto-agility tries to tackle the implied threat to information security by allowing swift deprecation of vulnerable primitives and repla
https://en.wikipedia.org/wiki/Virginia%20Cornish
Virginia Wood Cornish is the Helena Rubinstein Professor of Chemistry at Columbia University. Background and education Cornish received her BA in chemistry in 1991, working with professor Ronald Breslow. Her PhD research, on site-specific protein labeling and mutagenesis, was carried out with Peter Schultz. Cornish was an NSF postdoctoral fellow at MIT with Robert T. Sauer. She is the first female graduate from Columbia College to be hired to a full-time faculty position since the College became coeducational in 1983. Research Cornish and her lab group use the tools of systems biology, synthetic biology, and DNA encoding to produce desired chemical products from specific organismic hosts. In 2016, she was part of a notable group of genomic scientists calling for increased ethical study and self-regulation as the costs and effort of creating de novo genomes plummeted. As the "read" phase of the Human Genome Project was completed in 2004, this new effort was dubbed Genome Project-Write. Awards 2009 – Pfizer Award in Enzyme Chemistry 2009 – Irving Sigal Young Investigator Award 2003 – Sloan Foundation Fellow References Year of birth missing (living people) Living people American women chemists Columbia College (New York) alumni Massachusetts Institute of Technology alumni Synthetic biologists Human Genome Project scientists 21st-century American women
https://en.wikipedia.org/wiki/Developmental%20bias
In evolutionary biology, developmental bias refers to the production against or towards certain ontogenetic trajectories which ultimately influence the direction and outcome of evolutionary change by affecting the rates, magnitudes, directions and limits of trait evolution. Historically, the term was synonymous with developmental constraint, however, the latter has been more recently interpreted as referring solely to the negative role of development in evolution. The role of the embryo In modern evolutionary biology, the idea of developmental bias is embedded into a current of thought called Structuralism, which emphasizes the role of the organism as a causal force of evolutionary change. In the Structuralist view, phenotypic evolution is the result of the action of natural selection on previously ‘filtered’ variation during the course of ontogeny. It contrasts with the Functionalist (also “adaptationist”, “pan-selectionist” or “externalist”) view in which phenotypic evolution results only from the interaction between the deterministic action of natural selection and variation caused by mutation. The rationale behind the role of the organism, or more specifically the embryo, as a causal force in evolution and for the existence of bias is as follows: The traditional, neo-Darwinian, approach to explain the process behind evolutionary change is natural selection acting upon heritable variation caused by genetic mutations. However, natural selection acts on phenotypes and mutation does not in itself produce phenotypic variation, thus, there is a conceptual gap regarding the connection between a mutation and the potential change in phenotype. For a mutation to readily alter a phenotype, and hence be visible to natural selection, it has to modify the ontogenetic trajectory, a process referred to as developmental reprogramming. Some kinds of reprogramming are more likely to occur than others given the nature of the genotype–phenotype map, which determines the propensi
https://en.wikipedia.org/wiki//e/%20%28operating%20system%29
/e/ (also known as /e/ OS and /e/OS, formerly Eelo) is a fork of LineageOS, an Android-based mobile operating system, and associated online services. /e/ is presented as privacy software that does not contain proprietary Google apps or services, and challenges the public to "find any parts of the system or default applications that are still leaking data to Google." Software /e/ is a fork of LineageOS, which is a fork of the CyanogenMod and Android operating systems. /e/ uses MicroG, "an open source project that hijacks Google API calls." according to Ron Amadeo of ars Technica, as an alternative for Google Play Services, and Mozilla Location Service for geolocation. Some /e/ applications and sources are proprietary. As of June 2022, /e/ includes a proprietary maps app. A privacy app was proprietary when first developed, then open source after release. Sources for some devices are not publicly available, according to foundation staff. History In 2017, Mandrake Linux creator Gaël Duval proposed the concept of an operating system without privacy-invasive software as a "non-profit project 'in the public interest'". Duval wrote, "Apple, Google, Facebook etc., business models are harmful for our economical and social environments". The operating system was initially called Eelo; the name was inspired by moray eels, which Duval saw as "fish that can hide in the sea". Duval launched a Kickstarter crowdfunding campaign with an initial goal of €25,000, and received at least €71,000 from contributors. Eelo was renamed to /e/ in July 2018 due to a conflict with the "eelloo" trademark, which was owned by a human resources company. In a March 2020 interview, Duval stated the /e/ name would be abandoned "for something else quite soon". Beta versions of /e/ were released for 20 to 30 smartphone models in September 2018. As of November 2019 /e/ supported 89 smartphone models. As of April 2020, /e/ was teaming with Fairphone to sell phones. Corporations and organizations EC
https://en.wikipedia.org/wiki/School%20of%20Life%20Sciences%20%28University%20of%20Dundee%29
The School of Life Sciences at the University of Dundee conducts research into the molecular and cellular mechanisms underlying human health and disease. History Life Sciences research at the university began within the Department of Physiology. Following a campaign by Robert P. Cook who was a lecturer in Physiological Chemistry, the Department of Biochemistry was formed in 1965.... In 1970, the University of Dundee appointed Peter Garland CBE as its first Chair of Biochemistry based within the Department. Peter's arrival coincided with the opening of the Medical Sciences Institute (MSI) and Biological Sciences Institute (BSI) that provided up to date facilities for research to take place. The Department of Biochemistry moved from a converted stable block into the MSI alongside the Department of Anatomy and Physiology. In 1971, Peter recruited Sir Philip Cohen to Dundee, where he has remained ever since and who has played an instrumental role in the expansion of life sciences research in Dundee. The growth of the department was supported by Principal Adam Neville who redirected funds from other areas of the university. Peter had the ethos of employing the best is required to achieve the best, a strategy that exists to this day. Alongside his appointment of Philip, Peter also recruited David Lilley, Grahame Hardie and Chris Higgins. Further key appointments continued, this time led by Graham Warren; Peter's successor. In 1989, the Cancer Research Campaign supported the establishment of laboratories for David Glover, Birgitt Lane and David Lane. In 1990, the Medical Research Council set up the Protein Phosphorylation Unit which still exists today as the MRC Protein Phosphorylation and Ubiquitylation Unit (MRC PPU). Expansion of research facilities In 1994, the Wellcome Trust awarded £10 million to build the Wellcome Trust Biocentre which opened in 1997 and allowed further expansion in the research with several key appointments. As part of the fundraising efforts f
https://en.wikipedia.org/wiki/Residue-to-product%20ratio
In climate engineering, the residue-to-product ratio (RPR) is used to calculate how much unused crop residue might be left after harvesting a particular crop. Also called the residue yield or straw/grain ratio, the equation takes the mass of residue divided by the mass of crop produced, and the result is dimensionless. The RPR can be used to project costs and benefits of bio-energy projects, and is crucial in determining financial sustainability. The RPR is particularly important for estimating the production of biochar, a beneficial farm input obtained from crop residues through pyrolysis. However, it is important to note that RPR values are rough estimates taken from broad production statistics, and can vary greatly depending on crop variety, climate, processing, and residual moisture content. See also Carbon sequestration Biomass Biochar Biofuel Pyrolysis References Climate engineering Crops Biofuels
https://en.wikipedia.org/wiki/Carrier%20aggregation
In wireless communication, carrier aggregation is a technique used to increase the data rate per user, whereby multiple frequency blocks (called component carriers) are assigned to the same user. The maximum possible data rate per user is increased the more frequency blocks are assigned to a user. The sum data rate of a cell is increased as well because of a better resource utilization. In addition, load balancing is possible with carrier aggregation. Channel selection schemes for CA systems taking into account the optimal values for the training length and power, the number of the probed sub-channels and the feedback threshold such that the sum rate is also important for optimal achievable capacity. Types of carrier aggregation Depending on the positions of the component carriers three cases of carrier aggregation are distinguished: The case where the component carriers are contiguous in the same frequency band is called intra-band contiguous carrier aggregation. If the component carriers are in the same frequency band but are separated by a gap the carrier aggregation is called intra-band non-contiguous. The most complex case is when the component carriers lie in different frequency band. This is called inter-band carrier aggregation applied to heterogeneous networks. There is no difference between these three cases from a baseband perspective. However, the complexity from an radio frequency (RF) point of view is increased in the case inter-band carrier aggregation. Applications UMTS/HSPA+ The channel bandwidth for UMTS/HSPA+ is about 3.8 MHz with a carrier spacing of 5 MHz. Carrier aggregation is also called Dual Cell in the context of UMTS/HSPA+. Through carrier aggregation (part of the UMTS extension HSPA+) two downlink carriers may be assigned to one user since Release 8. Release 10 supports four-carrier aggregation and eight-carrier-aggregation is supported since Release 11. 3GPP standardized carrier aggregation for HSPA+ for the uplink for up to two
https://en.wikipedia.org/wiki/Truncated%20triakis%20octahedron
The truncated triakis octahedron, or more precisely an order-8 truncated triakis octahedron, is a convex polyhedron with 30 faces: 8 sets of 3 pentagons arranged in an octahedral arrangement, with 6 octagons in the gaps. Triakis octahedron It is constructed from taking a triakis octahedron by truncating the order-8 vertices. This creates 6 regular octagon faces, and leaves 24 mirror-symmetric pentagons. Octakis truncated cube The dual of the order-8 truncated triakis octahedron is called a octakis truncated cube. It can be seen as a truncated cube with octagonal pyramids augmented to the faces. See also Truncated triakis tetrahedron Truncated tetrakis cube Truncated triakis icosahedron External links George Hart's Polyhedron generator - "t8kO" (Conway polyhedron notation) Polyhedra Truncated tilings
https://en.wikipedia.org/wiki/Truncated%20triakis%20icosahedron
The truncated triakis icosahedron, or more precisely an order-10 truncated triakis icosahedron, is a convex polyhedron with 72 faces: 10 sets of 3 pentagons arranged in an icosahedral arrangement, with 12 decagons in the gaps. Triakis icosahedron It is constructed from taking a triakis icosahedron by truncating the order-10 vertices. This creates 12 regular decagon faces, and leaves 60 mirror-symmetric pentagons. Decakis truncated dodecahedron The dual of the truncated triakis icosahedron is called a decakis truncated dodecahedron. It can be seen as a truncated dodecahedron with decagonal pyramids augmented to the faces. See also Truncated triakis tetrahedron Truncated triakis octahedron Truncated tetrakis cube External links George Hart's Polyhedron generator - "t10kI" (Conway polyhedron notation) Polyhedra Truncated tilings
https://en.wikipedia.org/wiki/METI%20International
METI International, known simply as METI, is a non-profit research organization founded in July 2015 by Douglas Vakoch that creates and transmits interstellar messages to attempt to communicate with extraterrestrial civilizations. It is based in San Francisco, California. Overview METI targets nearby stars and researches the nature of the messages to send. On October 16, 17, and 18, 2017, it sent a message consisting of a scientific and mathematical tutorial to the red dwarf Luyten's Star, just over 12 light years from Earth. The message was sent from a radio transmitter at the EISCAT research facility in Tromsø, Norway. One of METI's missions is to examine people's views of transmitting interstellar messages. METI's aim is to build an interdisciplinary community to design interstellar messages, within the context of the evolution of intelligence and language. In May 2016, it convened the meeting “The Intelligence Of SETI: Cognition And Communication In Extraterrestrial Intelligence” in Puerto Rico. In May 2018 in Los Angeles, it held “Language in the Cosmos” in conjunction with the International Space Development Conference. to examine the connection between astrobiology and linguistics. On March 22, 2017, it held a workshop in Paris examining the question "What is life?" from an extraterrestrial perspective. METI also conducts an optical search of extraterrestrial intelligence (SETI). Its optical observatory in Panama looks for laser pulses from advanced civilizations. It has examined anomalous stars like the nearby red dwarf star Ross 128, as well as HD 164595, 94 light years from Earth. None of the searches has yielded evidence of artificial signals. Criticism American scientist and science-fiction author David Brin has questioned "whether small groups of zealots should bypass all institutions, peer critique, risk appraisal or public opinion, to shout ‘yoohoo’ into a potentially hazardous cosmos" and so force a fait accompli on humanity. Numerous other aut
https://en.wikipedia.org/wiki/Cyber%20Intelligence%20House
Cyber Intelligence House, formerly Kinkayo, is a Singapore-based cyber intelligence agency which specializes in cyber exposure. The company detects and monitors cyber exposure via dark web, deep web, and data breaches to alert individuals and businesses about potential cyber threats. Products The Cyber Exposure Index (CEI), the company's research-driven project, is the world's first proprietary global scoring system that calculates the exposure index of listed companies. Scores are based on exposed credentials, hacker-group activity, and leaked sensitive information. CEI is a cyber risk score given to public companies listed on stock exchanges hosted in 11 countries (Australia, Finland, Germany, Hong Kong, Indonesia, Italy, South Africa, UK, US, Malaysia, and Singapore). The CEI scores around 6,000 listed companies on their levels of exposure, after analyzing data collected on each company from publicly available sources in the dark web and deep web, and published data breaches. Evaluated companies are given a score ranging from 0 to 300+, with 0 indicating no exposure over the past 12 months and 300+ indicating that the company is among the top 10% most exposed companies globally. The index was first launched in October 2017. The results concluded that publicly listed companies in Australia, Hong Kong and Singapore were the least exposed to cyber threats. In October 2018, the results were updated and Singapore, Malaysia and Indonesia were earmarked as the least exposed countries. This has led to some speculation that cyber threat exposure is very location-centric, as Asia Pacific firms tend to have lower exposure than their American or EU counterparts. Hacked is a cyber exposure monitoring app for individuals, launched in June 2018. It is available on Android, and provides free email account monitoring plus additional paid monitoring options. Controversies The CEI faced some negative press in South Africa, when it was ranked as the third most exposed coun
https://en.wikipedia.org/wiki/Messaging%20Layer%20Security
Messaging Layer Security (MLS) is a security layer for end-to-end encrypting messages in arbitrarily sized groups. It is maintained by the MLS working group of the Internet Engineering Task Force to provide an efficient and practical security mechanism. Security properties Security properties of MLS include message confidentiality, message integrity and authentication, membership authentication, asynchronicity, forward secrecy, post-compromise security, and scalability. History The idea was born in 2016 and first discussed in an unofficial meeting during IETF 96 in Berlin with attendees from Wire, Mozilla and Cisco. Initial ideas were based on pairwise encryption for secure 1:1 and group communication. In 2017, an academic paper introducing Asynchronous Ratcheting Trees was published by the University of Oxford and Facebook setting the focus on more efficient encryption schemes. The first BoF took place in February 2018 at IETF 101 in London. The founding members are Mozilla, Facebook, Wire, Google, Twitter, University of Oxford, and INRIA. As of March 29, 2023, the IETF has approved publication of Messaging Layer Security (MLS) as a new standard. It was officially published on July 19, 2023. Implementations OpenMLS: language: Rust, license: MIT MLS++: language: C++, license: BSD-2 References External links RFC 9420 The Messaging Layer Security (MLS) Protocol Cryptography Internet privacy Secure communication
https://en.wikipedia.org/wiki/Molecular%20layer%20deposition
Molecular layer deposition (MLD) is a vapour phase thin film deposition technique based on self-limiting surface reactions carried out in a sequential manner. Essentially, MLD resembles the well established technique of atomic layer deposition (ALD) but, whereas ALD is limited to exclusively inorganic coatings, the precursor chemistry in MLD can use small, bifunctional organic molecules as well. This enables, as well as the growth of organic layers in a process similar to polymerization, the linking of both types of building blocks together in a controlled way to build up organic-inorganic hybrid materials. Even though MLD is a known technique in the thin film deposition sector, due to its relative youth it is not as explored as its inorganic counterpart, ALD, and a wide sector development is expected in the upcoming years. History Molecular layer deposition is a sister technique of atomic layer deposition. While the history of atomic layer deposition dates back to the 1970s, thanks to the independent work of Valentin Borisovich Aleskovskii. and Tuomo Suntola, the first MLD experiments with organic molecules were not published until 1991, when an article from Tetsuzo Yoshimura and co-workers appeared regarding the synthesis of polyimides using amines and anhydrides as reactants. After some work on organic compounds along the 1990s, the first papers related to hybrid materials emerged, after combining both ALD and MLD techniques. Since then, the number of articles submitted per year on molecular layer deposition has increased steadily, and a more diverse range of deposited layers have been observed, including polyamides, polyimines, polyurea, polythiourea and some copolymers, with special interest in the deposition of hybrid films. Reaction mechanism In similar fashion to an atomic layer deposition process, during an MLD process the reactants are pulsed on a sequential, cyclical manner, and all gas-solid reactions are self-limiting on the sample substrate. Each of
https://en.wikipedia.org/wiki/M%20equilibrium
M equilibrium is a set valued solution concept in game theory that relaxes the rational choice assumptions of perfect maximization (“no mistakes”) and perfect beliefs (“no surprises”). The concept can be applied to any normal-form game with finite and discrete strategies. M equilibrium was first introduced by Jacob K. Goeree and Philippos Louis. Background A large body of work in experimental game theory has documented systematic departures from Nash equilibrium, the cornerstone of classic game theory. The lack of empirical support for Nash equilibrium led Nash himself to return to doing research in pure mathematics. Selten, who shared the 1994 Nobel Prize with Nash, likewise concluded that “game theory is for proving theorems, not for playing games”. M equilibrium is motivated by the desire for an empirically relevant game theory. M equilibrium accomplishes this by replacing the two main assumptions underlying classical game theory, perfect maximization and rational expectations, with the weaker notions of ordinal monotonicity –players' choice probabilities are ranked the same as the expected payoffs based on their beliefs – and ordinal consistency – players' beliefs yield the same ranking of expected payoffs as their choices. M equilibria do not follow from the fixed-points that follow by imposing rational expectations and that have long dominated economics. Instead, the mathematical machinery used to characterize M equilibria is semi-algebraic geometry. Interestingly, some of this machinery was developed by Nash himself. The characterization of M equilibria as semi-algebraic sets allows for mathematically precise and empirically testable predictions. Definition M equilibrium is based on the following two conditions; Ordinal monotonicity: choice probabilities are ranked the same as the expected payoffs based on players’ beliefs. This replaces the assumption of “perfect maximization”. Ordinal consistency: player’s beliefs yield the same ranking of expected
https://en.wikipedia.org/wiki/Terminal%20investment%20hypothesis
The terminal investment hypothesis is the idea in life history theory that as an organism's residual reproductive value (or the total reproductive value minus the reproductive value of the current breeding attempt) decreases, its reproductive effort will increase. Thus, as an organism's prospects for survival decreases (through age or an immune challenge, for example), it will invest more in reproduction. This hypothesis is generally supported in animals, although results contrary to it do exist. Definition The terminal investment hypothesis posits that as residual reproductive value (measured as the total reproductive value minus the reproductive value of the current breeding attempt) decreases, reproductive effort increases. This is based on the cost of reproduction hypothesis, which says that an increase in resources dedicated to current reproduction decreases the potential for future reproduction. But, as the residual reproductive value decreases, the importance of this trade-off decreases, leading to increased investment in the current reproductive attempt. This terminal investment hypothesis can be illustrated by the equation , where is the total reproductive value, the reproductive value of the current breeding attempt, the proportionate increase in resulting from a positive decision (where a yes-no decision must be made regarding whether or not to increase reproductive effort), the cost of a positive decision where there is no selective pressure for either a positive decision or negative decision (this variable is also known as the "barely-justified cost"). The variable is the proportionate loss in from a negative decision. The barely-justified cost is thus inversely proportional to the residual reproductive value. When the level of reproductive investment has not reached the point where the equation above is true, more positive decisions about reproductive effort will be made. Thus, as the residual reproductive value decreases, more positive decis
https://en.wikipedia.org/wiki/Sharp%20MZ%20character%20set
Sharp MZ character set is a character set is made by Sharp Corporation for Sharp MZ. Character set The following tables show the Sharp MZ-700 character sets. Each character is shown with a potential Unicode equivalent. Space and control characters are represented by the abbreviations for their names. References Character sets
https://en.wikipedia.org/wiki/Earth%27s%20circumference
Earth's circumference is the distance around Earth. Measured around the equator, it is . Measured around the poles, the circumference is . Measurement of Earth's circumference has been important to navigation since ancient times. The first known scientific measurement and calculation was done by Eratosthenes, by comparing altitudes of the mid-day sun at two places a known north–south distance apart. He achieved a great degree of precision in his computation. Treating the Earth as a sphere, its circumference would be its single most important measurement. Earth deviates from spherical by about 0.3%, as characterized by flattening. In modern times, Earth's circumference has been used to define fundamental units of measurement of length: the nautical mile in the seventeenth century and the metre in the eighteenth. Earth's polar circumference is very near to 21,600 nautical miles because the nautical mile was intended to express one minute of latitude (see meridian arc), which is 21,600 partitions of the polar circumference (that is 60 minutes × 360 degrees). The polar circumference is also close to 40,000 kilometres because the metre was originally defined to be one ten millionth (i.e., a kilometre is one ten thousandth) of the arc from pole to equator (quarter meridian). The accuracy of measuring the circumference has improved since then, but the physical length of each unit of measure had remained close to what it was determined to be at the time, so the Earth's circumference is no longer a round number in metres or nautical miles. History Posidonius Posidonius calculated the Earth's circumference by reference to the position of the star Canopus. As explained by Cleomedes, Posidonius observed Canopus on but never above the horizon at Rhodes, while at Alexandria he saw it ascend as far as degrees above the horizon (the meridian arc between the latitude of the two locales is actually 5 degrees 14 minutes). Since he thought Rhodes was 5,000 stadia due north of Al
https://en.wikipedia.org/wiki/Einstein%20Institute%20of%20Mathematics
The Einstein Institute of Mathematics () is a centre for scientific research in mathematics at the Hebrew University of Jerusalem, founded in 1925 with the opening of the university. A leading research institute, the institute's faculty has included recipients of the Nobel Prize, Fields Medal, Wolf Prize, and Israel Prize. History About a year before the official inauguration of the Hebrew University, a Jewish-American philanthropist, Philip Wattenberg, endowed the new university with $190,000 (equivalent to $ million in ) for a research institute in the name of theoretical physicist Albert Einstein. The Einstein Mathematics-Physics Institute in 1925. Its inaugural lecture was given by Edmund Landau (on problems from number theory), the first lecture in higher mathematics to be delivered in modern Hebrew. The Institute moved to the Philip Wattenburg Building in 1928, designed by Benjamin Chaikin and Sir Frank Mears, where it remained until the Hebrew University lost access to Mount Scopus in 1948. Edmund Landau served as the university's first Professor of Mathematics, and negotiated the transfer of Felix Klein's private library from Göttingen to Jerusalem, which served as the basis for the new mathematical library in Jerusalem. Other early faculty members included Binyamin Amirà, Abraham Fraenkel, and Michael Fekete. A number of researchers arrived at the institute during the rise of the Nazi regime in Germany, such as Issai Schur and Otto Toeplitz. The Israel Journal of Mathematics was founded at the institute in 1963 as a continuation of the Bulletin of the Research Council of Israel (Section F). A division for computer science was formed within the institute in 1969, which became the independent Institute for Computer Science in 1992. Notable members Current members Karim Adiprasito (1988– ); New Horizons Prize, EMS Prize (2019) Shmuel Agmon (1922– ); Israel Prize (1991) Robert Aumann (1930– ); Israel Prize (1994), Nobel Prize (2005) Ehud de Shalit (1
https://en.wikipedia.org/wiki/Fauna%20of%20Romania
The fauna of Romania comprises all the animal species inhabiting the country of Romania and its coastal territory in the Black Sea. Invertebrates Vertebrates According to a systematic list of the Romanian vertebrate fauna, there are 732 species of vertebrates living in Romania. When grouped into classes, the largest number of these species are birds, with 382 species, followed by fish with 184. 110 of these species are mammals, 31 are reptiles, 20 are amphibians, while only four belong to the Cyclostomata class of jawless fish. Cyclostomata The cyclostomata superclass of vertebrates is represented in Romania by four species of lampreys that live in fast, mountains streams. They are found particularly in Transylvania, in rivers such as Criș, Mureș, Someș and Vișeu. Fish Freshwater Romania's rivers, lakes and ponds are home to numerous species of freshwater fish, most importantly carp, Prussian carp, chub, trout, perch, zander, bream, pike, roach and the Wels catfish. Additionally, six species of sturgeon live in the Black Sea, but travel upriver on the Danube in order to mate. Five of the six Danube sturgeon species are critically endangered, with only the sterlet being considered vulnerable. The most well known of these six species is probably the beluga sturgeon, which is heavily fished for the female's valuable roe – known as beluga caviar. Romania is also home to the asprete, a critically endangered species described by the media as a living fossil and "Europe's rarest fish". Once common in the waters of the Argeș river and its tributaries, Râul Doamnei and Vâlsan, by the 2000s it was only found to be extant in a 1 km stretch of the Vâlsan, with conservationist fearing that less than 10 individuals remained. Increased conservation and survey efforts led to a resurgence in the numbers of identified individuals, with a 2022 survey finding 58 asprete across a 15 km stretch of the Vâlsan. Saltwater The saltwater fish of Romania are the Black Sea species o
https://en.wikipedia.org/wiki/Pairing%20strategy
In a positional game, a pairing strategy is a strategy that a player can use to guarantee victory, or at least force a draw. It is based on dividing the positions on the game-board into disjoint pairs. Whenever the opponent picks a position in a pair, the player picks the other position in the same pair. Example Consider the 5-by-5 variant of Tic-tac-toe. We can create 12 pairwise-disjoint pairs of board positions, denoted by 1,...,12 below: Note that the central element (denoted by *) does not belong to any pair; it is not needed in this strategy. Each horizontal, vertical or diagonal line contains at least one pair. Therefore the following pairing strategy can be used to force a draw: "whenever your opponent chooses an element of pair i, choose the other element of pair i". At the end of the game, you have an element of each winning-line. Therefore, you guarantee that the other player cannot win. Since both players can use this strategy, the game is a draw. This example is generalized below for an arbitrary Maker-Breaker game. In such a game, the goal of Maker is to occupy an entire winning-set, while the goal of Breaker is to prevent this by owning an element in each winning-set. Pairing strategy for Maker A pairing-strategy for Maker requires a set of element-pairs such that: All pairs are pairwise-disjoint; Every set that contains at least one element from each pair, contains some winning-set. Whenever Breaker picks an element of a pair, Maker picks the other element of the same pair. At the end, Maker's set contains at least one element from each pair; by condition 2, he occupies an entire winning-set (this is true even when Maker plays second). As an example, consider a game-board containing all vertices in a perfect binary tree except the root. The winning-sets are all the paths from the leaf to one of the two children of the root. We can partition the elements into pairs by pairing each element with its sibling. The pairing-strategy guarantee
https://en.wikipedia.org/wiki/Link%20flap
Link flap is a condition where a communications link alternates between up and down states. Link flap can be caused by end station reboots, power-saving features, incorrect duplex configuration or marginal connections and signal integrity issues on the link. References Computer networking
https://en.wikipedia.org/wiki/Sequential%20bargaining
Sequential bargaining (also known as alternate-moves bargaining, alternating-offers protocol, etc.) is a structured form of bargaining between two participants, in which the participants take turns in making offers. Initially, person #1 has the right to make an offer to person #2. If person #2 accepts the offer, then an agreement is reached and the process ends. If person #2 rejects the offer, then the participants switch turns, and now it is the turn of person #2 to make an offer (which is often called a counter-offer). The people keep switching turns until either an agreement is reached, or the process ends with a disagreement due to a certain end condition. Several end conditions are common, for example: There is a pre-specified limit on the number of turns; after that many turns, the process ends. There is a pre-specified limit on the negotiation time; when time runs out, the process ends. The number of possible offers is finite, and the protocol rules disallow to offer the same agreement twice. Hence, if the number of possible offers is finite, at some point all them are exhausted, and the negotiation ends without an agreement. Several settings of sequential bargaining have been studied. Dividing the Dollar: two people should decide how to split a given amount of money between them. If they do not reach an agreement, they get nothing. This setting can represent a buyer and a seller bargaining on the price of an item, where the valuations of both players are known. In this case, the amount of money is the difference between the buyer's value and the seller's value. Buyer and Seller: a buyer and the seller bargain over the price of an item, and their valuations of the item are not known. A general outcome set: there is an arbitrary finite set of possible outcomes, each of which yields a different payment to each of the two players. This setting can represent, for example, two parties who have to choose an agreed arbitrator from a given set of candidates.
https://en.wikipedia.org/wiki/Take-back%20system
A take-back system or simply takeback is one of the primary channels of waste collection, especially for e-waste, besides municipal sites. Take-back is the idea that manufacturers and sellers "take back" the products that are at the end of their lives. Take-back is aimed to reduce a business' environmental impacts on the earth and also increase efficiency and lower costs for their business models. "Take-back regulations have targeted a wide array of products including packaging, batteries, automobiles, and electronics" and economic value can be found from recycling or re-manufacturing such products. "The programs benefit municipalities by lowering their overall waste disposal costs and reducing the burden on landfill sites". Although for certain companies, the take-back system is mandatory under legislation, many do it voluntarily. Take-back can be split up into: Store retailer take-back and Producer take-back. It can be further split into two types: Recycling Re-manufacturing Recycling One major option of the take-back system includes store retailers or producers taking back the products that have been distributed to their consumers in order to recycle the materials of these products. The take-back system encourages businesses to redesign their products into ones that are easily recyclable, reducing the burden that virgin materials have on the environment for the present and the future. This also gives companies an alternative supply of raw minerals. Re-manufacturing The other major option of the take-back system includes the store retailers or producers taking back their products in order to create new ones. This process uses the older products in order to restore it into products that are of the same quality as the new ones. By this process, companies save up to 85% of energy that would have been used to manufacture brand new products. Economic effect The take-back system has shown economic effects for many companies that have adapted it. An example
https://en.wikipedia.org/wiki/SimThyr
SimThyr is a free continuous dynamic simulation program for the pituitary-thyroid feedback control system. The open-source program is based on a nonlinear model of thyroid homeostasis. In addition to simulations in the time domain the software supports various methods of sensitivity analysis. Its simulation engine is multi-threaded and supports multiple processor cores. SimThyr provides a GUI, which allows for visualising time series, modifying constant structure parameters of the feedback loop (e.g. for simulation of certain diseases), storing parameter sets as XML files (referred to as "scenarios" in the software) and exporting results of simulations in various formats that are suitable for statistical software. SimThyr is intended for both educational purposes and in-silico research. Mathematical model The underlying model of thyroid homeostasis is based on fundamental biochemical, physiological and pharmacological principles, e.g. Michaelis-Menten kinetics, non-competitive inhibition and empirically justified kinetic parameters. The model has been validated in healthy controls and in cohorts of patients with hypothyroidism and thyrotoxicosis. Scientific uses Multiple studies have employed SimThyr for in silico research on the control of thyroid function. The original version was developed to check hypotheses about the generation of pulsatile TSH release. Later and expanded versions of the software were used to develop the hypothesis of the TSH-T3 shunt in the hypothalamus-pituitary-thyroid axis, to assess the validity of calculated parameters of thyroid homeostasis (including SPINA-GT and SPINA-GD) and to study allostatic mechanisms leading to non-thyroidal illness syndrome. SimThyr was also used to show that the release rate of thyrotropin is controlled by multiple factors other than T4 and that the relation between free T4 and TSH may be different in euthyroidism, hypothyroidism and thyrotoxicosis. Public perception, reception and discussion of the sof
https://en.wikipedia.org/wiki/Paul%20H.%20Silverman
Paul Hyman Silverman (October 8, 1924 – July 15, 2004) was an American medical researcher in the fields of immunology, epidemiology, and parasitology. He was recognized for his research on stem cells and on the human genome. Early life and education Silverman was born on October 8, 1924, in Minneapolis, Minnesota. Growing up, he became fascinated with reading, and he won a local prize for his reading comprehension ability. He attended the University of Minnesota as a pre-medical student while also working three part-time jobs. He went on to serve in a MASH in the United States Army during World War II. He received a bachelor's degree from Roosevelt University. In 1951, Silverman received his M.S. from Northwestern University, after which he moved to Israel with his family. In Israel, he began research on malaria, which he continued to study for many years thereafter. In 1953, he and his family moved again, this time to England. There, he began studying at the Liverpool School of Tropical Medicine, from which he received his Ph.D. in parasitology and epidemiology in 1955. Academic career Silverman returned to the United States when he was 39 years old. He then accepted a position at the University of Illinois at Urbana–Champaign before moving to the University of New Mexico in 1972. At the University of New Mexico, he and his team developed a killed malaria vaccine based on Jonas Salk's polio vaccine. He became Vice President for Research and Graduate Studies at the University of New Mexico in 1975, and joined the State University of New York as their Provost for Research and Graduate Studies in 1978. In 1980, he became president of the University of Maine, a position he held until 1984. At the University of Maine, he was credited with expanding the scope of research activities. In 1984, he returned to research as a senior scientist at the Lawrence Berkeley National Laboratory. He also served at the University of California, Berkeley as Associate Laboratory Directo
https://en.wikipedia.org/wiki/G.%20Peter%20Lepage
G. Peter Lepage (born 13 April 1952) is a Canadian American theoretical physicist and an academic administrator. He was the Harold Tanner Dean of the College of Arts and Sciences at Cornell University from 2003 to 2013. Early life and education Gerard Peter Lepage was born in Canada in 1952. Lepage studied at McGill University and graduated with a bachelor's degree in honours physics in 1972 and the University of Cambridge with a master's degree (M.A.St - Part III of the Mathematical Tripos) in 1973. In 1978, he received his PhD in theoretical physics from Stanford University. Academic career Lepage was a research associate at the Stanford Linear Accelerator Center in 1978. He was a postdoctoral research associate at the Laboratory of Nuclear Studies, Cornell University from 1978 to 1980. In 1980, he joined the physics faculty at Cornell University where he became a professor. He received academic tenure in 1984 after only four years on the university faculty. From 1999 to 2003, he was the chair of Cornell's physics department. He was appointed the Harold Tanner Dean of the College of Arts and Sciences, serving from 2003 to 2013. He is a Fellow of the American Academy of Arts and Sciences and a Fellow of the American Physical Society. He was previously an Alfred P. Sloan Fellow (1983–85; 1990) and John Simon Guggenheim Fellow (1996–97) Since 2012 he has been a member of the National Science Board. G. Peter Lepage has been a visiting scholar at a number of institutions: the Institute for Advanced Study, Princeton; Department of Applied Mathematics and Theoretical Physics, Cambridge; the University of California Institute of Theoretical Physics, Santa Barbara, the Fermi National Accelerator Center near Chicago, and the Institute for Nuclear Theory, Seattle. He was on the editorial board of Physical Review D and Physical Review Letters and received the Outstanding Referee Award from the APS in 2009. He has served on the scientific program committees for the Stan
https://en.wikipedia.org/wiki/Emotive%20Internet
Emotive Internet (also Emotive Web, Emotional Internet) is a conceptualization of the Internet as an emergent emotional public space, such as how it serves as a space for the social sharing of emotions. It can also denote the quality of the Internet that allows it to be used to communicate in an emotive fashion or with emotional intent. Since it is an expressive medium, it also enables users to construct and represent their identities online. This is evident in the way emotional responses have been integrated in online communication and interactions. The concept is also linked to emotional analytics and emotion-sensing applications, particularly those technologies that power the Internet of Things (IoTs) - the smart home devices that have the capability to store and process the user's emotional profile to deliver services. Concept It is recognized that the Internet has the capability to allow its users a genuine display of emotions. This is demonstrated in the video conferencing platform, which represents the closest form of synchronous computer-mediated communication (CMC) to face-to-face communication because it tends to reproduce, on a technological level, relationship and communicative experience that feature complex sensorial channels. It transmits a considerable amount of information so that it sustains interpersonal relationships and exploit shared context that allows for mutual understanding. This phenomenon is also no longer confined to communication between users but also between users and smart devices or users and applications. An application of the emotive quality of the Internet involves emerging technologies that fall within the affective computing field. These include those that use sensor technologies and computer algorithms to enable smart machines to detect, recognize, and share human emotions. The display of emotions in the Internet can be observed in platforms where users interact such as online forums. A Polish study, which analyzed thousa
https://en.wikipedia.org/wiki/Extrinsic%20mortality
Extrinsic mortality is the sum of the effects of external factors, such as predation, starvation and other environmental factors not under control of the individual that cause death. This is opposed to intrinsic mortality, which is the sum of the effects of internal factors contributing to normal, chronologic aging, such as, for example, mutations due to DNA replication errors, and which determined species maximum lifespan. Extrinsic mortality plays a significant role in evolutionary theories of aging, as well as the discussion of health barriers across socioeconomic borders. Evolutionary theories of aging Extrinsic mortality is implicit in both classical theories of aging and non-classical studies of aging. In both cases, its existence causes a selective pressure for either longer lifespans and later reproductive periods or shorter lifespans and earlier reproductive periods. Classical theories of aging include: Mutation Accumulation Theory of Aging - Because extrinsic mortality is so high in the natural world, few animals survive to old age. As a result, random harmful germ line mutations accumulate with little to no selective disadvantage. Antagonistic Pleiotropy Hypothesis - Because pleiotropic genes that express phenotypes that increase fitness in early life and other phenotypes that decrease fitness in later life are generally selected for, harmful phenotypes contribute to senescence and an evolutionary tradeoff between lifespan and reproduction develops. Disposable Soma Theory of Aging - Because resources are often limited and the influence of natural selection declines with age for similar reasons as the Mutation Accumulation theory of aging, organisms that invest greater amounts of energy into earlier reproduction rather than long term body maintenance are more successful. The lack of somatic maintenance and anti-aging mechanisms leads to senescence. These classical evolutionary theories of aging postulate that quantities of extrinsic mortality factor
https://en.wikipedia.org/wiki/Ilse%20Fischer
Ilse Fischer (born 29 June 1975) is an Austrian mathematician whose research concerns enumerative combinatorics and algebraic combinatorics, connecting these topics to representation theory and statistical mechanics. She is a professor of mathematics at the University of Vienna. Education and career Fischer was born in Klagenfurt. She studied at the University of Vienna beginning in 1993, earning a master's degree (mag. rer. nat.), doctorate (dr. rer. nat.), and habilitation there respectively in 1998, 2000, and 2006. Her doctoral dissertation, Enumeration of perfect matchings: Rhombus tilings and Pfaffian graphs, was jointly supervised by Christian Krattenthaler and Franz Rendl, and her habilitation thesis was A polynomial method for the enumeration of plane partitions and alternating sign matrices. She worked as an assistant at Alpen-Adria-Universität Klagenfurt from 1999 to 2004, with a year of postdoctoral research at the Massachusetts Institute of Technology in 2001. She moved to the University of Vienna in 2004, and at Vienna she was promoted to associate professor in 2011 and to full professor in 2017. Recognition Fischer won the 2006 Dr. Maria Schaumayer Prize, and the 2009 Start-Preis of the Austrian Science Fund. With Roger Behrend and Matjaž Konvalinka, Fischer is a winner of the 2019 David P. Robbins Prize of the American Mathematical Society and Mathematical Association of America, for their joint research on alternating sign matrices. References External links 1975 births Living people Scientists from Klagenfurt Austrian mathematicians Women mathematicians Combinatorialists University of Vienna alumni Academic staff of the University of Vienna
https://en.wikipedia.org/wiki/Three-photon%20microscopy
Three-photon microscopy (3PEF) is a high-resolution fluorescence microscopy based on nonlinear excitation effect. Different from two-photon excitation microscopy, it uses three exciting photons. It typically uses 1300 nm or longer wavelength lasers to excite the fluorescent dyes with three simultaneously absorbed photons. The fluorescent dyes then emit one photon whose energy is (slightly smaller than) three times the energy of each incident photon. Compared to two-photon microscopy, three-photon microscopy reduces the fluorescence away from the focal plane by , which is much faster than that of two-photon microscopy by . In addition, three-photon microscopy employs near-infrared light with less tissue scattering effect. This causes three-photon microscopy to have higher resolution than conventional microscopy. Concept Three-photon excited fluorescence was first observed by Singh and Bradley in 1964 when they estimated the three-photon absorption cross section of naphthalene crystals. In 1996, Stefan W. Hell designed experiments to validate the feasibility of applying three-photon excitation to scanning fluorescence microscopy, which further proved the concept of three-photon excited fluorescence. Three-photon microscopy shares a few similarities with Two-photon excitation microscopy. Both of them employ the point scanning method. Both are able to image 3D samples by adjusting the position of the focus lens along the axial and lateral directions. The structures of both systems do not require a pinhole to block out-focus light. However, three-photon microscopy differs from Two-photon excitation microscopy in their Point spread function, resolution, penetration depth, resistance to out-of-focus light and strength of photobleaching. In three-photon excitation, the fluorophore absorbs three photons almost simultaneously. The wavelength of the excitation laser is about 1200 nm or more in three photon microscopy with the emission wavelength slightly longer than one-th
https://en.wikipedia.org/wiki/Triammonium%20citrate
Triammonium citrate is a chemical compound whose molecular formula is C6H17N3O7. Synopsis It was patented some date prior to 1986. This substance causes serious eye irritation, causes skin irritation and may cause respiratory irritation. It is known in the European E number food additive series as E380. It is known in the United States as "an indirect food additive for use only as a component of adhesives", and as a "substance added directly to human food affirmed as generally recognized as safe (GRAS)." References E-number additives Food additives Ammonium compounds Citrates
https://en.wikipedia.org/wiki/Octenylsuccinic%20acid
Octenylsuccinic acid is a manufactured ingredient in foods. It is used to treat gum arabic, and also in waxy maize production and for rice starch production. See also Octenyl succinic acid modified gum arabic References food additives Dicarboxylic acids
https://en.wikipedia.org/wiki/Spanish%20Astrobiology%20Center
Spanish Astrobiology Center ( (CAB)) is a state-run institute in Spain dedicated to astrobiology research, and it is part of the National Institute of Aerospace Technology (INTA) as well as the Spanish National Research Council (CSIC). It was created in 1999 and it is affiliated with NASA Astrobiology Institute. Its main objective is "understanding life as a consequence of the evolution of the matter and energy in the Universe." History The foundation of Spain's Astrobiology Center (CAB) had its beginnings in 1998 when a group of Spanish scientists led by Juan Pérez-Mercader, presented a proposal of affiliation to the newly created NASA Astrobiology Institute (NAI). The affiliation was accepted and the center was officially created on 19 November 1999. It operated from offices at the National Institute of Aerospace Technology (INTA) until it moved to its own building inaugurated in January 2003. Organization The Astrobiology Center is based in Madrid, Spain, its Director is Víctor Parro García, and the Vicedirector is Francisco Najarro. The center is organized into several research and support units, and some of these are associated to Spanish universities, including the University of Valladolid and the Autonomous University of Madrid. The center is part of the National Institute of Aerospace Technology (INTA) as well as the Spanish National Research Council (CSIC). The center is structured in several departments: Astrophysics Department, Molecular Evolution Department, Planetary Science and Habitability Department, Advanced Instrumentations Department, as well as several support units. Research CAB has contributed to NASA in its mission to better characterize and find conditions for life in the Universe, and has prioritized Martian weather research and endurance of some extremophile microorganisms. CAB has developed instruments for multiple missions: Rover Environmental Monitoring Station (REMS) for the Curiosity rover Temperature and Winds for InSight (
https://en.wikipedia.org/wiki/Hazards%20of%20synthetic%20biology
The hazards of synthetic biology include biosafety hazards to workers and the public, biosecurity hazards stemming from deliberate engineering of organisms to cause harm, and hazards to the environment. The biosafety hazards are similar to those for existing fields of biotechnology, mainly exposure to pathogens and toxic chemicals; however, novel synthetic organisms may have novel risks. For biosecurity, there is concern that synthetic or redesigned organisms could theoretically be used for bioterrorism. Potential biosecurity risks include recreating known pathogens from scratch, engineering existing pathogens to be more dangerous, and engineering microbes to produce harmful biochemicals. Lastly, environmental hazards include adverse effects on biodiversity and ecosystem services, including potential changes to land use resulting from agricultural use of synthetic organisms. In general, existing hazard controls, risk assessment methodologies, and regulations developed for traditional genetically modified organisms (GMOs) also apply to synthetic organisms. "Extrinsic" biocontainment methods used in laboratories include biosafety cabinets and gloveboxes, as well as personal protective equipment. In agriculture, they include isolation distances and pollen barriers, similar to methods for biocontainment of GMOs. Synthetic organisms might potentially offer increased hazard control because they can be engineered with "intrinsic" biocontainment methods that limit their growth in an uncontained environment, or prevent horizontal gene transfer to natural organisms. Examples of intrinsic biocontainment include auxotrophy, biological kill switches, inability of the organism to replicate or to pass synthetic genes to offspring, and the use of xenobiological organisms using alternative biochemistry, for example using artificial xeno nucleic acids (XNA) instead of DNA. Existing risk analysis systems for GMOs are generally applicable to synthetic organisms, although there may be
https://en.wikipedia.org/wiki/Space%20Data%20Integrator
Space Data Integrator is a process/service platform or tool being developed by the US FAA to integrate space launch and reentry into the US National Airspace System. History The project was initiated in 2015. No funds for SDI were included in the FAA 2018 budget request. In March 2018 the FAA initiated a Market Survey on the requirements for SDI. References Further reading Improving the Integration of Launch and Reentry Operations into the National Airspace System Mazzotta and Murray. 2015? Aerospace engineering
https://en.wikipedia.org/wiki/NTCA%20-%20The%20Rural%20Broadband%20Association
NTCA - The Rural Broadband Association (NTCA) is a membership association with the goal of improving communications services in rural America. With a membership comprising nearly 850 independent rural American telecommunications companies in forty-four states, NTCA provides training and employee benefit packages to its members. It also advocates rural issues to legislatures, including universal service, rural infrastructure, cybersecurity, telemedicine, and consumer protection. History In 1949, the Rural Electrification Administration (REA) loan program was established to give long-term, low-interest loans to rural telephone systems. In response, the National Rural Electric Cooperative Association (NRECA) created a committee of representatives from emerging joint electric-telephone cooperative organizations. On June 1, 1954, the eight committee members founded the National Telephone Cooperative Association (NTCA) as a separate national organization that would represent telephone cooperatives. In 1956, NTCA successfully advocated for the maintenance of the REA telephone loan program, which the Eisenhower Administration had attempted to terminate. The organization soon entered into an agreement with NRECA to make the NRECA's insurance and benefit programs available to the employees of NTCA member organizations. By the end of 1956, NTCA’s membership had grown from its original eight members to sixty members, with that figure growing to nearly one hundred members during the 1960s. Throughout the 1960s, NTCA worked to improve the availability of financing to its members by supporting the REA and advocating for the creation of a supplemental bank for rural telephone systems. In 1971, Congress established the Rural Telephone Bank to provide financing to rural telephone companies. In 1970 NTCA allowed locally-owned and controlled commercial telecoms to join the association as non-voting members and in 1971 held its first legislative conference. During the 1970s, NTCA’s me
https://en.wikipedia.org/wiki/Direct%20air%20capture
Direct air capture (DAC) is the use of chemical or physical processes to extract carbon dioxide directly from the ambient air. If the extracted is then sequestered in safe long-term storage (called direct air carbon capture and sequestration (DACCS)), the overall process will achieve carbon dioxide removal and be a "negative emissions technology" (NET). As of 2023, DAC has yet to become profitable because the cost of using DAC to sequester carbon dioxide is several times the carbon price. The carbon dioxide () is captured directly from the ambient air; this is contrast to carbon capture and storage (CCS) which captures from point sources, such as a cement factory or a bioenergy plant. After the capture, DAC generates a concentrated stream of for sequestration or utilization or production of carbon-neutral fuel and windgas. Carbon dioxide removal is achieved when ambient air makes contact with chemical media, typically an aqueous alkaline solvent or sorbents. These chemical media are subsequently stripped of CO2 through the application of energy (namely heat), resulting in a CO2 stream that can undergo dehydration and compression, while simultaneously regenerating the chemical media for reuse. When combined with long-term storage of , DAC is known as direct air carbon capture and storage (DACCS or DACS). It would require renewable energies to power since approximately 400kJ of energy is needed per mole of CO2 capture. DACCS can act as a carbon dioxide removal mechanism (or a carbon negative technology), although it has yet to be profitable because the cost per tonne of carbon dioxide is several times the carbon price. DAC was suggested in 1999 and is still in development. Several commercial plants are planned or in operation in Europe and the US. Large-scale DAC deployment may be accelerated when connected with economical applications or policy incentives. In contrast to carbon capture and storage (CCS) which captures emissions from a point source such as a
https://en.wikipedia.org/wiki/Windisch%E2%80%93Kolbach%20unit
°WK or degrees Windisch-Kolbach is a unit for measuring the diastatic power of malt, named after the German brewer Wilhelm Windisch and the Luxembourg brewer Paul Kolbach. It is a common unit in beer brewing (especially in Europe) that measures the ability of enzymes in malt to reduce starch to sugar (maltose). It is defined as the amount of maltose formed by 100 g of malt in 30 min at 20 °C. Degrees Lintner is a unit used in the United States for the same purpose. The conversion is as follows: . 334 °WK = 3.014×10−7 Katal References W. Diemair: Analytik der Lebensmittel Nachweis und Bestimmung von Lebensmittel-Inhaltsstoffen, Springer Verlag, Berlin/Heidelberg/New York 1967, S. 255–256. Units of measurement Brewing
https://en.wikipedia.org/wiki/Martin%20Grohe
Martin Grohe (born 1967) is a German mathematician and computer scientist known for his research on parameterized complexity, mathematical logic, finite model theory, the logic of graphs, database theory, and descriptive complexity theory. He is a University Professor of Computer Science at RWTH Aachen University, where he holds the Chair for Logic and Theory of Discrete Systems. Education Grohe earned his doctorate (dr. rer. nat.) at the University of Freiburg in 1994. His dissertation, The Structure of Fixed-Point Logics, was supervised by Heinz-Dieter Ebbinghaus. After postdoctoral research at the University of California, Santa Cruz and Stanford University, he earned his habilitation at the University of Freiburg in 1998. Books Grohe is the author of Descriptive Complexity, Canonisation, and Definable Graph Structure Theory (Lecture Notes in Logic 47, Cambridge University Press, 2017). In 2011, Grohe and Johann A. Makowsky published as editors the 558th proceedings of the AMS-ASL special session on Model Theoretic Methods in Finite Combinatorics, which was held on January 5-8 2009 in Washington, DC. With Jörg Flum, he is the co-author of Parameterized Complexity Theory (Springer, 2006). Recognition Grohe won the Heinz Maier–Leibnitz Prize awarded by the German Research Foundation in 1999. He was elected as an ACM Fellow in 2018 for "contributions to logic in computer science, database theory, algorithms, and computational complexity". References External links 1967 births Living people German computer scientists 20th-century German mathematicians Mathematical logicians University of Freiburg alumni Academic staff of RWTH Aachen University 21st-century German mathematicians
https://en.wikipedia.org/wiki/Qualcomm%20Wi-Fi%20SON
Qualcomm Wi-Fi SON (Self-Organizing Network) is a solution developed by Qualcomm for Wi-Fi networks to simply and automatically select and link different wireless networking devices together, using the concept of "mesh networking". It is supposed to improve network coverage in different corners in a house or an apartment, and also provide improved security. Note that the technology behind the solution is more like mesh networking than cellular self-organizing network which would have included dynamical adjustment between different access points and client devices. References Qualcomm Wi-Fi Mesh networking
https://en.wikipedia.org/wiki/Compucolor%20II%20character%20set
Compucolor II is a character set developed by Compucolor Corporation for their Compucolor computers. These used a SMC CRT5027 video controller, a Japanese-produced version of the Texas Instruments TMS 9927, programmed to provide a screen format of 32 lines with 64 characters per line. Character set The following table shows the Compucolor II character set: � Not in Unicode, most are pieces designed to make 2x2 character large letters See also Compucolor Intelligent Systems Intecolor/Compucolor 8001 character set References Character sets
https://en.wikipedia.org/wiki/Airport%20City
Airport City is a free-to-play city-building simulation game developed and published by Game Insight. Airport City was first launched on the Facebook social media platform on September 1, 2011. The game was released for Google Play on February 11, 2012, and for iOS on August 16, 2012, followed by Amazon Appstore on October 26, 2012, and Microsoft Windows on April 3, 2014. The game features mobile cross-platform play, allowing players on any supported device to cooperate and compete with each other. As of September 2019, Game Insight has reported over 75 million Airport City registered players worldwide across all platforms. Gameplay Airport City tasks the player with expanding a small airport into a major one, while developing a nearby town to support the airport operations with passengers from residential buildings and taxes from commercial properties. As both city and airport grow, the space program becomes available that challenges the player to launch space missions either solo or together with other players. Another group-based activity is airline alliances that players can create to complete missions together with team members and to compete against other alliances. New content for the game involves the addition of new aircraft, buildings, destinations, and collectible items to bring back from those flights, taking the form of either expansion packs or limited time special events released roughly every month. Seasonal events either reflect real world events, such as Halloween, Christmas, and Easter, or fictitious, such as a UFO crash landing near the city or loosely based on the events from Adventures in Wonderland novel. Reception Critical response Pocket-lints Ian Morris called Airport City "the stupidest, most annoying game we've ever played on Android" that manages to be quite addictive with something to do all the time, and "perfecty possible, and enjoyable" to play for free despite its repetitive nature. Pete Davison writing for the Adweek nam
https://en.wikipedia.org/wiki/Crisper%20drawer
A crisper drawer (also known as a crisper) is a compartment within a refrigerator designed to prolong the freshness of stored produce. Crisper drawers have a different level of humidity from the rest of the refrigerator, optimizing freshness in fruits and vegetables. Some can be adjusted to either prevent the loss of moisture from produce, or to allow ethylene gas produced by certain fruits to escape in order to prevent them from rotting quickly. Etymology The first known use of the word "crisper" in relation to a crisper drawer was in 1835. Design and operation Crisper drawers operate by creating an environment of greater humidity than the rest of the refrigerator. Many crisper drawers have a separate humidity control which closes or opens a vent in the drawer. When the vent is in the closed position, airflow is shut off, creating greater humidity in the drawer. High humidity is optimal for the storage of leafy or thin-skinned vegetables. When the vent is in the open position, airflow keeps humidity in the crisper drawer low, which is beneficial for the storage of fruits. Additionally, because some fruits emit high levels of ethylene gas, the open vent allows the ethylene gas to escape, preventing these foods from rotting. The ability to separate low-humidity fruits from high-humidity vegetables using the different crisper drawers also prevents ethylene gas from damaging the latter. Crisper drawers which do not have a humidity control are, by default, high humidity crispers. Reported public confusion Appliance manufacturers have reported that many refrigerator owners are unaware of the purpose or operation of crisper drawers. A 2010 survey commissioned by Robert Bosch GmbH found that 55 percent of surveyed Americans "admit to not knowing how to use their crisper drawer controls". In the UK, sources often use the term "crisper drawer" in conjunction with a nearby explanation, like the Vegetable Expert advice website calling them "special compartments or 'crisper
https://en.wikipedia.org/wiki/Interim%20Register%20of%20Marine%20and%20Nonmarine%20Genera
The Interim Register of Marine and Nonmarine Genera (IRMNG) is a taxonomic database which attempts to cover published genus names for all domains of life from 1758 in zoology (1753 in botany) up to the present, arranged in a single, internally consistent taxonomic hierarchy, for the benefit of Biodiversity Informatics initiatives plus general users of biodiversity (taxonomic) information. In addition to containing over 490,000 published genus name instances as at March 2020 (also including subgeneric names in zoology), the database holds over 1.7 million species names (1.3 million listed as "accepted"), although this component of the data is not maintained in as current or complete state as the genus-level holdings. IRMNG can be queried online for access to the latest version of the dataset and is also made available as periodic snapshots or data dumps for import/upload into other systems as desired. Description IRMNG contains scientific names (only) of the genera, a subset of species, and principal higher ranks of most plants, animals and other kingdoms, both living and extinct, within a standardized taxonomic hierarchy, with associated machine-readable information on habitat (e.g. marine/nonmarine) and extant/fossil status for the majority of entries. The database aspires to provide complete coverage of both accepted and unaccepted genus names across all kingdoms, with a subset only of species names included as a secondary activity. In its March 2020 release, IRMNG contained 492,620 genus names, of which 232,093 were listed as "accepted", 121,389 "unaccepted", 7,462 of "other" status i.e. interim unpublished, nomen dubium, nomen nudum, taxon inquirendum or temporary name, and 131,676 as "uncertain" (unassessed for taxonomic status at this time). The data originate from a range of (frequently domain-specific) print, online and database sources, including (among others) Nomenclator Zoologicus for animals and Index Nominum Genericorum for plants, and are reorganised
https://en.wikipedia.org/wiki/Vera%20Fischer%20%28mathematician%29
Vera V. Fischer is an Austrian mathematician specializing in set theory, mathematical logic, and infinitary combinatorics. She is a privatdozent in the Kurt Gödel Research Center for Mathematical Logic at the University of Vienna. Education and career Fischer completed her doctorate in 2008 at York University in Canada. Her dissertation, The Consistency of Arbitrarily Large Spread between the Bounding and the Splitting Numbers, was supervised by Juris Steprāns. Before joining the Kurt Gödel Research Center, she worked at TU Wien from 2014 to 2015, where she led a project under the Lise Meitner Programme of the Austrian Science Fund. Recognition In 2017, Fischer won the Start-Preis of the Austrian Science Fund. In 2018, she won the Prize of the Austrian Mathematical Society. References External links Home page Year of birth missing (living people) Living people Austrian mathematicians Women mathematicians Mathematical logicians Women logicians Combinatorialists York University alumni Academic staff of TU Wien Academic staff of the University of Vienna
https://en.wikipedia.org/wiki/Dallol%20%28hydrothermal%20system%29
Dallol is a unique, terrestrial hydrothermal system around a cinder cone volcano in the Danakil Depression, northeast of the Erta Ale Range in Ethiopia. It is known for its unearthly colors and mineral patterns, and the very acidic fluids that discharge from its hydrothermal springs. Etymology The term Dallol was coined by the Afar people and means dissolution or disintegration, describing a landscape of green acid ponds and geysers (pH-values less than 1) and iron oxide, sulfur and salt desert plains. The area somewhat resembles the hot springs areas of Yellowstone National Park. Description Dallol mountain has an area of about , and rises about above the surrounding salt plains. A circular depression near the centre is probably a collapsed crater. The southwestern slopes have water-eroded salt canyons, pillars, and blocks. There are numerous saline springs and fields of small fumaroles. Numerous hot springs discharge brine and acidic liquid here. Small, widespread, temporary geysers produce cones of salt. The Dallol deposits include significant bodies of potash found directly at the surface. The yellow, ochre and brown colourings are the result of the presence of iron and other impurities. Older, inactive springs tend to be dark brown because of oxidation processes. Formation It was formed by the intrusion of basaltic magma into Miocene salt deposits and subsequent hydrothermal activity. Phreatic eruptions took place here in 1926, forming Dallol Volcano; numerous other eruption craters dot the salt flats nearby. These craters are the lowest known subaerial volcanic vents in the world, at or more below sea level. In October 2004 the shallow magma chamber beneath Dallol deflated and fed a magma intrusion southwards beneath the rift. The most recent signs of activity occurred in January 2011 in what may have been a degassing event from deep below the surface. Physical properties Dallol lies in the evaporitic plain of the Danakil Depression at the Afar Triangle
https://en.wikipedia.org/wiki/Agata%20Ciabattoni
Agata Ciabattoni is an Italian mathematical logician specializing in non-classical logic. She is a full professor at the Institute of Logic and Computation of the Faculty of Informatics at the Vienna University of Technology (TU Wien), and a co-chair of the Vienna Center for Logic and Algorithms of TU Wien (VCLA). Education and career Ciabattoni is originally from Ripatransone. She studied computer science at the University of Bologna, and completed her Ph.D. in 2000 at the University of Milan. Her dissertation, Proof-theory in many-valued logics, was supervised by Daniele Mundici. She moved to Vienna in 2000 with the support of an EU Marie Curie Fellowship, and In 2007, she earned her habilitation at TU Wien. She remains affiliated with TU Wien, as a professor in the faculty of informatics. She also serves as the Collegium Logicum lecture series chair for the Kurt Gödel Society. Contributions One of Ciabattoni's projects at TU Wien involves using mathematical logic to formalize the ethical reasoning in the Vedas, a body of Indian sacred texts. Recognition In 2011, Ciabattoni won the Start-Preis of the Austrian Science Fund, the only woman to win the prize that year. References External links Home page Year of birth missing (living people) Living people Austrian mathematicians Women mathematicians Mathematical logicians Women logicians University of Bologna alumni University of Milan alumni Academic staff of TU Wien
https://en.wikipedia.org/wiki/Human%20uses%20of%20bats
Human uses of bats include economic uses such as bushmeat or in traditional medicine. Bats are also used symbolically in religion, mythology, superstition, and the arts. Perceived medical uses of bats include treating epilepsy in South America, night blindness in China, rheumatism, asthma, chest pain, and fever in South Asia. Bat meat is consumed in Oceania, Australia, Asia, and Africa, with about 13% of all species hunted for food. Other economic uses of bats include using their teeth as currency on the island of Makira. Bats are widely represented in the arts, with inclusion in epic poems, plays, fables, and comic books. Though frequently associated with malevolence in Western art, bats are symbols of happiness in China. Economic uses Traditional medicine Live bats are sold in Bolivia for purported medicinal uses. Specifically, consuming the bats' blood is believed to treat epilepsy. A 2010 study documented that per month, 3,000 bats were sold in markets in four Bolivian cities. Species sold in these markets include Seba's short-tailed bats, mouse-eared bats, and common vampire bats. Bat excrement (guano) is used in traditional Chinese medicine as a treatment for night blindness. The Romans believed that bat blood was an antidote for snake venom. Flying foxes are killed for use in traditional medicine. The Indian flying fox, for example, has many perceived medical uses. Some believe that its fat is a treatment for rheumatism. Tribes in the Attappadi region of India eat the cooked flesh of the Indian flying fox to treat asthma and chest pain. Healers of the Kanda Tribe of Bangladesh use hair from Indian flying foxes to create treatments for "fever with shivering." Meat Bats are consumed for their meat in several regions, including Oceania, Australia, Southeast Asia, China, and West and Central Africa. Bats have been used as a food source for humans for thousands of years. At least 167 species of bats are hunted around the world, or about 13% of all bat speci
https://en.wikipedia.org/wiki/Eraser%20%28software%29
Eraser is an open-source secure file erasure tool available for the Windows operating system. It supports both file and volume wiping. Eraser securely erases data by overwriting it such that the data is irrecoverable. It supports a variety of data destruction standards, including British HMG IS5 (Infosec Standard 5), American DoD 5220.22-M, and the Gutmann method which features a 35-pass overwrite. The tool has been recommended in TechAdvisor, The Guardian, and PC World, and is a tool suggested by the United States government Computer Emergency Readiness Team. See also BleachBit CCleaner Data erasure Shred (Unix) References External links Official website Documentation Data erasure software 2003 software Utilities for Windows
https://en.wikipedia.org/wiki/Trigonometric%20Rosen%E2%80%93Morse%20potential
The trigonometric Rosen–Morse potential, named after the physicists Nathan Rosen and Philip M. Morse, is among the exactly solvable quantum mechanical potentials. Definition In dimensionless units and modulo additive constants, it is defined as where is a relative distance, is an angle rescaling parameter, and is so far a matching length parameter. Another parametrization of same potential is which is the trigonometric version of a one-dimensional hyperbolic potential introduced in molecular physics by Nathan Rosen and Philip M. Morse and given by, a parallelism that explains the potential's name. The most prominent application concerns the parametrization, with non-negative integer, and is due to Schrödinger who intended to formulate the hydrogen atom problem on Albert Einstein's closed universe, , the direct product of a time line with a three-dimensional closed space of positive constant curvature, the hypersphere , and introduced it on this geometry in his celebrated equation as the counterpart to the Coulomb potential, a mathematical problem briefly highlighted below. The case: Four-dimensional rigid rotator in inertial quantum motion on the three dimensional hypersphere The hypersphere is a surface in a four-dimensional Euclidean space, , and is defined as, where , , , and are the Cartesian coordinates of a vector in , and is termed to as hyper-radius. Correspondingly, Laplace operator in is given by, In now switching to polar coordinates, one finds the Laplace operator expressed as Here, stands for the squared angular momentum operator in four dimensions, while is the standard three-dimensional squared angular momentum operator. Considering now the hyper-spherical radius as a constant, one encounters the Laplace-Beltrami operator on as With that the free wave equation on takes the form The solutions, , to this equation are the so-called four-dimensional hyper-spherical harmonics defined as where are the Gegenbauer polynomials
https://en.wikipedia.org/wiki/KeepSolid
KeepSolid VPN Unlimited is a personal virtual private network software product available for iOS, macOS, Android, Windows, and Linux. VPN Unlimited is developed by KeepSolid Inc, an American company with headquarters in New York City, the United States. Founded in 2013 in New York, KeepSolid provides cybersecurity software. VPN Unlimited has 500 servers about 80 locations around the world and operates under the jurisdiction of the United States. History KeepSolid was established in 2013 as a software development company and launched VPN Unlimited on iOS with two servers. In May 2016, VPN Unlimited received the Editor's Choice Award by PCMag. In November 2018, VPN Unlimited released a lite version of freemium apps for Android and iOS. In 2019, VPN Unlimited added an open-source WireGuard protocol to the list of available VPN protocols, in addition to IKEv2 and OpenVPN. As of 2021, VPN Unlimited operates 500 servers in more than 80 locations worldwide. Technology VPN Unlimited has desktop applications for Windows, macOS, and Linux, and mobile apps for iOS and Android. The app is also available as an extension for Mozilla Firefox, Opera, Edge, and Google Chrome browsers. VPN Unlimited supports protocols such as IKEv2, L2TP, OpenVPN, PPTP, and a proprietary protocol called KeepSolid Wise. The technology routes the VPN traffic via TCP 443 and UDP 33434 ports, which makes the traffic harder to detect and block. Additional features Besides general-use VPN servers, the provider offers servers for specific purposes, including personal VPN servers, P2P sharing, and torrenting. Some additional features include, kill switch which disconnects the connection in case of any failure and privacy feature like no-log policy. VPN Unlimited offers four subscription plans: monthly, yearly, tri-yearly, and lifetime. Servers KeepSolid VPN Unlimited maintains over 500 servers in more than 80 locations in different countries, including Australia, Austria, Belgium, Bulgaria, B