text
stringlengths
26
3.6k
page_title
stringlengths
1
71
source
stringclasses
1 value
token_count
int64
10
512
id
stringlengths
2
8
url
stringlengths
31
117
topic
stringclasses
4 values
section
stringlengths
4
49
sublist
stringclasses
9 values
Other species are important in commercial microbial fermentations. For example, alcoholic beverages such as Japanese sake are often made from rice or other starchy ingredients (like manioc), rather than from grapes or malted barley. Typical microorganisms used to make alcohol, such as yeasts of the genus Saccharomyces, cannot ferment these starches. Therefore, koji mold such as Aspergillus oryzae is used to first break down the starches into simpler sugars. Members of the genus are also sources of natural products that can be used in the development of medications to treat human disease. Aspergillus spp. are known to produce anthraquinone which has commercial importance due to its antibacterial and antifungal properties. Perhaps the largest application of Aspergillus niger is as the major source of citric acid; this organism accounts for over 99% of global citric acid production, or more than 1.4 million tonnes (>1.5 million US tons) per year. A. niger is also commonly used for the production of native and foreign enzymes, including glucose oxidase, lysozyme, and lactase. In these instances, the culture is rarely grown on a solid substrate, although this is still common practice in Japan, but is more often grown as a submerged culture in a bioreactor. In this way, the most important parameters can be strictly controlled, and maximal productivity can be achieved. This process also makes it far easier to separate the chemical or enzyme of importance from the medium, and is therefore far more cost-effective. Research
Aspergillus
Wikipedia
335
1529518
https://en.wikipedia.org/wiki/Aspergillus
Biology and health sciences
Basics
Plants
A. nidulans (Emericella nidulans) has been used as a research organism for many years and was used by Guido Pontecorvo to demonstrate parasexuality in fungi. Recently, A. nidulans was one of the pioneering organisms to have its genome sequenced by researchers at the Broad Institute. As of 2008, a further seven Aspergillus species have had their genomes sequenced: the industrially useful A. niger (two strains), A. oryzae, and A. terreus, and the pathogens A. clavatus, A. fischerianus (Neosartorya fischeri), A. flavus, and A. fumigatus (two strains). A. fischerianus is hardly ever pathogenic, but is very closely related to the common pathogen A. fumigatus; it was sequenced in part to better understand A. fumigatus pathogenicity. Sexual reproduction Of the 250 species of aspergilli, about 64% have no known sexual state. However, many of these species likely have an as yet unidentified sexual stage. Sexual reproduction occurs in two fundamentally different ways in fungi. These are outcrossing (in heterothallic fungi) in which two different individuals contribute nuclei, and self-fertilization or selfing (in homothallic fungi) in which both nuclei are derived from the same individual. In recent years, sexual cycles have been discovered in numerous species previously thought to be asexual. These discoveries reflect recent experimental focus on species of particular relevance to humans. A. fumigatus is the most common species to cause disease in immunodeficient humans. In 2009, A. fumigatus was shown to have a heterothallic, fully functional sexual cycle. Isolates of complementary mating types are required for sex to occur. A. flavus is the major producer of carcinogenic aflatoxins in crops worldwide. It is also an opportunistic human and animal pathogen, causing aspergillosis in immunocompromised individuals. In 2009, a sexual state of this heterothallic fungus was found to arise when strains of opposite mating types were cultured together under appropriate conditions.
Aspergillus
Wikipedia
472
1529518
https://en.wikipedia.org/wiki/Aspergillus
Biology and health sciences
Basics
Plants
A. lentulus is an opportunistic human pathogen that causes invasive aspergillosis with high mortality rates. In 2013, A. lentulus was found to have a heterothallic functional sexual breeding system. A. terreus is commonly used in industry to produce important organic acids and enzymes, and was the initial source for the cholesterol-lowering drug lovastatin. In 2013, A. terreus was found to be capable of sexual reproduction when strains of opposite mating types were crossed under appropriate culture conditions. These findings with Aspergillus species are consistent with accumulating evidence, from studies of other eukaryotic species, that sex was likely present in the common ancestor of all eukaryotes. A. nidulans, a homothallic fungus, is capable of self-fertilization. Selfing involves activation of the same mating pathways characteristic of sex in outcrossing species, i.e. self-fertilization does not bypass required pathways for outcrossing sex, but instead requires activation of these pathways within a single individual. Among those Aspergillus species that exhibit a sexual cycle, the overwhelming majority in nature are homothallic (self-fertilizing). This observation suggests Aspergillus species can generally maintain sex though little genetic variability is produced by homothallic self-fertilization. A. fumigatus, a heterothallic (outcrossing) fungus that occurs in areas with widely different climates and environments, also displays little genetic variability either within geographic regions or on a global scale, again suggesting sex, in this case outcrossing sex, can be maintained even when little genetic variability is produced.
Aspergillus
Wikipedia
348
1529518
https://en.wikipedia.org/wiki/Aspergillus
Biology and health sciences
Basics
Plants
Genomics The simultaneous publication of three Aspergillus genome manuscripts in Nature in December 2005 established the genus as the leading filamentous fungal genus for comparative genomic studies. Like most major genome projects, these efforts were collaborations between a large sequencing centre and the respective community of scientists. For example, the Institute for Genome Research (TIGR) worked with the A. fumigatus community. A. nidulans was sequenced at the Broad Institute. A. oryzae was sequenced in Japan at the National Institute of Advanced Industrial Science and Technology. The Joint Genome Institute of the Department of Energy has released sequence data for a citric acid-producing strain of A. niger. TIGR, now renamed the J. Craig Venter Institute, is currently spearheading a project on the A. flavus genome. Aspergillus is characterized by high levels of genetic diversity and, using protostome divergence as a scale, is as diverse as the Vertebrates phylum although both inter and intra-specific genome structure is relatively plastic. The genomes of some Aspergillus species, such as A. flavus and A. oryzae, are more rich and around 20% larger than others, such as A. nidulans and A. fumigatus. Several mechanisms could explain this difference, although the combination of segmental duplication, genome duplication, and horizontal gene transfer acting in a piecemeal fashion is well-supported. Genome sizes for sequenced species of Aspergillus range from about 29.3 Mb for A. fumigatus to 37.1 Mb for A. oryzae, while the numbers of predicted genes vary from about 9926 for A. fumigatus to about 12,071 for A. oryzae. The genome size of an enzyme-producing strain of A. niger is of intermediate size at 33.9 Mb.
Aspergillus
Wikipedia
402
1529518
https://en.wikipedia.org/wiki/Aspergillus
Biology and health sciences
Basics
Plants
Pathogens Some Aspergillus species cause serious disease in humans and animals. The most common pathogenic species are A. fumigatus and A. flavus, which produces aflatoxin which is both a toxin and a carcinogen, and which can contaminate foods such as nuts. The most common species causing allergic disease are A. fumigatus and A. clavatus. Other species are important as agricultural pathogens. Aspergillus spp. cause disease on many grain crops, especially maize, and some variants synthesize mycotoxins, including aflatoxin. Aspergillus can cause neonatal infections. A. fumigatus (the most common species) infections are primary pulmonary infections and can potentially become a rapidly necrotizing pneumonia with a potential to disseminate. The organism can be differentiated from other common mold infections based on the fact that it takes on a mold form both in the environment and in the host (unlike Candida albicans which is a dimorphic mold in the environment and a yeast in the body). Aspergillosis Aspergillosis is the group of diseases caused by Aspergillus. The most common species among paranasal sinus infections associated with aspergillosis is A. fumigatus. The symptoms include fever, cough, chest pain, or breathlessness, which also occur in many other illnesses, so diagnosis can be difficult. Usually, only patients with already weakened immune systems or who suffer other lung conditions are susceptible. In humans, the major forms of disease are: Acute invasive aspergillosis, a form that grows into surrounding tissue, more common in those with weakened immune systems such as AIDS or chemotherapy patients Allergic bronchopulmonary aspergillosis, which affects patients with respiratory diseases such as asthma, cystic fibrosis, and sinusitis Aspergilloma, a "fungus ball" that can form within cavities such as the lung Disseminated invasive aspergillosis, an infection spread widely through the body Fungal infections from Aspergillus spores remain one theory of sickness and untimely death of some early Egyptologists and tomb explorers. Ancient spores which grew on the remains of food offerings and mummies sealed in tombs and chambers may have been blown around and inhaled by the excavators, ultimately linked to the notion of the curse of the pharaohs.
Aspergillus
Wikipedia
494
1529518
https://en.wikipedia.org/wiki/Aspergillus
Biology and health sciences
Basics
Plants
Aspergillosis of the air passages is also frequently reported in birds, and certain species of Aspergillus have been known to infect insects. Most people inhale Aspergillus into their lungs everyday, but generally only the immuno-compromised become sick with aspergillosis.
Aspergillus
Wikipedia
61
1529518
https://en.wikipedia.org/wiki/Aspergillus
Biology and health sciences
Basics
Plants
The caddisflies (order Trichoptera) are a group of insects with aquatic larvae and terrestrial adults. There are approximately 14,500 described species, most of which can be divided into the suborders Integripalpia and Annulipalpia on the basis of the adult mouthparts. Integripalpian larvae construct a portable casing to protect themselves as they move around looking for food, while annulipalpian larvae make themselves a fixed retreat in which they remain, waiting for food to come to them. The affinities of the small third suborder Spicipalpia are unclear, and molecular analysis suggests it may not be monophyletic. Also called sedge-flies or rail-flies, the adults are small moth-like insects with two pairs of hairy membranous wings. They are closely related to the Lepidoptera (moths and butterflies) which have scales on their wings; the two orders together form the superorder Amphiesmenoptera. The aquatic larvae are found in a wide variety of habitats such as streams, rivers, lakes, ponds, spring seeps and temporary waters (vernal pools), and even the ocean. The larvae of many species use silk to make protective cases, which are often strengthened with gravel, sand, twigs, bitten-off pieces of plants, or other debris. The larvae exhibit various feeding strategies, with different species being predators, leaf shredders, algal grazers, or collectors of particles from the water column and benthos. Most adults have short lives during which they do not feed. In fly fishing, artificial flies called dry flies are tied to imitate adults, while larvae and pupae are imitated with artificial flies called wet flies or nymphs. It is also possible to use them as bait, though this is not as common as artificial flies and is known as bait fishing. Common and widespread genera such as Helicopsyche and Hydropsyche are important in the sport, where caddisflies are known as "sedges". Caddisflies are useful as bioindicators, as they are sensitive to water pollution and are large enough to be assessed in the field. In art, the French artist Hubert Duprat has created works by providing caddis larvae with small grains of gold and precious stones for them to build into decorative cases.
Caddisfly
Wikipedia
485
1529533
https://en.wikipedia.org/wiki/Caddisfly
Biology and health sciences
Insects: General
Animals
Etymology The name of the order "Trichoptera" derives from the Greek: (, "hair"), genitive trichos + (, "wing"), and refers to the fact that the wings of these insects are bristly. The origin of the word "caddis" is unclear, but it dates back to at least as far as Izaak Walton's 1653 book The Compleat Angler, where "cod-worms or caddis" were mentioned as being used as bait. The term cadyss was being used in the fifteenth century for silk or cotton cloth, and "cadice-men" were itinerant vendors of such materials, but a connection between these words and the insects has not been established. Evolution and phylogeny Fossil history Fossil caddisflies have been found in rocks dating back to the Triassic. The largest numbers of fossilised remains are those of larval cases, which are made of durable materials that preserve well. Body fossils of caddisflies are extremely rare, the oldest being from the Early and Middle Triassic, some 230 million years ago, and wings are another source of fossils. The evolution of the group to one with fully aquatic larvae seems to have taken place sometime during the Triassic. The finding of fossils resembling caddisfly larval cases in marine deposits in Brazil may push back the origins of the order to the Early Permian period. Evolution Nearly all adult caddisflies are terrestrial, but their larvae and pupae are aquatic. They share this characteristic with several distantly-related groups, namely the dragonflies, mayflies, stoneflies, alderflies and lacewings. The ancestors of all these groups were terrestrial, with open tracheal systems, convergently evolving different types of gills for their aquatic larvae as they took to the water to avoid predation. Caddisflies was the only group of these insects to use silk as part of their lifestyle, which has been a contributing factor to their success and why they are the most species-rich order of aquatic insects.
Caddisfly
Wikipedia
418
1529533
https://en.wikipedia.org/wiki/Caddisfly
Biology and health sciences
Insects: General
Animals
About 14,500 species of caddisfly in 45 families have been recognised worldwide, but many more species remain to be described. Most can be divided into the suborders Integripalpia and Annulipalpia on the basis of the adult mouthparts. The characteristics of adults depend on the palps, wing venation and genitalia of both sexes. The latter two characters have undergone such extensive differentiation among the different superfamilies that the differences between the suborders is not clear-cut. The larvae of Annulipalpians are campodeiform (free-living, well sclerotized, long legged predators with dorso-ventrally flattened bodies and protruding mouthparts). The larvae of Integripalpians are polypod (poorly sclerotized detritivores, with abdominal prolegs in addition to thoracic legs, living permanently in tight-fitting cases). The affinities of the third suborder, Spicipalpia, are unclear; the larvae are free-living with no cases, instead creating net-like traps from silk. Phylogeny The cladogram of external relationships, based on molecular analysis, shows the order as a clade, sister to the Lepidoptera, and more distantly related to the Diptera (true flies) and Mecoptera (scorpionflies). The cladogram of relationships within the order is based on a 2002 molecular phylogeny using ribosomal RNA, a nuclear elongation factor gene, and mitochondrial cytochrome oxidase. The Annulipalpia and Integripalpia are clades, but the relationships within the Spicipalpia are unclear. Distribution Caddisflies are found worldwide, with the greater diversity being in warmer regions. They are associated with bodies of freshwater, the larvae being found in lakes, ponds, rivers, streams and other water bodies. The land caddis, Enoicyla pusilla (family: Limnephilidae), lives in the damp litter of the woodland floor. In the United Kingdom it is found in and around the county of Worcestershire in oakwoods. Ecology
Caddisfly
Wikipedia
441
1529533
https://en.wikipedia.org/wiki/Caddisfly
Biology and health sciences
Insects: General
Animals
Caddisfly larvae can be found in all feeding guilds in freshwater habitats. Most early stage larvae and some late stage ones are collector-gatherers, picking up fragments of organic matter from the benthos. Other species are collector-filterers, sieving organic particles from the water using silken nets, or hairs on their legs. Some species are scrapers, feeding on the film of algae and other periphyton that grows on underwater objects in sunlight. Others are shredder-herbivores, chewing fragments off living plant material while others are shredder-detritivores, gnawing at rotting wood or chewing dead leaves that have been pre-processed by bacteria and fungi; most of the nutrients of the latter group come from consumption of the bacteria and fungi. The predatory species either actively hunt their prey, typically other insects, tiny crustaceans and worms, or lie in wait for unwary invertebrates to come too close. A few species feed opportunistically on dead animals or fish, and some Leptoceridae larvae feed on freshwater sponges. One such opportunistic species is Gumaga nigricula (family: Sericostomatidae) which has been observed scavenging fish carcasses and even bits of deer flesh. This particular family of caddisflies is typically classified among the shredders, suggesting caution when classifying macroinvertebrates into strict ecological functional groups, as some may shift their diets opportunistically. Like mayflies, stoneflies and dragonflies, but to a somewhat lesser extent, caddisflies are an indicator of good water quality; they die out of streams with polluted waters. They are an important part of the food web, both larvae and adults being eaten by many fish. The newly hatched adult is particularly vulnerable as it struggles to the surface after emerging from the submerged pupa, and as it dries its wings. The fish find these new adults easy pickings, and fishing flies resembling them can be successful for anglers at the right time of year.
Caddisfly
Wikipedia
425
1529533
https://en.wikipedia.org/wiki/Caddisfly
Biology and health sciences
Insects: General
Animals
The adult stage of a caddisfly may only survive for a few weeks; many species do not feed as adults and die soon after breeding, but some species are known to feed on nectar. The winged insects are nocturnal and provide food for night-flying birds, bats, small mammals, amphibians and arthropods. The larval stage lasts much longer, often for one or more years, and has a bigger impact on the environment. They form an important part of the diet of fish such as the trout. The fish acquire them by two means, either plucking them off vegetation or the stream-bed as the larvae move about, or during the daily behavioural drift; this drift happens during the night for many species of aquatic larvae, or around midday for some cased caddisfly species, and may result from population pressures or be a dispersal device. The larvae may drift in great numbers either close to the bottom, in mid-water or just below the surface. The fish swallow them whole, case and all. Underwater structures Cases Caddisflies are best known for the portable cases created by their larvae. About thirty families of caddisfly, members of the suborder Integripalpia, adopt this stratagem. These larvae eat detritus, largely decaying vegetable material, and the dead leaf fragments on which they feed tend to accumulate in hollows, in slow-moving sections of streams and behind stones and tree roots. The cases provide protection to the larvae as they make their way between these resources. The case is a tubular structure made of silk, secreted from salivary glands near the mouth of the larva, and is started soon after the egg hatches. Various reinforcements may be incorporated into its structure, the nature of the materials and design depending on the larva's genetic makeup; this means that caddisfly larvae can be recognised by their cases down to family, and even genus level. The materials used include grains of sand, larger fragments of rock, bark, sticks, leaves, seeds and mollusc shells. These are neatly arranged and stuck onto the outer surface of the silken tube. As the larva grows, more material is added at the front, and the larva can turn round in the tube and trim the rear end so that it does not drag along the substrate.
Caddisfly
Wikipedia
476
1529533
https://en.wikipedia.org/wiki/Caddisfly
Biology and health sciences
Insects: General
Animals
Caddisfly cases are open at both ends, the larvae drawing oxygenated water through the posterior end, over their gills, and pumping it out of the wider, anterior end. The larvae move around inside the tubes and this helps maintain the water current; the lower the oxygen content of the water, the more active the larvae need to be. This mechanism enable caddisfly larvae to live in waters too low in oxygen content to support stonefly and mayfly larvae. Fixed retreats In contrast to larvae that have portable cases, members of the Annulipalpia have a completely different feeding strategy. They make fixed retreats in which they remain stationary, waiting for food to come to them. Members of the Psychomyiidae, Ecnomidae and Xiphocentronidae families construct simple tubes of sand and other particles held together by silk and anchored to the bottom, and feed on the accumulations of silt formed when suspended material is deposited. The tube can be lengthened when the growing larva needs to feed in new areas. More complex tubes, short and flattened, are built by Polycentropodidae larvae in hollows in rocks or other submerged objects, sometimes with strands of silk suspended across the nearby surface. These larvae are carnivorous, resembling spiders in their feeding habits and rushing out of their retreat to attack any unwary small prey crawling across the surface. Silk domes Larvae of members of the family Glossosomatidae in the suborder Spicipalpia create dome-shaped enclosures of silk which enables them to graze on the periphyton, the biological film that grows on stones and other objects, while carrying their enclosure around like turtles. In the family Philopotamidae, the nets are sac-like, with intricate structure and tiny mesh. The larvae have specialised mouthparts to scrape off the microflora that get trapped in the net as water flows through. Nets The larvae of other species of caddisfly make nets rather than cases. These are silken webs stretching between aquatic vegetation and over stones. These net-making larvae usually live in running water, different species occupying different habitats with varying water speeds. There is a constant drift of invertebrates washed downstream by the current, and these animals, and bits of debris, accumulate in the nets which serve both as food traps and as retreats. Development and morphology
Caddisfly
Wikipedia
478
1529533
https://en.wikipedia.org/wiki/Caddisfly
Biology and health sciences
Insects: General
Animals
Caddisfly larvae are aquatic, with six pairs of tracheal gills on the underside of the abdomen. The eggs are laid above water on emergent twigs or vegetation or on the water surface although females of some species enter water to choose sites. Although most species lay eggs, a few in the genus Triplectides are ovoviviparous. Some species lay eggs on land and although most are associated with freshwater, a few like Symphitoneuria are found in coastal saline water. Philanisus plebeius females lay their eggs into the coelomic cavity of intertidal starfish. The larvae are long and roughly cylindrical, very similar to those of lepidoptera but lacking prolegs. In case-bearing species, the heads are heavily sclerotised while the abdomen is soft; the antennae are short and the mouthparts adapted for biting. Each of the usually ten abdominal segments bears a pair of legs with a single tarsal joint. In case-bearing species, the first segment bears three papillae, one above and two at the sides, which anchor the larva centrally in the tube. The posterior segment bears a pair of hooks for grappling. There are five to seven larval instars, followed by an aquatic pupa which has functional mandibles (to cut through the case), gills, and swimming legs. The pupal cocoon is spun from silk, but like the larval case, often has other materials attached. When pupating, species that build portable cases attach them to some underwater object, seal the front and back apertures against predators while still allowing water to flow through, and pupate within it. Once fully developed, most pupal caddisflies cut through their cases with a special pair of mandibles, swim up to the water surface, moult using the exuviae as a floating platform, and emerge as fully formed adults. They can often fly immediately after breaking from their pupal cuticle. Emergence is mainly univoltine (once per year) with all the adults of a species emerging at the same time. Development is within a year in warm places, but takes over a year in high latitudes and at high elevation in mountain lakes and streams.
Caddisfly
Wikipedia
461
1529533
https://en.wikipedia.org/wiki/Caddisfly
Biology and health sciences
Insects: General
Animals
The adult caddisfly is a medium-sized insect with membranous, hairy wings, which are held in a tent-wise fashion when the insect is at rest. The antennae are fairly long and threadlike, the mouthparts are reduced in size and the legs have five tarsi (lower leg joints). Adults are nocturnal and are attracted to light. Some species are strong fliers and can disperse to new localities, but many fly only weakly. Adults are usually short-lived, most being non-feeders and equipped only to breed. Once mated, the female caddisfly lays eggs in a gelatinous mass, attaching them above or below the water surface depending on species. The eggs hatch in a few weeks. Relationship with humans In angling Adult caddisflies are called sedges by anglers. Individual species emerge en masse at different times, and are used one after the other, often for only a few days each year, as models for artificial fishing flies for fly fishing in trout streams. A mass emergence is known as a hatch. Each type has its own angling name, so for example Mystacides is the dancer; Sericostoma the caperer; Leptocerus the silverhorn; Phryganea the murragh or great red sedge; Brachycentrus subnubilis the grannom; Lepidostoma the silver sedge; Oecetis the longhorn sedge; Cheumatopsyche the little sister sedge; Helicopsyche the speckled Peter, an important fishing fly in North America; and Hydropsyche the specked sedge, perhaps the most important caddisfly genus for anglers with over 50 species of net-makers. As bioindicators Caddisflies are useful as bioindicators (of good water quality), since they are sensitive to water pollution, and are large enough to be assessed conveniently in the field. Some species indicate undisturbed habitat, and some indicate degraded habitat. Although caddisflies may be found in waterbodies of varying qualities, species-rich caddisfly assemblages are generally thought to indicate clean water bodies, such as lakes, ponds, and marshes. Together with stoneflies and mayflies, caddisflies feature importantly in bioassessment surveys of streams and other water bodies.
Caddisfly
Wikipedia
483
1529533
https://en.wikipedia.org/wiki/Caddisfly
Biology and health sciences
Insects: General
Animals
In art While caddisflies in the wild construct their cases out of twigs, sand, aquatic plants, and rocks, the French artist Hubert Duprat makes art by providing wild caddisflies with precious stones and other materials. He collected caddisfly larvae from the wild and put them in climate-controlled tanks. He removes the larvae from their original cases and adds precious and semi-precious items such as grains of gold into the tank. The larvae then build new cases out of precious items, creating a unique form of artwork. The resulting works are sold across the world. As food In Japan the larvae of Stenopsyche marmorata are eaten as a delicacy called Zazamushi. Taxonomy There are roughly 16,266 extant species in 618 genera and 51 families worldwide.
Caddisfly
Wikipedia
161
1529533
https://en.wikipedia.org/wiki/Caddisfly
Biology and health sciences
Insects: General
Animals
Suborder Annulipalpia Superfamily Hydropsychoidea Family Hydropsychidae Superfamily Psychomyioidea Family Dipseudopsidae Family Ecnomidae Family †Electralbertidae Family Polycentropodidae Family Psychomyiidae Family Xiphocentronidae Superfamily Philopotamoidea Family Philopotamidae Family Stenopsychidae Suborder Integripalpia Superfamily Leptoceroidea Family Atriplectididae Family Calamoceratidae Family Molannidae Family Leptoceridae Family Limnocentropodidae Family Odontoceridae Family Philorheithridae Superfamily Limnephiloidea Family Apataniidae Family Brachycentridae Family Goeridae Family Limnephilidae Family Lepidostomatidae Family Oeconesidae Family Pisuliidae Family Rossianidae Family †Taymyrelectronidae Family Uenoidae Superfamily †Necrotaulioidea Family †Necrotauliidae Superfamily Phyrganeoidea Family †Baissoferidae Family †Dysoneuridae Family †Kalophryganeidae Family Phryganeidae Family Phryganopsychidae Family Plectrotarsidae Superfamily Sericostomatoidea Family Anomalopsychidae Family Antipodoeciidae Family Barbarochthonidae Family Beraeidae Family Calocidae Family Chathamiidae Family Conoesucidae Family Helicophidae Family Helicopsychidae Family Hydrosalpingidae Family Kokiriidae Family Petrothrincidae Family Sericostomatidae Superfamily Tasimioidea Family Tasimiidae Superfamily †Vitimotaulioidea Family †Vitimotauliidae Family †Cladochoristidae Family †Microptysmatidae Family †Prosepididontidae Family †Protomeropidae Family †Uraloptysmatidae Suborder Spicipalpia Superfamily Hydroptiloidea Family Glossosomatidae Family Hydroptilidae Family Ptilocolepidae Superfamily Rhyacophiloidea Family Hydrobiosidae Family Rhyacophilidae
Caddisfly
Wikipedia
437
1529533
https://en.wikipedia.org/wiki/Caddisfly
Biology and health sciences
Insects: General
Animals
A conidium ( ; : conidia), sometimes termed an asexual chlamydospore or chlamydoconidium (: chlamydoconidia), is an asexual, non-motile spore of a fungus. The word conidium comes from the Ancient Greek word for dust, (). They are also called mitospores due to the way they are generated through the cellular process of mitosis. They are produced exogenously. The two new haploid cells are genetically identical to the haploid parent, and can develop into new organisms if conditions are favorable, and serve in biological dispersal. Asexual reproduction in ascomycetes (the phylum Ascomycota) is by the formation of conidia, which are borne on specialized stalks called conidiophores. The morphology of these specialized conidiophores is often distinctive between species and, before the development of molecular techniques at the end of the 20th century, was widely used for identification of (e.g. Metarhizium) species. The terms microconidia and macroconidia are sometimes used. Conidiogenesis There are two main types of conidium development: Blastic conidiogenesis, where the spore is already evident before it separates from the conidiogenic hypha which is giving rise to it, and Thallic conidiogenesis, where first a cross-wall appears and thus the created cell develops into a spore. Conidia germination A conidium may form germ tubes (germination tubes) and/or conidial anastomosis tubes (CATs) in specific conditions. These two are some of the specialized hyphae that are formed by fungal conidia. The germ tubes will grow to form the hyphae and fungal mycelia. The conidial anastomosis tubes are morphologically and physiologically distinct from germ tubes. After conidia are induced to form conidial anastomosis tubes, they grow homing toward each other, and they fuse. Once fusion happens, the nuclei can pass through fused CATs. These are events of fungal vegetative growth and not sexual reproduction. Fusion between these cells seems to be important for some fungi during early stages of colony establishment. The production of these cells has been suggested to occur in 73 different species of fungi.
Conidium
Wikipedia
493
1530205
https://en.wikipedia.org/wiki/Conidium
Biology and health sciences
Fungal morphology and anatomy
Biology
Germination in Aspergillus As evidenced by recent literature, conidia germination of Aspergillus, a common mold, specifically is of interest. Aspergillus is not only a familiar fungus found across various different settings in the world, but it poses a danger for immunocompromised individuals, as inhaled Aspergillus conidia could germinate inside the respiratory tract and cause aspergillosis, a form of pulmonary infection, and continual developments of aspergillosis such as new risk groups and the resistance against antifungal drugs. Stages of Germination: Dormancy Germination in Aspergillus follows a sequence of three different stages: dormancy, isotropic growth, and polarized growth. The dormant conidia are able to germinate even after an year of remaining at room temperature, due to their resilient intracellular and extracellular characteristics, which enable them to undergo harsh conditions like dehydration, variation in osmotic pressure, oxidation, and temperature, and change in UV exposure and acidity levels. These abilities of the dormant conidia are dictated by a few central regulatory proteins, which are the main drivers of the conidia and conidiophore formation. One of these proteins, the developmental regulatory protein wetA, has been found to be particularly essential; in wetA-defective mutants have reduced tolerance to external factors mentioned above, and exhibit weak synthesization of the conidial cell wall. In addition to these central regulators, some notable groups of genes/proteins include other regulatory proteins like the velvet regulator proteins, which contribute to fungal growth, and other molecules that target specific unfavorable intra and extracellular conditions, like heat shock proteins.
Conidium
Wikipedia
356
1530205
https://en.wikipedia.org/wiki/Conidium
Biology and health sciences
Fungal morphology and anatomy
Biology
Stages of Germination: Isotropic and Polarized Growth The phases following dormancy include isotropic growth, in which increased intracellular osmotic pressure and water uptake causes swelling of the conidia and increased cellular diameter, and polarized growth, in which the swelling from isotropic growth directs the growth to one side of the cell, and leads to the formation of a germ tube. First, however, the conidia must go through the stage of breaking dormancy. In some species of Aspergillus, dormancy is broken when the dormant conidia is introduced to a carbon source in the presence of water and air, while in other species, the mere presence of glucose is enough to trigger it. The dense outer layer of the dormant conidia is shed and the growth of the hyphae cells begins, which has a significantly different composition compared to the dormant conidia cell. Breaking of dormancy involves transcription, but not translation; protein synthesis inhibitors prevent isotropic growth, while DNA and RNA synthesis inhibitors do not, and the start of breaking of dormancy is accompanied by and increase in transcripts for genes for biosynthesis of proteins, and immediate protein synthesis. Following the expansion of the cell via isotropic growth, studies have observed many new proteins emerging from the processes in the breaking of dormancy and transcripts associated with remodeling of the cell wall, suggesting that remodeling of the cell wall is a central process during isotropic growth. In the polarized growth stage, upregulated and overexpressed proteins and transcripts included ones involved in synthesis of chitin (a major component of the fungal cell wall), mitosis and DNA processing, remodeling of cell morphology, and ones in germ tube formation pertaining to infection and virulence factors. Structures for release of conidia Conidiogenesis is an important mechanism of spread of plant pathogens. In some cases, specialized macroscopic fruiting structures perhaps 1 mm or so in diameter containing masses of conidia are formed under the skin of the host plant and then erupt through the surface, allowing the spores to be distributed by wind and rain. One of these structures is called a conidioma (plural: conidiomata).
Conidium
Wikipedia
464
1530205
https://en.wikipedia.org/wiki/Conidium
Biology and health sciences
Fungal morphology and anatomy
Biology
Two important types of conidiomata, distinguished by their form, are: pycnidia (singular: pycnidium), which are flask-shaped, and acervuli (singular: acervulus), which have a simpler cushion-like form. Pycnidial conidiomata or pycnidia form in the fungal tissue itself, and are shaped like a bulging vase. The conidia are released through a small opening at the apex, the ostiole. Acervular conidiomata, or acervuli, are cushion-like structures that form within the tissues of a host organism: subcuticular, lying under the outer layer of the plant (the cuticle), intraepidermal, inside the outer cell layer (the epidermis), subepidermal, under the epidermis, or deeper inside the host. Mostly they develop a flat layer of relatively short conidiophores which then produce masses of spores. The increasing pressure leads to the splitting of the epidermis and cuticle and allows release of the conidia from the tissue. Health issues Conidia are always present in the air, but levels fluctuate from day to day and with the seasons. An average person inhales at least 40 conidia per hour. Exposure to conidia from certain species, such as those of Cryptostroma corticale, is known to cause hypersensitivity pneumonitis, an occupational hazard for forest workers and paper mill employees. Conidia are often the method by which some normally harmless but heat-tolerating (thermotolerant), common fungi establish infection in certain types of severely immunocompromised patients (usually acute leukemia patients on induction chemotherapy, AIDS patients with superimposed B-cell lymphoma, bone marrow transplantation patients (taking immunosuppressants), or major organ transplant patients with graft versus host disease). Their immune system is not strong enough to fight off the fungus, and it may, for example, colonise the lung, resulting in a pulmonary infection. Especially with species of the Aspergillus genus, germination in the respiratory tract can lead to aspergillosis, which is quite common, can vary in severity, and has shown signs of developing new risk groups and antifungal drug resistance.
Conidium
Wikipedia
499
1530205
https://en.wikipedia.org/wiki/Conidium
Biology and health sciences
Fungal morphology and anatomy
Biology
Bioturbation is defined as the reworking of soils and sediments by animals or plants. It includes burrowing, ingestion, and defecation of sediment grains. Bioturbating activities have a profound effect on the environment and are thought to be a primary driver of biodiversity. The formal study of bioturbation began in the 1800s by Charles Darwin experimenting in his garden. The disruption of aquatic sediments and terrestrial soils through bioturbating activities provides significant ecosystem services. These include the alteration of nutrients in aquatic sediment and overlying water, shelter to other species in the form of burrows in terrestrial and water ecosystems, and soil production on land. Bioturbators are deemed ecosystem engineers because they alter resource availability to other species through the physical changes they make to their environments. This type of ecosystem change affects the evolution of cohabitating species and the environment, which is evident in trace fossils left in marine and terrestrial sediments. Other bioturbation effects include altering the texture of sediments (diagenesis), bioirrigation, and displacement of microorganisms and non-living particles. Bioturbation is sometimes confused with the process of bioirrigation, however these processes differ in what they are mixing; bioirrigation refers to the mixing of water and solutes in sediments and is an effect of bioturbation. Walruses, salmon, and pocket gophers are examples of large bioturbators. Although the activities of these large macrofaunal bioturbators are more conspicuous, the dominant bioturbators are small invertebrates, such as earthworms, polychaetes, ghost shrimp, mud shrimp, and midge larvae. The activities of these small invertebrates, which include burrowing and ingestion and defecation of sediment grains, contribute to mixing and the alteration of sediment structure.
Bioturbation
Wikipedia
374
1530478
https://en.wikipedia.org/wiki/Bioturbation
Physical sciences
Sedimentology
Earth science
Functional groups Bioturbators have been organized by a variety of functional groupings based on either ecological characteristics or biogeochemical effects. While the prevailing categorization is based on the way bioturbators transport and interact with sediments, the various groupings likely stem from the relevance of a categorization mode to a field of study (such as ecology or sediment biogeochemistry) and an attempt to concisely organize the wide variety of bioturbating organisms in classes that describe their function. Examples of categorizations include those based on feeding and motility, feeding and biological interactions, and mobility modes. The most common set of groupings are based on sediment transport and are as follows: Gallery-diffusers create complex tube networks within the upper sediment layers and transport sediment through feeding, burrow construction, and general movement throughout their galleries. Gallery-diffusers are heavily associated with burrowing polychaetes, such as Nereis diversicolor and Marenzelleria spp. Biodiffusers transport sediment particles randomly over short distances as they move through sediments. Animals mostly attributed to this category include bivalves such as clams, and amphipod species, but can also include larger vertebrates, such as bottom-dwelling fish and rays that feed along the sea floor. Biodiffusers can be further divided into two subgroups, which include epifaunal (organisms that live on the surface sediments) biodiffusers and surface biodiffusers. This subgrouping may also include gallery-diffusers, reducing the number of functional groups. Upward-conveyors are oriented head-down in sediments, where they feed at depth and transport sediment through their guts to the sediment surface. Major upward-conveyor groups include burrowing polychaetes like the lugworm, Arenicola marina, and thalassinid shrimps. Downward-conveyor species are oriented with their heads towards the sediment-water interface and defecation occurs at depth. Their activities transport sediment from the surface to deeper sediment layers as they feed. Notable downward-conveyors include those in the peanut worm family, Sipunculidae. Regenerators are categorized by their ability to release sediment to the overlying water column, which is then dispersed as they burrow. After regenerators abandon their burrows, water flow at the sediment surface can push in and collapse the burrow. Examples of regenerator species include fiddler and ghost crabs. Ecological roles
Bioturbation
Wikipedia
512
1530478
https://en.wikipedia.org/wiki/Bioturbation
Physical sciences
Sedimentology
Earth science
The evaluation of the ecological role of bioturbators has largely been species-specific. However, their ability to transport solutes, such as dissolved oxygen, enhance organic matter decomposition and diagenesis, and alter sediment structure has made them important for the survival and colonization by other macrofaunal and microbial communities. Microbial communities are greatly influenced by bioturbator activities, as increased transport of more energetically favorable oxidants, such as oxygen, to typically highly reduced sediments at depth alters the microbial metabolic processes occurring around burrows. As bioturbators burrow, they also increase the surface area of sediments across which oxidized and reduced solutes can be exchanged, thereby increasing the overall sediment metabolism. This increase in sediment metabolism and microbial activity further results in enhanced organic matter decomposition and sediment oxygen uptake. In addition to the effects of burrowing activity on microbial communities, studies suggest that bioturbator fecal matter provides a highly nutritious food source for microbes and other macrofauna, thus enhancing benthic microbial activity. This increased microbial activity by bioturbators can contribute to increased nutrient release to the overlying water column. Nutrients released from enhanced microbial decomposition of organic matter, notably limiting nutrients, such as ammonium, can have bottom-up effects on ecosystems and result in increased growth of phytoplankton and bacterioplankton.
Bioturbation
Wikipedia
292
1530478
https://en.wikipedia.org/wiki/Bioturbation
Physical sciences
Sedimentology
Earth science
Burrows offer protection from predation and harsh environmental conditions. For example, termites (Macrotermes bellicosus) burrow and create mounds that have a complex system of air ducts and evaporation devices that create a suitable microclimate in an unfavorable physical environment. Many species are attracted to bioturbator burrows because of their protective capabilities. The shared use of burrows has enabled the evolution of symbiotic relationships between bioturbators and the many species that utilize their burrows. For example, gobies, scale-worms, and crabs live in the burrows made by innkeeper worms. Social interactions provide evidence of co-evolution between hosts and their burrow symbionts. This is exemplified by shrimp-goby associations. Shrimp burrows provide shelter for gobies and gobies serve as a scout at the mouth of the burrow, signaling the presence of potential danger. In contrast, the blind goby Typhlogobius californiensis lives within the deep portion of Callianassa shrimp burrows where there is not much light. The blind goby is an example of a species that is an obligate commensalist, meaning their existence depends on the host bioturbator and its burrow. Although newly hatched blind gobies have fully developed eyes, their eyes become withdrawn and covered by skin as they develop. They show evidence of commensal morphological evolution because it is hypothesized that the lack of light in the burrows where the blind gobies reside is responsible for the evolutionary loss of functional eyes.
Bioturbation
Wikipedia
320
1530478
https://en.wikipedia.org/wiki/Bioturbation
Physical sciences
Sedimentology
Earth science
Bioturbators can also inhibit the presence of other benthic organisms by smothering, exposing other organisms to predators, or resource competition. While thalassinidean shrimps can provide shelter for some organisms and cultivate interspecies relationships within burrows, they have also been shown to have strong negative effects on other species, especially those of bivalves and surface-grazing gastropods, because thalassinidean shrimps can smother bivalves when they resuspend sediment. They have also been shown to exclude or inhibit polychaetes, cumaceans, and amphipods. This has become a serious issue in the northwestern United States, as ghost and mud shrimp (thalassinidean shrimp) are considered pests to bivalve aquaculture operations. The presence of bioturbators can have both negative and positive effects on the recruitment of larvae of conspecifics (those of the same species) and those of other species, as the resuspension of sediments and alteration of flow at the sediment-water interface can affect the ability of larvae to burrow and remain in sediments. This effect is largely species-specific, as species differences in resuspension and burrowing modes have variable effects on fluid dynamics at the sediment-water interface. Deposit-feeding bioturbators may also hamper recruitment by consuming recently settled larvae. Biogeochemical effects Since its onset around 539 million years ago, bioturbation has been responsible for changes in ocean chemistry, primarily through nutrient cycling. Bioturbators played, and continue to play, an important role in nutrient transport across sediments. For example, bioturbating animals are hypothesized to have affected the cycling of sulfur in the early oceans. According to this hypothesis, bioturbating activities had a large effect on the sulfate concentration in the ocean. Around the Cambrian-Precambrian boundary (539 million years ago), animals begin to mix reduced sulfur from ocean sediments to overlying water causing sulfide to oxidize, which increased the sulfate composition in the ocean. During large extinction events, the sulfate concentration in the ocean was reduced. Although this is difficult to measure directly, seawater sulfur isotope compositions during these times indicates bioturbators influenced the sulfur cycling in the early Earth.
Bioturbation
Wikipedia
477
1530478
https://en.wikipedia.org/wiki/Bioturbation
Physical sciences
Sedimentology
Earth science
Bioturbators have also altered phosphorus cycling on geologic scales. Bioturbators mix readily available particulate organic phosphorus (P) deeper into ocean sediment layers which prevents the precipitation of phosphorus (mineralization) by increasing the sequestration of phosphorus above normal chemical rates. The sequestration of phosphorus limits oxygen concentrations by decreasing production on a geologic time scale. This decrease in production results in an overall decrease in oxygen levels, and it has been proposed that the rise of bioturbation corresponds to a decrease in oxygen levels of that time. The negative feedback of animals sequestering phosphorus in the sediments and subsequently reducing oxygen concentrations in the environment limits the intensity of bioturbation in this early environment. Organic contaminants Bioturbation can either enhance or reduce the flux of contaminants from the sediment to the water column, depending on the mechanism of sediment transport. In polluted sediments, bioturbating animals can mix the surface layer and cause the release of sequestered contaminants into the water column. Upward-conveyor species, like polychaete worms, are efficient at moving contaminated particles to the surface. Invasive animals can remobilize contaminants previously considered to be buried at a safe depth. In the Baltic Sea, the invasive Marenzelleria species of polychaete worms can burrow to 35-50 centimeters which is deeper than native animals, thereby releasing previously sequestered contaminants. However, bioturbating animals that live in the sediment (infauna) can also reduce the flux of contaminants to the water column by burying hydrophobic organic contaminants into the sediment. Burial of uncontaminated particles by bioturbating organisms provides more absorptive surfaces to sequester chemical pollutants in the sediments. Ecosystem impacts Nutrient cycling is still affected by bioturbation in the modern Earth. Some examples in the terrestrial and aquatic ecosystems are below. Terrestrial
Bioturbation
Wikipedia
398
1530478
https://en.wikipedia.org/wiki/Bioturbation
Physical sciences
Sedimentology
Earth science
Plants and animals utilize soil for food and shelter, disturbing the upper soil layers and transporting chemically weathered rock called saprolite from the lower soil depths to the surface. Terrestrial bioturbation is important in soil production, burial, organic matter content, and downslope transport. Tree roots are sources of soil organic matter, with root growth and stump decay also contributing to soil transport and mixing. Death and decay of tree roots first delivers organic matter to the soil and then creates voids, decreasing soil density. Tree uprooting causes considerable soil displacement by producing mounds, mixing the soil, or inverting vertical sections of soil. Burrowing animals, such as earth worms and small mammals, form passageways for air and water transport which changes the soil properties, such as the vertical particle-size distribution, soil porosity, and nutrient content. Invertebrates that burrow and consume plant detritus help produce an organic-rich topsoil known as the soil biomantle, and thus contribute to the formation of soil horizons. Small mammals such as pocket gophers also play an important role in the production of soil, possibly with an equal magnitude to abiotic processes. Pocket gophers form above-ground mounds, which moves soil from the lower soil horizons to the surface, exposing minimally weathered rock to surface erosion processes, speeding soil formation. Pocket gophers are thought to play an important role in the downslope transport of soil, as the soil that forms their mounds is more susceptible to erosion and subsequent transport. Similar to tree root effects, the construction of burrows-even when backfilled- decreases soil density. The formation of surface mounds also buries surface vegetation, creating nutrient hotspots when the vegetation decomposes, increasing soil organic matter. Due to the high metabolic demands of their burrow-excavating subterranean lifestyle, pocket gophers must consume large amounts of plant material. Though this has a detrimental effect on individual plants, the net effect of pocket gophers is increased plant growth from their positive effects on soil nutrient content and physical soil properties.
Bioturbation
Wikipedia
422
1530478
https://en.wikipedia.org/wiki/Bioturbation
Physical sciences
Sedimentology
Earth science
Freshwater Important sources of bioturbation in freshwater ecosystems include benthivorous (bottom-dwelling) fish, macroinvertebrates such as worms, insect larvae, crustaceans and molluscs, and seasonal influences from anadromous (migrating) fish such as salmon. Anadromous fish migrate from the sea into fresh-water rivers and streams to spawn. Macroinvertebrates act as biological pumps for moving material between the sediments and water column, feeding on sediment organic matter and transporting mineralized nutrients into the water column. Both benthivorous and anadromous fish can affect ecosystems by decreasing primary production through sediment re-suspension, the subsequent displacement of benthic primary producers, and recycling nutrients from the sediment back into the water column. Lakes and ponds The sediments of lake and pond ecosystems are rich in organic matter, with higher organic matter and nutrient contents in the sediments than in the overlying water. Nutrient re-regeneration through sediment bioturbation moves nutrients into the water column, thereby enhancing the growth of aquatic plants and phytoplankton (primary producers). The major nutrients of interest in this flux are nitrogen and phosphorus, which often limit the levels of primary production in an ecosystem. Bioturbation increases the flux of mineralized (inorganic) forms of these elements, which can be directly used by primary producers. In addition, bioturbation increases the water column concentrations of nitrogen and phosphorus-containing organic matter, which can then be consumed by fauna and mineralized. Lake and pond sediments often transition from the aerobic (oxygen containing) character of the overlaying water to the anaerobic (without oxygen) conditions of the lower sediment over sediment depths of only a few millimeters, therefore, even bioturbators of modest size can affect this transition of the chemical characteristics of sediments. By mixing anaerobic sediments into the water column, bioturbators allow aerobic processes to interact with the re-suspended sediments and the newly exposed bottom sediment surfaces.
Bioturbation
Wikipedia
404
1530478
https://en.wikipedia.org/wiki/Bioturbation
Physical sciences
Sedimentology
Earth science
Macroinvertebrates including chironomid (non-biting midges) larvae and tubificid worms (detritus worms) are important agents of bioturbation in these ecosystems and have different effects based on their respective feeding habits. Tubificid worms do not form burrows, they are upward conveyors. Chironomids, on the other hand, form burrows in the sediment, acting as bioirrigators and aerating the sediments and are downward conveyors. This activity, combined with chironomid's respiration within their burrows, decrease available oxygen in the sediment and increase the loss of nitrates through enhanced rates of denitrification. The increased oxygen input to sediments by macroinvertebrate bioirrigation coupled with bioturbation at the sediment-water interface complicates the total flux of phosphorus . While bioturbation results in a net flux of phosphorus into the water column, the bio-irrigation of the sediments with oxygenated water enhances the adsorption of phosphorus onto iron-oxide compounds, thereby reducing the total flux of phosphorus into the water column. The presence of macroinvertebrates in sediment can initiate bioturbation due to their status as an important food source for benthivorous fish such as carp. Of the bioturbating, benthivorous fish species, carp in particular are important ecosystem engineers and their foraging and burrowing activities can alter the water quality characteristics of ponds and lakes. Carp increase water turbidity by the re-suspension of benthic sediments. This increased turbidity limits light penetration and coupled with increased nutrient flux from the sediment into the water column, inhibits the growth of macrophytes (aquatic plants) favoring the growth of phytoplankton in the surface waters. Surface phytoplankton colonies benefit from both increased suspended nutrients and from recruitment of buried phytoplankton cells released from the sediments by the fish bioturbation. Macrophyte growth has also been shown to be inhibited by displacement from the bottom sediments due to fish burrowing. Rivers and streams River and stream ecosystems show similar responses to bioturbation activities, with chironomid larvae and tubificid worm macroinvertebrates remaining as important benthic agents of bioturbation. These environments can also be subject to strong season bioturbation effects from anadromous fish.
Bioturbation
Wikipedia
492
1530478
https://en.wikipedia.org/wiki/Bioturbation
Physical sciences
Sedimentology
Earth science
Salmon function as bioturbators on both gravel to sand-sized sediment and a nutrient scale, by moving and re-working sediments in the construction of redds (gravel depressions or "nests" containing eggs buried under a thin layer of sediment) in rivers and streams and by mobilization of nutrients. The construction of salmon redds functions to increase the ease of fluid movement (hydraulic conductivity) and porosity of the stream bed. In select rivers, if salmon congregate in large enough concentrations in a given area of the river, the total sediment transport from redd construction can equal or exceed the sediment transport from flood events. The net effect on sediment movement is the downstream transfer of gravel, sand and finer materials and enhancement of water mixing within the river substrate. The construction of salmon redds increases sediment and nutrient fluxes through the hyporheic zone (area between surface water and groundwater) of rivers and effects the dispersion and retention of marine derived nutrients (MDN) within the river ecosystem. MDN are delivered to river and stream ecosystems by the fecal matter of spawning salmon and the decaying carcasses of salmon that have completed spawning and died. Numerical modeling suggests that residence time of MDN within a salmon spawning reach is inversely proportional to the amount of redd construction within the river. Measurements of respiration within a salmon-bearing river in Alaska further suggest that salmon bioturbation of the river bed plays a significant role in mobilizing MDN and limiting primary productivity while salmon spawning is active. The river ecosystem was found to switch from a net autotrophic to heterotrophic system in response to decreased primary production and increased respiration. The decreased primary production in this study was attributed to the loss of benthic primary producers who were dislodged due to bioturbation, while increased respiration was thought to be due to increased respiration of organic carbon, also attributed to sediment mobilization from salmon redd construction. While marine derived nutrients are generally thought to increase productivity in riparian and freshwater ecosystems, several studies have suggested that temporal effects of bioturbation should be considered when characterizing salmon influences on nutrient cycles. Marine Major marine bioturbators range from small infaunal invertebrates to fish and marine mammals. In most marine sediments, however, they are dominated by small invertebrates, including polychaetes, bivalves, burrowing shrimp, and amphipods. Shallow and coastal
Bioturbation
Wikipedia
495
1530478
https://en.wikipedia.org/wiki/Bioturbation
Physical sciences
Sedimentology
Earth science
Coastal ecosystems, such as estuaries, are generally highly productive, which results in the accumulation of large quantities of detritus (organic waste). These large quantities, in addition to typically small sediment grain size and dense populations, make bioturbators important in estuarine respiration. Bioturbators enhance the transport of oxygen into sediments through irrigation and increase the surface area of oxygenated sediments through burrow construction. Bioturbators also transport organic matter deeper into sediments through general reworking activities and production of fecal matter. This ability to replenish oxygen and other solutes at sediment depth allows for enhanced respiration by both bioturbators as well as the microbial community, thus altering estuarine elemental cycling. The effects of bioturbation on the nitrogen cycle are well-documented. Coupled denitrification and nitrification are enhanced due to increased oxygen and nitrate delivery to deep sediments and increased surface area across which oxygen and nitrate can be exchanged. The enhanced nitrification-denitrification coupling contributes to greater removal of biologically available nitrogen in shallow and coastal environments, which can be further enhanced by the excretion of ammonium by bioturbators and other organisms residing in bioturbator burrows. While both nitrification and denitrification are enhanced by bioturbation, the effects of bioturbators on denitrification rates have been found to be greater than that on rates of nitrification, further promoting the removal of biologically available nitrogen. This increased removal of biologically available nitrogen has been suggested to be linked to increased rates of nitrogen fixation in microenvironments within burrows, as indicated by evidence of nitrogen fixation by sulfate-reducing bacteria via the presence of nifH (nitrogenase) genes. Bioturbation by walrus feeding is a significant source of sediment and biological community structure and nutrient flux in the Bering Sea. Walruses feed by digging their muzzles into the sediment and extracting clams through powerful suction. By digging through the sediment, walruses rapidly release large amounts of organic material and nutrients, especially ammonium, from the sediment to the water column. Additionally, walrus feeding behavior mixes and oxygenates the sediment and creates pits in the sediment which serve as new habitat structures for invertebrate larvae.
Bioturbation
Wikipedia
476
1530478
https://en.wikipedia.org/wiki/Bioturbation
Physical sciences
Sedimentology
Earth science
Deep sea Bioturbation is important in the deep sea because deep-sea ecosystem functioning depends on the use and recycling of nutrients and organic inputs from the photic zone. In low energy regions (areas with relatively still water), bioturbation is the only force creating heterogeneity in solute concentration and mineral distribution in the sediment. It has been suggested that higher benthic diversity in the deep sea could lead to more bioturbation which, in turn, would increase the transport of organic matter and nutrients to benthic sediments. Through the consumption of surface-derived organic matter, animals living on the sediment surface facilitate the incorporation of particulate organic carbon (POC) into the sediment where it is consumed by sediment dwelling animals and bacteria. Incorporation of POC into the food webs of sediment dwelling animals promotes carbon sequestration by removing carbon from the water column and burying it in the sediment. In some deep-sea sediments, intense bioturbation enhances manganese and nitrogen cycling. Mathematical modelling The role of bioturbators in sediment biogeochemistry makes bioturbation a common parameter in sediment biogeochemical models, which are often numerical models built using ordinary and partial differential equations. Bioturbation is typically represented as DB, or the biodiffusion coefficient, and is described by a diffusion and, sometimes, an advective term. This representation and subsequent variations account for the different modes of mixing by functional groups and bioirrigation that results from them. The biodiffusion coefficient is usually measured using radioactive tracers such as Pb210, radioisotopes from nuclear fallout, introduced particles including glass beads tagged with radioisotopes or inert fluorescent particles, and chlorophyll a. Biodiffusion models are then fit to vertical distributions (profiles) of tracers in sediments to provide values for DB. Parameterization of bioturbation, however, can vary, as newer and more complex models can be used to fit tracer profiles. Unlike the standard biodiffusion model, these more complex models, such as expanded versions of the biodiffusion model, random walk, and particle-tracking models, can provide more accuracy, incorporate different modes of sediment transport, and account for more spatial heterogeneity.
Bioturbation
Wikipedia
468
1530478
https://en.wikipedia.org/wiki/Bioturbation
Physical sciences
Sedimentology
Earth science
Evolution The onset of bioturbation had a profound effect on the environment and the evolution of other organisms. Bioturbation is thought to have been an important co-factor of the Cambrian Explosion, during which most major animal phyla appeared in the fossil record over a short time. Predation arose during this time and promoted the development of hard skeletons, for example bristles, spines, and shells, as a form of armored protection. It is hypothesized that bioturbation resulted from this skeleton formation. These new hard parts enabled animals to dig into the sediment to seek shelter from predators, which created an incentive for predators to search for prey in the sediment (see Evolutionary Arms Race). Burrowing species fed on buried organic matter in the sediment which resulted in the evolution of deposit feeding (consumption of organic matter within sediment). Prior to the development of bioturbation, laminated microbial mats were the dominant biological structures of the ocean floor and drove much of the ecosystem functions. As bioturbation increased, burrowing animals disturbed the microbial mat system and created a mixed sediment layer with greater biological and chemical diversity. This greater biological and chemical diversity is thought to have led to the evolution and diversification of seafloor-dwelling species. An alternate, less widely accepted hypothesis for the origin of bioturbation exists. The trace fossil Nenoxites is thought to be the earliest record of bioturbation, predating the Cambrian Period. The fossil is dated to 555 million years, which places it in the Ediacaran Period. The fossil indicates a 5 centimeter depth of bioturbation in muddy sediments by a burrowing worm. This is consistent with food-seeking behavior, as there tended to be more food resources in the mud than the water column. However, this hypothesis requires more precise geological dating to rule out an early Cambrian origin for this specimen. The evolution of trees during the Devonian Period enhanced soil weathering and increased the spread of soil due to bioturbation by tree roots. Root penetration and uprooting also enhanced soil carbon storage by enabling mineral weathering and the burial of organic matter. Fossil record
Bioturbation
Wikipedia
432
1530478
https://en.wikipedia.org/wiki/Bioturbation
Physical sciences
Sedimentology
Earth science
Patterns or traces of bioturbation are preserved in lithified rock. The study of such patterns is called ichnology, or the study of "trace fossils", which, in the case of bioturbators, are fossils left behind by digging or burrowing animals. This can be compared to the footprint left behind by these animals. In some cases bioturbation is so pervasive that it completely obliterates sedimentary structures, such as laminated layers or cross-bedding. Thus, it affects the disciplines of sedimentology and stratigraphy within geology. The study of bioturbator ichnofabrics uses the depth of the fossils, the cross-cutting of fossils, and the sharpness (or how well defined) of the fossil to assess the activity that occurred in old sediments. Typically the deeper the fossil, the better preserved and well defined the specimen. Important trace fossils from bioturbation have been found in marine sediments from tidal, coastal and deep sea sediments. In addition sand dune, or Eolian, sediments are important for preserving a wide variety of fossils. Evidence of bioturbation has been found in deep-sea sediment cores including into long records, although the act extracting the core can disturb the signs of bioturbation, especially at shallower depths. Arthropods, in particular are important to the geologic record of bioturbation of Eolian sediments. Dune records show traces of burrowing animals as far back as the lower Mesozoic (250 Million years ago), although bioturbation in other sediments has been seen as far back as 550 Ma.
Bioturbation
Wikipedia
329
1530478
https://en.wikipedia.org/wiki/Bioturbation
Physical sciences
Sedimentology
Earth science
Research history Bioturbation's importance for soil processes and geomorphology was first realized by Charles Darwin, who devoted his last scientific book to the subject (The Formation of Vegetable Mould through the Action of Worms). Darwin spread chalk dust over a field to observe changes in the depth of the chalk layer over time. Excavations 30 years after the initial deposit of chalk revealed that the chalk was buried 18 centimeters under the sediment, which indicated a burial rate of 6 millimeters per year. Darwin attributed this burial to the activity of earthworms in the sediment and determined that these disruptions were important in soil formation. In 1891, geologist Nathaniel Shaler expanded Darwin's concept to include soil disruption by ants and trees. The term "bioturbation" was later coined by Rudolf Richter in 1952 to describe structures in sediment caused by living organisms. Since the 1980s, the term "bioturbation" has been widely used in soil and geomorphology literature to describe the reworking of soil and sediment by plants and animals.
Bioturbation
Wikipedia
210
1530478
https://en.wikipedia.org/wiki/Bioturbation
Physical sciences
Sedimentology
Earth science
Fungal infection, also known as mycosis, is a disease caused by fungi. Different types are traditionally divided according to the part of the body affected; superficial, subcutaneous, and systemic. Superficial fungal infections include common tinea of the skin, such as tinea of the body, groin, hands, feet and beard, and yeast infections such as pityriasis versicolor. Subcutaneous types include eumycetoma and chromoblastomycosis, which generally affect tissues in and beneath the skin. Systemic fungal infections are more serious and include cryptococcosis, histoplasmosis, pneumocystis pneumonia, aspergillosis and mucormycosis. Signs and symptoms range widely. There is usually a rash with superficial infection. Fungal infection within the skin or under the skin may present with a lump and skin changes. Pneumonia-like symptoms or meningitis may occur with a deeper or systemic infection. Fungi are everywhere, but only some cause disease. Fungal infection occurs after spores are either breathed in, come into contact with skin or enter the body through the skin such as via a cut, wound or injection. It is more likely to occur in people with a weak immune system. This includes people with illnesses such as HIV/AIDS, and people taking medicines such as steroids or cancer treatments. Fungi that cause infections in people include yeasts, molds and fungi that are able to exist as both a mold and yeast. The yeast Candida albicans can live in people without producing symptoms, and is able to cause both superficial mild candidiasis in healthy people, such as oral thrush or vaginal yeast infection, and severe systemic candidiasis in those who cannot fight infection themselves. Diagnosis is generally based on signs and symptoms, microscopy, culture, sometimes requiring a biopsy and the aid of medical imaging. Some superficial fungal infections of the skin can appear similar to other skin conditions such as eczema and lichen planus. Treatment is generally performed using antifungal medicines, usually in the form of a cream or by mouth or injection, depending on the specific infection and its extent. Some require surgically cutting out infected tissue. Fungal infections have a world-wide distribution and are common, affecting more than one billion people every year. An estimated 1.7 million deaths from fungal disease were reported in 2020. Several, including sporotrichosis, chromoblastomycosis and mycetoma are neglected.
Fungal infection
Wikipedia
510
1530575
https://en.wikipedia.org/wiki/Fungal%20infection
Biology and health sciences
Infectious disease
null
A wide range of fungal infections occur in other animals, and some can be transmitted from animals to people. Classification Mycoses are traditionally divided into superficial, subcutaneous, or systemic, where infection is deep, more widespread and involving internal body organs. They can affect the nails, vagina, skin and mouth. Some types such as blastomycosis, cryptococcus, coccidioidomycosis and histoplasmosis, affect people who live in or visit certain parts of the world. Others such as aspergillosis, pneumocystis pneumonia, candidiasis, mucormycosis and talaromycosis, tend to affect people who are unable to fight infection themselves. Mycoses might not always conform strictly to the three divisions of superficial, subcutaneous and systemic. Some superficial fungal infections can cause systemic infections in people who are immunocompromised. Some subcutaneous fungal infections can invade into deeper structures, resulting in systemic disease. Candida albicans can live in people without producing symptoms, and is able to cause both mild candidiasis in healthy people and severe invasive candidiasis in those who cannot fight infection themselves. ICD-11 codes ICD-11 codes include: 1F20 Aspergillosis 1F21 Basidiobolomycosis 1F22 Blastomycosis 1F23 Candidosis 1F24 Chromoblastomycosis 1F25 Coccidioidomycosis 1F26 Conidiobolomycosis 1F27 Cryptococcosis 1F28 Dermatophytosis 1F29 Eumycetoma 1F2A Histoplasmosis 1F2B Lobomycosis 1F2C Mucormycosis 1F2D Non-dermatophyte superficial dermatomycoses 1F2E Paracoccidioidomycosis 1F2F Phaeohyphomycosis 1F2G Pneumocystosis 1F2H Scedosporiosis 1F2J Sporotrichosis 1F2K Talaromycosis 1F2L Emmonsiosis Superficial mycoses Superficial mycoses include candidiasis in healthy people, common tinea of the skin, such as tinea of the body, groin, hands, feet and beard, and malassezia infections such as pityriasis versicolor. Subcutaneous
Fungal infection
Wikipedia
510
1530575
https://en.wikipedia.org/wiki/Fungal%20infection
Biology and health sciences
Infectious disease
null
Subcutaneous fungal infections include sporotrichosis, chromoblastomycosis, and eumycetoma. Systemic Systemic fungal infections include histoplasmosis, cryptococcosis, coccidioidomycosis, blastomycosis, mucormycosis, aspergillosis, pneumocystis pneumonia and systemic candidiasis. Systemic mycoses due to primary pathogens originate normally in the lungs and may spread to other organ systems. Organisms that cause systemic mycoses are inherently virulent.. Systemic mycoses due to opportunistic pathogens are infections of people with immune deficiencies who would otherwise not be infected. Examples of immunocompromised conditions include AIDS, alteration of normal flora by antibiotics, immunosuppressive therapy, and metastatic cancer. Examples of opportunistic mycoses include Candidiasis, Cryptococcosis and Aspergillosis. Signs and symptoms Most common mild mycoses often present with a rash. Infections within the skin or under the skin may present with a lump and skin changes. Less common deeper fungal infections may present with pneumonia like symptoms or meningitis. Causes Mycoses are caused by certain fungi; yeasts, molds and some fungi that can exist as both a mold and yeast. They are everywhere and infection occurs after spores are either breathed in, come into contact with skin or enter the body through the skin such as via a cut, wound or injection. Candida albicans is the most common cause of fungal infection in people, particularly as oral or vaginal thrush, often following taking antibiotics. Risk factors Fungal infections are more likely in people with weak immune systems. This includes people with illnesses such as HIV/AIDS, and people taking medicines such as steroids or cancer treatments. People with diabetes also tend to develop fungal infections. Very young and very old people, also, are groups at risk. Individuals being treated with antibiotics are at higher risk of fungal infections. Children whose immune systems are not functioning properly (such as children with cancer) are at risk of invasive fungal infections.
Fungal infection
Wikipedia
436
1530575
https://en.wikipedia.org/wiki/Fungal%20infection
Biology and health sciences
Infectious disease
null
COVID-19 During the COVID-19 pandemic some fungal infections have been associated with COVID-19. Fungal infections can mimic COVID-19, occur at the same time as COVID-19 and more serious fungal infections can complicate COVID-19. A fungal infection may occur after antibiotics for a bacterial infection which has occurred following COVID-19. The most common serious fungal infections in people with COVID-19 include aspergillosis and invasive candidiasis. COVID-19–associated mucormycosis is generally less common, but in 2021 was noted to be significantly more prevalent in India. Mechanism Fungal infections occur after spores are either breathed in, come into contact with skin or enter the body through a wound. Diagnosis Diagnosis is generally by signs and symptoms, microscopy, biopsy, culture and sometimes with the aid of medical imaging. Differential diagnosis Some tinea and candidiasis infections of the skin can appear similar to eczema and lichen planus. Pityriasis versicolor can look like seborrheic dermatitis, pityriasis rosea, pityriasis alba and vitiligo. Some fungal infections such as coccidioidomycosis, histoplasmosis, and blastomycosis can present with fever, cough, and shortness of breath, thereby resembling COVID-19. Prevention Keeping the skin clean and dry, as well as maintaining good hygiene, will help larger topical mycoses. Because some fungal infections are contagious, it is important to wash hands after touching other people or animals. Sports clothing should also be washed after use. Treatment Treatment depends on the type of fungal infection, and usually requires topical or systemic antifungal medicines. Pneumocystosis that does not respond to anti-fungals is treated with co-trimoxazole. Sometimes, infected tissue needs to be surgically cut away. Epidemiology Worldwide, every year fungal infections affect more than one billion people. An estimated 1.6 million deaths from fungal disease were reported in 2017. The figure has been rising, with an estimated 1.7 million deaths from fungal disease reported in 2020. Fungal infections also constitute a significant cause of illness and mortality in children.
Fungal infection
Wikipedia
464
1530575
https://en.wikipedia.org/wiki/Fungal%20infection
Biology and health sciences
Infectious disease
null
According to the Global Action Fund for Fungal Infections, every year there are over 10 million cases of fungal asthma, around 3 million cases of long-term aspergillosis of lungs, 1 million cases of blindness due to fungal keratitis, more than 200,000 cases of meningitis due to cryptococcus, 700,000 cases of invasive candidiasis, 500,000 cases of pneumocystosis of lungs, 250,000 cases of invasive aspergillosis, and 100,000 cases of histoplasmosis. History In 500BC, an apparent account of ulcers in the mouth by Hippocrates may have been thrush. The Hungarian microscopist based in Paris David Gruby first reported that human disease could be caused by fungi in the early 1840s. SARS 2003 During the 2003 SARS outbreak, fungal infections were reported in 14.8–33% of people affected by SARS, and it was the cause of death in 25–73.7% of people with SARS. Other animals A wide range of fungal infections occur in other animals, and some can be transmitted from animals to people, such as Microsporum canis from cats.
Fungal infection
Wikipedia
247
1530575
https://en.wikipedia.org/wiki/Fungal%20infection
Biology and health sciences
Infectious disease
null
Sodium permanganate is the inorganic compound with the formula NaMnO4. It is closely related to the more commonly encountered potassium permanganate, but it is generally less desirable, because it is more expensive to produce. It is mainly available as the monohydrate. This salt absorbs water from the atmosphere and has a low melting point. Being about 15 times more soluble than KMnO4, sodium permanganate finds some applications where very high concentrations of MnO4− are sought. Preparation and properties Sodium permanganate cannot be prepared analogously to the route to KMnO4 because the required intermediate manganate salt, Na2MnO4, does not form. Thus less direct routes are used including conversion from KMnO4. Sodium permanganate behaves similarly to potassium permanganate. It dissolves readily in water to give deep purple solutions, evaporation of which gives prismatic purple-black glistening crystals of the monohydrate NaMnO4·H2O. The potassium salt does not form a hydrate. Because of its hygroscopic nature, it is less useful in analytical chemistry than its potassium counterpart. It can be prepared by the reaction of manganese dioxide with sodium hypochlorite: 2 MnO2 + 3 NaClO + 2 NaOH → 2 NaMnO4 + 3 NaCl + H2O Applications Because of its high solubility, its aqueous solutions are used as a drilled hole debris remover and etchant in printed circuitry, with a limited utility though. It is gaining popularity in water treatment for taste, odor, and zebra mussel control. The V-2 rocket used it in combination with hydrogen peroxide to drive a steam turbopump. As an oxidizer, sodium permanganate is used in environmental remediation of soil and groundwater contaminated with chlorinated solvents using the remediation technology in situ chemical oxidation, also referred to as ISCO.
Sodium permanganate
Wikipedia
409
8676888
https://en.wikipedia.org/wiki/Sodium%20permanganate
Physical sciences
Metallic oxyanions
Chemistry
A zero-day (also known as a 0-day) is a vulnerability in software or hardware that is typically unknown to the vendor and for which no patch or other fix is available. The vendor thus has zero days to prepare a patch, as the vulnerability has already been described or exploited. Despite developers' goal of delivering a product that works entirely as intended, virtually all software and hardware contains bugs. Many of these impair the security of the system and are thus vulnerabilities. Although the basis of only a minority of cyberattacks, zero-days are considered more dangerous than known vulnerabilities because there are fewer countermeasures possible. States are the primary users of zero-day vulnerabilities, not only because of the high cost of finding or buying them, but also the significant cost of writing the attack software. Many vulnerabilities are discovered by hackers or security researchers, who may disclose them to the vendor (often in exchange for a bug bounty) or sell them to states or criminal groups. The use of zero-days increased after many popular software companies began to encrypt messages and data, meaning that the unencrypted data could only be obtained by hacking into the software before it was encrypted. Definition Despite developers' goal of delivering a product that works entirely as intended, virtually all software and hardware contain bugs. If a bug creates a security risk, it is called a vulnerability. Vulnerabilities vary in their ability to be exploited by malicious actors. Some are not usable at all, while others can be used to disrupt the device with a denial of service attack. The most valuable allow the attacker to inject and run their own code, without the user being aware of it. Although the term "zero-day" initially referred to the time since the vendor had become aware of the vulnerability, zero-day vulnerabilities can also be defined as the subset of vulnerabilities for which no patch or other fix is available. A zero-day exploit is any exploit that takes advantage of such a vulnerability.
Zero-day vulnerability
Wikipedia
426
16001916
https://en.wikipedia.org/wiki/Zero-day%20vulnerability
Technology
Computer security
null
Exploits An exploit is the delivery mechanism that takes advantage of the vulnerability to penetrate the target's systems, for such purposes as disrupting operations, installing malware, or exfiltrating data. Researchers Lillian Ablon and Andy Bogart write that "little is known about the true extent, use, benefit, and harm of zero-day exploits". Exploits based on zero-day vulnerabilities are considered more dangerous than those that take advantage of a known vulnerability. However, it is likely that most cyberattacks use known vulnerabilities, not zero-days. States are the primary users of zero-day exploits, not only because of the high cost of finding or buying vulnerabilities, but also the significant cost of writing the attack software. Nevertheless, anyone can use a vulnerability, and according to research by the RAND Corporation, "any serious attacker can always get an affordable zero-day for almost any target". Many targeted attacks and most advanced persistent threats rely on zero-day vulnerabilities. The average time to develop an exploit from a zero-day vulnerability was estimated at 22 days. The difficulty of developing exploits has been increasing over time due to increased anti-exploitation features in popular software. Window of vulnerability Zero-day vulnerabilities are often classified as alive—meaning that there is no public knowledge of the vulnerability—and dead—the vulnerability has been disclosed, but not patched. If the software's maintainers are actively searching for vulnerabilities, it is a living vulnerability; such vulnerabilities in unmaintained software are called immortal. Zombie vulnerabilities can be exploited in older versions of the software but have been patched in newer versions. Even publicly known and zombie vulnerabilities are often exploitable for an extended period. Security patches can take months to develop, or may never be developed. A patch can have negative effects on the functionality of software and users may need to test the patch to confirm functionality and compatibility. Larger organizations may fail to identify and patch all dependencies, while smaller enterprises and personal users may not install patches. Research suggests that risk of cyberattack increases if the vulnerability is made publicly known or a patch is released. Cybercriminals can reverse engineer the patch to find the underlying vulnerability and develop exploits, often faster than users install the patch.
Zero-day vulnerability
Wikipedia
479
16001916
https://en.wikipedia.org/wiki/Zero-day%20vulnerability
Technology
Computer security
null
According to research by RAND Corporation published in 2017, zero-day exploits remain usable for 6.9 years on average, although those purchased from a third party only remain usable for 1.4 years on average. The researchers were unable to determine if any particular platform or software (such as open-source software) had any relationship to the life expectancy of a zero-day vulnerability. Although the RAND researchers found that 5.7 percent of a stockpile of secret zero-day vulnerabilities will have been discovered by someone else within a year, another study found a higher overlap rate, as high as 10.8 percent to 21.9 percent per year. Countermeasures Because, by definition, there is no patch that can block a zero-day exploit, all systems employing the software or hardware with the vulnerability are at risk. This includes secure systems such as banks and governments that have all patches up to date. Security systems are designed around known vulnerabilities, and repeated exploitations of a zero-day exploit could continue undetected for an extended period of time. Although there have been many proposals for a system that is effective at detecting zero-day exploits, this remains an active area of research in 2023. Many organizations have adopted defense-in-depth tactics so that attacks are likely to require breaching multiple levels of security, which makes it more difficult to achieve. Conventional cybersecurity measures such as training and access control such as multifactor authentication, least-privilege access, and air-gapping makes it harder to compromise systems with a zero-day exploit. Since writing perfectly secure software is impossible, some researchers argue that driving up the cost of exploits is a good strategy to reduce the burden of cyberattacks. Market
Zero-day vulnerability
Wikipedia
357
16001916
https://en.wikipedia.org/wiki/Zero-day%20vulnerability
Technology
Computer security
null
Zero-day exploits can fetch millions of dollars. There are three main types of buyers: White: the vendor, or to third parties such as the Zero Day Initiative that disclose to the vendor. Often such disclosure is in exchange for a bug bounty. Not all companies respond positively to disclosures, as they can cause legal liability and operational overhead. It is not uncommon to receive cease-and-desist letters from software vendors after disclosing a vulnerability for free. Gray: the largest and most lucrative. Government or intelligence agencies buy zero-days and may use it in an attack, stockpile the vulnerability, or notify the vendor. The United States federal government is one of the largest buyers. As of 2013, the Five Eyes (United States, United Kingdom, Canada, Australia, and New Zealand) captured the plurality of the market and other significant purchasers included Russia, India, Brazil, Malaysia, Singapore, North Korea, and Iran. Middle Eastern countries were poised to become the biggest spenders. Black: organized crime, which typically prefers exploit software rather than just knowledge of a vulnerability. These users are more likely to employ "half-days" where a patch is already available. In 2015, the markets for government and crime were estimated at at least ten times larger than the white market. Sellers are often hacker groups that seek out vulnerabilities in widely used software for financial reward. Some will only sell to certain buyers, while others will sell to anyone. White market sellers are more likely to be motivated by non pecuniary rewards such as recognition and intellectual challenge. Selling zero day exploits is legal. Despite calls for more regulation, law professor Mailyn Fidler says there is little chance of an international agreement because key players such as Russia and Israel are not interested. The sellers and buyers that trade in zero-days tend to be secretive, relying on non-disclosure agreements and classified information laws to keep the exploits secret. If the vulnerability becomes known, it can be patched and its value consequently crashes. Because the market lacks transparency, it can be hard for parties to find a fair price. Sellers might not be paid if the vulnerability was disclosed before it was verified, or if the buyer declined to purchase it but used it anyway. With the proliferation of middlemen, sellers could never know to what use the exploits could be put. Buyers could not guarantee that the exploit was not sold to another party. Both buyers and sellers advertise on the dark web.
Zero-day vulnerability
Wikipedia
500
16001916
https://en.wikipedia.org/wiki/Zero-day%20vulnerability
Technology
Computer security
null
Research published in 2022 based on maximum prices paid as quoted by a single exploit broker found a 44 percent annualized inflation rate in exploit pricing. Remote zero-click exploits could fetch the highest price, while those that require local access to the device are much cheaper. Vulnerabilities in widely used software are also more expensive. They estimated that around 400 to 1,500 people sold exploits to that broker and they made around $5,500 to $20,800 annually. Disclosure and stockpiling , there is an ongoing debate as to whether the United States should disclose the vulnerabilities it is aware of, so that they can be patched, or keep them secret for its own use. Reasons that states keep an vulnerability secret include wanting to use it offensively, or defensively in penetration testing. Disclosing the vulnerability reduces the risk that consumers and all users of the software will be victimized by malware or data breaches. The phases of zero-day vulnerability disclosure, along with a typical timeline, are as follows: Discovery: A researcher identifies the vulnerability, marking "Day 0." Reporting: The researcher notifies the vendor or a third party, starting remediation efforts. Patch Development: The vendor develops a fix, which can take weeks to months depending on the complexity. Public Disclosure: Once a patch is released, details are shared publicly. If no patch is issued within an agreed period (commonly 90 days), some researchers disclose it to push for action. History Zero-day exploits increased in significance after services such as Apple, Google, Facebook, and Microsoft encrypted servers and messages, meaning that the most feasible way to access a user's data was to intercept it at the source before it was encrypted. One of the best-known use of zero-day exploits was the Stuxnet worm, which used four zero-day vulnerabilities to damage Iran's nuclear program in 2010. The worm showed what could be achieved by zero-day exploits, unleashing an expansion in the market.
Zero-day vulnerability
Wikipedia
419
16001916
https://en.wikipedia.org/wiki/Zero-day%20vulnerability
Technology
Computer security
null
The United States National Security Agency (NSA) increased its search for zero-day vulnerabilities after large tech companies refused to install backdoors into the software, tasking the Tailored Access Operations (TAO) with discovering and purchasing zero-day exploits. In 2007, former NSA employee Charlie Miller publicly revealed for the first time that the United States government was buying zero-day exploits. Some information about the NSA involvement with zero-days was revealed in the documents leaked by NSA contractor Edward Snowden in 2013, but details were lacking. Reporter Nicole Perlroth concluded that "either Snowden’s access as a contractor didn’t take him far enough into the government’s systems for the intel required, or some of the government’s sources and methods for acquiring zero-days were so confidential, or controversial, that the agency never dared put them in writing". One of the most infamous vulnerabilities discovered after 2013, Heartbleed (CVE-2014-0160), was not a zero-day when publicly disclosed but underscored the critical impact that software bugs can have on global cybersecurity. This flaw in the OpenSSL cryptographic library could have been exploited as a zero-day prior to its discovery, allowing attackers to steal sensitive information such as private keys and passwords. In 2016 the hacking group known as Shadow Brokers released a trove of sophisticated zero-day exploits reportedly stolen from the United States National Security Agency (NSA). These included tools such as EternalBlue, which leveraged a vulnerability in Microsoft Windows' Server Message Block (SMB) protocol. EternalBlue was later weaponized in high-profile attacks like WannaCry and NotPetya, causing widespread global damage and highlighting the risks of stockpiling vulnerabilities. The year 2020 saw one of the most sophisticated cyber espionage campaigns to date, in which attackers exploited multiple vulnerabilities, including zero-day vulnerabilities, to compromise SolarWinds' Orion software. This allowed access to numerous government and corporate networks. In 2021 Chinese state-sponsored group, Hafnium, exploited zero-day vulnerabilities in Microsoft Exchange Server to conduct cyber espionage. Known as ProxyLogon, these flaws allowed attackers to bypass authentication and execute arbitrary code, compromising thousands of systems globally.
Zero-day vulnerability
Wikipedia
472
16001916
https://en.wikipedia.org/wiki/Zero-day%20vulnerability
Technology
Computer security
null
In 2022 the spyware Pegasus, developed by Israel's NSO Group, was found to exploit zero-click vulnerabilities in messaging apps like iMessage and WhatsApp. These exploits allowed attackers to access targets' devices without requiring user interaction, heightening concerns over surveillance and privacy.
Zero-day vulnerability
Wikipedia
62
16001916
https://en.wikipedia.org/wiki/Zero-day%20vulnerability
Technology
Computer security
null
The domesticated hedgehog kept as a pet is typically the African pygmy hedgehog (Atelerix albiventris). Other species kept as pets include the long-eared hedgehog (Hemiechinus auritus) and the Indian long-eared hedgehog (Hemiechinus collaris). In the ancient era Although ancient humans were familiar with the hedgehog, hunting it for food and using its spines in the processing of wool, it was not likely kept as a pet. Aristotle described the behaviour of "pet" hedgehogs kept in the home as a means for predicting weather by someone in Byzantium—Plutarch describes the same, but refers to the man as living in Cyzicus—but this is probably an unusual situation, as hedgehogs were generally not regarded as valuable animals. Other sources suggest that the Ancient Greeks may have kept hedgehogs around the home for their potential to eat beetles and other pests. The Guinness World Records describe the Romans as having domesticated a relative of the Algerian hedgehog in the 4th century BCE, to use for meat and quills as well as pets. The Romans did use the quill-covered hedgehog skins to clean their shawls, making them important to commerce, which resulted in the Roman Senate regulating the trade in hedgehog skins. The quills were used in the training of other animals, such as keeping a calf from suckling after it had been weaned. Modern domestication In the early 1980s, hedgehog domestication became popular in the United States. Some U.S. states, however, ban them, or require a license to own one. Since domestication restarted, several new colors of hedgehogs have been cultivated or become common, including albino and pinto hedgehogs. "Pinto" is a color pattern, rather than a color: A total lack of color on the quills and skin beneath, in distinct patches. Currently, the species most common among domestic hedgehogs are African, from warm climates (above ). They do not hibernate in the wild, and if one of these African hedgehogs begins hibernation in response to lowered body temperature, the result can be its death. The process is easily reversed by warming, if caught within a few days of onset. Legality Because a hedgehog is commonly kept in a cage or similar enclosure, it is allowed in some residences where cats and dogs are not allowed.
Domesticated hedgehog
Wikipedia
504
2177275
https://en.wikipedia.org/wiki/Domesticated%20hedgehog
Biology and health sciences
Eulipotyphla
Animals
It is illegal to own a hedgehog as a pet in some jurisdictions in North America, and a license is needed to legally breed them. These restrictions may have been enacted due to the ability of some hedgehog species to carry foot and mouth disease, a highly contagious disease of cloven-hooved animals. The European hedgehog is a protected species in all countries that have signed the Berne Convention; this includes all member states of the Council of Europe, as well as the European Union and a small number of other states. In these countries, it is illegal to capture the European hedgehog or keep it as a pet.
Domesticated hedgehog
Wikipedia
129
2177275
https://en.wikipedia.org/wiki/Domesticated%20hedgehog
Biology and health sciences
Eulipotyphla
Animals
Legal status internationally Austria: European hedgehogs are protected and cannot be kept as pets. Four-toed hedgehogs (African Pygmy hedgehogs) may legally be kept as pets. Australia: All hedgehogs are classified as exotic pets that are illegal to import. Canada: In Quebec – European hedgehogs are illegal. Four-toed hedgehogs are legal. In Ontario – European hedgehogs are protected and cannot be kept as pets. Four-toed hedgehogs may legally be kept as pets. Denmark: European hedgehogs are protected and cannot be kept as pets. Four-toed hedgehogs may legally be kept as pets. Finland: European hedgehogs are protected and cannot be kept as pets. Four-toed hedgehogs may legally be kept as pets. Germany: European hedgehogs are protected and cannot be kept as pets. They may be removed from their habitat if injured or sick, but only for the purposes of restoring their health. If, for some reason, they cannot be released back into the wild (due to neurological conditions or permanent damage to limbs, for example) keeping them is still illegal and they must be put to sleep by a qualified vet. Four-toed hedgehogs may legally be kept as pets. Italy: European hedgehogs are protected and cannot be kept as pets. Four-toed hedgehogs may legally be kept as pets. Latvia: European hedgehogs are protected and cannot be kept as pets. Four-toed hedgehogs may legally be kept as pets. Netherlands: European hedgehogs are protected and cannot be kept as pets. Since 01-01-2024, the selling, trading, and breeding of four-toed hedgehogs has become illegal. Poland: European hedgehogs are protected and cannot be kept as pets. Four-toed hedgehogs may legally be kept as pets. France: European hedgehogs are protected, no specie of hedgehog can be kept as pets. Spain: European hedgehogs are protected and cannot be kept as pets. Four-toed hedgehogs are illegal and considered an exotic invasive species. Sweden: European hedgehogs are protected and cannot be kept as pets. Four-toed hedgehogs may legally be kept as pets. United Kingdom: European hedgehogs are protected and cannot be kept as pets. Four-toed hedgehogs may legally be kept as pets. United States:
Domesticated hedgehog
Wikipedia
508
2177275
https://en.wikipedia.org/wiki/Domesticated%20hedgehog
Biology and health sciences
Eulipotyphla
Animals
In Idaho and Oregon – European hedgehogs cannot be kept as pets. Four-toed hedgehogs may legally be kept as pets. In New Jersey and Wyoming – a permit is required. In Wisconsin – an import permit from the state department of agriculture is required to bring a hedgehog into the state. In Fairfax County, Virginia, it became legal to keep hedgehogs as pets in 2019. In Pennsylvania – hedgehogs may not be imported into the state, but hedgehogs in the state as of 1992 and their descendants are allowed. It is currently illegal to own a hedgehog in California, Georgia, Hawaii, New York City, and Washington, D.C. Singapore: Hedgehogs of all kinds are illegal, along with other exotic pets such as iguanas, tarantulas, scorpions, and snakes. Turkey: European hedgehogs are protected and cannot be kept as pets, and four-toed hedgehogs may also not legally be kept as pets.
Domesticated hedgehog
Wikipedia
204
2177275
https://en.wikipedia.org/wiki/Domesticated%20hedgehog
Biology and health sciences
Eulipotyphla
Animals
Enclosures In the wild, a hedgehog will cover many miles each night. A hedgehog with insufficient range may show signs of depression, such as excessive sleeping, refusal to eat, repetitious behaviour, and self-mutilation. Hedgehogs require a fair amount of exercise to avoid liver problems due to excess weight. Therefore, a domesticated hedgehog must have access to a running wheel. Running wheels must be selected carefully to avoid foot injury. Running wheels made of solid material that are approximately 1 foot (12 inches) in diameter are recommended. Food In the wild, a hedgehog is opportunistic and will eat many things, but the majority of the diet comprises insects. As insectivores, hedgehogs need a diet that is high in protein and low in fat. They also require chitin, which comes from the exoskeleton of insects; fiber in the diet may be a substitute for the chitin component. There are prepared foods specifically for pet hedgehogs and insectivores, including foods made from insect components. Also available are alimentary powders to sprinkle on other food which provide chitin and other nutrients. Pet hedgehogs may eat such table foods as cooked, lean chicken, turkey, beef or pork. They will often eat small amounts of vegetables and fruit. Hedgehogs are lactose-intolerant and will have stomach problems after consuming most dairy products, though occasional plain low-fat yogurt or cottage cheese seem to be well tolerated. Allergies Hedgehogs produce very little danger to their owners and handlers. It is possible to be allergic to items surrounding the hedgehog, such as the hedgehog's food or bedding, but it is rare that a person would be allergic to the hedgehog itself. After handling hedgehogs, some have claimed that pink dots on their hands is an allergic reaction. This is more likely caused by small pricks from the hedgehog's spines. If a hedgehog is not clean, the pricks can become infected. The infection is from contaminants on the hedgehog or on the surface of the hands, not from an allergic reaction to the hedgehog. As is true with most animal handling, one should wash their hands after handling a hedgehog. Hedgehogs are commonly allergic to wood oils. Wood bedding should be avoided, specifically cedar. The oil found in cedar can cause severe upper respiratory problems. Aspen however is widely accepted as a safe substitute.
Domesticated hedgehog
Wikipedia
511
2177275
https://en.wikipedia.org/wiki/Domesticated%20hedgehog
Biology and health sciences
Eulipotyphla
Animals
Diseases Hedgehogs can easily become obese; if they can no longer roll completely into a ball, it is a clear sign of obesity. Conversely, hedgehogs often stop eating under situations of stress, such as when adjusting to a new home. Hedgehogs are prone to many diseases, including cancer, which spreads quickly in hedgehogs, and wobbly hedgehog syndrome (WHS), a neurological problem. Some symptoms of WHS resemble those of multiple sclerosis (MS) in humans, therefore the condition the animal experiences can be compared with what MS patients experience. A possible cause of WHS is a genetic flaw allowing a virus to attack the hedgehog's nervous system. The nose can display a variety of symptoms of a troubled hedgehog, especially respiratory illnesses, such as pneumonia. In many cases, the form of pneumonia that affects hedgehogs is bacterial in nature. If acted upon quickly, antibiotics can have a very positive effect. Signs to watch for include bubbles, excessive dripping, or constant sneezing. Hedgehogs usually react to stress with temporary digestive disorders that include vomiting and green feces.
Domesticated hedgehog
Wikipedia
231
2177275
https://en.wikipedia.org/wiki/Domesticated%20hedgehog
Biology and health sciences
Eulipotyphla
Animals
Cosmic dustalso called extraterrestrial dust, space dust, or star dustis dust that occurs in outer space or has fallen onto Earth. Most cosmic dust particles measure between a few molecules and , such as micrometeoroids (<30 μm) and meteoroids (>30 μm). Cosmic dust can be further distinguished by its astronomical location: intergalactic dust, interstellar dust, interplanetary dust (as in the zodiacal cloud), and circumplanetary dust (as in a planetary ring). There are several methods to obtain space dust measurement. In the Solar System, interplanetary dust causes the zodiacal light. Solar System dust includes comet dust, planetary dust (like from Mars), asteroidal dust, dust from the Kuiper belt, and interstellar dust passing through the Solar System. Thousands of tons of cosmic dust are estimated to reach Earth's surface every year, with most grains having a mass between 10−16 kg (0.1 pg) and 10−4 kg (0.1 g). The density of the dust cloud through which the Earth is traveling is approximately 10−6 dust grains/m3. Cosmic dust contains some complex organic compounds (amorphous organic solids with a mixed aromatic–aliphatic structure) that could be created naturally, and rapidly, by stars. A smaller fraction of dust in space is "stardust" consisting of larger refractory minerals that condensed as matter left by stars. Interstellar dust particles were collected by the Stardust spacecraft and samples were returned to Earth in 2006. Study and importance Cosmic dust was once solely an annoyance to astronomers, as it obscures objects they wished to observe. When infrared astronomy began, the dust particles were observed to be significant and vital components of astrophysical processes. Their analysis can reveal information about phenomena like the formation of the Solar System. For example, cosmic dust can drive the mass loss when a star is nearing the end of its life, play a part in the early stages of star formation, and form planets. In the Solar System, dust plays a major role in the zodiacal light, Saturn's B Ring spokes, the outer diffuse planetary rings at Jupiter, Saturn, Uranus and Neptune, and comets.
Cosmic dust
Wikipedia
460
2178570
https://en.wikipedia.org/wiki/Cosmic%20dust
Physical sciences
Basics_2
Astronomy
The interdisciplinary study of dust brings together different scientific fields: physics (solid-state, electromagnetic theory, surface physics, statistical physics, thermal physics), fractal mathematics, surface chemistry on dust grains, meteoritics, as well as every branch of astronomy and astrophysics. These disparate research areas can be linked by the following theme: the cosmic dust particles evolve cyclically; chemically, physically and dynamically. The evolution of dust traces out paths in which the Universe recycles material, in processes analogous to the daily recycling steps with which many people are familiar: production, storage, processing, collection, consumption, and discarding. Observations and measurements of cosmic dust in different regions provide an important insight into the Universe's recycling processes; in the clouds of the diffuse interstellar medium, in molecular clouds, in the circumstellar dust of young stellar objects, and in planetary systems such as the Solar System, where astronomers consider dust as in its most recycled state. The astronomers accumulate observational ‘snapshots’ of dust at different stages of its life and, over time, form a more complete movie of the Universe's complicated recycling steps. Parameters such as the particle's initial motion, material properties, intervening plasma and magnetic field determined the dust particle's arrival at the dust detector. Slightly changing any of these parameters can give significantly different dust dynamical behavior. Therefore, one can learn about where that object came from, and what is (in) the intervening medium. Detection methods A wide range of methods is available to study cosmic dust. Cosmic dust can be detected by remote sensing methods that utilize the radiative properties of cosmic dust particles, c.f. Zodiacal light measurement. Cosmic dust can also be detected directly ('in-situ') using a variety of collection methods and from a variety of collection locations. Estimates of the daily influx of extraterrestrial material entering the Earth's atmosphere range between 5 and 300 tonnes. NASA collects samples of star dust particles in the Earth's atmosphere using plate collectors under the wings of stratospheric-flying airplanes. Dust samples are also collected from surface deposits on the large Earth ice-masses (Antarctica and Greenland/the Arctic) and in deep-sea sediments.
Cosmic dust
Wikipedia
459
2178570
https://en.wikipedia.org/wiki/Cosmic%20dust
Physical sciences
Basics_2
Astronomy
Don Brownlee at the University of Washington in Seattle first reliably identified the extraterrestrial nature of collected dust particles in the latter 1970s. Another source is the meteorites, which contain stardust extracted from them. Stardust grains are solid refractory pieces of individual presolar stars. They are recognized by their extreme isotopic compositions, which can only be isotopic compositions within evolved stars, prior to any mixing with the interstellar medium. These grains condensed from the stellar matter as it cooled while leaving the star. In interplanetary space, dust detectors on planetary spacecraft have been built and flown, some are presently flying, and more are presently being built to fly. The large orbital velocities of dust particles in interplanetary space (typically 10–40 km/s) make intact particle capture problematic. Instead, in-situ dust detectors are generally devised to measure parameters associated with the high-velocity impact of dust particles on the instrument, and then derive physical properties of the particles (usually mass and velocity) through laboratory calibration (i.e., impacting accelerated particles with known properties onto a laboratory replica of the dust detector). Over the years dust detectors have measured, among others, the impact light flash, acoustic signal and impact ionisation. Recently the dust instrument on Stardust captured particles intact in low-density aerogel. Dust detectors in the past flew on the HEOS 2, Helios, Pioneer 10, Pioneer 11, Giotto, Galileo, Ulysses and Cassini space missions, on the Earth-orbiting LDEF, EURECA, and Gorid satellites, and some scientists have utilized the Voyager 1 and 2 spacecraft as giant Langmuir probes to directly sample the cosmic dust. Presently dust detectors are flying on the Ulysses, Proba, Rosetta, Stardust, and the New Horizons spacecraft. The collected dust at Earth or collected further in space and returned by sample-return space missions is then analyzed by dust scientists in their respective laboratories all over the world. One large storage facility for cosmic dust exists at the NASA Houston JSC.
Cosmic dust
Wikipedia
427
2178570
https://en.wikipedia.org/wiki/Cosmic%20dust
Physical sciences
Basics_2
Astronomy
Infrared light can penetrate cosmic dust clouds, allowing us to peer into regions of star formation and the centers of galaxies. NASA's Spitzer Space Telescope was the largest infrared space telescope, before the launch of the James Webb Space Telescope. During its mission, Spitzer obtained images and spectra by detecting the thermal radiation emitted by objects in space between wavelengths of 3 and 180 micrometres. Most of this infrared radiation is blocked by the Earth's atmosphere and cannot be observed from the ground. Findings from the Spitzer have revitalized the studies of cosmic dust. One report showed some evidence that cosmic dust is formed near a supermassive black hole. Another detection mechanism is polarimetry. Dust grains are not spherical and tend to align to interstellar magnetic fields, preferentially polarizing starlight that passes through dust clouds. In nearby interstellar space, where interstellar reddening is not intense enough to be detected, high precision optical polarimetry has been used to glean the structure of dust within the Local Bubble. In 2019, researchers found interstellar dust in Antarctica which they relate to the Local Interstellar Cloud. The detection of interstellar dust in Antarctica was done by the measurement of the radionuclides iron-60 and manganese-53 by highly sensitive Accelerator mass spectrometry. Radiation properties A dust particle interacts with electromagnetic radiation in a way that depends on its cross section, the wavelength of the electromagnetic radiation, and on the nature of the grain: its refractive index, size, etc. The radiation process for an individual grain is called its emissivity, dependent on the grain's efficiency factor. Further specifications regarding the emissivity process include extinction, scattering, absorption, or polarisation. In the radiation emission curves, several important signatures identify the composition of the emitting or absorbing dust particles. Dust particles can scatter light nonuniformly. Forward scattered light is light that is redirected slightly off its path by diffraction, and back-scattered light is reflected light. The scattering and extinction ("dimming") of the radiation gives useful information about the dust grain sizes. For example, if the in one's data is many times brighter in forward-scattered visible light than in back-scattered visible light, then it is understood that a significant fraction of the particles are about a micrometer in diameter.
Cosmic dust
Wikipedia
480
2178570
https://en.wikipedia.org/wiki/Cosmic%20dust
Physical sciences
Basics_2
Astronomy
The scattering of light from dust grains in long exposure visible photographs is quite noticeable in reflection nebulae, and gives clues about the individual particle's light-scattering properties. In X-ray wavelengths, many scientists are investigating the scattering of X-rays by interstellar dust, and some have suggested that astronomical X-ray sources would possess diffuse haloes, due to the dust. Presolar grains Presolar grains are contained within meteorites, from which they are extracted in terrestrial laboratories. The term "stardust" or "presolar stardust" is sometimes used to distinguish grains from a single star in comparison to aggregated interstellar dust particles, though this distinction is not universally applied. Presolar material was a component of the dust in the interstellar medium before its incorporation into meteorites. The meteorites have stored those presolar grains ever since the meteorites first assembled within the planetary accretion disk more than four billion years ago. Carbonaceous chondrites are especially fertile reservoirs of presolar material. Presolar grains definitionally existed before the Earth was formed. Presolar grain (and, less frequently, "stardust" or "presolar stardust") is the scientific term referring to refractory dust grains that condensed from cooling ejected gases from individual presolar stars and incorporated into the cloud from which the Solar System condensed. Many different types of presolar grains have been identified by laboratory measurements of the highly unusual isotopic composition of the chemical elements that comprise each presolar grain. These refractory mineral grains may earlier have been coated with volatile compounds, but those are lost in the dissolving of meteorite matter in acids, leaving only insoluble refractory minerals. Finding the grain cores without dissolving most of the meteorite has been possible, but difficult and labor-intensive. Many new aspects of nucleosynthesis have been discovered from the isotopic ratios within the presolar grains. An important property of presolar is the hard, refractory, high-temperature nature of the grains. Prominent are silicon carbide, graphite, aluminium oxide, aluminium spinel, and other such solids that would condense at high temperature from a cooling gas, such as in stellar winds or in the decompression of the inside of a supernova. They differ greatly from the solids formed at low temperature within the interstellar medium.
Cosmic dust
Wikipedia
497
2178570
https://en.wikipedia.org/wiki/Cosmic%20dust
Physical sciences
Basics_2
Astronomy
Also important are their extreme isotopic compositions, which are expected to exist nowhere in the interstellar medium. This also suggests that the presolar grains condensed from the gases of individual stars before the isotopes could be diluted by mixing with the interstellar medium. These allow the source stars to be identified. For example, the heavy elements within the silicon carbide (SiC) grains are almost pure S-process isotopes, fitting their condensation within AGB star red giant winds inasmuch as the AGB stars are the main source of S-process nucleosynthesis and have atmospheres observed by astronomers to be highly enriched in dredged-up s process elements. Another dramatic example is given by supernova condensates, usually shortened by acronym to SUNOCON (from SUperNOva CONdensate) to distinguish them from other grains condensed within stellar atmospheres. SUNOCONs contain in their calcium an excessively large abundance of 44Ca, demonstrating that they condensed containing abundant radioactive 44Ti, which has a 65-year half-life. The outflowing 44Ti nuclei were thus still "alive" (radioactive) when the SUNOCON condensed near one year within the expanding supernova interior, but would have become an extinct radionuclide (specifically 44Ca) after the time required for mixing with the interstellar gas. Its discovery proved the prediction from 1975 that it might be possible to identify SUNOCONs in this way. The SiC SUNOCONs (from supernovae) are only about 1% as numerous as are SiC stardust from AGB stars. Stardust itself (SUNOCONs and AGB grains that come from specific stars) is but a modest fraction of the condensed cosmic dust, forming less than 0.1% of the mass of total interstellar solids. The high interest in presolar grains derives from new information that it has brought to the sciences of stellar evolution and nucleosynthesis. Laboratories have studied solids that existed before the Earth was formed. This was once thought impossible, especially in the 1970s when cosmochemists were confident that the Solar System began as a hot gas virtually devoid of any remaining solids, which would have been vaporized by high temperature. The existence of presolar grains proved this historic picture incorrect.
Cosmic dust
Wikipedia
467
2178570
https://en.wikipedia.org/wiki/Cosmic%20dust
Physical sciences
Basics_2
Astronomy
Some bulk properties Cosmic dust is made of dust grains and aggregates into dust particles. These particles are irregularly shaped, with porosity ranging from fluffy to compact. The composition, size, and other properties depend on where the dust is found, and conversely, a compositional analysis of a dust particle can reveal much about the dust particle's origin. General diffuse interstellar medium dust, dust grains in dense clouds, planetary rings dust, and circumstellar dust, are each different in their characteristics. For example, grains in dense clouds have acquired a mantle of ice and on average are larger than dust particles in the diffuse interstellar medium. Interplanetary dust particles (IDPs) are generally larger still. Most of the influx of extraterrestrial matter that falls onto the Earth is dominated by meteoroids with diameters in the range 50 to 500 micrometers, of average density 2.0 g/cm3 (with porosity about 40%). The total influx rate of meteoritic sites of most IDPs captured in the Earth's stratosphere range between 1 and 3 g/cm3, with an average density at about 2.0 g/cm3. Other specific dust properties: in circumstellar dust, astronomers have found molecular signatures of CO, silicon carbide, amorphous silicate, polycyclic aromatic hydrocarbons, water ice, and polyformaldehyde, among others (in the diffuse interstellar medium, there is evidence for silicate and carbon grains). Cometary dust is generally different (with overlap) from asteroidal dust. Asteroidal dust resembles carbonaceous chondritic meteorites. Cometary dust resembles interstellar grains which can include silicates, polycyclic aromatic hydrocarbons, and water ice. In September 2020, evidence was presented of solid-state water in the interstellar medium, and particularly, of water ice mixed with silicate grains in cosmic dust grains. Dust grain formation
Cosmic dust
Wikipedia
409
2178570
https://en.wikipedia.org/wiki/Cosmic%20dust
Physical sciences
Basics_2
Astronomy
The large grains in interstellar space are probably complex, with refractory cores that condensed within stellar outflows topped by layers acquired during incursions into cold dense interstellar clouds. That cyclic process of growth and destruction outside of the clouds has been modeled to demonstrate that the cores live much longer than the average lifetime of dust mass. Those cores mostly start with silicate particles condensing in the atmospheres of cool, oxygen-rich red-giants and carbon grains condensing in the atmospheres of cool carbon stars. Red giants have evolved or altered off the main sequence and have entered the giant phase of their evolution and are the major source of refractory dust grain cores in galaxies. Those refractory cores are also called stardust (section above), which is a scientific term for the small fraction of cosmic dust that condensed thermally within stellar gases as they were ejected from the stars. Several percent of refractory grain cores have condensed within expanding interiors of supernovae, a type of cosmic decompression chamber. Meteoriticists who study refractory stardust (extracted from meteorites) often call it presolar grains but that within meteorites is only a small fraction of all presolar dust. Stardust condenses within the stars via considerably different condensation chemistry than that of the bulk of cosmic dust, which accretes cold onto preexisting dust in dark molecular clouds of the galaxy. Those molecular clouds are very cold, typically less than 50K, so that ices of many kinds may accrete onto grains, in cases only to be destroyed or split apart by radiation and sublimation into a gas component. Finally, as the Solar System formed many interstellar dust grains were further modified by coalescence and chemical reactions in the planetary accretion disk. The history of the various types of grains in the early Solar System is complicated and only partially understood. Astronomers know that the dust is formed in the envelopes of late-evolved stars from specific observational signatures. In infrared light, emission at 9.7 micrometres is a signature of silicate dust in cool evolved oxygen-rich giant stars. Emission at 11.5 micrometres indicates the presence of silicon carbide dust in cool evolved carbon-rich giant stars. These help provide evidence that the small silicate particles in space came from the ejected outer envelopes of these stars.
Cosmic dust
Wikipedia
492
2178570
https://en.wikipedia.org/wiki/Cosmic%20dust
Physical sciences
Basics_2
Astronomy
Conditions in interstellar space are generally not suitable for the formation of silicate cores. This would take excessive time to accomplish, even if it might be possible. The arguments are that: given an observed typical grain diameter a, the time for a grain to attain a, and given the temperature of interstellar gas, it would take considerably longer than the age of the Universe for interstellar grains to form. On the other hand, grains are seen to have recently formed in the vicinity of nearby stars, in nova and supernova ejecta, and in R Coronae Borealis variable stars which seem to eject discrete clouds containing both gas and dust. So mass loss from stars is unquestionably where the refractory cores of grains formed. Most dust in the Solar System is highly processed dust, recycled from the material out of which the Solar System formed and subsequently collected in the planetesimals, and leftover solid material such as comets and asteroids, and reformed in each of those bodies' collisional lifetimes. During the Solar System's formation history, the most abundant element was (and still is) H2. The metallic elements: magnesium, silicon, and iron, which are the principal ingredients of rocky planets, condensed into solids at the highest temperatures of the planetary disk. Some molecules such as CO, N2, NH3, and free oxygen, existed in a gas phase. Some molecules, for example, graphite (C) and SiC would condense into solid grains in the planetary disk; but carbon and SiC grains found in meteorites are presolar based on their isotopic compositions, rather than from the planetary disk formation. Some molecules also formed complex organic compounds and some molecules formed frozen ice mantles, of which either could coat the "refractory" (Mg, Si, Fe) grain cores. Stardust once more provides an exception to the general trend, as it appears to be totally unprocessed since its thermal condensation within stars as refractory crystalline minerals. The condensation of graphite occurs within supernova interiors as they expand and cool, and do so even in gas containing more oxygen than carbon, a surprising carbon chemistry made possible by the intense radioactive environment of supernovae. This special example of dust formation has merited specific review.
Cosmic dust
Wikipedia
469
2178570
https://en.wikipedia.org/wiki/Cosmic%20dust
Physical sciences
Basics_2
Astronomy
Planetary disk formation of precursor molecules was determined, in large part, by the temperature of the solar nebula. Since the temperature of the solar nebula decreased with heliocentric distance, scientists can infer a dust grain's with knowledge of the grain's materials. Some materials could only have been formed at high temperatures, while other grain materials could only have been formed at much lower temperatures. The materials in a single interplanetary dust particle often show that the grain elements formed in different locations and at different times in the solar nebula. Most of the matter present in the original solar nebula has since disappeared; drawn into the Sun, expelled into interstellar space, or reprocessed, for example, as part of the planets, asteroids or comets. Due to their highly processed nature, IDPs (interplanetary dust particles) are fine-grained mixtures of thousands to millions of mineral grains and amorphous components. We can picture an IDP as a "matrix" of material with embedded elements which were formed at different times and places in the solar nebula and before the solar nebula's formation. Examples of embedded elements in cosmic dust are GEMS, chondrules, and CAIs. From the solar nebula to Earth The arrows in the adjacent diagram show one possible path from a collected interplanetary dust particle back to the early stages of the solar nebula.
Cosmic dust
Wikipedia
277
2178570
https://en.wikipedia.org/wiki/Cosmic%20dust
Physical sciences
Basics_2
Astronomy
We can follow the trail to the right in the diagram to the IDPs that contain the most volatile and primitive elements. The trail takes us first from interplanetary dust particles to chondritic interplanetary dust particles. Planetary scientists classify chondritic IDPs in terms of their diminishing degree of oxidation so that they fall into three major groups: the carbonaceous, the ordinary, and the enstatite chondrites. As the name implies, the carbonaceous chondrites are rich in carbon, and many have anomalies in the isotopic abundances of H, C, N, and O. From the carbonaceous chondrites, we follow the trail to the most primitive materials. They are almost completely oxidized and contain the lowest condensation temperature elements ("volatile" elements) and the largest amount of organic compounds. Therefore, dust particles with these elements are thought to have been formed in the early life of the Solar System. The volatile elements have never seen temperatures above about 500 K, therefore, the IDP grain "matrix" consists of some very primitive Solar System material. Such a scenario is true in the case of comet dust. The provenance of the small fraction that is stardust (see above) is quite different; these refractory interstellar minerals thermally condense within stars, become a small component of interstellar matter, and therefore remain in the presolar planetary disk. Nuclear damage tracks are caused by the ion flux from solar flares. Solar wind ions impacting on the particle's surface produce amorphous radiation damaged rims on the particle's surface. And spallogenic nuclei are produced by galactic and solar cosmic rays. A dust particle that originates in the Kuiper Belt at 40 AU would have many more times the density of tracks, thicker amorphous rims and higher integrated doses than a dust particle originating in the main-asteroid belt. Based on 2012 computer model studies, the complex organic molecules necessary for life (extraterrestrial organic molecules) may have formed in the protoplanetary disk of dust grains surrounding the Sun before the formation of the Earth. According to the computer studies, this same process may also occur around other stars that acquire planets.
Cosmic dust
Wikipedia
457
2178570
https://en.wikipedia.org/wiki/Cosmic%20dust
Physical sciences
Basics_2
Astronomy
In September 2012, NASA scientists reported that polycyclic aromatic hydrocarbons (PAHs), subjected to interstellar medium (ISM) conditions, are transformed, through hydrogenation, oxygenation and hydroxylation, to more complex organics – "a step along the path toward amino acids and nucleotides, the raw materials of proteins and DNA, respectively". Further, as a result of these transformations, the PAHs lose their spectroscopic signature which could be one of the reasons "for the lack of PAH detection in interstellar ice grains, particularly the outer regions of cold, dense clouds or the upper molecular layers of protoplanetary disks." In February 2014, NASA announced a greatly upgraded database for detecting and monitoring polycyclic aromatic hydrocarbons (PAHs) in the universe. According to NASA scientists, over 20% of the carbon in the Universe may be associated with PAHs, possible starting materials for the formation of life. PAHs seem to have been formed shortly after the Big Bang, are abundant in the Universe, and are associated with new stars and exoplanets. In March 2015, NASA scientists reported that, for the first time, complex DNA and RNA organic compounds of life, including uracil, cytosine and thymine, have been formed in the laboratory under outer space conditions, using starting chemicals, such as pyrimidine, found in meteorites. Pyrimidine, like polycyclic aromatic hydrocarbons (PAHs), the most carbon-rich chemical found in the Universe, may have been formed in red giants or in interstellar dust and gas clouds, according to the scientists. Some "dusty" clouds in the universe The Solar System has its own interplanetary dust cloud, as do extrasolar systems. There are different types of nebulae with different physical causes and processes: diffuse nebula, infrared (IR) reflection nebula, supernova remnant, molecular cloud, HII regions, photodissociation regions, and dark nebula. Distinctions between those types of nebula are that different radiation processes are at work. For example, H II regions, like the Orion Nebula, where a lot of star-formation is taking place, are characterized as thermal emission nebulae. Supernova remnants, on the other hand, like the Crab Nebula, are characterized as nonthermal emission (synchrotron radiation).
Cosmic dust
Wikipedia
487
2178570
https://en.wikipedia.org/wiki/Cosmic%20dust
Physical sciences
Basics_2
Astronomy
Some of the better known dusty regions in the Universe are the diffuse nebulae in the Messier catalog, for example: M1, M8, M16, M17, M20, M42, M43. Some larger dust catalogs are Sharpless (1959) A Catalogue of HII Regions, Lynds (1965) Catalogue of Bright Nebulae, Lynds (1962) Catalogue of Dark Nebulae, van den Bergh (1966) Catalogue of Reflection Nebulae, Green (1988) Rev. Reference Cat. of Galactic SNRs, The National Space Sciences Data Center (NSSDC), and CDS Online Catalogs. Dust sample return The Discovery program's Stardust mission, was launched on 7 February 1999 to collect samples from the coma of comet Wild 2, as well as samples of cosmic dust. It returned samples to Earth on 15 January 2006. In 2007, the recovery of particles of interstellar dust from the samples was announced. Dust particles on Earth In 2017, Genge et al published a paper about "urban collection" of dust particles on Earth. The team were able to collect 500 micrometeorites from rooftops. Dust was collected in Oslo and in Paris, and "all particles are silicate-dominated (S type) cosmic spherules with subspherical shapes that form by melting during atmospheric entry and consist of quench crystals of magnesian olivine, relict crystals of forsterite, and iron-bearing olivine within glass". In the UK, scientists look for micrometeorites on the rooftops of cathedrals, like Canterbury Cathedral and Rochester Cathedral. Currently 40,000 tons of cosmic dust falls to Earth each year.
Cosmic dust
Wikipedia
346
2178570
https://en.wikipedia.org/wiki/Cosmic%20dust
Physical sciences
Basics_2
Astronomy
Deinotheriidae ("terrible beasts") is a family of prehistoric elephant-like proboscideans that lived during the Cenozoic era, first appearing in Africa during the Oligocene then spreading across Europe and the lower latitudes of Asia during the Miocene epoch. Their most distinctive features were their lack of upper tusks and downward-curving tusks on the lower jaw. Deinotheres were not very diverse; the only three known genera are Chilgatherium, Prodeinotherium, and Deinotherium. These form an evolutionary succession, with each new genus replacing the preceding one. Deinotheres were relatively conservative and showed little morphological change over their evolution, aside from a progressive increase in body size. Some species of Deinotherium are among the largest known land mammals ever, considerably exceeding modern elephants in size. The last members of Deinotherium persisted until the end of the Early Pleistocene in Africa, around 1 million years ago. Description The body shape and proportions of deinotheres were very much like those of modern elephants. The legs were long, like modern elephants, but the skull was rather flatter than that of true elephants. The upper jaw lacked incisor and canine teeth, but possessed five low-crowned molars on each side, with the same number in the lower jaw. Deinotheres used their front teeth for crushing their food, and the back teeth for shearing (slicing) the plant material. The front part of the lower jaw was turned downwards and bore the two tusk-like incisors. These curved downwards and backwards in a sort of huge hook and constituted the most distinct feature of the deinotheres. The tusks were used to strip vegetation rather than for digging. While the earliest deinothere Chilgatherium probably weighed only around and was less than tall, some species of Deinotherium represent among the largest known proboscideans, with shoulder heights of over and body masses around , considerably exceeding living African bush elephants in body size, making them among the largest land mammals ever.
Deinotheriidae
Wikipedia
429
2179965
https://en.wikipedia.org/wiki/Deinotheriidae
Biology and health sciences
Proboscidea
Animals
Ecology Deinotheres were "shearing browsers" adapted for feeding on plants above ground level. The way they chewed their food was probably similar to that of modern tapirs, with the front teeth being used to crush the food, while the second and third molars have a strong vertical shearing action, with little lateral (side-to-side) movement. This chewing action differs from both that of gomphotheres (lateral grinding) and elephants (horizontal shearing). Deinothere molars show little wear, indicating a diet of soft, non-gritty, forest vegetation, with the down-turned lower tusks being used for stripping bark or other vegetation. Deinotherium giganteum has a more elongated lower fore limb than early and middle Miocene Prodeinotherium, indicating a more efficient stride as an adaptation to the spread of savannas in Europe during the late Miocene. Deinotheres probably migrated from forest to forest, traversing the wide and (to them) useless grasslands. Evolutionary history Deinotheriids are thought to have diverged away from the ancestors of Elephantiformes during the Eocene, over 40 millon years ago, based on the presence of primitive Elephantiformes in Lutetian deposits. Phylogeny of Proboscidea showing placement of Deinotheriidae, following Hautier et al. 2021:The oldest known deinothere is Chilgatherium harrisi from the late Oligocene, around 27-28 million years ago. Its fossil remains have been found in the district of Chilga in Ethiopia (hence the name). It is primarily known from tooth remains. By the early Miocene, deinotheres had grown to the size of a small elephant and had migrated to Eurasia. Several species are known, all belonging to the genus Prodeinotherium. During the late middle Miocene, these modest-seized proboscideans were replaced by much larger forms across Eurasia. In Europe, Prodeinotherium bavaricum appeared in the early Miocene mammal faunal zone MN 4, but was soon replaced by Deinotherium giganteum in the middle Miocene. Likewise in Asia, Prodeinotherium is known from the early Miocene strata in the Bugti Hills, and continued into the middle Miocene Chinji Formation, where it was replaced by D. indicum.
Deinotheriidae
Wikipedia
489
2179965
https://en.wikipedia.org/wiki/Deinotheriidae
Biology and health sciences
Proboscidea
Animals
While these Miocene deinotheres were dispersed widely and evolved to huge elephant sizes, they were not as common as the contemporary (but smaller) Elephantoidea. Fossil remains of this age are known from the France, Germany, Greece, Malta, and northern India and Pakistan. These consist chiefly of teeth and the bones of the skull. After the extinction of the paraceratheres at the Oligocene-Miocene transition, the deinotheres were (and remained) the largest animals walking the Earth. The late Miocene was the heyday of the giant deinotheres. D. giganteum was common from Vallesian and Turolian localities in Europe. Prodeinotherium, which was reasonably well represented in the early Miocene of Africa, was succeeded by D. bozasi at the beginning of the late Miocene. And in Asia, D. indicum was most common in the late-Miocene Dhok Pathan Formation. Fossil teeth of D. giganteum, from the late-Miocene Sinap Formation at the Turkish site of Kayadibi are larger than those from older localities, such as Eppelsheim, Wissberg, and Montredon, indicating a tendency for increasing size of members of the species over time. These were the biggest animals of their day, protected from both predators and rival herbivores by virtue of their huge bulk. The largest mammoths did not approach them in size until the Pleistocene. With the end of the Miocene, deinothere fortunes declined. D. indicum died out about 7 million years ago, possibly driven to extinction by the same process of climate change that had previously eliminated the even more enormous Paraceratherium. While in Europe, D. giganteum continued, albeit with dwindling numbers, until the middle Pliocene; the most recent specimen is from Romania. In its original African homeland, Deinotherium continued to flourish throughout the Pliocene, and fossils have been uncovered at several of the African sites where remains of hominids have also been found. The last deinothere species to become extinct was D. bozasi. The youngest known specimens are from the Kanjera Formation, Kenya, about 1 million years ago (early Pleistocene). The causes of the extinction of such a successful and long-lived animal are not known, although a small number of other species of African megafauna also died out at this time.
Deinotheriidae
Wikipedia
496
2179965
https://en.wikipedia.org/wiki/Deinotheriidae
Biology and health sciences
Proboscidea
Animals
In statistical mechanics, Boltzmann's equation (also known as the Boltzmann–Planck equation) is a probability equation relating the entropy , also written as , of an ideal gas to the multiplicity (commonly denoted as or ), the number of real microstates corresponding to the gas's macrostate: where is the Boltzmann constant (also written as simply ) and equal to 1.380649 × 10−23 J/K, and is the natural logarithm function (or log base e, as in the image above). In short, the Boltzmann formula shows the relationship between entropy and the number of ways the atoms or molecules of a certain kind of thermodynamic system can be arranged. History The equation was originally formulated by Ludwig Boltzmann between 1872 and 1875, but later put into its current form by Max Planck in about 1900. To quote Planck, "the logarithmic connection between entropy and probability was first stated by L. Boltzmann in his kinetic theory of gases". A 'microstate' is a state specified in terms of the constituent particles of a body of matter or radiation that has been specified as a macrostate in terms of such variables as internal energy and pressure. A macrostate is experimentally observable, with at least a finite extent in spacetime. A microstate can be instantaneous, or can be a trajectory composed of a temporal progression of instantaneous microstates. In experimental practice, such are scarcely observable. The present account concerns instantaneous microstates. The value of was originally intended to be proportional to the Wahrscheinlichkeit (the German word for probability) of a macroscopic state for some probability distribution of possible microstates—the collection of (unobservable microscopic single particle) "ways" in which the (observable macroscopic) thermodynamic state of a system can be realized by assigning different positions and momenta to the respective molecules.
Boltzmann's entropy formula
Wikipedia
417
9328562
https://en.wikipedia.org/wiki/Boltzmann%27s%20entropy%20formula
Physical sciences
Thermodynamics
Physics
There are many instantaneous microstates that apply to a given macrostate. Boltzmann considered collections of such microstates. For a given macrostate, he called the collection of all possible instantaneous microstates of a certain kind by the name monode, for which Gibbs' term ensemble is used nowadays. For single particle instantaneous microstates, Boltzmann called the collection an ergode. Subsequently, Gibbs called it a microcanonical ensemble, and this name is widely used today, perhaps partly because Bohr was more interested in the writings of Gibbs than of Boltzmann. Interpreted in this way, Boltzmann's formula is the most basic formula for the thermodynamic entropy. Boltzmann's paradigm was an ideal gas of identical particles, of which are in the -th microscopic condition (range) of position and momentum. For this case, the probability of each microstate of the system is equal, so it was equivalent for Boltzmann to calculate the number of microstates associated with a macrostate. was historically misinterpreted as literally meaning the number of microstates, and that is what it usually means today. can be counted using the formula for permutations where ranges over all possible molecular conditions and "" denotes factorial. The "correction" in the denominator is due to the fact that identical particles in the same condition are indistinguishable. is sometimes called the "thermodynamic probability" since it is an integer greater than one, while mathematical probabilities are always numbers between zero and one. Introduction of the natural logarithm In Boltzmann’s 1877 paper, he clarifies molecular state counting to determine the state distribution number introducing the logarithm to simplify the equation. Boltzmann writes: “The first task is to determine the permutation number, previously designated by 𝒫 , for any state distribution. Denoting by J the sum of the permutations 𝒫 for all possible state distributions, the quotient 𝒫 /J is the state distribution’s probability, henceforth denoted by W. We would first like to calculate the permutations 𝒫 for the state distribution characterized by w0 molecules with kinetic energy 0, w1 molecules with kinetic energy ϵ, etc. …
Boltzmann's entropy formula
Wikipedia
479
9328562
https://en.wikipedia.org/wiki/Boltzmann%27s%20entropy%20formula
Physical sciences
Thermodynamics
Physics
“The most likely state distribution will be for those w0, w1 … values for which 𝒫 is a maximum or since the numerator is a constant, for which the denominator is a minimum. The values w0, w1 must simultaneously satisfy the two constraints (1) and (2). Since the denominator of 𝒫 is a product, it is easiest to determine the minimum of its logarithm, …” Therefore, by making the denominator small, he maximizes the number of states. So to simplify the product of the factorials, he uses their natural logarithm to add them. This is the reason for the natural logarithm in Boltzmann’s entropy formula. Generalization Boltzmann's formula applies to microstates of a system, each possible microstate of which is presumed to be equally probable. But in thermodynamics, the universe is divided into a system of interest, plus its surroundings; then the entropy of Boltzmann's microscopically specified system can be identified with the system entropy in classical thermodynamics. The microstates of such a thermodynamic system are not equally probable—for example, high energy microstates are less probable than low energy microstates for a thermodynamic system kept at a fixed temperature by allowing contact with a heat bath. For thermodynamic systems where microstates of the system may not have equal probabilities, the appropriate generalization, called the Gibbs entropy, is: This reduces to equation () if the probabilities pi are all equal. Boltzmann used a formula as early as 1866. He interpreted as a density in phase space—without mentioning probability—but since this satisfies the axiomatic definition of a probability measure we can retrospectively interpret it as a probability anyway. Gibbs gave an explicitly probabilistic interpretation in 1878. Boltzmann himself used an expression equivalent to () in his later work and recognized it as more general than equation (). That is, equation () is a corollary of equation ()—and not vice versa. In every situation where equation () is valid, equation () is valid also—and not vice versa. Boltzmann entropy excludes statistical dependencies
Boltzmann's entropy formula
Wikipedia
479
9328562
https://en.wikipedia.org/wiki/Boltzmann%27s%20entropy%20formula
Physical sciences
Thermodynamics
Physics
The term Boltzmann entropy is also sometimes used to indicate entropies calculated based on the approximation that the overall probability can be factored into an identical separate term for each particle—i.e., assuming each particle has an identical independent probability distribution, and ignoring interactions and correlations between the particles. This is exact for an ideal gas of identical particles that move independently apart from instantaneous collisions, and is an approximation, possibly a poor one, for other systems. The Boltzmann entropy is obtained if one assumes one can treat all the component particles of a thermodynamic system as statistically independent. The probability distribution of the system as a whole then factorises into the product of N separate identical terms, one term for each particle; and when the summation is taken over each possible state in the 6-dimensional phase space of a single particle (rather than the 6N-dimensional phase space of the system as a whole), the Gibbs entropy simplifies to the Boltzmann entropy . This reflects the original statistical entropy function introduced by Ludwig Boltzmann in 1872. For the special case of an ideal gas it exactly corresponds to the proper thermodynamic entropy. For anything but the most dilute of real gases, leads to increasingly wrong predictions of entropies and physical behaviours, by ignoring the interactions and correlations between different molecules. Instead one must consider the ensemble of states of the system as a whole, called by Boltzmann a holode, rather than single particle states. Gibbs considered several such kinds of ensembles; relevant here is the canonical one.
Boltzmann's entropy formula
Wikipedia
324
9328562
https://en.wikipedia.org/wiki/Boltzmann%27s%20entropy%20formula
Physical sciences
Thermodynamics
Physics
An induction generator or asynchronous generator is a type of alternating current (AC) electrical generator that uses the principles of induction motors to produce electric power. Induction generators operate by mechanically turning their rotors faster than synchronous speed. A regular AC induction motor usually can be used as a generator, without any internal modifications. Because they can recover energy with relatively simple controls, induction generators are useful in applications such as mini hydro power plants, wind turbines, or in reducing high-pressure gas streams to lower pressure. An induction generator draws reactive excitation current from an external source. Induction generators have an AC rotor and cannot bootstrap using residual magnetization to black start a de-energized distribution system as synchronous machines do. Power factor correcting capacitors can be added externally to neutralize a constant amount of the variable reactive excitation current. After starting, an induction generator can use a capacitor bank to produce reactive excitation current, but the isolated power system's voltage and frequency are not self-regulating and destabilize readily. Principle of Operation An induction generator produces electrical power when its rotor is turned faster than the synchronous speed. For a four-pole motor (two pairs of poles on stator) powered by a 60 Hz source, the synchronous speed is 1800 rotations per minute (rpm) and 1500 RPM powered at 50 Hz. The motor always turns slightly slower than the synchronous speed. The difference between synchronous and operating speed is called "slip" and is often expressed as percent of the synchronous speed. For example, a motor operating at 1450 RPM that has a synchronous speed of 1500 RPM is running at a slip of +3.3%. In operation as a motor, the stator flux rotation is at the synchronous speed, which is faster than the rotor speed. This causes the stator flux to cycle at the slip frequency inducing rotor current through the mutual inductance between the stator and rotor. The induced current create a rotor flux with magnetic polarity opposite to the stator. In this way, the rotor is dragged along behind stator flux, with the currents in the rotor induced at the slip frequency. The motor runs at the speed where the induced rotor current gives rise to torque equal to the shaft load.
Induction generator
Wikipedia
488
9331066
https://en.wikipedia.org/wiki/Induction%20generator
Technology
Power generation
null
In generator operation, a prime mover (turbine or engine) drives the rotor above the synchronous speed (negative slip). The stator flux induces current in the rotor, but the opposing rotor flux is now cutting the stator coils, a current is induced in the stator coils 270° behind the magnetizing current, in phase with magnetizing voltage. The motor delivers real (in-phase) power to the power system. Excitation An induction motor requires an externally supplied current to the stator windings in order to induce a current in the rotor. Because the current in an inductor is integral of the voltage with respect to time, for a sinusoidal voltage waveform the current lags the voltage by 90°, and the induction motor always consumes reactive power, regardless of whether it is consuming electrical power and delivering mechanical power as a motor or consuming mechanical power and delivering electrical power to the system. A source of excitation current for magnetizing flux (reactive power) for the stator is still required, to induce rotor current. This can be supplied from the electrical grid or, once it starts producing power, from a capacitive reactance. The generating mode for induction motors is complicated by the need to excite the rotor, which being induced by an alternating current is demagnetized at shutdown with no residual magnetization to bootstrap a cold start. It is necessary to connect an external source of magnetizing current to initialize production. The power frequency and voltage are not self regulating. The generator is able to supply current out of phase with the voltage requiring more external equipment to build a functional isolated power system. Similar is the operation of the induction motor in parallel with a synchronous motor serving as a power factor compensator. A feature in the generator mode in parallel to the grid is that the rotor speed is higher than in the driving mode. Then active energy is being given to the grid. Another disadvantage of induction motor generator is that it consumes a significant magnetizing current I0 = (20-35)%.
Induction generator
Wikipedia
428
9331066
https://en.wikipedia.org/wiki/Induction%20generator
Technology
Power generation
null
Active Power Active power delivered to the line is proportional to slip above the synchronous speed. Full rated power of the generator is reached at very small slip values (motor dependent, typically 3%). At synchronous speed of 1800 RPM, generator will produce no power. When the driving speed is increased to 1860 RPM (typical example), full output power is produced. If the prime mover is unable to produce enough power to fully drive the generator, speed will remain somewhere between 1800 and 1860 RPM range. Required Capacitance A capacitor bank must supply reactive power to the motor when used in stand-alone mode. The reactive power supplied should be equal or greater than the reactive power that the generator normally draws when operating as a motor. Torque vs. Slip The basic fundamental of induction generators is the conversion from mechanical energy to electrical energy. This requires an external torque applied to the rotor to turn it faster than the synchronous speed. However, indefinitely increasing torque doesn't lead to an indefinite increase in power generation. The rotating magnetic field torque excited from the armature works to counter the motion of the rotor and prevent over speed because of induced motion in the opposite direction. As the speed of the motor increases the counter torque reaches a max value of torque (breakdown torque) that it can operate until before the operating conditions become unstable. Ideally, induction generators work best in the stable region between the no-load condition and maximum torque region. Rating Current The maximum power that can be produced by an induction motor operated as a generator is limited by the rated current of the generator's windings. Grid and stand-alone connections In induction generators, the reactive power required to establish the air gap magnetic flux is provided by a capacitor bank connected to the machine in case of stand-alone system and in case of grid connection it draws reactive power from the grid to maintain its air gap flux. For a grid-connected system, frequency and voltage at the machine will be dictated by the electric grid, since it is very small compared to the whole system. For stand-alone systems, frequency and voltage are complex function of machine parameters, capacitance used for excitation, and load value and type. Uses Induction generators are often used in wind turbines and some micro hydro installations due to their ability to produce useful power at varying rotor speeds. Induction generators are mechanically and electrically simpler than other generator types. They are also more rugged, requiring no brushes or commutators.
Induction generator
Wikipedia
506
9331066
https://en.wikipedia.org/wiki/Induction%20generator
Technology
Power generation
null
Limitations An induction generator connected to a capacitor system can generate sufficient reactive power to operate independently. When the load current exceeds the capability of the generator to supply both magnetization reactive power and load power the generator will immediately cease to produce power. The load must be removed and the induction generator restarted with either an external DC motor or if present, residual magnetism in the core. Induction generators are particularly suitable for wind generating stations as in this case speed is always a variable factor. Unlike synchronous motors, induction generators are load-dependent and cannot be used alone for grid frequency control. Example application As an example, consider the use of a 10 hp, 1760 r/min, 440 V, three-phase induction motor (a.k.a. induction electrical machine in an asynchronous generator regime) as asynchronous generator. The full-load current of the motor is 10 A and the full-load power factor is 0.8. Required capacitance per phase if capacitors are connected in delta: Apparent power Active power Reactive power For a machine to run as an asynchronous generator, capacitor bank must supply minimum 4567 / 3 phases = 1523 VAR per phase. Voltage per capacitor is 440 V because capacitors are connected in delta. Capacitive current Ic = Q/E = 1523/440 = 3.46 A Capacitive reactance per phase Xc = E/Ic = 127 Ω Minimum capacitance per phase: C = 1 / (2*π*f*Xc) = 1 / (2 * 3.141 * 60 * 127) = 21 μF. If the load also absorbs reactive power, capacitor bank must be increased in size to compensate. Prime mover speed should be used to generate frequency of 60 Hz: Typically, slip should be similar to full-load value when machine is running as motor, but negative (generator operation): if Ns = 1800, one can choose N=Ns+40 rpm Required prime mover speed N = 1800 + 40 = 1840 rpm.
Induction generator
Wikipedia
434
9331066
https://en.wikipedia.org/wiki/Induction%20generator
Technology
Power generation
null
A/B testing (also known as bucket testing, split-run testing, or split testing) is a user experience research method. A/B tests consist of a randomized experiment that usually involves two variants (A and B), although the concept can be also extended to multiple variants of the same variable. It includes application of statistical hypothesis testing or "two-sample hypothesis testing" as used in the field of statistics. A/B testing is a way to compare multiple versions of a single variable, for example by testing a subject's response to variant A against variant B, and determining which of the variants is more effective. Multivariate testing or multinomial testing is similar to A/B testing, but may test more than two versions at the same time or use more controls. Simple A/B tests are not valid for observational, quasi-experimental or other non-experimental situations—commonplace with survey data, offline data, and other, more complex phenomena. Definition "A/B testing" is a shorthand for a simple randomized controlled experiment, in which a number of samples (e.g. A and B) of a single vector-variable are compared. A/B tests are widely considered the simplest form of controlled experiment, especially when they only involve two variants. However, by adding more variants to the test, its complexity grows. The following example illustrates an A/B test with a single variable: Suppose a company has a customer database of 2,000 people and decides to create an email campaign with a discount code in order to generate sales through its website. The company creates two versions of the email with different call to action (the part of the copy which encourages customers to do something — in the case of a sales campaign, make a purchase) and identifying promotional code. To 1,000 people it sends the email with the call to action stating, "Offer ends this Saturday! Use code A1", To the remaining 1,000 people, it sends the email with the call to action stating, "Offer ends soon! Use code B1". All other elements of the emails' copy and layout are identical.
A/B testing
Wikipedia
437
9332179
https://en.wikipedia.org/wiki/A/B%20testing
Mathematics
Statistics
null
The company then monitors which campaign has the higher success rate by analyzing the use of the promotional codes. The email using the code A1 has a 5% response rate (50 of the 1,000 people emailed used the code to buy a product), and the email using the code B1 has a 3% response rate (30 of the recipients used the code to buy a product). The company therefore determines that in this instance, the first Call To Action is more effective and will use it in future sales. A more nuanced approach would involve applying statistical testing to determine if the differences in response rates between A1 and B1 were statistically significant (that is, highly likely that the differences are real, repeatable, and not due to random chance). In the example above, the purpose of the test is to determine which is the more effective way to encourage customers to make a purchase. If, however, the aim of the test had been to see which email would generate the higher click-ratethat is, the number of people who actually click onto the website after receiving the emailthen the results might have been different. For example, even though more of the customers receiving the code B1 accessed the website, because the Call To Action didn't state the end-date of the promotion many of them may feel no urgency to make an immediate purchase. Consequently, if the purpose of the test had been simply to see which email would bring more traffic to the website, then the email containing code B1 might well have been more successful. An A/B test should have a defined outcome that is measurable such as number of sales made, click-rate conversion, or number of people signing up/registering. Common test statistics Two-sample hypothesis tests are appropriate for comparing the two samples where the samples are divided by the two control cases in the experiment. Z-tests are appropriate for comparing means under stringent conditions regarding normality and a known standard deviation. Student's t-tests are appropriate for comparing means under relaxed conditions when less is assumed. Welch's t test assumes the least and is therefore the most commonly used test in a two-sample hypothesis test where the mean of a metric is to be optimized. While the mean of the variable to be optimized is the most common choice of estimator, others are regularly used. For a comparison of two binomial distributions such as a click-through rate one would use Fisher's exact test.
A/B testing
Wikipedia
502
9332179
https://en.wikipedia.org/wiki/A/B%20testing
Mathematics
Statistics
null
Segmentation and targeting A/B tests most commonly apply the same variant (e.g., user interface element) with equal probability to all users. However, in some circumstances, responses to variants may be heterogeneous. That is, while a variant A might have a higher response rate overall, variant B may have an even higher response rate within a specific segment of the customer base. For instance, in the above example, the breakdown of the response rates by gender could have been: In this case, we can see that while variant A had a higher response rate overall, variant B actually had a higher response rate with men. As a result, the company might select a segmented strategy as a result of the A/B test, sending variant B to men and variant A to women in the future. In this example, a segmented strategy would yield an increase in expected response rates from to – constituting a 30% increase. If segmented results are expected from the A/B test, the test should be properly designed at the outset to be evenly distributed across key customer attributes, such as gender. That is, the test should both (a) contain a representative sample of men vs. women, and (b) assign men and women randomly to each “variant” (variant A vs. variant B). Failure to do so could lead to experiment bias and inaccurate conclusions to be drawn from the test. This segmentation and targeting approach can be further generalized to include multiple customer attributes rather than a single customer attributefor example, customers' age and genderto identify more nuanced patterns that may exist in the test results. Tradeoffs Positives The results of A/B tests are simple to interpret and use to get a clear idea of what users prefer, since it is directly testing one option over another. It is based on real user behavior, so the data can be very helpful especially when determining what works better between two options. A/B tests can also provide answers to highly specific design questions. One example of this is Google's A/B testing with hyperlink colors. In order to optimize revenue, they tested dozens of different hyperlink hues to see which color the users tend to click more on. Negatives
A/B testing
Wikipedia
456
9332179
https://en.wikipedia.org/wiki/A/B%20testing
Mathematics
Statistics
null
A/B tests are sensitive to variance; they require a large sample size in order to reduce standard error and produce a statistically significant result. In applications where active users are abundant, such as popular online social media platforms, obtaining a large sample size is trivial. In other cases, large sample sizes are obtained by increasing the experiment enrollment period. However, using a technique coined by Microsoft as Controlled-experiment Using Pre-Experiment Data (CUPED), variance from before the experiment start can be taken into account so that fewer samples are required to produce a statistically significant result. Due to its nature as an experiment, running an A/B test introduces the risk of wasted time and resources if the test produces unwanted results, such as negative or no impact to business metrics. In December 2018, representatives with experience in large-scale A/B testing from thirteen different organizations (Airbnb, Amazon, Booking.com, Facebook, Google, LinkedIn, Lyft, Microsoft, Netflix, Twitter, Uber, and Stanford University) summarized the top challenges in a SIGKDD Explorations paper. The challenges can be grouped into four areas: Analysis, Engineering and Culture, Deviations from Traditional A/B tests, and Data quality. History It is difficult to definitively establish when A/B testing was first used. The first randomized double-blind trial, to assess the effectiveness of a homeopathic drug, occurred in 1835. Experimentation with advertising campaigns, which has been compared to modern A/B testing, began in the early twentieth century. The advertising pioneer Claude Hopkins used promotional coupons to test the effectiveness of his campaigns. However, this process, which Hopkins described in his Scientific Advertising, did not incorporate concepts such as statistical significance and the null hypothesis, which are used in statistical hypothesis testing. Modern statistical methods for assessing the significance of sample data were developed separately in the same period. This work was done in 1908 by William Sealy Gosset when he altered the Z-test to create Student's t-test.
A/B testing
Wikipedia
413
9332179
https://en.wikipedia.org/wiki/A/B%20testing
Mathematics
Statistics
null
With the growth of the internet, new ways to sample populations have become available. Google engineers ran their first A/B test in the year 2000 in an attempt to determine what the optimum number of results to display on its search engine results page would be. The first test was unsuccessful due to glitches that resulted from slow loading times. Later A/B testing research would be more advanced, but the foundation and underlying principles generally remain the same, and in 2011, 11 years after Google's first test, Google ran over 7,000 different A/B tests. In 2012, a Microsoft employee working on the search engine Microsoft Bing created an experiment to test different ways of displaying advertising headlines. Within hours, the alternative format produced a revenue increase of 12% with no impact on user-experience metrics. Today, major software companies such as Microsoft and Google each conduct over 10,000 A/B tests annually. A/B testing has been claimed by some to be a change in philosophy and business-strategy in certain niches, though the approach is identical to a between-subjects design, which is commonly used in a variety of research traditions. A/B testing as a philosophy of web development brings the field into line with a broader movement toward evidence-based practice. Many companies now use the "designed experiment" approach to making marketing decisions, with the expectation that relevant sample results can improve positive conversion results. It is an increasingly common practice as the tools and expertise grow in this area. Applications A/B testing in online social media A/B tests have been used by large social media sites like LinkedIn, Facebook, and Instagram to understand user engagement and satisfaction of online features, such as a new feature or product. A/B tests have also been used to conduct complex experiments on subjects such as network effects when users are offline, how online services affect user actions, and how users influence one another. A/B testing for e-commerce On an e-commerce website, the purchase funnel is typically a good candidate for A/B testing, since even marginal-decreases in drop-off rates can represent a significant gain in sales. Significant improvements can be sometimes seen through testing elements like copy text, layouts, images and colors, but not always. In these tests, users only see one of two versions, since the goal is to discover which of the two versions is preferable.
A/B testing
Wikipedia
486
9332179
https://en.wikipedia.org/wiki/A/B%20testing
Mathematics
Statistics
null
A/B testing for product pricing A/B testing can be used to determine the right price for the product, as this is perhaps one of the most difficult tasks when a new product or service is launched. A/B testing (especially valid for digital goods) is an excellent way to find out which price-point and offering maximize the total revenue. Political A/B testing A/B tests have also been used by political campaigns. In 2007, Barack Obama's presidential campaign used A/B testing as a way to garner online attraction and understand what voters wanted to see from the presidential candidate. For example, Obama's team tested four distinct buttons on their website that led users to sign up for newsletters. Additionally, the team used six different accompanying images to draw in users. Through A/B testing, staffers were able to determine how to effectively draw in voters and garner additional interest. HTTP Routing and API feature testing A/B testing is very common when deploying a newer version of an API. For real-time user experience testing, an HTTP Layer-7 Reverse proxy is configured in such a way that, N% of the HTTP traffic goes into the newer version of the backend instance, while the remaining 100-N% of HTTP traffic hits the (stable) older version of the backend HTTP application service. This is usually done for limiting the exposure of customers to a newer backend instance such that, if there is a bug on the newer version, only N% of the total user agents or clients get affected while others get routed to a stable backend, which is a common ingress control mechanism.
A/B testing
Wikipedia
330
9332179
https://en.wikipedia.org/wiki/A/B%20testing
Mathematics
Statistics
null
A loupe ( ) is a simple, small magnification device used to see small details more closely. They generally have higher magnification than a magnifying glass, and are designed to be held or worn close to the eye. A loupe does not have an attached handle, and its focusing lens(es) are contained in an opaque cylinder or cone. On some loupes this cylinder folds into an enclosing housing that protects the lenses when not in use. Optics Three basic types of loupes exist: Simple lenses, generally used for low-magnification designs because of high optical aberration. Compound lenses, generally used for higher magnifications to control optical aberration. Prismatic, multiple lenses with prisms. Uses Loupes are used in many professions where magnification enables precision work to be done with greater efficiency and ease. Examples include surgery, dentistry, ophthalmology, the jewelry trade, gemology, questioned document examination and watchmaking. Loupes are also sometimes used in photography and printing. Jewellers and gemologists Jewellers typically use a monocular, handheld loupe to magnify gemstones and other jewelry that they wish to inspect. A 10× magnification is good to use for inspecting jewelry and hallmarks and is the Gemological Institute of America's standard for grading diamond clarity. Stones will sometimes be inspected at higher magnifications than 10×, although the depth of field and field of view become too small to be instructive. The accepted standard for grading diamonds is therefore that inclusions and blemishes visible at 10× impact the clarity grade. The inclusions in VVS diamonds are hard to find even at 10×. Watchmaking Loupes are employed to assist watchmakers in assembling mechanical watches. Many aspects require the use of the loupe, in particular the assembly of the watch mechanism itself, the assembly and details of the watch dial, as well as the formation of the watch strap and installation of precious stones onto the watch face.
Loupe
Wikipedia
415
11817652
https://en.wikipedia.org/wiki/Loupe
Technology
Optical instruments
null
Photography Analog (film) photographers use loupes to review, edit or analyze negatives and slides on a light table. Typical magnifications for viewing slides full-frame depend on image format; 35 mm frames (24×36 mm slides to 38×38 mm superslides) are best viewed at ca. 5×, while ca. 3× is optimal for viewing medium format slides (6×4.5 cm / 6×6 cm / 6×7 cm). Often, a 10× loupe is used to examine critical sharpness. Photographers using large format cameras may use a loupe to view the ground glass image to aid in focusing. Users of digital single-lens reflex cameras use loupes to help to identify dust and other particles on the sensor, in preparation for sensor cleaning. Dentistry Dentists, hygienists, and dental therapists typically use binocular loupe glasses since they need both hands free when performing dental procedures. The magnification helps with accurate diagnoses of oral conditions and enhances surgical precision when completing treatment. Additionally, loupes can improve dentists' posture which can decrease occupational strain. Some dental loupes are flip-type, which take the form of two small cylinders, one in front of each lens of the glasses. Other types are inset within the lens of the glasses. Dental caries, also known as cavities, are most accurately identified by visual and tactile examination of a clean, dry tooth. Magnification enables dentists to improve their ability to differentiate between a stain and a cavity. Cavities are rated and scored based on their visual presentation. If magnification is too high diagnosis becomes difficult due to the small field of view. Ideal magnification for diagnostic purposes is up to 2×. Treatment of dental caries, periodontal disease, and pulpal disease are all aided by magnification. The dental specialty of endodontics has performed the vast majority of research regarding magnification in dentistry. Because the identification of accessory canals in addition to the primary pulp canals is essential to complete nonsurgical root canal therapy, magnification provides dentists enhanced visualization to locate and treat more obscured canals.
Loupe
Wikipedia
451
11817652
https://en.wikipedia.org/wiki/Loupe
Technology
Optical instruments
null
Treatment of periodontal disease is achieved by removing calculus deposits, plaque and therefore bacteria which causes inflammation and subsequently bone destruction. In severe cases, surgery to reduce pocket depth is indicated. Periodontists and hygienists must visualize plaque and calculus to remove it. Magnification can assist dentists and hygienists with identification and removal of plaque and calculus in addition to improving visualization for periodontal surgery. Ergonomics have long been a pain point for doctors who need to physically strain, bending over and looking down, to treat their patients. Over time this posture results in discomfort, pain, and even neuromuscular disease. Some modern loupes address this by incorporating refractive prisms which alter the course of the light through the telescopes, so that the dentist can maintain a neutral, upright position with eyes relaxed and looking straight ahead. A typical magnification for use in dentistry is 2.5×, but dental loupes can be anywhere in the range from 2× to 8×. Optimal magnification is a function of the type of work the doctor does - namely, how much detail he or she needs to see, taking into consideration that when magnification increases, the field of view decreases. As a tool that sits on the face and is used for hours at a time, weight is also a significant factor in considering the type of loupes to use. Together with proper access to the oral cavity, light is an important part of performing precision dentistry. Because a dentist's head often eclipses the overhead dental lamp, loupes may be fitted with a light source. Loupe-mounted lights used to be fed by fiber optic cables that connected to either a wall-mounted or table-top light source. Newer models feature a more convenient LED lamp within the loupe-mounted light and an electric cord coming from either the conventional wall-mounted or table-top light source or a belt clip rechargeable battery pack. Options for loupe-mounted cameras and video recorders are also available. Surgery
Loupe
Wikipedia
417
11817652
https://en.wikipedia.org/wiki/Loupe
Technology
Optical instruments
null
Surgeons in many specialties commonly use loupes when doing surgery on delicate structures. The loupes used by surgeons are mounted in the lenses of glasses and are custom made for the individual surgeon, taking into account their corrected vision, interpupillary distance and desired focal distance. Multiple magnification powers are available. They are most commonly used in otolaryngology, neurosurgery, ophthalmology, plastic surgery, cardiac surgery, orthopedic surgery, and vascular surgery. Geology The loupe is a vital geological field tool used to identify small mineral crystals and structures in rocks. Collectables Loupes are an essential tool in both numismatics, the study of currency, and the related practice of coin collection. Coin collectors frequently employ loupes for better evaluation of the quality of their coins, since identifying surface wear is vital when attempting to classify the grade of a coin. Uncirculated coins (coins without wear) can command a substantial premium over coins with slight wear. This wear cannot always be seen with the naked eye. Numismatists can also employ loupes to identify some counterfeit coins that would pass a naked-eye visual inspection. Loupes are similarly used for evaluating other collectable objects, such as trading cards and antiques. Archival conservation Conservators often use hand held loupes or head-mounted binocular magnifiers such as the Optivisor to examine artifacts and documents requiring cleaning or repair.
Loupe
Wikipedia
299
11817652
https://en.wikipedia.org/wiki/Loupe
Technology
Optical instruments
null
In chemistry, a space-filling model, also known as a calotte model, is a type of three-dimensional (3D) molecular model where the atoms are represented by spheres whose radii are proportional to the radii of the atoms and whose center-to-center distances are proportional to the distances between the atomic nuclei, all in the same scale. Atoms of different chemical elements are usually represented by spheres of different colors. Space-filling calotte models are also referred to as CPK models after the chemists Robert Corey, Linus Pauling, and Walter Koltun, who over a span of time developed the modeling concept into a useful form. They are distinguished from other 3D representations, such as the ball-and-stick and skeletal models, by the use of the "full size" space-filling spheres for the atoms. The models are tactile and manually rotatable. They are useful for visualizing the effective shape and relative dimensions of a molecule, and (because of the rotatability) the shapes of the surface of the various conformers. On the other hand, these models mask the chemical bonds between the atoms, and make it difficult to see the structure of the molecule that is obscured by the atoms nearest to the viewer in a particular pose. For this reason, such models are of greater utility if they can be used dynamically, especially when used with complex molecules (e.g., see the greater understanding of the molecules shape given when the THC model is clicked on to rotate). History
Space-filling model
Wikipedia
312
7181923
https://en.wikipedia.org/wiki/Space-filling%20model
Physical sciences
Substance
Chemistry
Space-filling models arise out of a desire to represent molecules in ways that reflect the electronic surfaces that molecules present, that dictate how they interact, one with another (or with surfaces, or macromolecules such as enzymes, etc.). Crystallographic data are the starting point for understanding static molecular structure, and these data contain the information rigorously required to generate space-filling representations (e.g., see these crystallographic models); most often, however, crystallographers present the locations of atoms derived from crystallography via "thermal ellipsoids" whose cut-off parameters are set for convenience both to show the atom locations (with anisotropies), and to allow representation of the covalent bonds or other interactions between atoms as lines. In short, for reasons of utility, crystallographic data historically have appeared in presentations closer to ball-and-stick models. Hence, while crystallographic data contain the information to create space-filling models, it remained for individuals interested in modeling an effective static shape of a molecule, and the space it occupied, and the ways in which it might present a surface to another molecule, to develop the formalism shown above. In 1952, Robert Corey and Linus Pauling described accurate scale models of molecules which they had built at Caltech. In their models, they envisioned the surface of the molecule as being determined by the van der Waals radius of each atom of the molecule, and crafted atoms as hardwood spheres of diameter proportional to each atom's van der Waals radius, in the scale 1 inch = 1 Å. To allow bonds between atoms a portion of each sphere was cut away to create a pair of matching flat faces, with the cuts dimensioned so that the distance between sphere centers was proportional to the lengths of standard types of chemical bonds. A connector was designed—a metal bushing that threaded into each sphere at the center of each flat face. The two spheres were then firmly held together by a metal rod inserted into the pair of opposing bushing (with fastening by screws). The models also had special features to allow representation of hydrogen bonds.
Space-filling model
Wikipedia
440
7181923
https://en.wikipedia.org/wiki/Space-filling%20model
Physical sciences
Substance
Chemistry
In 1965, Walter L. Koltun designed and patented a simplified system with molded plastic atoms of various colours, which were joined by specially designed snap connectors; this simpler system accomplished essentially the same ends as the Corey-Pauling system, and allowed for the development of the models as a popular way of working with molecules in training and research environments. Such colour-coded, bond length-defined, van der Waals-type space-filling models are now commonly known as CPK models, after these three developers of the specific concept. In modern research efforts, attention returned to use of data-rich crystallographic models in combination with traditional and new computational methods to provide space-filling models of molecules, both simple and complex, where added information such as which portions of the surface of the molecule were readily accessible to solvent, or how the electrostatic characteristics of a space-filling representation—which in the CPK case is almost fully left to the imagination—could be added to the visual models created. The two closing images give examples of the latter type of calculation and representation, and its utility.
Space-filling model
Wikipedia
222
7181923
https://en.wikipedia.org/wiki/Space-filling%20model
Physical sciences
Substance
Chemistry
In photometry, luminous flux or luminous power is the measure of the perceived power of light. It differs from radiant flux, the measure of the total power of electromagnetic radiation (including infrared, ultraviolet, and visible light), in that luminous flux is adjusted to reflect the varying sensitivity of the human eye to different wavelengths of light. Units The SI unit of luminous flux is the lumen (lm). One lumen is defined as the luminous flux of light produced by a light source that emits one candela of luminous intensity over a solid angle of one steradian. In other systems of units, luminous flux may have units of power. Weighting The luminous flux accounts for the sensitivity of the eye by weighting the power at each wavelength with the luminosity function, which represents the eye's response to different wavelengths. The luminous flux is a weighted sum of the power at all wavelengths in the visible band. Light outside the visible band does not contribute. The ratio of the total luminous flux to the radiant flux is called the luminous efficacy. This model of the human visual brightness perception, is standardized by the CIE and ISO. Context Luminous flux is often used as an objective measure of the useful light emitted by a light source, and is typically reported on the packaging for light bulbs, although it is not always prominent. Consumers commonly compare the luminous flux of different light bulbs since it provides an estimate of the apparent amount of light the bulb will produce, and a lightbulb with a higher ratio of luminous flux to consumed power is more efficient. Luminous flux is not used to compare brightness, as this is a subjective perception which varies according to the distance from the light source and the angular spread of the light from the source. Measurement Luminous flux of artificial light sources is typically measured using an integrating sphere, or a goniophotometer outfitted with a photometer or a spectroradiometer. Relationship to luminous intensity
Luminous flux
Wikipedia
391
581888
https://en.wikipedia.org/wiki/Luminous%20flux
Physical sciences
Optics
Physics
Luminous flux (in lumens) is a measure of the total amount of light a lamp puts out. The luminous intensity (in candelas) is a measure of how bright the beam in a particular direction is. If a lamp has a 1 lumen bulb and the optics of the lamp are set up to focus the light evenly into a 1 steradian beam, then the beam would have a luminous intensity of 1 candela. If the optics were changed to concentrate the beam into 1/2 steradian then the source would have a luminous intensity of 2 candela. The resulting beam is narrower and brighter, however the luminous flux remains the same. Examples
Luminous flux
Wikipedia
137
581888
https://en.wikipedia.org/wiki/Luminous%20flux
Physical sciences
Optics
Physics
An antenna tuner, a matchbox, transmatch, antenna tuning unit (ATU), antenna coupler, or feedline coupler is a device connected between a radio transmitter or receiver and its antenna to improve power transfer between them by matching the impedance of the radio to the antenna's feedline. Antenna tuners are particularly important for use with transmitters. Transmitters feed power into a resistive load, very often 50 ohms, for which the transmitter is optimally designed for power output, efficiency, and low distortion. If the load seen by the transmitter departs from this design value due to improper tuning of the antenna/feedline combination the power output will change, distortion may occur and the transmitter may overheat. ATUs are a standard part of almost all radio transmitters; they may be a circuit included inside the transmitter itself or a separate piece of equipment connected between the transmitter and the antenna. In transmitters in which the antenna is mounted separate from the transmitter and connected to it by a transmission line (feedline), there may be a second ATU (or matching network) at the antenna to match the impedance of the antenna to the transmission line. In low power transmitters with attached antennas, such as cell phones and walkie-talkies, the ATU is fixed to work with the antenna. In high power transmitters like radio stations, the ATU is adjustable to accommodate changes in the antenna or transmitter, and adjusting the ATU to match the transmitter to the antenna is an important procedure done after any changes to these components have been made. This adjustment is done with an instrument called a SWR meter. In radio receivers ATUs are not so important, because in the low frequency part of the radio spectrum the signal to noise ratio (SNR) is dominated by atmospheric noise. It does not matter if the impedance of the antenna and receiver are mismatched so some of the incoming power from the antenna is reflected and does not reach the receiver, because the signal can be amplified to make up for it. However in high frequency receivers the receiver's SNR is dominated by noise in the receiver's front end, so it is important that the receiving antenna is impedance-matched to the receiver to give maximum signal amplitude in the front end stages, to overcome noise.
Antenna tuner
Wikipedia
467
582127
https://en.wikipedia.org/wiki/Antenna%20tuner
Technology
Broadcasting
null
Overview An antenna's impedance is different at different frequencies. An antenna tuner matches a radio with a fixed impedance (typically 50 Ohms for modern transceivers) to the combination of the feedline and the antenna; useful when the impedance seen at the input end of the feedline is unknown, complex, or otherwise different from the transceiver. Coupling through an ATU allows the use of one antenna on a broad range of frequencies. However, despite its name, an antenna tuner ' actually matches the transmitter only to the complex impedance reflected back to the input end of the feedline. If both tuner and transmission line were lossless, tuning at the transmitter end would indeed produce a match at every point in the transmitter-feedline-antenna system. However, in practical systems feedline losses limit the ability of the antenna 'tuner' to match the antenna or change its resonant frequency. If the loss of power is very low in the line carrying the transmitter's signal into the antenna, a tuner at the transmitter end can produce a worthwhile degree of matching and tuning for the antenna and feedline network as a whole. With lossy feedlines (such as commonly used 50 Ohm coaxial cable) maximum power transfer only occurs if matching is done at both ends of the line. If there is still a high SWR (multiple reflections) in the feedline beyond the ATU, any loss in the feedline is multiplied several times by the transmitted waves reflecting back and forth between the tuner and the antenna, heating the wire instead of sending out a signal. Even with a matching unit at both ends of the feedline – the near ATU matching the transmitter to the feedline and the remote ATU matching the feedline to the antenna – losses in the circuitry of the two ATUs will reduce power delivered to the antenna. Therefore, operating an antenna far from its design frequency and compensating with a transmatch between the transmitter and the feedline is not as efficient as using a resonant antenna with a matched-impedance feedline, nor as efficient as a matched feedline from the transmitter to a remote antenna tuner attached directly to the antenna.
Antenna tuner
Wikipedia
455
582127
https://en.wikipedia.org/wiki/Antenna%20tuner
Technology
Broadcasting
null
Broad band matching methods Transformers, autotransformers, and baluns are sometimes incorporated into the design of narrow band antenna tuners and antenna cabling connections. They will all usually have little effect on the resonant frequency of either the antenna or the narrow band transmitter circuits, but can widen the range of impedances that the antenna tuner can match, and/or convert between balanced and unbalanced cabling where needed. Ferrite transformers Solid-state power amplifiers operating from 1–30 MHz typically use one or more wideband transformers wound on ferrite cores. MOSFETs and bipolar junction transistors are designed to operate into a low impedance, so the transformer primary typically has a single turn, while the 50 Ohm secondary will have 2 to 4 turns. This feedline system design has the advantage of reducing the retuning required when the operating frequency is changed. A similar design can match an antenna to a transmission line; For example, many TV antennas have a 300 Ohm impedance and feed the signal to the TV via a 75 Ohm coaxial line. A small ferrite core transformer makes the broad band impedance transformation. This transformer does not need, nor is it capable of adjustment. For receive-only use in a TV the small SWR variation with frequency is not a major problem. It should be added that many ferrite based transformers perform a balanced to unbalanced transformation along with the impedance change. When the balanced to unbalanced function is present these transformers are called a balun (otherwise an unun). The most common baluns have either a 1:1 or a 1:4 impedance transformation.
Antenna tuner
Wikipedia
349
582127
https://en.wikipedia.org/wiki/Antenna%20tuner
Technology
Broadcasting
null
Autotransformers There are several designs for impedance matching using an autotransformer, which is a single-wire transformer with different connection points or taps spaced along the windings. They are distinguished mainly by their impedance transform ratio (1:1, 1:4, 1:9, etc., the square of the winding ratio), and whether the input and output sides share a common ground, or are matched from a cable that is grounded on one side (unbalanced) to an ungrounded (usually balanced) cable. When autotransformers connect balanced and unbalanced lines they are called baluns, just as two-winding transformers. When two differently-grounded cables or circuits must be connected but the grounds kept independent, a full, two-winding transformer with the desired ratio is used instead. The circuit pictured at the right has three identical windings wrapped in the same direction around either an "air" core (for very high frequencies) or ferrite core (for middle, or low frequencies). The three equal windings shown are wired for a common ground shared by two unbalanced lines (so this design is called an unun), and can be used as 1:1, 1:4, or 1:9 impedance match, depending on the tap chosen. (The same windings could be connected differently to make a balun instead.) For example, if the right-hand side is connected to a resistive load of 10 Ohms, the user can attach a source at any of the three ungrounded terminals on the left side of the autotransformer to get a different impedance. Notice that on the left side, the line with more windings measures greater impedance for the same 10 Ohm load on the right. Narrow band design The "narrow-band" methods described below cover a very much smaller span of frequencies, by comparison with the broadband methods described above.
Antenna tuner
Wikipedia
404
582127
https://en.wikipedia.org/wiki/Antenna%20tuner
Technology
Broadcasting
null
Antenna matching methods that use transformers tend to cover a wide range of frequencies. A single, typical, commercially available balun can cover frequencies from 3.5–30.0 MHz, or nearly the entire shortwave radio band. Matching to an antenna using a cut segment of transmission line (described below) is perhaps the most efficient of all matching schemes in terms of electrical power, but typically can only cover a range about 3.5–3.7 MHz wide – a very small range indeed, compared to a broadband balun. Antenna coupling or feedline matching circuits are also narrowband for any single setting, but can be re-tuned more conveniently. However they are perhaps the least efficient in terms of power-loss (aside from having no impedance matching at all!). Transmission line antenna tuning methods The insertion of a special section of transmission line, whose characteristic impedance differs from that of the main line, can be used to match the main line to the antenna. An inserted line with the proper impedance and connected at the proper location can perform complicated matching effects with very high efficiency, but spans a very limited frequency range. The simplest example this method is the quarter-wave impedance transformer formed by a section of mismatched transmission line. If a quarter-wavelength of 75 Ohm coaxial cable is linked to a 50 Ohm load, the SWR in the 75 Ohm quarter wavelength of line can be calculated as 75Ω / 50Ω = 1.5; the quarter-wavelength of line transforms the mismatched impedance to 112.5 Ohms (75 Ohms × 1.5 = 112.5 Ohms). Thus this inserted section matches a 112 Ohm antenna to a 50 Ohm main line. The  wavelength coaxial transformer is a useful way to match 50 to 75 Ohms using the same general method. The theoretical basis is discussion by the inventor, and wider application of the method is found here: Branham, P. (1959). A Convenient Transformer for matching Co-axial lines. Geneva: CERN.
Antenna tuner
Wikipedia
424
582127
https://en.wikipedia.org/wiki/Antenna%20tuner
Technology
Broadcasting
null
A second common method is the use of a stub: A shorted, or open section of line is connected in parallel with the main line. With coax this is done using a 'T'-connector. The length of the stub and its location can be chosen so as to produce a matched line below the stub, regardless of the complex impedance or SWR of the antenna itself. The J-pole antenna is an example of an antenna with a built-in stub match. Basic lumped circuit matching using the L network The basic circuit required when lumped capacitances and inductors are used is shown below. This circuit is important in that many automatic antenna tuners use it, and also because more complex circuits can be analyzed as groups of L-networks. This is called an L network not because it contains an inductor, (in fact some L-networks consist of two capacitors), but because the two components are at right angles to each other, having the shape of a rotated and sometimes reversed English letter 'L'. The 'T' ("Tee") network and the π ("Pi") network also have a shape similar to the English and Greek letters they are named after. This basic network is able to act as an impedance transformer. If the output has an impedance consisting of resistance Rload and reactance j Xload, while the input is to be attached to a source which has an impedance of Rsource resistance and j Xsource reactance, then and . In this example circuit, XL and XC can be swapped. All the ATU circuits below create this network, which exists between systems with different impedances. For instance, if the source has a resistive impedance of 50 Ω and the load has a resistive impedance of 1000 Ω : If the frequency is 28 MHz, As, then, So, While as, then, Theory and practice A parallel network, consisting of a resistive element (1000 Ω) and a reactive element (−j 229.415 Ω), will have the same impedance and power factor as a series network consisting of resistive (50 Ω) and reactive elements (−j 217.94 Ω). By adding another element in series (which has a reactive impedance of +j 217.94 Ω), the impedance is 50 Ω (resistive).
Antenna tuner
Wikipedia
496
582127
https://en.wikipedia.org/wiki/Antenna%20tuner
Technology
Broadcasting
null
Types of L networks and their use The L-network can have eight different configurations, six of which are shown here. The two missing configurations are the same as the bottom row, but with the parallel element (wires vertical) on the right side of the series element (wires horizontal), instead of on the left, as shown. In discussion of the diagrams that follows the in connector comes from the transmitter or "source"; the out connector goes to the antenna or "load". The general rule (with some exceptions, described below) is that the series element of an L-network goes on the side with the lowest impedance. So for example, the three circuits in the left column and the two in the bottom row have the series (horizontal) element on the out side are generally used for stepping up from a low-impedance input (transmitter) to a high-impedance output (antenna), similar to the example analyzed in the section above. The top two circuits in the right column, with the series (horizontal) element on the in side, are generally useful for stepping down from a higher input to a lower output impedance. The general rule only applies to loads that are mainly resistive, with very little reactance. In cases where the load is highly reactive – such as an antenna fed with a signals whose frequency is far away from any resonance – the opposite configuration may be required. If far from resonance, the bottom two step down (high-in to low-out) circuits would instead be used to connect for a step up (low-in to high-out that is mostly reactance). The low- and high-pass versions of the four circuits shown in the top two rows use only one inductor and one capacitor. Normally, the low-pass would be preferred with a transmitter, in order to attenuate harmonics, but the high-pass configuration may be chosen if the components are more conveniently obtained, or if the radio already contains an internal low-pass filter, or if attenuation of low frequencies is desirable – for example when a local AM station broadcasting on a medium frequency may be overloading a high frequency receiver.
Antenna tuner
Wikipedia
445
582127
https://en.wikipedia.org/wiki/Antenna%20tuner
Technology
Broadcasting
null