text stringlengths 0 473k |
|---|
[SOURCE: https://en.wikipedia.org/wiki/Maple_(software)] | [TOKENS: 1100] |
Contents Maple (software) Maple is a symbolic and numeric computing environment as well as a multi-paradigm programming language. It covers several areas of technical computing, such as symbolic mathematics, numerical analysis, data processing, visualization, and others. A toolbox, MapleSim, adds functionality for multidomain physical modeling and code generation. Maple's capacity for symbolic computing include those of a general-purpose computer algebra system. For instance, it can manipulate mathematical expressions and find symbolic solutions to certain problems, such as those arising from ordinary and partial differential equations. Maple is developed commercially by the Canadian software company Maplesoft. The name 'Maple' is a reference to the software's Canadian heritage. Overview Users can enter mathematics in traditional mathematical notation. Custom user interfaces can also be created. There is support for numeric computations, to arbitrary precision, as well as symbolic computation and visualization. Examples of symbolic computations are given below. Maple incorporates a dynamically typed imperative-style programming language (resembling Pascal), which permits variables of lexical scope. There are also interfaces to other languages (C, C#, Fortran, Java, MATLAB, and Visual Basic), as well as to Microsoft Excel. Maple supports MathML 2.0, which is a W3C format for representing and interpreting mathematical expressions, including their display in web pages. There is also functionality for converting expressions from traditional mathematical notation to markup suitable for the typesetting system LaTeX. Maple is based on a small kernel, written in C, which provides the Maple language. Most functionality is provided by libraries, which come from a variety of sources. Most of the libraries are written in the Maple language; these have viewable source code. Many numerical computations are performed by the NAG Numerical Libraries, ATLAS libraries, or GMP libraries. Different functionality in Maple requires numerical data in different formats. Symbolic expressions are stored in memory as directed acyclic graphs. The standard interface and calculator interface are written in Java. History The first concept of Maple arose from a meeting in late 1980 at the University of Waterloo. Researchers at the university wished to purchase a computer powerful enough to run the Lisp-based computer algebra system Macsyma. Instead, they opted to develop their own computer algebra system, named Maple, that would run on lower cost computers. Aiming for portability, they began writing Maple in programming languages from the BCPL family (initially using a subset of B and C, and later on only C). A first limited version appeared after three weeks, and fuller versions entered mainstream use beginning in 1982. By the end of 1983, over 50 universities had copies of Maple installed on their machines.[citation needed] In 1984, the research group arranged with Watcom Products Inc to license and distribute the first commercially available version, Maple 3.3. In 1988 Waterloo Maple Inc. (Maplesoft) was founded. The company's original goal was to manage the distribution of the software, but eventually it grew to have its own R&D department, where most of Maple's development takes place today (the remainder being done at various university laboratories). In 1989, the first graphical user interface for Maple was developed and included with version 4.3 for the Macintosh. X11 and Windows versions of the new interface followed in 1990 with Maple V. In 1992, Maple V Release 2 introduced the Maple "worksheet" that combined text, graphics, and input and typeset output. In 1994 a special issue of a newsletter created by Maple developers called MapleTech was published. In 1999, with the release of Maple 6, Maple included some of the NAG Numerical Libraries. In 2003, the current "standard" interface was introduced with Maple 9. This interface is primarily written in Java (although portions, such as the rules for typesetting mathematical formulae, are written in the Maple language). The Java interface was criticized for being slow; improvements have been made in later versions, although the Maple 11 documentation recommends the previous ("classic") interface for users with less than 500 MB of physical memory. Between 1995 and 2005 Maple lost significant market share to competitors due to a weaker user interface. With Maple 10 in 2005, Maple introduced a new "document mode" interface, which has since been further developed across several releases. In September 2009 Maple and Maplesoft were acquired by the Japanese software retailer Cybernet Systems. Version history Features Features of Maple include: Examples of Maple code The following code, which computes the factorial of a nonnegative integer, is an example of an imperative programming construct within Maple: Simple functions can also be defined using the "maps to" arrow notation: Find Output: Compute the determinant of a matrix. The following code numerically calculates the roots of a high-order polynomial: The same command can also solve systems of equations: Plot x sin ( x ) {\displaystyle x\sin(x)} with x ranging from -10 to 10: Plot x 2 + y 2 {\displaystyle x^{2}+y^{2}} with x and y ranging from -1 to 1: Find functions f that satisfy the integral equation Use of the Maple engine The Maple engine is used within several other products from Maplesoft: Listed below are third-party commercial products that no longer use the Maple engine: See also References External links |
======================================== |
[SOURCE: https://en.wikipedia.org/wiki/Life_Investigation_For_Enceladus] | [TOKENS: 491] |
Contents Life Investigation For Enceladus Life Investigation For Enceladus (LIFE) was a proposed astrobiology mission concept that would capture icy particles from Saturn's moon Enceladus and return them to Earth, where they could be studied in detail for signs of life such as biomolecules. The LIFE orbiter concept was proposed by a team led by Peter Tsou to NASA's 13th Discovery Mission solicitation, but the mission was not selected by NASA for Phase-A design study. Mission concept Enceladus is a small icy moon with jets or geysers of water erupting from its surface that might be connected to active hydrothermal vents at its subsurface water ocean floor, where the moon's ocean meets the underlying rock, a prime habitat for life. The geysers could provide easy access for sampling the moon's subsurface ocean, and if there is microbial life in it, ice particles from the sea could contain the evidence astrobiologists need to identify them. The 15-year LIFE mission would use a 'Tanpopo' aerogel collector similar to the one NASA used in the Stardust sample return mission to return cometary dust in 2006. The proposed spacecraft would enter into Saturn orbit and enable multiple flybys through Enceladus's icy plumes. After spending about two years in orbit as slow as 2 km/s around Saturn, LIFE would use its propulsion system to escape Saturn and begin the ~4.5 year long voyage back to Earth with the collected particles in a return capsule. The spacecraft may sample Enceladus's plume, the E ring of Saturn, and the Titan upper atmosphere. In December 2014, NASA announced that it would be selecting finalists in June 2015 to submit proposals for a future Discovery Program mission, and selecting a winning proposal in September 2016. The selected mission must launch by the end of 2021. The mission would have a $425 million development cost cap, and it would reach Saturn after a series of gravity assists past Venus and the Earth. Samples from Enceladus's plume would make it to Earth about 14 years later. In September 2016, NASA announced that five proposals had been selected for further study.[needs update] Science payload As currently envisioned, the probe's science payload would consist of: The samples collected would be returned to Earth for extensive analyses. See also References |
======================================== |
[SOURCE: https://en.wikipedia.org/wiki/Foreach#Python] | [TOKENS: 1909] |
Contents Loop (statement) In computer programming, a loop is a control flow statement that allows code to be executed repeatedly, usually with minor alterations between repetitions. Loops can be used to perform a repeated action on all items in a collection, or to implement a long lived program. Overview Loops are a feature of high-level programming languages. In low-level programming languages the same functionality is achieved using jumps. When a program is compiled to machine code, looping may be achieved using jumps; but some loops can be optimized to run without jumping. Usually, loops are expected to run for a finite number of iteration.[citation needed] Without proper care, loops may accidentally be created that have no possibility of terminating. Such loops are called infinite loops. The problem of determining whether a program contains an infinite loop is known as the halting problem. Conditional loop A conditional loop (also known as an indeterminate loop) is a loop that determines whether to terminate based on a logical condition. These loops are flexible, but there exact behavior can be difficult to reason about.[citation needed] A conditional loop is usually composed of two parts: a condition and a body. The condition is a logical statement depending on the state of the program and the body is a block of code that runs as long as the condition holds. A common misconception is that the execution of the body terminates as soon as the condition does not hold anymore; but this is usually not the case.: 368 In most programming languages, the condition is checked once for every execution of the body. When the condition is checked is not standardized and some programming languages contain multiple conditional looping structures with different rules about when the condition is assessed.[citation needed] A pre-test loop is a conditional loop where the condition is checked before the body is executed. More precisely, the condition is checked and if it holds the body is execute. Afterwards, the condition is checked again, and if it holds the body is executed again. This process repeats until the condition does not hold. Many programming languages call this loop a while loop and refer to it with the keyword while. They are commonly formatted in manner similar to Instead of the keywords do and repeat others methods are sometime use to indicate where the body begins and ends, such as curly braces or whitespace. For example, the following code fragment first checks whether x is less than five, which it is, so the body is entered. There, x is displayed and then incremented by one. After executing the statements in the body, the condition is checked again, and the loop is executed again. This process repeats until x has the value five. A post-test loop is a conditional loop where the condition is checked after the body is executed. More precisely, the body is executed and afterwards the condition is checked. If it holds the body is run again and then the condition is checked. This is repeated until the condition does not hold. This is sometimes called a do-while loop due to the syntax used in various programming languages, although this can be confusing since Fortran and PL/I use the syntax "DO WHILE" for pre-test loops. Post-test loops are commonly formatted in manner similar to Instead of the keywords do and repeat others methods are sometime use to indicate where the body begins and ends, such as curly braces. Some languages may use a different naming convention for this type of loop. For example, the Pascal and Lua languages have a "repeat until" loop, which continues to run until the control expression is true and then terminates. A three-part for loop, popularized by C, has two additional parts: initialization (loop variant), and increment, both of which are blocks of code. The initialization is intended as code that prepares the loop and is run once in the beginning and increment is used to update the state of the program after each iteration of the loop. Otherwise, the three-part for loop is a pre-test loop. They are commonly formatted in manner similar to This syntax came from B and was originally invented by Stephen C. Johnson. The following C code is an example of a three part loop that prints the numbers from 0 to 4. Assuming there is a function called do_work() that does some work, the following are equivalent.[citation needed] As long as the continue statement is not used, the above is technically equivalent to the following (though these examples are not typical or modern style used in everyday computers): or Enumeration An enumeration (also known as an determinate loop) is a loop intended to iterate over all the items of a collection. It is not as flexible as a conditional loop; but it is more predictable.[citation needed] For example, it is easier to guarantee that enumerations terminate and they avoid potential off-by-one errors.[citation needed] Enumerations can be implemented using an iterator, whether implicitly or explicitly. They are commonly formatted in manner similar to Depending on the programming language, various keywords are used to invoke enumerations. For example, descendants of ALGOL use for, while descendants of Fortran use do and COBOL uses PERFORM VARYING. Enumerations are sometimes called "for loops," for example in Zig and Rust. This can be confusing since many of the most popular programming languages, such as C, C++, and Java, use that term for the three-part for loop, which is not an enumeration. Other programming languages, such as Perl and C#, avoid this confusion by using the term "foreach loop." The order in which the items in the collection are iterated through depends on the programming language. Fortran 95 has a loop, invoked using the keyword FORALL, that is independent of this order. It has the effect of executing each iteration of the loop at the same time. This feature was made obsolescent in Fortran 2018. Loops in functional programming In most functional programming languages, recursion is used instead of traditional loops. This is due to the fact that variables are immutable, and therefore the increment step of a loop cannot occur. To avoid running into stack overflow errors for long loops, functional programming languages implement tail call optimisation, which allows the same stack frame to be used for each iteration of the loop, compiling to effectively the same code as a while or for loop. Some languages, such as Haskell, have a special syntax known as a list comprehension, which is similar to enumeration, iterating over the contents of a list and transforming it into a new list. Loop counter A loop counter is a control variable that controls the iterations of a loop. Loop counters change with each iteration of a loop, providing a unique value for each iteration. The loop counter is used to decide when the loop should terminate. It is so named because most uses of this construct result in the variable taking on a range of integer values. A common identifier naming convention is for the loop counter to use the variable names i, j, and k (and so on if needed), where i would be the most outer loop, j the next inner loop, etc. This style is generally agreed to have originated from the early programming of Fortran[citation needed], where these variable names beginning with these letters were implicitly declared as having an integer type, and so were obvious choices for loop counters that were only temporarily required. The practice dates back further to mathematical notation where indices for sums and multiplications are often i, j, and k. Using terse names for loop counters, like i and j, is discouraged by some since the purpose of the variables is not as clear as if they were given a longer more descriptive name.: 383–382 Different languages specify different rules for what value the loop counter will hold on termination of its loop, and indeed some hold that it becomes undefined. This permits a compiler to generate code that leaves any value in the loop counter, or perhaps even leaves it unchanged because the loop value was held in a register and never stored in memory. Actual behavior may even vary according to the compiler's optimization settings. Modifying the loop counter within the body of the loop can lead to unexpected consequences. To prevent such problems, some languages make the loop counter immutable.[citation needed] However, only overt changes are likely to be detected by the compiler. Situations where the address of the loop counter is passed as an argument to a subroutine, make it very difficult to check because the routine's behavior is in general unknowable to the compiler unless the language supports procedure signatures and argument intents.[citation needed] Early exit and continuation Some languages may also provide supporting statements for altering how a loop's iteration proceeds. Common among these are the break statement, which terminates the current loop the program is in, and the continue statement, which skips to the next iteration of the current loop.: 379 These statements may have other names; For example in Fortran 90, they are called exit and cycle.[citation needed] A loop can also be terminated by returning from the function within which it is being executed. In the case of nested loops, the break and continue statements apply to the inner most loop. Some languages allow loops to be labelled. These statements can then be applied to any of the loops in which the program is nested. Infinite loop An infinite loop is a loop which never terminates. This can be intentional, or the result of a logic error. Systematically detecting infinite loops is known as the halting problem. Infinite loops are useful in applications which need to perform a repeated calculation until a program terminates, such as web servers. See also References |
======================================== |
[SOURCE: https://en.wikipedia.org/wiki/Corral_de_comedias] | [TOKENS: 1445] |
Contents Corral de comedias Corral de comedias (lit. 'theatrical courtyard') is a type of open-air theatre specific to Spain. In Spanish all secular plays were called comedias, which embraced three genres: tragedy, drama, and comedy itself. During the Spanish Golden Age, corrals became popular sites for theatrical presentations in the early 16th century when the theatre took on a special importance in the country. The performance was held in the afternoon and lasted two to three hours, there being no intermission, and few breaks. The entertainment was continuous, including complete shows with parts sung and danced. All spectators were placed according to their sex and social status. History In modern times, the first buildings devoted to the theatre in Spain appeared in the 16th century. Representations of comedias were instead held in the courtyard of houses or inns where a stage with background scenery was improvised along one of the sides. The three remaining sides served as public galleries to the wealthy, with the remaining spectators watching the play from the open courtyard. The courtyard structure was maintained in permanent theatres built for the purpose from the end of the sixteenth century, called corrales de comedias, which used the open-air enclosed rectangular courtyard typical of a block of houses. Playwrights and dramatists such as Lope de Vega, Juan Pérez de Montalbán, Tirso de Molina, and Pedro Calderón de la Barca created works which were performed in corrales de comedias. The first permanent theater of this type, Corral de la Cruz, was constructed in Madrid in 1579. The number of theaters increased rapidly after 1600, responding to the public's enthusiasm for this new form of entertainment. The oldest surviving corral, albeit significantly altered, is the Corral de comedias de Alcalá de Henares. This corral, formerly a courtyard theatre, has been roofed and used as a teatro romántico and a cinema, leading to major changes in the building's architecture. The last known such courtyard theatre to be built in Spain, Corral de comedias de Almagro, in Almagro, Castile-La Mancha, is a purpose-built theater that dates to 1628. This only functioning courtyard theater still standing, once one among the many, annually celebrates the Festival Internacional de Teatro Clásico (International Classical Theatre Festival); There are recently found remains of a corral in Torralba de Calatrava, which the municipality wished to rebuild in 2006. The theatre type specific to Spain, was extended to Mexico when a corral de comedias was built in Tecali de Herrera around 1540. The corrales present some parallels to Elizabethan theatre where productions were held in galleried inns. The George Inn, Southwark is a partially surviving example of such a structure. As well as similarities as regards the type of buildings used, there were similarities in the subject matter of the plays: Spanish literature was translated into English in Shakespeare's time, and in 1613 his company mounted a lost play called Cardenio which appears to have been based on an episode in Don Quixote. Although Shakespeare was not translated into Spanish until the eighteenth century, his work has since been performed in the corrales; for example in 2016, which saw the quatercentenary of the deaths of Cervantes and Shakespeare, there were Shakespearean productions in Amalgro and Alcalá de Henares. Architecture and fittings The stage was installed at one end of the court, against the back wall. In front of the stage was the outdoor patio at the end of which sat the so-called musketeers. The balconies and windows of the adjoining houses formed the quarters reserved for men and women of nobility. In Madrid, above the cazuela, were the quarters of the councilors and other authorities, such as the chairman of the Council of Castile. On the upper floors were the desvanes (attics), very small quarters, among which stood the tertulia of the church and a second cazuela. The stage and lateral galleries were protected by an overhang. An awning, hung from hooks, protected the men of the common public who sat in the patio from the sun, avoiding contrast between sunlit and shaded areas, such as was found on stage and in the courtyard. That probably also improved the acoustics of the venue, avoiding straining the voices of actors. This provision was similar to the Elizabethan theatres from the same period in England. In the earlier built corrals, there were no toilets: with the advent of "new enlightened governments" during the reign of Philip V, some corrales were closed due to hygiene issues, risk of fire or disorder. With the arrival of a bourgeois class who did not want to watch the plays in awkward spaces such as these, larger theatrical structures were constructed, which required confined spaces and specific acoustic treatment. Performances During the Spanish Golden Age, any theatrical event was known as comedia. The public came in masses for entertainments like this, whether comedy, drama or tragedy. The season of performances usually began on Easter Sunday, ending on Ash Wednesday. Smoking was forbidden because of the risk of fire, and from October to April the comedia began at two in the afternoon, in the spring at three and at four during summer, in order for all to finish before sunset. The performance's duration was approximately four to six hours, structured in six different rounds: the first act or loa, the opening round, then an appetizer, the second round, the masquerade or jácaras, a third round and the final act. Men and women could not sit together; men occupied the courtyard, side stands, the benches or the central stands, and the women watched the performance from their cazuelas above. The only place where they were allowed to be together was in the chamber corridors. Children were not allowed to attend. The audience paid fees at different points: at the entrance, then a tip to the "brotherhood" or beneficiary, and a third one for the privilege of a seat so they could watch the play comfortably. The theatrical company rarely received as much as 20% of the total. In university towns, it was forbidden to perform on weekdays, so the students would not be distracted. Two characters were instantly recognized in the corrales: the mozo, maintainer of order, equipped with a big garrote to calm the excited spectator, and the "spacer", that is, the one in charge of finding a suitable place for an individual in between two others. The first regulation on the operation of corrals was published by the Royal Council of Castile for the corrales of Madrid, later extended to the whole kingdom. Among its provisions, was the presence of a bailiff whose function was to ensure that no noise, tumults, or scandals ensued and that men and women were kept separated in their respective seating by the required entrances and exits. References Bibliography |
======================================== |
[SOURCE: https://en.wikipedia.org/wiki/Fresh_water#Aquatic_organisms] | [TOKENS: 2805] |
Contents Fresh water Fresh water or freshwater is any naturally occurring liquid or frozen water containing low concentrations of dissolved salts and other total dissolved solids. The term excludes seawater and brackish water, but it does include non-salty mineral-rich waters, such as chalybeate springs. Fresh water may encompass frozen and meltwater in ice sheets, ice caps, glaciers, snowfields and icebergs, natural precipitations such as rainfall, snowfall, hail/sleet and graupel, and surface runoffs that form inland bodies of water such as wetlands, ponds, lakes, rivers, streams, as well as groundwater contained in aquifers, subterranean rivers and lakes. Water is critical to the survival of all living organisms. Many organisms can thrive on salt water, but the great majority of vascular plants and most insects, amphibians, reptiles, mammals and birds need fresh water to survive. Fresh water is the water resource that is of the most and immediate use to humans. Fresh water is not always potable water, that is, water safe to drink by humans. Much of the earth's fresh water (on the surface and groundwater) is to a substantial degree unsuitable for human consumption without treatment. Fresh water can easily become polluted by human activities or due to naturally occurring processes, such as erosion. Fresh water makes up less than 3% of the world's water resources, and just 1% of that is readily available. About 70% of the world's freshwater reserves are frozen in Antarctica. Just 3% of it is extracted for human consumption. Agriculture uses roughly two thirds of all fresh water extracted from the environment. Fresh water is a renewable and variable, but finite natural resource. Fresh water is replenished through the process of the natural water cycle, in which water from seas, lakes, forests, land, rivers and reservoirs evaporates, forms clouds, and returns inland as precipitation. Locally, however, if more fresh water is consumed through human activities than is naturally restored, this may result in reduced fresh water availability (or water scarcity) from surface and underground sources and can cause serious damage to surrounding and associated environments. Water pollution also reduces the availability of fresh water. Where available water resources are scarce, humans have developed technologies like desalination and wastewater recycling to stretch the available supply further. However, given the high cost (both capital and running costs) and - especially for desalination - energy requirements, those remain mostly niche applications. A non-sustainable alternative is using so-called "fossil water" from underground aquifers. As some of those aquifers formed hundreds of thousands or even millions of years ago when local climates were wetter (e.g. from one of the Green Sahara periods) and are not appreciably replenished under current climatic conditions - at least compared to drawdown, these aquifers form essentially non-renewable resources comparable to peat or lignite, which are also continuously formed in the current era but orders of magnitude slower than they are mined. Definitions Fresh water can be defined as water with less than 500 parts per million (ppm) of dissolved salts. Other sources give higher upper salinity limits for fresh water, e.g. 1,000 ppm or 3,000 ppm. Fresh water habitats are classified as either lentic systems, which are the stillwaters including ponds, lakes, swamps and mires; lotic which are running-water systems; or groundwaters which flow in rocks and aquifers. There is, in addition, a zone which bridges between groundwater and lotic systems, which is the hyporheic zone, which underlies many larger rivers and can contain substantially more water than is seen in the open channel. It may also be in direct contact with the underlying underground water. Sources The original source of almost all fresh water is precipitation from the atmosphere, in the form of mist, rain and snow. Fresh water falling as mist, rain or snow contains materials dissolved from the atmosphere and material from the sea and land over which the rain bearing clouds have traveled. The precipitation leads eventually to the formation of water bodies that humans can use as sources of freshwater: ponds, lakes, rainfall, rivers, streams, and groundwater contained in underground aquifers. In coastal areas fresh water may contain significant concentrations of salts derived from the sea if windy conditions have lifted drops of seawater into the rain-bearing clouds. This can give rise to elevated concentrations of sodium, chloride, magnesium and sulfate as well as many other compounds in smaller concentrations. In desert areas, or areas with impoverished or dusty soils, rain-bearing winds can pick up sand and dust and this can be deposited elsewhere in precipitation and causing the freshwater flow to be measurably contaminated both by insoluble solids but also by the soluble components of those soils. Significant quantities of iron may be transported in this way including the well-documented transfer of iron-rich rainfall falling in Brazil derived from sand-storms in the Sahara in north Africa. In Africa, it was revealed that groundwater controls are complex and do not correspond directly to a single factor. Groundwater showed greater resilience to climate change than expected, and areas with an increasing threshold between 0.34 and 0.39 aridity index exhibited significant sensitivity to climate change. Land-use could affect infiltration and runoff processes. The years of most recharge coincided with the most precipitation anomalies, such as during El Niño and La Niña events. Three precipitation-recharge sensitivities were distinguished: in super arid areas with more than 0.67 aridity index, there was constant recharge with little variation with precipitation; in most sites (arid, semi-arid, humid), annual recharge increased as annual precipitation remained above a certain threshold; and in complex areas down to 0.1 aridity index (focused recharge), there was very inconsistent recharge (low precipitation but high recharge). Understanding these relationships can lead to the development of sustainable strategies for water collection. This understanding is particularly crucial in Africa, where water resources are often scarce and climate change poses significant challenges. Water distribution Saline water in oceans, seas and saline groundwater make up about 97% of all the water on Earth. Only 2.5–2.75% is fresh water, including 1.75–2% frozen in glaciers, ice and snow, 0.5–0.75% as fresh groundwater. The water table is the level below which all spaces are filled with water, while the area above this level, where spaces in the rock and soil contain both air and water, is known as the unsaturated zone. The water in this unsaturated zone is referred to as soil moisture. Below the water table, the entire region is known as the saturated zone, and the water in this zone is called groundwater. Groundwater plays a crucial role as the primary source of water for various purposes including drinking, washing, farming, and manufacturing, and even when not directly used as a drinking water supply it remains vital to protect due to its ability to carry contaminants and pollutants from the land into lakes and rivers, which constitute a significant percentage of other people's freshwater supply. It is almost ubiquitous underground, residing in the spaces between particles of rock and soil or within crevices and cracks in rock, typically within 100 m (330 ft) of the surface, and soil moisture, and less than 0.01% of it as surface water in lakes, swamps and rivers. Freshwater lakes contain about 87% of this fresh surface water, including 29% in the African Great Lakes, 22% in Lake Baikal in Russia, 21% in the North American Great Lakes, and 14% in other lakes. Swamps have most of the balance with only a small amount in rivers, most notably the Amazon River. The atmosphere contains 0.04% water. In areas with no fresh water on the ground surface, fresh water derived from precipitation may, because of its lower density, overlie saline ground water in lenses or layers. Most of the world's fresh water is frozen in ice sheets. Many areas have very little fresh water, such as deserts. Freshwater ecosystems Water is a critical issue for the survival of all living organisms. Some can use salt water but many organisms including the great majority of higher plants and most mammals must have access to fresh water to live. Some terrestrial mammals, especially desert rodents, appear to survive without drinking, but they do generate water through the metabolism of cereal seeds, and they also have mechanisms to conserve water to the maximum degree. Freshwater ecosystems are a subset of Earth's aquatic ecosystems that include the biological communities inhabiting freshwater waterbodies such as lakes, ponds, rivers, streams, springs, bogs, and wetlands. They can be contrasted with marine ecosystems, which have a much higher salinity. Freshwater habitats can be classified by different factors, including temperature, light penetration, nutrients, and vegetation. There are three basic types of freshwater ecosystems: lentic (slow moving water, including pools, ponds, and lakes), lotic (faster moving streams, for example creeks and rivers) and wetlands (semi-aquatic areas where the soil is saturated or inundated for at least part of the time). Freshwater ecosystems contain 41% of the world's known fish species. Challenges The increase in the world population and the increase in per capita water use puts increasing strains on the finite resources availability of clean fresh water. The response by freshwater ecosystems to a changing climate can be described in terms of three interrelated components: water quality, water quantity or volume, and water timing. A change in one often leads to shifts in the others as well. Water scarcity (closely related to water stress or water crisis) is the lack of any, local or economically viably transportable, sources of fresh water resources to meet the standard water demand in a region. There are two types of water scarcity. One is physical. The other is economic water scarcity.: 560 Physical water scarcity is where there is not enough water to meet all demands. This includes water needed for ecosystems to function. Regions with a desert climate often face physical water scarcity. Central Asia, West Asia, and North Africa are examples of arid areas. Economic water scarcity results from a lack of investment in infrastructure or technology to draw water from rivers, aquifers, or other water sources. It also results from weak human capacity to meet water demand.: 560 Many people in sub-Saharan Africa are living with economic water scarcity.: 11 There is and has always been enough physical supply of freshwater for current or near or distant future demand in a global scale. As such, water scarcity is caused by a mismatch between when and where people need water, and when and where it is available. This can happen due to an increase in the number of people in a region, changing living conditions and diets, and expansion of irrigated agriculture. Climate change (including droughts or floods), deforestation, water pollution and wasteful use of water can also mean there is not enough water. These variations in scarcity may also be a function of prevailing economic policy and planning approaches. An important concern for hydrological ecosystems is securing minimum streamflow, especially preserving and restoring instream water allocations. Fresh water is an important natural resource necessary for the survival of all ecosystems. Water pollution (or aquatic pollution) is the contamination of water bodies, with a negative impact on their uses.: 6 It is usually a result of human activities. Water bodies include lakes, rivers, oceans, aquifers, reservoirs and groundwater. Water pollution results when contaminants mix with these water bodies. Contaminants can come from one of four main sources. These are sewage discharges, industrial activities, agricultural activities, and urban runoff including stormwater. Water pollution may affect either surface water or groundwater. This form of pollution can lead to many problems. One is the degradation of aquatic ecosystems. Another is spreading water-borne diseases when people use polluted water for drinking or irrigation. Water pollution also reduces the ecosystem services such as drinking water provided by the water resource. Sources of water pollution are either point sources or non-point sources. Point sources have one identifiable cause, such as a storm drain, a wastewater treatment plant, or an oil spill. Non-point sources are more diffuse. An example is agricultural runoff. Pollution is the result of the cumulative effect over time. Pollution may take many forms. One would is toxic substances such as oil, metals, plastics, pesticides, persistent organic pollutants, and industrial waste products. Another is stressful conditions such as changes of pH, hypoxia or anoxia, increased temperatures, excessive turbidity, or changes of salinity). The introduction of pathogenic organisms is another. Contaminants may include organic and inorganic substances. A common cause of thermal pollution is the use of water as a coolant by power plants and industrial manufacturers. Society and culture Uses of water include agricultural, industrial, household, recreational and environmental activities. The Sustainable Development Goals are a collection of 17 interlinked global goals designed to be a "blueprint to achieve a better and more sustainable future for all". Targets on fresh water conservation are included in SDG 6 (Clean water and sanitation) and SDG 15 (Life on land). For example, Target 6.4 is formulated as "By 2030, substantially increase water-use efficiency across all sectors and ensure sustainable withdrawals and supply of freshwater to address water scarcity and substantially reduce the number of people suffering from water scarcity." Another target, Target 15.1, is: "By 2020, ensure the conservation, restoration and sustainable use of terrestrial and inland freshwater ecosystems and their services, in particular forests, wetlands, mountains and drylands, in line with obligations under international agreements." See also Notes Subnotes References External links |
======================================== |
[SOURCE: https://en.wikipedia.org/wiki/Black_hole#Accretion_disk] | [TOKENS: 13839] |
Contents Black hole A black hole is an astronomical body so compact that its gravity prevents anything, including light, from escaping. Albert Einstein's theory of general relativity predicts that a sufficiently compact mass will form a black hole. The boundary of no escape is called the event horizon. In general relativity, a black hole's event horizon seals an object's fate but produces no locally detectable change when crossed. General relativity also predicts that every black hole should have a central singularity, where the curvature of spacetime is infinite. In many ways, a black hole acts like an ideal black body, as it reflects no light. Quantum field theory in curved spacetime predicts that event horizons emit Hawking radiation, with the same spectrum as a black body of a temperature inversely proportional to its mass. This temperature is of the order of billionths of a kelvin for stellar black holes, making it essentially impossible to observe directly. Objects whose gravitational fields are too strong for light to escape were first considered in the 18th century by John Michell and Pierre-Simon Laplace. In 1916, Karl Schwarzschild found the first modern solution of general relativity that would characterise a black hole. Due to his influential research, the Schwarzschild metric is named after him. David Finkelstein, in 1958, first interpreted Schwarzschild's model as a region of space from which nothing can escape. Black holes were long considered a mathematical curiosity; it was not until the 1960s that theoretical work showed they were a generic prediction of general relativity. The first black hole known was Cygnus X-1, identified by several researchers independently in 1971. Black holes typically form when massive stars collapse at the end of their life cycle. After a black hole has formed, it can grow by absorbing mass from its surroundings. Supermassive black holes of millions of solar masses may form by absorbing other stars and merging with other black holes, or via direct collapse of gas clouds. There is consensus that supermassive black holes exist in the centres of most galaxies. The presence of a black hole can be inferred through its interaction with other matter and with electromagnetic radiation such as visible light. Matter falling toward a black hole can form an accretion disk of infalling plasma, heated by friction and emitting light. In extreme cases, this creates a quasar, some of the brightest objects in the universe. Merging black holes can also be detected by observation of the gravitational waves they emit. If other stars are orbiting a black hole, their orbits can be used to determine the black hole's mass and location. Such observations can be used to exclude possible alternatives such as neutron stars. In this way, astronomers have identified numerous stellar black hole candidates in binary systems and established that the radio source known as Sagittarius A*, at the core of the Milky Way galaxy, contains a supermassive black hole of about 4.3 million solar masses. History The idea of a body so massive that even light could not escape was first proposed in the late 18th century by English astronomer and clergyman John Michell and independently by French scientist Pierre-Simon Laplace. Both scholars proposed very large stars in contrast to the modern concept of an extremely dense object. Michell's idea, in a short part of a letter published in 1784, calculated that a star with the same density but 500 times the radius of the sun would not let any emitted light escape; the surface escape velocity would exceed the speed of light.: 122 Michell correctly hypothesized that such supermassive but non-radiating bodies might be detectable through their gravitational effects on nearby visible bodies. In 1796, Laplace mentioned that a star could be invisible if it were sufficiently large while speculating on the origin of the Solar System in his book Exposition du Système du Monde. Franz Xaver von Zach asked Laplace for a mathematical analysis, which Laplace provided and published in a journal edited by von Zach. In 1905, Albert Einstein showed that the laws of electromagnetism would be invariant under a Lorentz transformation: they would be identical for observers travelling at different velocities relative to each other. This discovery became known as the principle of special relativity. Although the laws of mechanics had already been shown to be invariant, gravity remained yet to be included.: 19 In 1907, Einstein published a paper proposing his equivalence principle, the hypothesis that inertial mass and gravitational mass have a common cause. Using the principle, Einstein predicted the redshift and half of the lensing effect of gravity on light; the full prediction of gravitational lensing required development of general relativity.: 19 By 1915, Einstein refined these ideas into his general theory of relativity, which explained how matter affects spacetime, which in turn affects the motion of other matter. This formed the basis for black hole physics. Only a few months after Einstein published the field equations describing general relativity, astrophysicist Karl Schwarzschild set out to apply the idea to stars. He assumed spherical symmetry with no spin and found a solution to Einstein's equations.: 124 A few months after Schwarzschild, Johannes Droste, a student of Hendrik Lorentz, independently gave the same solution. At a certain radius from the center of the mass, the Schwarzschild solution became singular, meaning that some of the terms in the Einstein equations became infinite. The nature of this radius, which later became known as the Schwarzschild radius, was not understood at the time. Many physicists of the early 20th century were skeptical of the existence of black holes. In a 1926 popular science book, Arthur Eddington critiqued the idea of a star with mass compressed to its Schwarzschild radius as a flaw in the then-poorly-understood theory of general relativity.: 134 In 1939, Einstein himself used his theory of general relativity in an attempt to prove that black holes were impossible. His work relied on increasing pressure or increasing centrifugal force balancing the force of gravity so that the object would not collapse beyond its Schwarzschild radius. He missed the possibility that implosion would drive the system below this critical value.: 135 By the 1920s, astronomers had classified a number of white dwarf stars as too cool and dense to be explained by the gradual cooling of ordinary stars. In 1926, Ralph Fowler showed that quantum-mechanical degeneracy pressure was larger than thermal pressure at these densities.: 145 In 1931, Subrahmanyan Chandrasekhar calculated that a non-rotating body of electron-degenerate matter below a certain limiting mass is stable, and by 1934 he showed that this explained the catalog of white dwarf stars.: 151 When Chandrasekhar announced his results, Eddington pointed out that stars above this limit would radiate until they were sufficiently dense to prevent light from exiting, a conclusion he considered absurd. Eddington and, later, Lev Landau argued that some yet unknown mechanism would stop the collapse. In the 1930s, Fritz Zwicky and Walter Baade studied stellar novae, focusing on exceptionally bright ones they called supernovae. Zwicky promoted the idea that supernovae produced stars with the density of atomic nuclei—neutron stars—but this idea was largely ignored.: 171 In 1939, based on Chandrasekhar's reasoning, J. Robert Oppenheimer and George Volkoff predicted that neutron stars below a certain mass limit, later called the Tolman–Oppenheimer–Volkoff limit, would be stable due to neutron degeneracy pressure. Above that limit, they reasoned that either their model would not apply or that gravitational contraction would not stop.: 380 John Archibald Wheeler and two of his students resolved questions about the model behind the Tolman–Oppenheimer–Volkoff (TOV) limit. Harrison and Wheeler developed the equations of state relating density to pressure for cold matter all the way through electron degeneracy and neutron degeneracy. Masami Wakano and Wheeler then used the equations to compute the equilibrium curve for stars, relating mass to circumference. They found no additional features that would invalidate the TOV limit. This meant that the only thing that could prevent black holes from forming was a dynamic process ejecting sufficient mass from a star as it cooled.: 205 The modern concept of black holes was formulated by Robert Oppenheimer and his student Hartland Snyder in 1939.: 80 In the paper, Oppenheimer and Snyder solved Einstein's equations of general relativity for an idealized imploding star, in a model later called the Oppenheimer–Snyder model, then described the results from far outside the star. The implosion starts as one might expect: the star material rapidly collapses inward. However, as the density of the star increases, gravitational time dilation increases and the collapse, viewed from afar, seems to slow down further and further until the star reaches its Schwarzschild radius, where it appears frozen in time.: 217 In 1958, David Finkelstein identified the Schwarzschild surface as an event horizon, calling it "a perfect unidirectional membrane: causal influences can cross it in only one direction". In this sense, events that occur inside of the black hole cannot affect events that occur outside of the black hole. Finkelstein created a new reference frame to include the point of view of infalling observers.: 103 Finkelstein's new frame of reference allowed events at the surface of an imploding star to be related to events far away. By 1962 the two points of view were reconciled, convincing many skeptics that implosion into a black hole made physical sense.: 226 The era from the mid-1960s to the mid-1970s was the "golden age of black hole research", when general relativity and black holes became mainstream subjects of research.: 258 In this period, more general black hole solutions were found. In 1963, Roy Kerr found the exact solution for a rotating black hole. Two years later, Ezra Newman found the cylindrically symmetric solution for a black hole that is both rotating and electrically charged. In 1967, Werner Israel found that the Schwarzschild solution was the only possible solution for a nonspinning, uncharged black hole, meaning that a Schwarzschild black hole would be defined by its mass alone. Similar identities were later found for Reissner-Nordstrom and Kerr black holes, defined only by their mass and their charge or spin respectively. Together, these findings became known as the no-hair theorem, which states that a stationary black hole is completely described by the three parameters of the Kerr–Newman metric: mass, angular momentum, and electric charge. At first, it was suspected that the strange mathematical singularities found in each of the black hole solutions only appeared due to the assumption that a black hole would be perfectly spherically symmetric, and therefore the singularities would not appear in generic situations where black holes would not necessarily be symmetric. This view was held in particular by Vladimir Belinski, Isaak Khalatnikov, and Evgeny Lifshitz, who tried to prove that no singularities appear in generic solutions, although they would later reverse their positions. However, in 1965, Roger Penrose proved that general relativity without quantum mechanics requires that singularities appear in all black holes. Astronomical observations also made great strides during this era. In 1967, Antony Hewish and Jocelyn Bell Burnell discovered pulsars and by 1969, these were shown to be rapidly rotating neutron stars. Until that time, neutron stars, like black holes, were regarded as just theoretical curiosities, but the discovery of pulsars showed their physical relevance and spurred a further interest in all types of compact objects that might be formed by gravitational collapse. Based on observations in Greenwich and Toronto in the early 1970s, Cygnus X-1, a galactic X-ray source discovered in 1964, became the first astronomical object commonly accepted to be a black hole. Work by James Bardeen, Jacob Bekenstein, Carter, and Hawking in the early 1970s led to the formulation of black hole thermodynamics. These laws describe the behaviour of a black hole in close analogy to the laws of thermodynamics by relating mass to energy, area to entropy, and surface gravity to temperature. The analogy was completed: 442 when Hawking, in 1974, showed that quantum field theory implies that black holes should radiate like a black body with a temperature proportional to the surface gravity of the black hole, predicting the effect now known as Hawking radiation. While Cygnus X-1, a stellar-mass black hole, was generally accepted by the scientific community as a black hole by the end of 1973, it would be decades before a supermassive black hole would gain the same broad recognition. Although, as early as the 1960s, physicists such as Donald Lynden-Bell and Martin Rees had suggested that powerful quasars in the center of galaxies were powered by accreting supermassive black holes, little observational proof existed at the time. However, the Hubble Space Telescope, launched decades later, found that supermassive black holes were not only present in these active galactic nuclei, but that supermassive black holes in the center of galaxies were ubiquitous: Almost every galaxy had a supermassive black hole at its center, many of which were quiescent. In 1999, David Merritt proposed the M–sigma relation, which related the dispersion of the velocity of matter in the center bulge of a galaxy to the mass of the supermassive black hole at its core. Subsequent studies confirmed this correlation. Around the same time, based on telescope observations of the velocities of stars at the center of the Milky Way galaxy, independent work groups led by Andrea Ghez and Reinhard Genzel concluded that the compact radio source in the center of the galaxy, Sagittarius A*, was likely a supermassive black hole. On 11 February 2016, the LIGO Scientific Collaboration and Virgo Collaboration announced the first direct detection of gravitational waves, named GW150914, representing the first observation of a black hole merger. At the time of the merger, the black holes were approximately 1.4 billion light-years away from Earth and had masses of 30 and 35 solar masses.: 6 In 2017, Rainer Weiss, Kip Thorne, and Barry Barish, who had spearheaded the project, were awarded the Nobel Prize in Physics for their work. Since the initial discovery in 2015, hundreds more gravitational waves have been observed by LIGO and another interferometer, Virgo. On 10 April 2019, the first direct image of a black hole and its vicinity was published, following observations made by the Event Horizon Telescope (EHT) in 2017 of the supermassive black hole in Messier 87's galactic centre. In 2022, the Event Horizon Telescope collaboration released an image of the black hole in the center of the Milky Way galaxy, Sagittarius A*; The data had been collected in 2017. In 2020, the Nobel Prize in Physics was awarded for work on black holes. Andrea Ghez and Reinhard Genzel shared one-half for their discovery that Sagittarius A* is a supermassive black hole. Penrose received the other half for his work showing that the mathematics of general relativity requires the formation of black holes. Cosmologists lamented that Hawking's extensive theoretical work on black holes would not be honored since he died in 2018. In December 1967, a student reportedly suggested the phrase black hole at a lecture by John Wheeler; Wheeler adopted the term for its brevity and "advertising value", and Wheeler's stature in the field ensured it quickly caught on, leading some to credit Wheeler with coining the phrase. However, the term was used by others around that time. Science writer Marcia Bartusiak traces the term black hole to physicist Robert H. Dicke, who in the early 1960s reportedly compared the phenomenon to the Black Hole of Calcutta, notorious as a prison where people entered but never left alive. The term was used in print by Life and Science News magazines in 1963, and by science journalist Ann Ewing in her article "'Black Holes' in Space", dated 18 January 1964, which was a report on a meeting of the American Association for the Advancement of Science held in Cleveland, Ohio. Definition A black hole is generally defined as a region of spacetime from which no information-carrying signals or objects can escape. However, verifying an object as a black hole by this definition would require waiting for an infinite time and at an infinite distance from the black hole to verify that indeed, nothing has escaped, and thus cannot be used to identify a physical black hole. Broadly, physicists do not have a precisely-agreed-upon definition of a black hole. Among astrophysicists, a black hole is a compact object with a mass larger than four solar masses. A black hole may also be defined as a reservoir of information: 142 or a region where space is falling inwards faster than the speed of light. Properties The no-hair theorem postulates that, once it achieves a stable condition after formation, a black hole has only three independent physical properties: mass, electric charge, and angular momentum; the black hole is otherwise featureless. If the conjecture is true, any two black holes that share the same values for these properties, or parameters, are indistinguishable from one another. The degree to which the conjecture is true for real black holes is currently an unsolved problem. The simplest static black holes have mass but neither electric charge nor angular momentum. According to Birkhoff's theorem, these Schwarzschild black holes are the only vacuum solution that is spherically symmetric. Solutions describing more general black holes also exist. Non-rotating charged black holes are described by the Reissner–Nordström metric, while the Kerr metric describes a non-charged rotating black hole. The most general stationary black hole solution known is the Kerr–Newman metric, which describes a black hole with both charge and angular momentum. The simplest static black holes have mass but neither electric charge nor angular momentum. Contrary to the popular notion of a black hole "sucking in everything" in its surroundings, from far away, the external gravitational field of a black hole is identical to that of any other body of the same mass. While a black hole can theoretically have any positive mass, the charge and angular momentum are constrained by the mass. The total electric charge Q and the total angular momentum J are expected to satisfy the inequality Q 2 4 π ϵ 0 + c 2 J 2 G M 2 ≤ G M 2 {\displaystyle {\frac {Q^{2}}{4\pi \epsilon _{0}}}+{\frac {c^{2}J^{2}}{GM^{2}}}\leq GM^{2}} for a black hole of mass M. Black holes with the maximum possible charge or spin satisfying this inequality are called extremal black holes. Solutions of Einstein's equations that violate this inequality exist, but they do not possess an event horizon. These are so-called naked singularities that can be observed from the outside. Because these singularities make the universe inherently unpredictable, many physicists believe they could not exist. The weak cosmic censorship hypothesis, proposed by Sir Roger Penrose, rules out the formation of such singularities, when they are created through the gravitational collapse of realistic matter. However, this theory has not yet been proven, and some physicists believe that naked singularities could exist. It is also unknown whether black holes could even become extremal, forming naked singularities, since natural processes counteract increasing spin and charge when a black hole becomes near-extremal. The total mass of a black hole can be estimated by analyzing the motion of objects near the black hole, such as stars or gas. All black holes spin, often fast—One supermassive black hole, GRS 1915+105 has been estimated to spin at over 1,000 revolutions per second. The Milky Way's central black hole Sagittarius A* rotates at about 90% of the maximum rate. The spin rate can be inferred from measurements of atomic spectral lines in the X-ray range. As gas near the black hole plunges inward, high energy X-ray emission from electron-positron pairs illuminates the gas further out, appearing red-shifted due to relativistic effects. Depending on the spin of the black hole, this plunge happens at different radii from the hole, with different degrees of redshift. Astronomers can use the gap between the x-ray emission of the outer disk and the redshifted emission from plunging material to determine the spin of the black hole. A newer way to estimate spin is based on the temperature of gasses accreting onto the black hole. The method requires an independent measurement of the black hole mass and inclination angle of the accretion disk followed by computer modeling. Gravitational waves from coalescing binary black holes can also provide the spin of both progenitor black holes and the merged hole, but such events are rare. A spinning black hole has angular momentum. The supermassive black hole in the center of the Messier 87 (M87) galaxy appears to have an angular momentum very close to the maximum theoretical value. That uncharged limit is J ≤ G M 2 c , {\displaystyle J\leq {\frac {GM^{2}}{c}},} allowing definition of a dimensionless spin magnitude such that 0 ≤ c J G M 2 ≤ 1. {\displaystyle 0\leq {\frac {cJ}{GM^{2}}}\leq 1.} Most black holes are believed to have an approximately neutral charge. For example, Michal Zajaček, Arman Tursunov, Andreas Eckart, and Silke Britzen found the electric charge of Sagittarius A* to be at least ten orders of magnitude below the theoretical maximum. A charged black hole repels other like charges just like any other charged object. If a black hole were to become charged, particles with an opposite sign of charge would be pulled in by the extra electromagnetic force, while particles with the same sign of charge would be repelled, neutralizing the black hole. This effect may not be as strong if the black hole is also spinning. The presence of charge can reduce the diameter of the black hole by up to 38%. The charge Q for a nonspinning black hole is bounded by Q ≤ G M , {\displaystyle Q\leq {\sqrt {G}}M,} where G is the gravitational constant and M is the black hole's mass. Classification Black holes can have a wide range of masses. The minimum mass of a black hole formed by stellar gravitational collapse is governed by the maximum mass of a neutron star and is believed to be approximately two-to-four solar masses. However, theoretical primordial black holes, believed to have formed soon after the Big Bang, could be far smaller, with masses as little as 10−5 grams at formation. These very small black holes are sometimes called micro black holes. Black holes formed by stellar collapse are called stellar black holes. Estimates of their maximum mass at formation vary, but generally range from 10 to 100 solar masses, with higher estimates for black holes progenated by low-metallicity stars. The mass of a black hole formed via a supernova has a lower bound: If the progenitor star is too small, the collapse may be stopped by the degeneracy pressure of the star's constituents, allowing the condensation of matter into an exotic denser state. Degeneracy pressure occurs from the Pauli exclusion principle—Particles will resist being in the same place as each other. Smaller progenitor stars, with masses less than about 8 M☉, will be held together by the degeneracy pressure of electrons and will become a white dwarf. For more massive progenitor stars, electron degeneracy pressure is no longer strong enough to resist the force of gravity and the star will be held together by neutron degeneracy pressure, which can occur at much higher densities, forming a neutron star. If the star is still too massive, even neutron degeneracy pressure will not be able to resist the force of gravity and the star will collapse into a black hole.: 5.8 Stellar black holes can also gain mass via accretion of nearby matter, often from a companion object such as a star. Black holes that are larger than stellar black holes but smaller than supermassive black holes are called intermediate-mass black holes, with masses of approximately 102 to 105 solar masses. These black holes seem to be rarer than their stellar and supermassive counterparts, with relatively few candidates having been observed. Physicists have speculated that such black holes may form from collisions in globular and star clusters or at the center of low-mass galaxies. They may also form as the result of mergers of smaller black holes, with several LIGO observations finding merged black holes within the 110-350 solar mass range. The black holes with the largest masses are called supermassive black holes, with masses more than 106 times that of the Sun. These black holes are believed to exist at the centers of almost every large galaxy, including the Milky Way. Some scientists have proposed a subcategory of even larger black holes, called ultramassive black holes, with masses greater than 109-1010 solar masses. Theoretical models predict that the accretion disc that feeds black holes will be unstable once a black hole reaches 50-100 billion times the mass of the Sun, setting a rough upper limit to black hole mass. Structure While black holes are conceptually invisible sinks of all matter and light, in astronomical settings, their enormous gravity alters the motion of surrounding objects and pulls nearby gas inwards at near-light speed, making the area around black holes the brightest objects in the universe. Some black holes have relativistic jets—thin streams of plasma travelling away from the black hole at more than one-tenth of the speed of light. A small faction of the matter falling towards the black hole gets accelerated away along the hole rotation axis. These jets can extend as far as millions of parsecs from the black hole itself. Black holes of any mass can have jets. However, they are typically observed around spinning black holes with strongly-magnetized accretion disks. Relativistic jets were more common in the early universe, when galaxies and their corresponding supermassive black holes were rapidly gaining mass. All black holes with jets also have an accretion disk, but the jets are usually brighter than the disk. Quasars, typically found in other galaxies, are believed to be supermassive black holes with jets; microquasars are believed to be stellar-mass objects with jets, typically observed in the Milky Way. The mechanism of formation of jets is not yet known, but several options have been proposed. One method proposed to fuel these jets is the Blandford-Znajek process, which suggests that the dragging of magnetic field lines by a black hole's rotation could launch jets of matter into space. The Penrose process, which involves extraction of a black hole's rotational energy, has also been proposed as a potential mechanism of jet propulsion. Due to conservation of angular momentum, gas falling into the gravitational well created by a massive object will typically form a disk-like structure around the object.: 242 As the disk's angular momentum is transferred outward due to internal processes, its matter falls farther inward, converting its gravitational energy into heat and releasing a large flux of x-rays. The temperature of these disks can range from thousands to millions of Kelvin, and temperatures can differ throughout a single accretion disk. Accretion disks can also emit in other parts of the electromagnetic spectrum, depending on the disk's turbulence and magnetization and the black hole's mass and angular momentum. Accretion disks can be defined as geometrically thin or geometrically thick. Geometrically thin disks are mostly confined to the black hole's equatorial plane and have a well-defined edge at the innermost stable circular orbit (ISCO), while geometrically thick disks are supported by internal pressure and temperature and can extend inside the ISCO. Disks with high rates of electron scattering and absorption, appearing bright and opaque, are called optically thick; optically thin disks are more translucent and produce fainter images when viewed from afar. Accretion disks of black holes accreting beyond the Eddington limit are often referred to as polish donuts due to their thick, toroidal shape that resembles that of a donut. Quasar accretion disks are expected to usually appear blue in color. The disk for a stellar black hole, on the other hand, would likely look orange, yellow, or red, with its inner regions being the brightest. Theoretical research suggests that the hotter a disk is, the bluer it should be, although this is not always supported by observations of real astronomical objects. Accretion disk colors may also be altered by the Doppler effect, with the part of the disk travelling towards an observer appearing bluer and brighter and the part of the disk travelling away from the observer appearing redder and dimmer. In Newtonian gravity, test particles can stably orbit at arbitrary distances from a central object. In general relativity, however, there exists a smallest possible radius for which a massive particle can orbit stably. Any infinitesimal inward perturbations to this orbit will lead to the particle spiraling into the black hole, and any outward perturbations will, depending on the energy, cause the particle to spiral in, move to a stable orbit further from the black hole, or escape to infinity. This orbit is called the innermost stable circular orbit, or ISCO. The location of the ISCO depends on the spin of the black hole and the spin of the particle itself. In the case of a Schwarzschild black hole (spin zero) and a particle without spin, the location of the ISCO is: r I S C O = 3 r s = 6 G M c 2 , {\displaystyle r_{\rm {ISCO}}=3\,r_{\text{s}}={\frac {6\,GM}{c^{2}}},} where r I S C O {\displaystyle r_{\rm {_{ISCO}}}} is the radius of the ISCO, r s {\displaystyle r_{\text{s}}} is the Schwarzschild radius of the black hole, G {\displaystyle G} is the gravitational constant, and c {\displaystyle c} is the speed of light. The radius of this orbit changes slightly based on particle spin. For charged black holes, the ISCO moves inwards. For spinning black holes, the ISCO is moved inwards for particles orbiting in the same direction that the black hole is spinning (prograde) and outwards for particles orbiting in the opposite direction (retrograde). For example, the ISCO for a particle orbiting retrograde can be as far out as about 9 r s {\displaystyle 9r_{\text{s}}} , while the ISCO for a particle orbiting prograde can be as close as at the event horizon itself. The photon sphere is a spherical boundary for which photons moving on tangents to that sphere are bent completely around the black hole, possibly orbiting multiple times. Light rays with impact parameters less than the radius of the photon sphere enter the black hole. For Schwarzschild black holes, the photon sphere has a radius 1.5 times the Schwarzschild radius; the radius for non-Schwarzschild black holes is at least 1.5 times the radius of the event horizon. When viewed from a great distance, the photon sphere creates an observable black hole shadow. Since no light emerges from within the black hole, this shadow is the limit for possible observations.: 152 The shadow of colliding black holes should have characteristic warped shapes, allowing scientists to detect black holes that are about to merge. While light can still escape from the photon sphere, any light that crosses the photon sphere on an inbound trajectory will be captured by the black hole. Therefore, any light that reaches an outside observer from the photon sphere must have been emitted by objects between the photon sphere and the event horizon. Light emitted towards the photon sphere may also curve around the black hole and return to the emitter. For a rotating, uncharged black hole, the radius of the photon sphere depends on the spin parameter and whether the photon is orbiting prograde or retrograde. For a photon orbiting prograde, the photon sphere will be 1-3 Schwarzschild radii from the center of the black hole, while for a photon orbiting retrograde, the photon sphere will be between 3-5 Schwarzschild radii from the center of the black hole. The exact location of the photon sphere depends on the magnitude of the black hole's rotation. For a charged, nonrotating black hole, there will only be one photon sphere, and the radius of the photon sphere will decrease for increasing black hole charge. For non-extremal, charged, rotating black holes, there will always be two photon spheres, with the exact radii depending on the parameters of the black hole. Near a rotating black hole, spacetime rotates similar to a vortex. The rotating spacetime will drag any matter and light into rotation around the spinning black hole. This effect of general relativity, called frame dragging, gets stronger closer to the spinning mass. The region of spacetime in which it is impossible to stay still is called the ergosphere. The ergosphere of a black hole is a volume bounded by the black hole's event horizon and the ergosurface, which coincides with the event horizon at the poles but bulges out from it around the equator. Matter and radiation can escape from the ergosphere. Through the Penrose process, objects can emerge from the ergosphere with more energy than they entered with. The extra energy is taken from the rotational energy of the black hole, slowing down the rotation of the black hole.: 268 A variation of the Penrose process in the presence of strong magnetic fields, the Blandford–Znajek process, is considered a likely mechanism for the enormous luminosity and relativistic jets of quasars and other active galactic nuclei. The observable region of spacetime around a black hole closest to its event horizon is called the plunging region. In this area it is no longer possible for free falling matter to follow circular orbits or stop a final descent into the black hole. Instead, it will rapidly plunge toward the black hole at close to the speed of light, growing increasingly hot and producing a characteristic, detectable thermal emission. However, light and radiation emitted from this region can still escape from the black hole's gravitational pull. For a nonspinning, uncharged black hole, the radius of the event horizon, or Schwarzschild radius, is proportional to the mass, M, through r s = 2 G M c 2 ≈ 2.95 M M ⊙ k m , {\displaystyle r_{\mathrm {s} }={\frac {2GM}{c^{2}}}\approx 2.95\,{\frac {M}{M_{\odot }}}~\mathrm {km,} } where rs is the Schwarzschild radius and M☉ is the mass of the Sun.: 124 For a black hole with nonzero spin or electric charge, the radius is smaller,[Note 1] until an extremal black hole could have an event horizon close to r + = G M c 2 , {\displaystyle r_{\mathrm {+} }={\frac {GM}{c^{2}}},} half the radius of a nonspinning, uncharged black hole of the same mass. Since the volume within the Schwarzschild radius increase with the cube of the radius, average density of a black hole inside its Schwarzschild radius is inversely proportional to the square of its mass: supermassive black holes are much less dense than stellar black holes. The average density of a 108 M☉ black hole is comparable to that of water. The defining feature of a black hole is the existence of an event horizon, a boundary in spacetime through which matter and light can pass only inward towards the center of the black hole. Nothing, not even light, can escape from inside the event horizon. The event horizon is referred to as such because if an event occurs within the boundary, information from that event cannot reach or affect an outside observer, making it impossible to determine whether such an event occurred.: 179 For non-rotating black holes, the geometry of the event horizon is precisely spherical, while for rotating black holes, the event horizon is oblate. To a distant observer, a clock near a black hole would appear to tick more slowly than one further from the black hole.: 217 This effect, known as gravitational time dilation, would also cause an object falling into a black hole to appear to slow as it approached the event horizon, never quite reaching the horizon from the perspective of an outside observer.: 218 All processes on this object would appear to slow down, and any light emitted by the object to appear redder and dimmer, an effect known as gravitational redshift. An object falling from half of a Schwarzschild radius above the event horizon would fade away until it could no longer be seen, disappearing from view within one hundredth of a second. It would also appear to flatten onto the black hole, joining all other material that had ever fallen into the hole. On the other hand, an observer falling into a black hole would not notice any of these effects as they cross the event horizon. Their own clocks appear to them to tick normally, and they cross the event horizon after a finite time without noting any singular behaviour. In general relativity, it is impossible to determine the location of the event horizon from local observations, due to Einstein's equivalence principle.: 222 Black holes that are rotating and/or charged have an inner horizon, often called the Cauchy horizon, inside of the black hole. The inner horizon is divided up into two segments: an ingoing section and an outgoing section. At the ingoing section of the Cauchy horizon, radiation and matter that fall into the black hole would build up at the horizon, causing the curvature of spacetime to go to infinity. This would cause an observer falling in to experience tidal forces. This phenomenon is often called mass inflation, since it is associated with a parameter dictating the black hole's internal mass growing exponentially, and the buildup of tidal forces is called the mass-inflation singularity or Cauchy horizon singularity. Some physicists have argued that in realistic black holes, accretion and Hawking radiation would stop mass inflation from occurring. At the outgoing section of the inner horizon, infalling radiation would backscatter off of the black hole's spacetime curvature and travel outward, building up at the outgoing Cauchy horizon. This would cause an infalling observer to experience a gravitational shock wave and tidal forces as the spacetime curvature at the horizon grew to infinity. This buildup of tidal forces is called the shock singularity. Both of these singularities are weak, meaning that an object crossing them would only be deformed a finite amount by tidal forces, even though the spacetime curvature would still be infinite at the singularity. This is as opposed to a strong singularity, where an object hitting the singularity would be stretched and squeezed by an infinite amount. They are also null singularities, meaning that a photon could travel parallel to the them without ever being intercepted. Ignoring quantum effects, every black hole has a singularity inside, points where the curvature of spacetime becomes infinite, and geodesics terminate within a finite proper time.: 205 For a non-rotating black hole, this region takes the shape of a single point; for a rotating black hole it is smeared out to form a ring singularity that lies in the plane of rotation.: 264 In both cases, the singular region has zero volume. All of the mass of the black hole ends up in the singularity.: 252 Since the singularity has nonzero mass in an infinitely small space, it can be thought of as having infinite density. Observers falling into a Schwarzschild black hole (i.e., non-rotating and not charged) cannot avoid being carried into the singularity once they cross the event horizon. As they fall further into the black hole, they will be torn apart by the growing tidal forces in a process sometimes referred to as spaghettification or the noodle effect. Eventually, they will reach the singularity and be crushed into an infinitely small point.: 182 However any perturbations, such as those caused by matter or radiation falling in, would cause space to oscillate chaotically near the singularity. Any matter falling in would experience intense tidal forces rapidly changing in direction, all while being compressed into an increasingly small volume. Alternative forms of general relativity, including addition of some quatum effects, can lead to regular, or nonsingular, black holes without singularities. For example, the fuzzball model, based on string theory, states that black holes are actually made up of quantum microstates and need not have a singularity or an event horizon. The theory of loop quantum gravity proposes that the curvature and density at the center of a black hole is large, but not infinite. Formation Black holes are formed by gravitational collapse of massive stars, either by direct collapse or during a supernova explosion in a process called fallback. Black holes can result from the merger of two neutron stars or a neutron star and a black hole. Other more speculative mechanisms include primordial black holes created from density fluctuations in the early universe, the collapse of dark stars, a hypothetical object powered by annihilation of dark matter, or from hypothetical self-interacting dark matter. Gravitational collapse occurs when an object's internal pressure is insufficient to resist the object's own gravity. At the end of a star's life, it will run out of hydrogen to fuse, and will start fusing more and more massive elements, until it gets to iron. Since the fusion of elements heavier than iron would require more energy than it would release, nuclear fusion ceases. If the iron core of the star is too massive, the star will no longer be able to support itself and will undergo gravitational collapse. While most of the energy released during gravitational collapse is emitted very quickly, an outside observer does not actually see the end of this process. Even though the collapse takes a finite amount of time from the reference frame of infalling matter, a distant observer would see the infalling material slow and halt just above the event horizon, due to gravitational time dilation. Light from the collapsing material takes longer and longer to reach the observer, with the delay growing to infinity as the emitting material reaches the event horizon. Thus the external observer never sees the formation of the event horizon; instead, the collapsing material seems to become dimmer and increasingly red-shifted, eventually fading away. Observations of quasars at redshift z ∼ 7 {\displaystyle z\sim 7} , less than a billion years after the Big Bang, has led to investigations of other ways to form black holes. The accretion process to build supermassive black holes has a limiting rate of mass accumulation and a billion years is not enough time to reach quasar status. One suggestion is direct collapse of nearly pure hydrogen gas (low metalicity) clouds characteristic of the young universe, forming a supermassive star which collapses into a black hole. It has been suggested that seed black holes with typical masses of ~105 M☉ could have formed in this way which then could grow to ~109 M☉. However, the very large amount of gas required for direct collapse is not typically stable to fragmentation to form multiple stars. Thus another approach suggests massive star formation followed by collisions that seed massive black holes which ultimately merge to create a quasar.: 85 A neutron star in a common envelope with a regular star can accrete sufficient material to collapse to a black hole or two neutron stars can merge. These avenues for the formation of black holes are considered relatively rare. In the current epoch of the universe, conditions needed to form black holes are rare and are mostly only found in stars. However, in the early universe, conditions may have allowed for black hole formations via other means. Fluctuations of spacetime soon after the Big Bang may have formed areas that were denser then their surroundings. Initially, these regions would not have been compact enough to form a black hole, but eventually, the curvature of spacetime in the regions become large enough to cause them to collapse into a black hole. Different models for the early universe vary widely in their predictions of the scale of these fluctuations. Various models predict the creation of primordial black holes ranging from a Planck mass (~2.2×10−8 kg) to hundreds of thousands of solar masses. Primordial black holes with masses less than 1015 g would have evaporated by now due to Hawking radiation. Despite the early universe being extremely dense, it did not re-collapse into a black hole during the Big Bang, since the universe was expanding rapidly and did not have the gravitational differential necessary for black hole formation. Models for the gravitational collapse of objects of relatively constant size, such as stars, do not necessarily apply in the same way to rapidly expanding space such as the Big Bang. In principle, black holes could be formed in high-energy particle collisions that achieve sufficient density, although no such events have been detected. These hypothetical micro black holes, which could form from the collision of cosmic rays and Earth's atmosphere or in particle accelerators like the Large Hadron Collider, would not be able to aggregate additional mass. Instead, they would evaporate in about 10−25 seconds, posing no threat to the Earth. Evolution Black holes can also merge with other objects such as stars or even other black holes. This is thought to have been important, especially in the early growth of supermassive black holes, which could have formed from the aggregation of many smaller objects. The process has also been proposed as the origin of some intermediate-mass black holes. Mergers of supermassive black holes may take a long time: As a binary of supermassive black holes approach each other, most nearby stars are ejected, leaving little for the remaining black holes to gravitationally interact with that would allow them to get closer to each other. This phenomenon has been called the final parsec problem, as the distance at which this happens is usually around one parsec. When a black hole accretes matter, the gas in the inner accretion disk orbits at very high speeds because of its proximity to the black hole. The resulting friction heats the inner disk to temperatures at which it emits vast amounts of electromagnetic radiation (mainly X-rays) detectable by telescopes. By the time the matter of the disk reaches the ISCO, between 5.7% and 42% of its mass will have been converted to energy, depending on the black hole's spin. About 90% of this energy is released within about 20 black hole radii. In many cases, accretion disks are accompanied by relativistic jets that are emitted along the black hole's poles, which carry away much of the energy. The mechanism for the creation of these jets is currently not well understood, in part due to insufficient data. Many of the universe's most energetic phenomena have been attributed to the accretion of matter on black holes. Active galactic nuclei and quasars are believed to be the accretion disks of supermassive black holes. X-ray binaries are generally accepted to be binary systems in which one of the two objects is a compact object accreting matter from its companion. Ultraluminous X-ray sources may be the accretion disks of intermediate-mass black holes. At a certain rate of accretion, the outward radiation pressure will become as strong as the inward gravitational force, and the black hole should unable to accrete any faster. This limit is called the Eddington limit. However, many black holes accrete beyond this rate due to their non-spherical geometry or instabilities in the accretion disk. Accretion beyond the limit is called Super-Eddington accretion and may have been commonplace in the early universe. Stars have been observed to get torn apart by tidal forces in the immediate vicinity of supermassive black holes in galaxy nuclei, in what is known as a tidal disruption event (TDE). Some of the material from the disrupted star forms an accretion disk around the black hole, which emits observable electromagnetic radiation. The correlation between the masses of supermassive black holes in the centres of galaxies with the velocity dispersion and mass of stars in their host bulges suggests that the formation of galaxies and the formation of their central black holes are related. Black hole winds from rapid accretion, particularly when the galaxy itself is still accreting matter, can compress gas nearby, accelerating star formation. However, if the winds become too strong, the black hole may blow nearly all of the gas out of the galaxy, quenching star formation. Black hole jets may also energize nearby cavities of plasma and eject low-entropy gas from out of the galactic core, causing gas in galactic centers to be hotter than expected. If Hawking's theory of black hole radiation is correct, then black holes are expected to shrink and evaporate over time as they lose mass by the emission of photons and other particles. The temperature of this thermal spectrum (Hawking temperature) is proportional to the surface gravity of the black hole, which is inversely proportional to the mass. Hence, large black holes emit less radiation than small black holes.: Ch. 9.6 A stellar black hole of 1 M☉ has a Hawking temperature of 62 nanokelvins. This is far less than the 2.7 K temperature of the cosmic microwave background radiation. Stellar-mass or larger black holes receive more mass from the cosmic microwave background than they emit through Hawking radiation and thus will grow instead of shrinking. To have a Hawking temperature larger than 2.7 K (and be able to evaporate), a black hole would need a mass less than the Moon. Such a black hole would have a diameter of less than a tenth of a millimetre. The Hawking radiation for an astrophysical black hole is predicted to be very weak and would thus be exceedingly difficult to detect from Earth. A possible exception is the burst of gamma rays emitted in the last stage of the evaporation of primordial black holes. Searches for such flashes have proven unsuccessful and provide stringent limits on the possibility of existence of low mass primordial black holes, with modern research predicting that primordial black holes must make up less than a fraction of 10−7 of the universe's total mass. NASA's Fermi Gamma-ray Space Telescope, launched in 2008, has searched for these flashes, but has not yet found any. The properties of a black hole are constrained and interrelated by the theories that predict these properties. When based on general relativity, these relationships are called the laws of black hole mechanics. For a black hole that is not still forming or accreting matter, the zeroth law of black hole mechanics states the black hole's surface gravity is constant across the event horizon. The first law relates changes in the black hole's surface area, angular momentum, and charge to changes in its energy. The second law says the surface area of a black hole never decreases on its own. Finally, the third law says that the surface gravity of a black hole is never zero. These laws are mathematical analogs of the laws of thermodynamics. They are not equivalent, however, because, according to general relativity without quantum mechanics, a black hole can never emit radiation, and thus its temperature must always be zero.: 11 Quantum mechanics predicts that a black hole will continuously emit thermal Hawking radiation, and therefore must always have a nonzero temperature. It also predicts that all black holes have entropy which scales with their surface area. When quantum mechanics is accounted for, the laws of black hole mechanics become equivalent to the classical laws of thermodynamics. However, these conclusions are derived without a complete theory of quantum gravity, although many potential theories do predict black holes having entropy and temperature. Thus, the true quantum nature of black hole thermodynamics continues to be debated.: 29 Observational evidence Millions of black holes with around 30 solar masses derived from stellar collapse are expected to exist in the Milky Way. Even a dwarf galaxy like Draco should have hundreds. Only a few of these have been detected. By nature, black holes do not themselves emit any electromagnetic radiation other than the hypothetical Hawking radiation, so astrophysicists searching for black holes must generally rely on indirect observations. The defining characteristic of a black hole is its event horizon. The horizon itself cannot be imaged, so all other possible explanations for these indirect observations must be considered and eliminated before concluding that a black hole has been observed.: 11 The Event Horizon Telescope (EHT) is a global system of radio telescopes capable of directly observing a black hole shadow. The angular resolution of a telescope is based on its aperture and the wavelengths it is observing. Because the angular diameters of Sagittarius A* and Messier 87* in the sky are very small, a single telescope would need to be about the size of the Earth to clearly distinguish their horizons using radio wavelengths. By combining data from several different radio telescopes around the world, the Event Horizon Telescope creates an effective aperture the diameter size of the Earth. The EHT team used imaging algorithms to compute the most probable image from the data in its observations of Sagittarius A* and M87*. Gravitational-wave interferometry can be used to detect merging black holes and other compact objects. In this method, a laser beam is split down two long arms of a tunnel. The laser beams reflect off of mirrors in the tunnels and converge at the intersection of the arms, cancelling each other out. However, when a gravitational wave passes, it warps spacetime, changing the lengths of the arms themselves. Since each laser beam is now travelling a slightly different distance, they do not cancel out and produce a recognizable signal. Analysis of the signal can give scientists information about what caused the gravitational waves. Since gravitational waves are very weak, gravitational-wave observatories such as LIGO must have arms several kilometers long and carefully control for noise from Earth to be able to detect these gravitational waves. Since the first measurements in 2016, multiple gravitational waves from black holes have been detected and analyzed. The proper motions of stars near the centre of the Milky Way provide strong observational evidence that these stars are orbiting a supermassive black hole. Since 1995, astronomers have tracked the motions of 90 stars orbiting an invisible object coincident with the radio source Sagittarius A*. In 1998, by fitting the motions of the stars to Keplerian orbits, the astronomers were able to infer that Sagittarius A* must be a 2.6×106 M☉ object must be contained within a radius of 0.02 light-years. Since then, one of the stars—called S2—has completed a full orbit. From the orbital data, astronomers were able to refine the calculations of the mass of Sagittarius A* to 4.3×106 M☉, with a radius of less than 0.002 light-years. This upper limit radius is larger than the Schwarzschild radius for the estimated mass, so the combination does not prove Sagittarius A* is a black hole. Nevertheless, these observations strongly suggest that the central object is a supermassive black hole as there are no other plausible scenarios for confining so much invisible mass into such a small volume. Additionally, there is some observational evidence that this object might possess an event horizon, a feature unique to black holes. The Event Horizon Telescope image of Sagittarius A*, released in 2022, provided further confirmation that it is indeed a black hole. X-ray binaries are binary systems that emit a majority of their radiation in the X-ray part of the electromagnetic spectrum. These X-ray emissions result when a compact object accretes matter from an ordinary star. The presence of an ordinary star in such a system provides an opportunity for studying the central object and to determine if it might be a black hole. By measuring the orbital period of the binary, the distance to the binary from Earth, and the mass of the companion star, scientists can estimate the mass of the compact object. The Tolman-Oppenheimer-Volkoff limit (TOV limit) dictates the largest mass a nonrotating neutron star can be, and is estimated to be about two solar masses. While a rotating neutron star can be slightly more massive, if the compact object is much more massive than the TOV limit, it cannot be a neutron star and is generally expected to be a black hole. The first strong candidate for a black hole, Cygnus X-1, was discovered in this way by Charles Thomas Bolton, Louise Webster, and Paul Murdin in 1972. Observations of rotation broadening of the optical star reported in 1986 lead to a compact object mass estimate of 16 solar masses, with 7 solar masses as the lower bound. In 2011, this estimate was updated to 14.1±1.0 M☉ for the black hole and 19.2±1.9 M☉ for the optical stellar companion. X-ray binaries can be categorized as either low-mass or high-mass; This classification is based on the mass of the companion star, not the compact object itself. In a class of X-ray binaries called soft X-ray transients, the companion star is of relatively low mass, allowing for more accurate estimates of the black hole mass. These systems actively emit X-rays for only several months once every 10–50 years. During the period of low X-ray emission, called quiescence, the accretion disk is extremely faint, allowing detailed observation of the companion star. Numerous black hole candidates have been measured by this method. Black holes are also sometimes found in binaries with other compact objects, such as white dwarfs, neutron stars, and other black holes. The centre of nearly every galaxy contains a supermassive black hole. The close observational correlation between the mass of this hole and the velocity dispersion of the host galaxy's bulge, known as the M–sigma relation, strongly suggests a connection between the formation of the black hole and that of the galaxy itself. Astronomers use the term active galaxy to describe galaxies with unusual characteristics, such as unusual spectral line emission and very strong radio emission. Theoretical and observational studies have shown that the high levels of activity in the centers of these galaxies, regions called active galactic nuclei (AGN), may be explained by accretion onto supermassive black holes. These AGN consist of a central black hole that may be millions or billions of times more massive than the Sun, a disk of interstellar gas and dust called an accretion disk, and two jets perpendicular to the accretion disk. Although supermassive black holes are expected to be found in most AGN, only some galaxies' nuclei have been more carefully studied in attempts to both identify and measure the actual masses of the central supermassive black hole candidates. Some of the most notable galaxies with supermassive black hole candidates include the Andromeda Galaxy, Messier 32, Messier 87, the Sombrero Galaxy, and the Milky Way itself. Another way black holes can be detected is through observation of effects caused by their strong gravitational field. One such effect is gravitational lensing: The deformation of spacetime around a massive object causes light rays to be deflected, making objects behind them appear distorted. When the lensing object is a black hole, this effect can be strong enough to create multiple images of a star or other luminous source. However, the distance between the lensed images may be too small for contemporary telescopes to resolve—this phenomenon is called microlensing. Instead of seeing two images of a lensed star, astronomers see the star brighten slightly as the black hole moves towards the line of sight between the star and Earth and then return to its normal luminosity as the black hole moves away. The turn of the millennium saw the first 3 candidate detections of black holes in this way, and in January 2022, astronomers reported the first confirmed detection of a microlensing event from an isolated black hole. This was also the first determination of an isolated black hole mass, 7.1±1.3 M☉. Alternatives While there is a strong case for supermassive black holes, the model for stellar-mass black holes assumes of an upper limit for the mass of a neutron star: objects observed to have more mass are assumed to be black holes. However, the properties of extremely dense matter are poorly understood. New exotic phases of matter could allow other kinds of massive objects. Quark stars would be made up of quark matter and supported by quark degeneracy pressure, a form of degeneracy pressure even stronger than neutron degeneracy pressure. This would halt gravitational collapse at a higher mass than for a neutron star. Even stronger stars called electroweak stars would convert quarks in their cores into leptons, providing additional pressure to stop the star from collapsing. If, as some extensions of the Standard Model posit, quarks and leptons are made up of the even-smaller fundamental particles called preons, a very compact star could be supported by preon degeneracy pressure. While none of these hypothetical models can explain all of the observations of stellar black hole candidates, a Q star is the only alternative which could significantly exceed the mass limit for neutron stars and thus provide an alternative for supermassive black holes.: 12 A few theoretical objects have been conjectured to match observations of astronomical black hole candidates identically or near-identically, but which function via a different mechanism. A dark energy star would convert infalling matter into vacuum energy; This vacuum energy would be much larger than the vacuum energy of outside space, exerting outwards pressure and preventing a singularity from forming. A black star would be gravitationally collapsing slowly enough that quantum effects would keep it just on the cusp of fully collapsing into a black hole. A gravastar would consist of a very thin shell and a dark-energy interior providing outward pressure to stop the collapse into a black hole or formation of a singularity; It could even have another gravastar inside, called a 'nestar'. Open questions According to the no-hair theorem, a black hole is defined by only three parameters: its mass, charge, and angular momentum. This seems to mean that all other information about the matter that went into forming the black hole is lost, as there is no way to determine anything about the black hole from outside other than those three parameters. When black holes were thought to persist forever, this information loss was not problematic, as the information can be thought of as existing inside the black hole. However, black holes slowly evaporate by emitting Hawking radiation. This radiation does not appear to carry any additional information about the matter that formed the black hole, meaning that this information is seemingly gone forever. This is called the black hole information paradox. Theoretical studies analyzing the paradox have led to both further paradoxes and new ideas about the intersection of quantum mechanics and general relativity. While there is no consensus on the resolution of the paradox, work on the problem is expected to be important for a theory of quantum gravity.: 126 Observations of faraway galaxies have found that ultraluminous quasars, powered by supermassive black holes, existed in the early universe as far as redshift z ≥ 7 {\displaystyle z\geq 7} . These black holes have been assumed to be the products of the gravitational collapse of large population III stars. However, these stellar remnants were not massive enough to produce the quasars observed at early times without accreting beyond the Eddington limit, the theoretical maximum rate of black hole accretion. Physicists have suggested a variety of different mechanisms by which these supermassive black holes may have formed. It has been proposed that smaller black holes may have also undergone mergers to produce the observed supermassive black holes. It is also possible that they were seeded by direct-collapse black holes, in which a large cloud of hot gas avoids fragmentation that would lead to multiple stars, due to low angular momentum or heating from a nearby galaxy. Given the right circumstances, a single supermassive star forms and collapses directly into a black hole without undergoing typical stellar evolution. Additionally, these supermassive black holes in the early universe may be high-mass primordial black holes, which could have accreted further matter in the centers of galaxies. Finally, certain mechanisms allow black holes to grow faster than the theoretical Eddington limit, such as dense gas in the accretion disk limiting outward radiation pressure that prevents the black hole from accreting. However, the formation of bipolar jets prevent super-Eddington rates. In fiction Black holes have been portrayed in science fiction in a variety of ways. Even before the advent of the term itself, objects with characteristics of black holes appeared in stories such as the 1928 novel The Skylark of Space with its "black Sun" and the "hole in space" in the 1935 short story Starship Invincible. As black holes grew to public recognition in the 1960s and 1970s, they began to be featured in films as well as novels, such as Disney's The Black Hole. Black holes have also been used in works of the 21st century, such as Christopher Nolan's science fiction epic Interstellar. Authors and screenwriters have exploited the relativistic effects of black holes, particularly gravitational time dilation. For example, Interstellar features a black hole planet with a time dilation factor of over 60,000:1, while the 1977 novel Gateway depicts a spaceship approaching but never crossing the event horizon of a black hole from the perspective of an outside observer due to time dilation effects. Black holes have also been appropriated as wormholes or other methods of faster-than-light travel, such as in the 1974 novel The Forever War, where a network of black holes is used for interstellar travel. Additionally, black holes can feature as hazards to spacefarers and planets: A black hole threatens a deep-space outpost in 1978 short story The Black Hole Passes, and a binary black hole dangerously alters the orbit of a planet in the 2018 Netflix reboot of Lost in Space. Notes References Further reading External links |
======================================== |
[SOURCE: https://en.wikipedia.org/wiki/Conditional_(computer_programming)#Else_if] | [TOKENS: 3185] |
Contents Conditional (computer programming) In computer programming, a conditional statement directs program control flow based on the value of a condition; a Boolean expression. A conditional expression evaluates to a value without the side-effect of changing control flow. Many programming languages (such as C) have distinct conditional statements and expressions. In pure functional programming, a conditional expression does not have side-effects, many functional programming languages with conditional expressions (such as Lisp) support side-effects. Conditional statement Although the syntax of an if-then-else statement varies by language, the general syntax is shown as pseudocode below. The part represented by the condition placeholder is an expression that evaluates to either true or false. If true, control passes to consequent and when complete to after end if. If false, control passes to alternative and when complete to after end if. As the else clause is optional, the else alternative part can be omitted. Typically, both consequent and alternative can be either a single statement or a block of statements. The following example, also in pseudocode, replaces placeholders with example logic. In early programming languages, especially dialects of BASIC, an if–then-else statement could only contain goto statements but this tended to result in hard-to-read spaghetti code. As a result, structured programming, which supports control flow via code blocks, gained in popularity, until it became the norm in most BASIC variants and all languages. Such mechanisms and principles were based on the ALGOL family of languages, including Pascal and Modula-2. While it is possible to use goto in a structured way, structured programming makes this easier. A structured if–then–else statement is one of the key elements of structured programming, and it is present in most popular languages such as C, Java, JavaScript and Visual Basic. The convention is that an else clause, like the then clause, responds to the nearest preceding if clause. However, the semantics of nested conditionals in some early languages such as ALGOL 60 were less than clear; the syntax was inadequate to always specify the same predicate if clause. Thus, the parser might randomly pair the else with any one of the perhaps manifold if clauses in the intended nested hierarchy. This is known as the dangling else problem. It is resolved in various ways, depending on the language (in some, by means of explicit block-ending syntax (such as end if ) or a block enclosure, such as curly brackets ( {⋯} ). Chaining conditionals is often provided in a language via an else-if construct. Only the statements following the first condition that is true are executed. Other statements are skipped. In placeholder pseudocode: In the following pseudocode, a shop offers as much as a 30% discount for an item. If the discount is 10%, then the first if statement is true and "you have to pay $30" is printed. All other statements below that first if statement are skipped. Only one end if is needed if one uses else if instead of else followed by if. In ALGOL 68, the 1968 “Draft Report” (circulated as a supplement to ALGOL Bulletin no. 26) still used the bold keyword elsf in “contracted” conditionals. The spelling elif was then standardized in the “Revised Report on the Algorithmic Language ALGOL 68” (1973), which lists both the bold words if ~ then ~ elif ~ else ~ fi and their “brief” symbols, where elif corresponds to |: in the compact form ( ~ | ~ |: ~ | ~ | ~ ). In Ada, the elseif keyword is syntactic sugar for the two words else if. PHP also supports an elseif keyword both for its curly brackets or colon syntaxes. Perl and Ruby provide the keyword elsif to avoid the large number of braces that would be required by multiple if and else statements. Python uses the special keyword elif because structure is denoted by indentation rather than braces, so a repeated use of else and if would require increased indentation after every condition. Visual Basic, supports ElseIf. Similarly, the earlier UNIX shells (later gathered up to the POSIX shell syntax) use elif too, but giving the choice of delimiting with spaces, line breaks, or both. However, in many languages more directly descended from Algol, such as Simula, Pascal, BCPL and C, this special syntax for the else if construct is not present, nor is it present in the many syntactical derivatives of C, such as Java, ECMAScript, and so on. This works because in these languages, any single statement (in this case if cond...) can follow a conditional without being enclosed in a block.[clarification needed] If all terms in the sequence of conditionals are testing the value of a single expression (e.g., if x = 0, else if x = 1, else if x = 2 ...), an alternative is the switch statement. In a language that does not have a switch statement, these can be encoded as a chained if-then-else. A switch statement supports multiway branching, often comparing the value of an expression with constant values and transferring control to the code of the first match. There is usually a provision for a default action if no match is found. An optimizing compiler may use a control table to implement the logic of a switch statement. In a dynamic language, the cases may not be limited to constant expressions, and might extend to pattern matching, as in the shell script example on the right, where the '*)' implements the default case as a regular expression matching any string. The Guarded Command Language (GCL) of Edsger Dijkstra supports conditional execution as a list of commands consisting of a Boolean-valued guard (corresponding to a condition) and its corresponding statement. In GCL, exactly one of the statements whose guards is true is evaluated, but which one is arbitrary. In this code the Gi's are the guards and the Si's are the statements. If none of the guards is true, the program's behavior is undefined. GCL is intended primarily for reasoning about programs, but similar notations have been implemented in Concurrent Pascal and occam. The earliest conditional statement in Fortran, up to Fortran 77, was the arithmetic if statement which jumped to one of three labels depending on whether a value (of type integer, real, or double precision) is <0, 0, or >0. In the following code, control passes to one of the labels based on the value of e. This is equivalent to the following sequence. As it acts like goto, arithmetic if is unstructured, not structured, programming. It was the only conditional statement in the original implementation of Fortran on the IBM 704 computer. On that computer, the test-and-branch op-code had three addresses for those three states. Other computers would have "flag" registers such as positive, zero, negative, even, overflow, carry, associated with the last arithmetic operations and would use instructions such as 'Branch if accumulator negative' then 'Branch if accumulator zero' or similar. Note that the expression is evaluated once only, and in cases such as integer arithmetic where overflow may occur, the overflow or carry flags would be considered also. The Arithmetic IF statement was listed as obsolescent starting with the Fortran 90 Standard. It was deleted from the Fortran 2018 Standard. Nonetheless most compilers continue to support it for compatibility with legacy codes. In contrast to other languages, in Smalltalk the conditional statement is not a language construct but defined in the class Boolean as an abstract method that takes two parameters, both closures. Boolean has two subclasses, True and False, which both define the method, True executing the first closure only, False executing the second closure only. JavaScript supports if-else statements similar to C syntax. The following example has conditional Math.random() < 0.5 which is true if the random float (value between 0 and 1) is greater than 0.5. The statement uses it to randomly choose between outputting You got Heads! or You got Tails!. Conditionals can be chained as shown below: In Lambda calculus, the concept of an if-then-else conditional can be expressed using the following expressions: Note: if ifThenElse is passed two functions as the left and right conditionals; it is necessary to also pass an empty tuple () to the result of ifThenElse in order to actually call the chosen function, otherwise ifThenElse will just return the function object without getting called. In a system where numbers can be used without definition (like Lisp, Traditional paper math, so on), the above can be expressed as a single closure below: Here, true, false, and ifThenElse are bound to their respective definitions which are passed to their scope at the end of their block. A working JavaScript analogy(using only functions of single variable for rigor) to this is as follows: The code above with multivariable functions looks like this: Another version of the earlier example without a system where numbers are assumed is below. The first example shows the first branch being taken, while second example shows the second branch being taken. Smalltalk uses a similar idea for its true and false representations, with True and False being singleton objects that respond to messages ifTrue/ifFalse differently. Haskell used to use this exact model for its Boolean type, but at the time of writing, most Haskell programs use syntactic sugar if a then b else c construct which unlike ifThenElse does not compose unless either wrapped in another function or re-implemented as shown in The Haskell section of this page. Conditional expression Many languages support a conditional expression, which unlike a statement evaluates to a value instead of controlling control flow. The concept of conditional expression was first developed by John McCarthy during his research into symbolic processing and LISP in the late 1950s. ALGOL 60 and some other members of the ALGOL family allow if–then–else as an expression. The idea of including conditional expressions was suggested by John McCarthy, though the ALGOL committee decided to use English words rather than McCarthy's mathematical syntax: Compound statements are all terminated (guarded) by distinctive closing brackets: This scheme not only avoids the dangling else problem but also avoids having to use BEGIN and END in embedded statement sequences. Choice clause example with Brief symbols: Conditional expressions have always been a fundamental part of Lisp . In pure LISP, the COND function is used. In dialects such as Scheme, Racket and Common Lisp : In Haskell 98, there is only an if expression, no if statement, and the else part is compulsory, as every expression must have some value. Logic that would be expressed with conditionals in other languages is usually expressed with pattern matching in recursive functions. Because Haskell is lazy, it is possible to write control structures, such as if, as ordinary expressions; the lazy evaluation means that an if function can evaluate only the condition and proper branch (where a strict language would evaluate all three). It can be written like this: C and related languages support a ternary operator that provides for conditional expressions like: If condition is true, then the expression evaluates to true-value; otherwise to false-value. In the following code, r is assigned to "foo" if x > 10, and to "bar" if not. To accomplish the same using an if-statement, this would take more than one statement, and require mentioning r twice: Some argue that the explicit if-then statement is easier to read and that it may compile to more efficient code than the ternary operator, while others argue that concise expressions are easier to read and better since they have less repeated clauses. In Visual Basic and some other languages, a function called IIf is provided, which can be used as a conditional expression. However, it does not behave like a true conditional expression, because both the true and false branches are always evaluated; it is just that the result of one of them is thrown away, while the result of the other is returned by the IIf function. In Tcl if is not a keyword but a function (in Tcl known as command or proc). For example invokes a function named if passing 2 arguments: The first one being the condition and the second one being the true branch. Both arguments are passed as strings (in Tcl everything within curly brackets is a string). In the above example the condition is not evaluated before calling the function. Instead, the implementation of the if function receives the condition as a string value and is responsible to evaluate this string as an expression in the callers scope. Such a behavior is possible by using uplevel and expr commands. Uplevel makes it possible to implement new control constructs as Tcl procedures (for example, uplevel could be used to implement the while construct as a Tcl procedure). Because if is actually a function it also returns a value. The return value from the command is the result of the body script that was executed, or an empty string if none of the expressions was non-zero and there was no bodyN. In Rust, if is always an expression. It evaluates to the value of whichever branch is executed, or to the unit type () if no branch is executed. If a branch does not provide a return value, it evaluates to () by default. To ensure the if expression's type is known at compile time, each branch must evaluate to a value of the same type. For this reason, an else branch is effectively compulsory unless the other branches evaluate to (), because an if without an else can always evaluate to () by default. The following assigns r to 1 or 2 depending on the value of x. Values can be omitted when not needed. Pattern matching Pattern matching is an alternative to conditional statements (such as if–then–else and switch). It is available in many languages with functional programming features, such as Wolfram Language, ML and many others. Here is a simple example written in the OCaml language: The power of pattern matching is the ability to concisely match not only actions but also values to patterns of data. Here is an example written in Haskell which illustrates both of these features: This code defines a function map, which applies the first argument (a function) to each of the elements of the second argument (a list), and returns the resulting list. The two lines are the two definitions of the function for the two kinds of arguments possible in this case – one where the list is empty (just return an empty list) and the other case where the list is not empty. Pattern matching is not strictly speaking always a choice construct, because it is possible in Haskell to write only one alternative, which is guaranteed to always be matched – in this situation, it is not being used as a choice construct, but simply as a way to bind names to values. However, it is frequently used as a choice construct in the languages in which it is available. Hash-based conditionals In programming languages that have associative arrays or comparable data structures, such as Python, Perl, PHP or Objective-C, it is idiomatic to use them to implement conditional assignment. In languages that have anonymous functions or that allow a programmer to assign a named function to a variable reference, conditional flow can be implemented by using a hash as a dispatch table. Branch predication An alternative to conditional branch instructions is branch predication. Predication is an architectural feature that enables instructions to be conditionally executed instead of modifying the control flow. Choice system cross reference This table refers to the most recent language specification of each language. For languages that do not have a specification, the latest officially released implementation is referred to. See also References External links |
======================================== |
[SOURCE: https://en.wikipedia.org/wiki/GDP] | [TOKENS: 6743] |
Contents Gross domestic product Heterodox Gross domestic product (GDP) is a monetary measure of the total market value of all of the final goods and services which are produced and rendered during a specific period of time by a country or countries. GDP is often used to measure the economic activity of a country or region. The major components of GDP are consumption, government spending, net exports (exports minus imports), and investment. Changing any of these factors can increase the size of the economy. For example, population growth through mass immigration can raise consumption and demand for public services, thereby contributing to GDP growth. However, GDP is not a measure of overall standard of living or well-being, as it does not account for how income is distributed among the population. A country may rank high in GDP but still experience jobless growth depending on its planned economic structure and strategies. Dividing total GDP by the population gives an idealized rough measure of GDP per capita. Several national and international economic organizations, such as the OECD and the International Monetary Fund, maintain their own definitions of GDP. GDP is often used as a metric for international comparisons as well as a broad measure of economic progress. It serves as a statistical indicator of national development and progress. Total GDP can also be broken down into the contribution of each industry or sector of the economy. Nominal GDP is useful when comparing national economies on the international market using current exchange rate. To compare economies over time inflation can be adjusted by comparing real instead of nominal values. For cross-country comparisons, GDP figures are often adjusted for differences in the cost of living using purchasing power parity (PPP). GDP per capita at purchasing power parity can be useful for comparing living standards between nations. GDP has been criticized for leaving out key externalities, such as resource extraction,[clarification needed] environmental impact and unpaid domestic work. Alternative economic indicators such as doughnut economics use other measures, such as the Human Development Index or Better Life Index, as better approaches to measuring the effect of the economy on human development and well-being. History Sir William Petty came up with a concept of GDP, to calculate the tax burden, and argue landlords were unfairly taxed during warfare between the Dutch and the English between 1652 and 1674. Charles Davenant developed the method further in 1695. The modern concept of GDP was first developed by Simon Kuznets for a 1934 U.S. Congress report, where he warned against its use as a measure of welfare (see below under limitations and criticisms). After the Bretton Woods Conference in 1944, GDP became the main tool for measuring a country's economy. At that time gross national product (GNP) was the preferred estimate, which differed from GDP in that it measured production by a country's citizens at home and abroad rather than its "resident institutional units" (see OECD definition above). The switch from GNP to GDP in the United States occurred in 1991. The role that measurements of GDP played in World War II was crucial to the subsequent political acceptance of GDP values as indicators of national development and progress. A crucial role was played here by the U.S. Department of Commerce under Milton Gilbert where ideas from Kuznets were embedded into institutions. The history of the concept of GDP should be distinguished from the history of changes in many ways of estimating it. The value added by firms is relatively easy to calculate from their accounts, but the value added by the public sector, by financial industries, and by intangible asset creation is more complex. These activities are increasingly important in developed economies, and the international conventions governing their estimation and their inclusion or exclusion in GDP regularly change in an attempt to keep up with industrial advances. In the words of one academic economist, "The actual number for GDP is, therefore, the product of a vast patchwork of statistics and a complicated set of processes carried out on the raw data to fit them to the conceptual framework." China officially adopted GDP in 1993 as its indicator of economic performance. Previously, China had relied on a Marxist-inspired national accounting system. Determining gross domestic product (GDP) GDP can be determined in three ways, all of which should, theoretically, give the same result. They are the production (or output or value added) approach, the income approach, and the speculated expenditure approach. It is representative of the total output and income within an economy. The most direct of the three is the production approach, which sums the outputs of every class of enterprise to arrive at the total. The expenditure approach works on the principle that all of the products must be bought by somebody, therefore the value of the total product must be equal to people's total expenditures in buying things. The income approach works on the principle that the incomes of the productive factors ("producers", colloquially) must be equal to the value of their product, and determines GDP by finding the sum of all producers' incomes. Also known as the Value Added Approach, it calculates how much value is contributed at each stage of production. This approach mirrors the OECD (Organisation for Economic Co-operation and Development) definition given above. Gross value added = gross value of output – value of intermediate consumption. Value of output = value of the total sales of goods and services plus the value of changes in the inventory. The sum of the gross value added in the various economic activities is known as "GDP at factor cost". GDP at factor cost plus indirect taxes less subsidies on products = "GDP at producer price". For measuring the output of domestic product, economic activities (i.e. industries) are classified into various sectors. After classifying economic activities, the output of each sector is calculated by any of the following two methods: The value of output of all sectors is then added to get the gross value of output at factor cost. Subtracting each sector's intermediate consumption from gross output value gives the GVA (=GDP) at factor cost. Adding indirect tax minus subsidies to GVA (GDP) at factor cost gives the "GVA (GDP) at producer prices". The second way of estimating GDP is to use "the sum of primary incomes distributed by resident producer units". If GDP is calculated this way it is sometimes called gross domestic income (GDI), or GDP (I). GDI should provide the same amount as the expenditure method described later. By definition, GDI is equal to GDP. In practice, however, measurement errors will make the two figures slightly off when reported by national statistical agencies. This method measures GDP by adding incomes that firms pay households for factors of production they hire – wages for labour, interest for capital, rent for land and profits for entrepreneurship. The US "National Income and Product Accounts" divide incomes into five categories: These five income components sum to net domestic income at factor cost. Two adjustments must be made to get GDP: Total income can be subdivided according to various schemes, leading to various formulae for GDP measured by the income approach. A common one is:[citation needed] The sum of COE, GOS and GMI is called total factor income; it is the income of all of the factors of production in society. It measures the value of GDP at factor (basic) prices. The difference between basic prices and final prices (those used in the expenditure calculation) is the total taxes and subsidies that the government has levied or paid on that production. So adding taxes less subsidies on production and imports converts GDP(I) at factor cost to GDP(I) at final prices. Total factor income is also sometimes expressed as: The third way to estimate GDP is to calculate the sum of the final uses of goods and services (all uses except intermediate consumption) measured in purchasers' prices. Market goods that are produced are purchased by someone. In the case where a good is produced and unsold, the standard accounting convention is that the producer has bought the good from themselves. Therefore, measuring the total expenditure used to buy things is a way of measuring production. This is known as the expenditure method of calculating GDP. GDP (Y) is the sum of consumption (C), investment (I), government expenditures (G) and net exports (X − M). Here is a description of each GDP component: C, I, and G are expenditures on final goods and services; expenditures on intermediate goods and services do not count. (Intermediate goods and services are those used by businesses to produce other goods and services within the accounting year.) So for example if a car manufacturer buys auto parts, assembles the car and sells it, only the final car sold is counted towards the GDP. Meanwhile, if a person buys replacement auto parts to install them on their car, those are counted towards the GDP. According to the U.S. Bureau of Economic Analysis, which is responsible for calculating the national accounts in the United States, "In general, the source data for the expenditures components are considered more reliable than those for the income components [see income method, above]." Encyclopedia Britannica records an alternate way of measuring exports minus imports: notating it as the single variable NX. The raw GDP figure given by the equations above is called the nominal, historical, or current GDP. When comparing GDP figures from one year to another, compensating for changes in the value of money—for the effects of inflation or deflation—is desirable. To make it more meaningful for year-to-year comparisons, a nominal GDP may be multiplied by the ratio between the value of money in the year the GDP was measured and the value of money in a base year. For example, suppose a country's GDP in 1990 was $100 million and its GDP in 2000 was $300 million. Suppose also that inflation had halved the value of its currency over that period. To meaningfully compare its GDP in 2000 to its GDP in 1990, we could multiply the GDP in 2000 by one-half, to make it relative to 1990 as a base year. The result would be that the GDP in 2000 equals $300 million × 1⁄2 = $150 million, in 1990 monetary terms. We would see that the country's GDP had realistically increased 50 percent over that period, not 200 percent, as it might appear from the raw GDP data. The GDP adjusted for changes in money value in this way is called the real GDP. The factor used to convert GDP from current to constant values in this way is called the GDP deflator. Unlike consumer price index, which measures inflation or deflation in the price of household consumer goods, the GDP deflator measures changes in the prices of all domestically produced goods and services in an economy including investment goods and government services, as well as household consumption goods. Within each country GDP is normally measured by a national government statistical agency, as private sector organizations normally do not have access to the information required (especially information on expenditure and production by governments). The international standard for measuring GDP is contained in the book System of National Accounts (2008), which was prepared by representatives of the International Monetary Fund, European Union, Organisation for Economic Co-operation and Development, United Nations and World Bank. The publication is normally referred to as SNA2008 to distinguish it from the previous edition published in 1993 (SNA93) or 1968 (called SNA68) SNA2008 provides a set of rules and procedures for the measurement of national accounts. The standards are designed to be flexible, to allow for differences in local statistical needs and conditions. A peer-reviewed study published in the Journal of Political Economy in October 2022 found signs of manipulation of economic growth statistics in the majority of countries. This manipulation was greatest in countries that were semi-authoritarian/authoritarian, or did not have effective separation of powers. The study took the annual growth in the brightness of lights at night, as measured by satellites, and compared it to officially reported economic growth. Authoritarian states consistently reported higher growth in GDP than the growth in night lights would suggest, an effect that could not be explained by economic structures, sector composition or other factors. Corporate havens can also have a distorted GDP. Lists of countries by GDP Economic growth Real GDP can be used to calculate the GDP growth rate, which indicates how much a country's production has increased (or decreased, if the growth rate is negative) compared to the previous year, typically expressed as percentage change. The economic growth can be expressed as: Relation to gross national income GDP can be contrasted with the gross national income (GNI) also known as gross national product (GNP). The difference is that GDP defines its scope according to location, while GNI defines its scope according to ownership. In a global context, world GDP and world GNI are, therefore, equivalent terms. GDP is a product produced within a country's borders; GNI is product produced by enterprises owned by a country's citizens. The two would be the same if all of the productive enterprises in a country were owned by its own citizens and those citizens did not own productive enterprises in any other countries. In practice, however, foreign ownership makes GDP and GNI non-identical. Production within a country's borders, but by an enterprise owned by somebody outside the country, counts as part of its GDP but not its GNI; on the other hand, production by an enterprise located outside the country, but owned by one of its citizens, counts as part of its GNI but not its GDP. For example, the GNI of the US is the value of output produced by American-owned firms, regardless of where the firms are located. Gross national income (GNI) equals GDP plus income receipts from the rest of the world minus income payments to the rest of the world. In 1991, the United States switched from using GNP to using GDP as its primary measure of production. The relationship between United States GDP and GNP is shown in table 1.7.5 of the National Income and Product Accounts. Other examples that amplify differences between GDP and GNI can be found by comparing indicators of developed and developing countries. The GDP of Japan for 2020 was 5.05559 trillion. Predictably, as a developed country, Japan has a higher GNI of 5.16915 trillion for the same year, an increase of 113.560 million. This is indicative of the production level in the country being higher than that of national production. On the other hand, the case with Armenia is the opposite with its GNI in 2023 being lower than its GDP by 3.85 billion. This shows countries that receive investments and foreign aid from abroad. Limitations and criticisms Simon Kuznets, the economist who developed the first comprehensive set of measures of national income, stated in his second report to the U.S. Congress in 1937, in a section titled "Uses and Abuses of National Income Measurements": The valuable capacity of the human mind to simplify a complex situation in a compact characterization becomes dangerous when not controlled in terms of definitely stated criteria. With quantitative measurements especially, the definiteness of the result suggests, often misleadingly, a precision and simplicity in the outlines of the object measured. Measurements of national income are subject to this type of illusion and resulting abuse, especially since they deal with matters that are the center of conflict of opposing social groups where the effectiveness of an argument is often contingent upon oversimplification. [...] All these qualifications upon estimates of national income as an index of productivity are just as important when income measurements are interpreted from the point of view of economic welfare. But in the latter case additional difficulties will be suggested to anyone who wants to penetrate below the surface of total figures and market values. Economic welfare cannot be adequately measured unless the personal distribution of income is known. And no income measurement undertakes to estimate the reverse side of income, that is, the intensity and unpleasantness of effort going into the earning of income. The welfare of a nation can, therefore, scarcely be inferred from a measurement of national income as defined above. In 1962, Kuznets stated: Distinctions must be kept in mind between quantity and quality of growth, between costs and returns, and between the short and long run. Goals for more growth should specify more growth of what and for what. GDP as initially defined includes spending on goods and services that would shrink if underlying problems were solved or reduced - for example, medical care, crime-fighting, and the military. During World War II, Kuznets came to argue that military spending should be excluded during peacetime. This idea did not become popular; these activities are tracked because they fit into macroeconomic models (e.g. military spending uses up capital and labor). Ever since the development of GDP, multiple observers have pointed out limitations of using GDP as the overarching measure of economic and social progress. Furthermore, the GDP does not consider human health nor the educational aspect of a population. Instances of GDP measures have been considered numbers that are artificial constructs. American politician Robert F. Kennedy criticized GDP (or GNI), listing many examples of bad things it does count and good things it does not count: Gross National Product counts air pollution and cigarette advertising, and ambulances to clear our highways of carnage. It counts special locks for our doors and the jails for the people who break them. It counts the destruction of the redwood and the loss of our natural wonder in chaotic sprawl. It counts napalm and counts nuclear warheads and armored cars for the police to fight the riots in our cities. It counts Whitman's rifle and Speck's knife, and the television programs which glorify violence in order to sell toys to our children. Yet the gross national product does not allow for the health of our children, the quality of their education or the joy of their play. It does not include the beauty of our poetry or the strength of our marriages, the intelligence of our public debate or the integrity of our public officials. It measures neither our wit nor our courage, neither our wisdom nor our learning, neither our compassion nor our devotion to our country, it measures everything in short, except that which makes life worthwhile. And it can tell us everything about America except why we are proud that we are Americans. Deficit spending increases GDP in case of positive fiscal multipliers. GDP as a metric can incentivize politicians to overspend. Counterfactual GDP under a balanced budget scenario can be estimated. If a country becomes increasingly in debt, and spends large amounts of income on debt interest expense, will be reflected in a decreased GNI[citation needed] but not a decreased GDP. Similarly, if a country sells off its resources to entities outside their country this will also be reflected over time in decreased GNI, but not decreased GDP. This would make the use of GDP more attractive for politicians in countries with increasing national debt and decreasing assets. GDP excludes the value of household and other unpaid work. Some, including Martha Nussbaum, argue that this value should be included in measuring GDP, as household labor is largely a substitute for goods and services that would otherwise be purchased with money. Even under conservative estimates, the value of unpaid labor in Australia has been calculated to be over 50% of the country's GDP. A later study analyzed this value in other countries, with results ranging from a low of about 15% in Canada (using conservative estimates) to high of nearly 70% in the United Kingdom (using more liberal estimates). For the United States, the value was estimated to be between about 20% on the low end to nearly 50% on the high end, depending on the methodology being used. Because many public policies are shaped by GDP calculations and by the related field of national accounts, public policy might differ if unpaid work were included in total GDP. Some economists have advocated for changes in the way public policies are formed and implemented. Some have pointed out that GDP did not adapt to sociotechnical changes to give a more accurate picture of the modern economy and does not encapsulate the value of new activities such as delivering price-free information and entertainment on social media. In 2017 Diane Coyle explained that GDP excludes much unpaid work, writing that "many people contribute free digital work such as writing open-source software that can substitute for marketed equivalents, and it clearly has great economic value despite a price of zero", which constitutes a common criticism "of the reliance on GDP as the measure of economic success" especially after the emergence of the digital economy. A 2025 study in the American Economic Journal devised a new GDP measurement (GDP-B) that accounts for the welfare value of new goods and free goods. In 2019, Erik Brynjolfsson and Avinash Collis argued that GDP does not reflect the growing value of many digital goods because they have zero price. Along with several coauthors, they proposed an alternative approach, GDP-B, which is based on measuring the benefits of goods and services, rather than their price or cost. In 2013 scientists reported that large improvements in health only lead to modest long-term increases in GDP per capita. After developing an abstract metric similar to GDP, the Center for Partnership Studies highlighted that GDP "and other metrics that reflect and perpetuate them" may not be useful for facilitating the production of products and provision of services that are useful – or comparatively more useful – to society, and instead may "actually encourage, rather than discourage, destructive activities". The number of obese adults was approximately 600 million (12%) in 2015.[non sequitur] Many environmentalists argue that GDP is a poor measure of social progress because it does not take into account harm to the environment. In the language of economics, everything comes down to its monetary value. In essence, GDP rewards behaviors that are detrimental to the environment. GDP also does not capture certain phenomena impacting citizens' well-being. For example, traffic jams could cause GDP to increase as there is a higher consumption of gasoline, however, GDP fails to consider citizens' well-being in terms of the quality of air due to air pollution from the traffic jams. Various alternatives have been developed(see below). A 2020 study found that "poor regions' GDP grows faster by attracting more polluting production after connection to China's expressway system. GDP may not be a tool capable of recognizing how much natural capital agents of the economy are building or protecting. In 2020 scientists, as part of a World Scientists' Warning to Humanity-associated series, warned that worldwide growth in affluence in terms of GDP-metrics has increased resource use and pollutant emissions with affluent citizens of the world – in terms of e.g. resource-intensive consumption – being responsible for most negative environmental impacts and central to a transition to safer, sustainable conditions. They summarised evidence, presented solution approaches and stated that far-reaching lifestyle changes need to complement technological advancements and that existing societies, economies and cultures incite consumption expansion and that the structural imperative for growth in competitive market economies inhibits societal change. Sarah Arnold, Senior Economist at the New Economics Foundation (NEF) stated that "GDP includes activities that are detrimental to our economy and society in the long term, such as deforestation, strip mining, overfishing and so on". The number of trees that are net lost annually is estimated to be approximately 10 billion. The global average annual deforested land in the 2015–2020 demi-decade was 10 million hectares and the average annual net forest area loss in the 2000–2010 decade 4.7 million hectares, according to the Global Forest Resources Assessment 2020. According to one study, depending on the level of wealth inequality, higher GDP-growth can be associated with more deforestation. In 2019 "agriculture and agribusiness" accounted for 24% of the GDP of Brazil, where a large share of annual net tropical forest loss occurred and is associated with sizable portions of this economic activity domain. Steve Cohen of the Earth Institute elucidated that while GDP does not distinguish between different activities (or lifestyles), "all consumption behaviors are not created equal and do not have the same impact on environmental sustainability". Johan Rockström, director of the Potsdam Institute for Climate Impact Research, noted that "it's difficult to see if the current G.D.P.-based model of economic growth can go hand-in-hand with rapid cutting of emissions", which nations have agreed to attempt under the Paris Agreement in order to mitigate real-world impacts of climate change. In 1989, John B. Cobb and Herman Daly introduced Index of Sustainable Economic Welfare (ISEW) by taking into account other factors such as consumption of nonrenewable resources and degradation of the environment. ISEW is roughly defined as: personal consumption + public non-defensive expenditures − private defensive expenditures + capital formation + services from domestic labour − costs of environmental degradation − depreciation of natural capital. In 2005, Med Jones, an American Economist, at the International Institute of Management, introduced the first secular Gross National Happiness Index a.k.a. Gross National Well-being framework and Index to complement GDP economics with additional seven dimensions, including environment, education, and government, work, social and health (mental and physical) indicators. The proposal was inspired by the King of Bhutan's GNH philosophy. In 2019, Serge Pierre Besanger published a "GDP 3.0" proposal which combines an expanded GNI formula which he calls GNIX, with a Palma ratio and a set of environmental metrics based on the Daly Rule. The UK's Natural Capital Committee highlighted the shortcomings of GDP in its advice to the UK Government in 2013, pointing out that GDP "focuses on flows, not stocks. As a result, an economy can run down its assets yet, at the same time, record high levels of GDP growth, until a point is reached where the depleted assets act as a check on future growth". They then went on to say that "it is apparent that the recorded GDP growth rate overstates the sustainable growth rate. Broader measures of wellbeing and wealth are needed for this and there is a danger that short-term decisions based solely on what is currently measured by national accounts may prove to be costly in the long-term".[citation needed] China launched the Gross Ecosystem Product (GEP) in 2020. It measures the contribution of ecosystems to the economy, including by regulating climate. It spread widely across the country. The first province to issue local rules about GEP was Zhejiang , and a year later it has already decided the fate of a project in the Deqing region. For example, the GEP of Chengtian Radon Spring Nature Reserve has been calculated as US$43 million. GDP changes with population change. GDP adjusted for population is called Per-capita GDP or GDP per person. This measures the average production of a person in the country. The major advantage of GDP per capita as an indicator of the standard of living is that it is measured frequently, widely, and consistently. It is measured frequently in that most countries provide information on GDP every quarter, allowing trends to be seen quickly. It is measured widely in that some measure of GDP is available for almost every country in the world, allowing inter-country comparisons. It is measured consistently in that the technical definition of GDP is relatively consistent among countries. It can be argued that GDP per capita is an indicator of standard of living. As a result, GDP per capita as a standard of living is a continued usage because most people have a fairly accurate idea of what it is and know it is tough to come up with quantitative measures for such constructs as happiness, quality of life, and well-being. From the perspective of environmental, social and governance (ESG) measures, GDP per capita trends can be influenced by factors such as gender parity and elements of regulatory quality. The change in number of MSMEs (Micro, Small, and Medium Enterprises) in the Philippines from 2008 through 2021 would be an example of elements such as the per capita gross domestic product and unemployment rate having significant effect on a developing country with mixed economy. Although a high or rising level of GDP per capita is often associated with increased economic and social progress, the opposite sometimes occurs. For example, Jean Drèze and Amartya Sen have pointed out that an increase in GDP or in GDP growth does not necessarily lead to a higher standard of living, particularly in areas such as healthcare and education. Another important area that does not necessarily improve along with GDP is political liberty, which is most notable in China, where GDP growth is strong yet political liberties are heavily restricted. GDP does not account for the distribution of income among the residents of a country, because GDP is merely an aggregate measure. An economy may be highly developed or growing rapidly, but also contain a wide gap between the rich and the poor in a society. These inequalities often occur on the lines of race, ethnicity, gender, religion, or other minority status within countries. This can lead to misleading characterizations of economic well-being if the income distribution is heavily skewed toward the high end, as the poorer residents will not directly benefit from the overall level of wealth and income generated in their country (their purchasing power can decline, even as the mean GDP per capita rises). GDP per capita measures (like aggregate GDP measures) do not account for income distribution (and tend to overstate the average income per capita). For example, South Africa during apartheid ranked high in terms of GDP per capita, but the benefits of this immense wealth and income were not shared equally among its citizens. The United Nations has aimed in its Sustainable Development Goals, amongst other global initiatives, to address wealth inequality. GDP does not include several factors that influence the standard of living. In particular, it fails to account for: In the 1980s, Amartya Sen and Martha Nussbaum developed the capability approach, which focuses on the functional capabilities enjoyed by people within a country, rather than the aggregate GDP within a country. These capabilities consist of the functions that a person is able to achieve. In 1990, Mahbub ul Haq, a Pakistani economist at the United Nations, introduced the Human Development Index (HDI). The HDI is a composite index of life expectancy at birth, adult literacy rate and standard of living measured as a logarithmic function of GDP, adjusted to purchasing power parity. In 2009 Professors Joseph Stiglitz, Amartya Sen, and Jean-Paul Fitoussi at the Commission on the Measurement of Economic Performance and Social Progress (CMEPSP), formed by French President, Nicolas Sarkozy published a proposal to overcome the limitation of GDP economics to expand the focus to well-being economics with a well-being framework consisting of health, environment, work, physical safety, economic safety, and political freedom. This has been adopted in a number of countries as a wellbeing economy policy. In 2008, the Centre for Bhutan Studies began publishing the Bhutan Gross National Happiness (GNH) Index, whose contributors to happiness include physical, mental, and spiritual health; time balance; social and community vitality; cultural vitality; education; living standards; good governance; and ecological vitality. In 2013, the OECD Better Life Index was published by the OECD. The dimensions of the index included health, economic, workplace, income, jobs, housing, civic engagement, and life satisfaction. Since 2012, John Helliwell, Richard Layard and Jeffrey Sachs have edited an annual World Happiness Report which reports a national measure of subjective well-being, derived from a single survey question on satisfaction with life. GDP explains some of the cross-national variation in life satisfaction, but more of it is explained by other, social variables. In the second decade of the 21st century the United Nations gave the task of creating a measurment of progress that takes “human wellbeing, sustainability and equity” into account to a group co-chaired by Kaushik Basu and Nora Lustig. In 2025 it published a report which argued that recent global shocks make the need for transformation more urgent. In January 2026, the UN held a conference named "Beyond GDP" attended by major economists. In February 2026, Antonio Guterres said: "We must place true value on the environment and go beyond gross domestic product as a measure of human progress and wellbeing." The initiative followed a debate about sustainable economic structures between green growth advocates, green Keynesians and advocates of doughnut economy, Wellbeing economy, steady-state economics, degrowth, and according to a recent survey, 73% of around 800 climate policy researchers support post-growth. According to Jason Hickel this initiative will not be enough, because: “A deeper system change is required. Specifically, we need to democratise control over production, which can enable us to change what we produce and for whom. “The dominance of GDP is not an accident, it occurs because GDP measures what is valuable to capital. It is the structure of capitalism that ultimately must be overcome.” See also References Further reading External links Global Data Articles and books *Top country subdivisions by GDP *Top country subdivisions by GDP per capita *Top country metropolitan by GDP |
======================================== |
[SOURCE: https://en.wikipedia.org/wiki/Class_(computer_science)] | [TOKENS: 5656] |
Contents Class (programming) In programming, a class is a syntactic entity structure used to create objects.: 1.3.3 The capabilities of a class differ between programming languages, but generally the shared aspects consist of state (variables) and behavior (methods) that are each either associated with a particular object or with all objects of that class. Object state can differ between each instance of the class whereas the class state is shared by all of them. The object methods include access to the object state (via an implicit or explicit parameter that references the object) whereas class methods do not. If the language supports inheritance, a class can be defined based on another class with all of its state and behavior plus additional state and behavior that further specializes the class. The specialized class is a sub-class, and the class it is based on is its superclass. In purely object-oriented programming languages, such as Java and C#, all classes might be part of an inheritance tree such that the root class is Object, meaning all objects instances are of Object or implicitly extend Object, which is called a top type. History The concept was primarily introduced in the OOP by the Simula language in 1960's and continuously being used by a large of object-oriented programming languages.: 1.3.3 Its creation was based in similar concept as block used in previous-based ALGOL programming language.: 1.3.2 Attributes As an instance of a class, an object is constructed from a class via instantiation. Memory is allocated and initialized for the object state and a reference to the object is provided to consuming code. The object is usable until it is destroyed – its state memory is de-allocated. Most languages allow for custom logic at lifecycle events via a constructor and a destructor. An object expresses data type as an interface – the type of each member variable and the signature of each member function (method). A class defines an implementation of an interface, and instantiating the class results in an object that exposes the implementation via the interface. In the terms of type theory, a class is an implementation—a concrete data structure and collection of subroutines—while a type is an interface. Different (concrete) classes can produce objects of the same (abstract) type (depending on type system). For example, the type (interface) Stack might be implemented by SmallStack that is fast for small stacks but scales poorly and ScalableStack that scales well but has high overhead for small stacks. A class contains data field syntactically described (or properties, fields, data members, or attributes). These are usually field types and names that will be associated with state variables at program run time; these state variables either belong to the class or specific instances of the class. In most languages, the structure defined by the class determines the layout of the memory used by its instances. Other implementations are possible: for example, objects in Python use associative key-value containers. Some programming languages such as Eiffel support specification of invariants as part of the definition of the class, and enforce them through the type system. Encapsulation of state is necessary for being able to enforce the invariants of the class. The behavior (or action) of a class or its instances is defined using methods. Methods are subroutines with the ability to operate on objects or classes. These operations may alter the state of an object or simply provide ways of accessing it. Many kinds of methods exist, but support for them varies across languages. Some types of methods are created and called by programmer code, while other special methods—such as constructors, destructors, and conversion operators—are created and called by compiler-generated code. A language may also allow the programmer to define and call these special methods. Every class implements (or realizes) an interface by providing structure and behavior. Structure consists of data and state, and behavior consists of code that specifies how methods are implemented. There is a distinction between the definition of an interface and the implementation of that interface; however, this line is blurred in many programming languages because class declarations both define and implement an interface. Some languages, however, provide features that separate interface and implementation. For example, an abstract class can define an interface without providing an implementation. Languages that support class inheritance also allow classes to inherit interfaces from the classes that they are derived from. For example, if "class Z" inherits from "class Y" and if "class Y" implements the interface "interface X" then "class Z" also implements the functionality(constants and methods declaration) provided by "interface X". In languages that support access specifiers, the interface of a class is considered to be the set of public members of the class, including both methods and attributes (via implicit getter and setter methods); any private members or internal data structures are not intended to be depended on by external code and thus are not part of the interface. Object-oriented programming methodology dictates that the operations of any interface of a class are to be independent of each other. It results in a layered design where clients of an interface use the methods declared in the interface. An interface places no requirements for clients to invoke the operations of one interface in any particular order. This approach has the benefit that client code can assume that the operations of an interface are available for use whenever the client has access to the object. The buttons on the front of your television set are the interface between you and the electrical wiring on the other side of its plastic casing. You press the "power" button to toggle the television on and off. In this example, your particular television is the instance, each method is represented by a button, and all the buttons together compose the interface (other television sets that are the same model as yours would have the same interface). In its most common form, an interface is a specification of a group of related methods without any associated implementation of the methods. A television set also has a myriad of attributes, such as size and whether it supports color, which together comprise its structure. A class represents the full description of a television, including its attributes (structure) and buttons (interface). Getting the total number of televisions manufactured could be a static method of the television class. This method is associated with the class, yet is outside the domain of each instance of the class. A static method that finds a particular instance out of the set of all television objects is another example. The following is a common set of access specifiers: Although many object-oriented languages support the above access specifiers, their semantics may differ. Object-oriented design uses the access specifiers in conjunction with careful design of public method implementations to enforce class invariants—constraints on the state of the objects. A common usage of access specifiers is to separate the internal data of a class from its interface: the internal structure is made private, while public accessor methods can be used to inspect or alter such private data. Access specifiers do not necessarily control visibility, in that even private members may be visible to client external code. In some languages, an inaccessible but visible member may be referred to at runtime (for example, by a pointer returned from a member function), but an attempt to use it by referring to the name of the member from the client code will be prevented by the type checker. The various object-oriented programming languages enforce member accessibility and visibility to various degrees, and depending on the language's type system and compilation policies, enforced at either compile time or runtime. For example, the Java language does not allow client code that accesses the private data of a class to compile. In the C++ language, private methods are visible, but not accessible in the interface; however, they may be made invisible by explicitly declaring fully abstract classes that represent the interfaces of the class. Some languages feature other accessibility schemes: Conceptually, a superclass is a superset of its subclasses. For example, GraphicObject could be a superclass of Rectangle and Ellipse, while Square would be a subclass of Rectangle. These are all subset relations in set theory as well, i.e., all squares are rectangles but not all rectangles are squares. A common conceptual error is to mistake a part of relation with a subclass. For example, a car and truck are both kinds of vehicles and it would be appropriate to model them as subclasses of a vehicle class. However, it would be an error to model the parts of the car as subclass relations. For example, a car is composed of an engine and body, but it would not be appropriate to model an engine or body as a subclass of a car. In object-oriented modeling these kinds of relations are typically modeled as object properties. In this example, the Car class would have a property called parts. parts would be typed to hold a collection of objects, such as instances of Body, Engine, Tires, etc. Object modeling languages such as UML include capabilities to model various aspects of "part of" and other kinds of relations – data such as the cardinality of the objects, constraints on input and output values, etc. This information can be utilized by developer tools to generate additional code besides the basic data definitions for the objects, such as error checking on get and set methods. One important question when modeling and implementing a system of object classes is whether a class can have one or more superclasses. In the real world with actual sets, it would be rare to find sets that did not intersect with more than one other set. However, while some systems such as Flavors and CLOS provide a capability for more than one parent to do so at run time introduces complexity that many in the object-oriented community consider antithetical to the goals of using object classes in the first place. Understanding which class will be responsible for handling a message can get complex when dealing with more than one superclass. If used carelessly this feature can introduce some of the same system complexity and ambiguity classes were designed to avoid. Most modern object-oriented languages such as Smalltalk and Java require single inheritance at run time. For these languages, multiple inheritance may be useful for modeling but not for an implementation. However, semantic web application objects do have multiple superclasses. The volatility of the Internet requires this level of flexibility and the technology standards such as the Web Ontology Language (OWL) are designed to support it. A similar issue is whether or not the class hierarchy can be modified at run time. Languages such as Flavors, CLOS, and Smalltalk all support this feature as part of their meta-object protocols. Since classes are themselves first-class objects, it is possible to have them dynamically alter their structure by sending them the appropriate messages. Other languages that focus more on strong typing such as Java and C++ do not allow the class hierarchy to be modified at run time. Semantic web objects have the capability for run time changes to classes. The rationale is similar to the justification for allowing multiple superclasses, that the Internet is so dynamic and flexible that dynamic changes to the hierarchy are required to manage this volatility. Although many class-based languages support inheritance, inheritance is not an intrinsic aspect of classes.[dubious – discuss][non sequitur] An object-based language (i.e. Classic Visual Basic) supports classes yet does not support inheritance. Inter-class relationships A programming language may support various class relationship features. Classes can be composed of other classes, thereby establishing a compositional relationship between the enclosing class and its embedded classes. Compositional relationship between classes is also commonly known as a has-a relationship. For example, a class Car could be composed of and contain a class Engine. Therefore, a Car has an Engine. One aspect of composition is containment, which is the enclosure of component instances by the instance that has them. If an enclosing object contains component instances by value, the components and their enclosing object have a similar lifetime. If the components are contained by reference, they may not have a similar lifetime. For example, in Objective-C 2.0: This Car class has an instance of NSString (a string object), Engine, and NSArray (an array object). Classes can be derived from one or more existing classes, thereby establishing a hierarchical relationship between the derived-from classes (base classes, parent classes or superclasses) and the derived class (child class or subclass) . The relationship of the derived class to the derived-from classes is commonly known as an is-a relationship. For example, a class 'Button' could be derived from a class 'Control'. Therefore, a Button is a Control. Structural and behavioral members of the parent classes are inherited by the child class. Derived classes can define additional structural members (data fields) and behavioral members (methods) in addition to those that they inherit and are therefore specializations of their superclasses. Also, derived classes can override inherited methods if the language allows. Not all languages support multiple inheritance. For example, Java allows a class to implement multiple interfaces, but only inherit from one class. If multiple inheritance is allowed, the hierarchy is a directed acyclic graph (or DAG for short), otherwise it is a tree. The hierarchy has classes as nodes and inheritance relationships as links. Classes in the same level are more likely to be associated than classes in different levels. The levels of this hierarchy are called layers or levels of abstraction. Example (Simplified Objective-C 2.0 code, from iPhone SDK): In this example, a UITableView is a UIScrollView is a UIView is a UIResponder is an NSObject. In object-oriented analysis and in Unified Modelling Language (UML), an association between two classes represents a collaboration between the classes or their corresponding instances. Associations have direction; for example, a bi-directional association between two classes indicates that both of the classes are aware of their relationship. Associations may be labeled according to their name or purpose. An association role is given end of an association and describes the role of the corresponding class. For example, a "subscriber" role describes the way instances of the class "Person" participate in a "subscribes-to" association with the class "Magazine". Also, a "Magazine" has the "subscribed magazine" role in the same association. Association role multiplicity describes how many instances correspond to each instance of the other class of the association. Common multiplicities are "0..1", "1..1", "1..*" and "0..*", where the "*" specifies any number of instances. Taxonomy There are many categories of classes, some of which overlap. In a language that supports inheritance, an abstract class, or abstract base class (ABC), is a class that cannot be directly instantiated. By contrast, a concrete class is a class that can be directly instantiated. Instantiation of an abstract class can occur only indirectly, via a concrete subclass. An abstract class is either labeled as such explicitly or it may simply specify abstract methods (or virtual methods). An abstract class may provide implementations of some methods, and may also specify virtual methods via signatures that are to be implemented by direct or indirect descendants of the abstract class. Before a class derived from an abstract class can be instantiated, all abstract methods of its parent classes must be implemented by some class in the derivation chain. Most object-oriented programming languages allow the programmer to specify which classes are considered abstract and will not allow these to be instantiated. For example, in Java, C# and PHP, the keyword abstract is used. In C++, an abstract class is a class having at least one abstract method given by the appropriate syntax in that language (a pure virtual function in C++ parlance). A class consisting of only pure virtual methods is called a pure abstract base class (or pure ABC) in C++ and is also known as an interface by users of the language. Other languages, notably Java and C#, support a variant of abstract classes called an interface via a keyword in the language. In these languages, multiple inheritance is not allowed, but a class can implement multiple interfaces. Such a class can only contain abstract publicly accessible methods. In some languages, classes can be declared in scopes other than the global scope. There are various types of such classes. An inner class is a class defined within another class. The relationship between an inner class and its containing class can also be treated as another type of class association. An inner class is typically neither associated with instances of the enclosing class nor instantiated along with its enclosing class. Depending on the language, it may or may not be possible to refer to the class from outside the enclosing class. A related concept is inner types, also known as inner data type or nested type, which is a generalization of the concept of inner classes. C++ is an example of a language that supports both inner classes and inner types (via typedef declarations). A local class is a class defined within a procedure or function. Such structure limits references to the class name to within the scope where the class is declared. Depending on the semantic rules of the language, there may be additional restrictions on local classes compared to non-local ones. One common restriction is to disallow local class methods to access local variables of the enclosing function. For example, in C++, a local class may refer to static variables declared within its enclosing function, but may not access the function's automatic variables. A metaclass is a class where instances are classes. A metaclass describes a common structure of a collection of classes and can implement a design pattern or describe particular kinds of classes. Metaclasses are often used to describe frameworks. In some languages, such as Python, Ruby or Smalltalk, a class is also an object; thus each class is an instance of a unique metaclass that is built into the language. The Common Lisp Object System (CLOS) provides metaobject protocols (MOPs) to implement those classes and metaclasses. A final class cannot be subclassed. It is basically the opposite of an abstract class, which must be subclassed to be used and cannot be instantiated directly. A final class is implicitly a concrete class, can be instantiated directly. A class is declared as final via the keyword final in Java, C++ or PHP, or sealed in C#. However, this concept should not be confused with classes in Java qualified with the keyword sealed, that only allow inheritance from selected subclasses. For example, Java's String class is marked as final. Final classes may allow a compiler to perform optimizations that are not available for classes that can be subclassed. A "sealed class" is a class that restricts inheritance to a selected list of classes. It should not be confused with the sealed keyword in C#, which denotes a final class. The list of permitted classes the sealed class may inherit is specified using a "permits" clause. non-sealed is another keyword used to declare that a class or interface which extends a sealed class can be extended by unknown classes. An open class can be changed. Typically, an executable program cannot be changed by customers. Developers can often change some classes, but typically cannot change standard or built-in ones. In Ruby, all classes are open. In Python, classes can be created at runtime, and all can be modified afterward. Objective-C categories permit the programmer to add methods to an existing class without the need to recompile that class or even have access to its source code. Some languages have special support for mixins, though, in any language with multiple inheritance, a mixin is simply a class that does not represent an is-a-type-of relationship. Mixins are typically used to add the same methods to multiple classes; for example, a class UnicodeConversionMixin might provide a method called unicode_to_ascii when included in classes FileReader and WebPageScraper that do not share a common parent. In languages supporting the feature, a partial class is a class whose definition may be split into multiple pieces, within a single source-code file or across multiple files. The pieces are merged at compile time, making compiler output the same as for a non-partial class. The primary motivation for the introduction of partial classes is to facilitate the implementation of code generators, such as visual designers. It is otherwise a challenge or compromise to develop code generators that can manage the generated code when it is interleaved within developer-written code. Using partial classes, a code generator can process a separate file or coarse-grained partial class within a file, and is thus alleviated from intricately interjecting generated code via extensive parsing, increasing compiler efficiency and eliminating the potential risk of corrupting developer code. In a simple implementation of partial classes, the compiler can perform a phase of precompilation where it "unifies" all the parts of a partial class. Then, compilation can proceed as usual. Other benefits and effects of the partial class feature include: Partial classes have existed in Smalltalk under the name of Class Extensions for considerable time. With the arrival of the .NET framework 2, Microsoft introduced partial classes, supported in both C# 2.0 and Visual Basic 2005. WinRT also supports partial classes. Uninstantiable classes allow programmers to group together per-class fields and methods that are accessible at runtime without an instance of the class. Indeed, instantiation is prohibited for this kind of class. For example, in C#, a class marked "static" can not be instantiated, can only have static members (fields, methods, other), may not have instance constructors, and is sealed. An unnamed class or anonymous class is not bound to a name or identifier upon definition. This is analogous to named versus unnamed functions. Benefits The benefits of organizing software into object classes fall into three categories: Object classes facilitate rapid development because they lessen the semantic gap between the code and the users. System analysts can talk to both developers and users using essentially the same vocabulary, talking about accounts, customers, bills, etc. Object classes often facilitate rapid development because most object-oriented environments come with powerful debugging and testing tools. Instances of classes can be inspected at run time to verify that the system is performing as expected. Also, rather than get dumps of core memory, most object-oriented environments have interpreted debugging capabilities so that the developer can analyze exactly where in the program the error occurred and can see which methods were called to which arguments and with what arguments. Object classes facilitate ease of maintenance via encapsulation. When developers need to change the behavior of an object they can localize the change to just that object and its component parts. This reduces the potential for unwanted side effects from maintenance enhancements. Software reuse is also a major benefit of using Object classes. Classes facilitate re-use via inheritance and interfaces. When a new behavior is required it can often be achieved by creating a new class and having that class inherit the default behaviors and data of its superclass and then tailoring some aspect of the behavior or data accordingly. Re-use via interfaces (also known as methods) occurs when another object wants to invoke (rather than create a new kind of) some object class. This method for re-use removes many of the common errors that can make their way into software when one program re-uses code from another. Runtime representation As a data type, a class is usually considered as a compile time construct. A language or library may also support prototype or factory metaobjects that represent runtime information about classes, or even represent metadata that provides access to reflective programming (reflection) facilities and ability to manipulate data structure formats at runtime. Many languages distinguish this kind of run-time type information about classes from a class on the basis that the information is not needed at runtime. Some dynamic languages do not make strict distinctions between runtime and compile time constructs, and therefore may not distinguish between metaobjects and classes. For example, if Human is a metaobject representing the class Person, then instances of class Person can be created by using the facilities of the Human metaobject. Class-based programming Class-based programming, or more commonly class-orientated, is a style of object-oriented programming which all objects are created by a class and without inheritance between them. The most popular and developed model of OOP is a class-based model, instead of an object-based model. In this model, objects are entities that combine state (i.e., data), behavior (i.e., procedures, or methods) and identity (unique existence among all other objects). The structure and behavior of an object are defined by a class, which is a syntactic structure, or blueprint, of all objects of a specific type. An object must be explicitly created based on a class and an object thus created is considered to be an instance of that class. An object is similar to a structure, with the addition of method pointers, member access control, and an implicit data member which locates instances of the class (i.e., objects of the class) in the class hierarchy (essential for runtime inheritance features). Encapsulation prevents users from breaking the invariants of the class, which is useful because it allows the implementation of a class of objects to be changed for aspects not exposed in the interface without impact to user code. The definitions of encapsulation focus on the grouping and packaging of related information (cohesion) rather than security issues. Class-based languages, or, to be more precise, typed languages, where subclassing is the only way of subtyping, have been criticized for mixing up implementations and interfaces—the essential principle in object-oriented programming. The critics say one might create a bag class that stores a collection of objects, then extend it to make a new class called a set class where the duplication of objects is eliminated. Now, a function that takes an object of the bag class may expect that adding two objects increases the size of a bag by two, yet if one passes an object of a set class, then adding two objects may or may not increase the size of a bag by two. The problem arises precisely because subclassing implies subtyping even in the instances where the principle of subtyping, known as the Liskov substitution principle, does not hold. Barbara Liskov and Jeannette Wing formulated the principle succinctly in a 1994 paper as follows: Subtype Requirement: Let ϕ ( x ) {\displaystyle \phi (x)} be a property provable about objects x {\displaystyle x} of type T {\displaystyle T} . Then ϕ ( y ) {\displaystyle \phi (y)} should be true for objects y {\displaystyle y} of type S {\displaystyle S} where S {\displaystyle S} is a subtype of T {\displaystyle T} . Thus, normally one must distinguish subtyping and subclassing. Most current object-oriented languages distinguish subtyping and subclassing, however some approaches to design do not. Also, another common example is that a person object created from a child class cannot become an object of parent class because a child class and a parent class inherit a person class but class-based languages mostly do not allow to change the kind of class of the object at runtime. For class-based languages, this restriction is essential in order to preserve unified view of the class to its users. The users should not need to care whether one of the implementations of a method happens to cause changes that break the invariants of the class. Such changes can be made by destroying the object and constructing another in its place. Polymorphism can be used to preserve the relevant interfaces even when such changes are done, because the objects are viewed as black box abstractions and accessed via object identity. However, usually the value of object references referring to the object is changed, which causes effects to client code. Although Simula introduced the class abstraction, the canonical example of a class-based language is Smalltalk. Others include PHP, C++, Java, C#, and Objective-C.[dubious – discuss][non sequitur] Prototype-based programming In contrast to creating an object from a class, some programming contexts support object creation by copying (cloning) a prototype object. See also Notes References Further reading |
======================================== |
[SOURCE: https://en.wikipedia.org/wiki/Method_(computing)] | [TOKENS: 1609] |
Contents Method (computer programming) A method in object-oriented programming (OOP) is a procedure associated with an object, and generally also a message. An object consists of state data and behavior; these compose an interface, which specifies how the object may be used. A method is a behavior of an object parametrized by a user. Data is represented as properties of the object, and behaviors are represented as methods. For example, a Window object could have methods such as open and close, while its state (whether it is open or closed at any given point in time) would be a property. In class-based programming, methods are defined within a class, and objects are instances of a given class. One of the most important capabilities that a method provides is method overriding - the same name (e.g., area) can be used for multiple different kinds of classes. This allows the sending objects to invoke behaviors and to delegate the implementation of those behaviors to the receiving object. A method in Java programming sets the behavior of a class object. For example, an object can send an area message to another object and the appropriate formula is invoked whether the receiving object is a Rectangle, Circle, Triangle, etc. Methods also provide the interface that other classes use to access and modify the properties of an object; this is known as encapsulation. Encapsulation and overriding are the two primary distinguishing features between methods and procedure calls. Overriding and overloading Method overriding and overloading are two of the most significant ways that a method differs from a conventional procedure or function call. Overriding refers to a subclass redefining the implementation of a method of its superclass. For example, findArea may be a method defined on a shape class, Triangle, etc. would each define the appropriate formula to calculate their area. The idea is to look at objects as "black boxes" so that changes to the internals of the object can be made with minimal impact on the other objects that use it. This is known as encapsulation and is meant to make code easier to maintain and re-use. Method overloading, on the other hand, refers to differentiating the code used to handle a message based on the parameters of the method. If one views the receiving object as the first parameter in any method then overriding is just a special case of overloading where the selection is based only on the first argument. The following simple Java example illustrates the difference: Accessor, mutator and manager methods Accessor methods are used to read the data values of an object. Mutator methods are used to modify the data of an object. Manager methods are used to initialize and destroy objects of a class, e.g. constructors and destructors. These methods provide an abstraction layer that facilitates encapsulation and modularity. For example, if a bank-account class provides a getBalance() accessor method to retrieve the current balance (rather than directly accessing the balance data fields), then later revisions of the same code can implement a more complex mechanism for balance retrieval (e.g., a database fetch), without the dependent code needing to be changed. The concepts of encapsulation and modularity are not unique to object-oriented programming. Indeed, in many ways the object-oriented approach is simply the logical extension of previous paradigms such as abstract data types and structured programming. A constructor is a method that is called at the beginning of an object's lifetime to create and initialize the object, a process called construction (or instantiation). Initialization may include an acquisition of resources. Constructors may have parameters but usually do not return values in most languages. See the following example in Java: A Destructor is a method that is called automatically at the end of an object's lifetime, a process called Destruction. Destruction in most languages does not allow destructor method arguments nor return values. Destructors can be implemented so as to perform cleanup chores and other tasks at object destruction. In garbage-collected languages, such as Java,: 26, 29 C#,: 208–209 and Python, destructors are known as finalizers. They have a similar purpose and function to destructors, but because of the differences between languages that utilize garbage-collection and languages with manual memory management, the sequence in which they are called is different. Abstract methods An abstract method is one with only a signature and no implementation body. It is often used to specify that a subclass must provide an implementation of the method, as in an abstract class. Abstract methods are used to specify interfaces in some programming languages. The following Java code shows an abstract class that needs to be extended: The following subclass extends the main class: If a subclass provides an implementation for an abstract method, another subclass can make it abstract again. This is called reabstraction. In practice, this is rarely used. In C#, a virtual method can be overridden with an abstract method. (This also applies to Java, where all non-private methods are virtual.) Interfaces' default methods can also be reabstracted, requiring subclasses to implement them. (This also applies to Java.) Class methods Class methods are methods that are called on a class rather than an instance. They are typically used as part of an object meta-model. I.e, for each class, defined an instance of the class object in the meta-model is created. Meta-model protocols allow classes to be created and deleted. In this sense, they provide the same functionality as constructors and destructors described above. But in some languages such as the Common Lisp Object System (CLOS) the meta-model allows the developer to dynamically alter the object model at run time: e.g., to create new classes, redefine the class hierarchy, modify properties, etc. Special methods Special methods are very language-specific and a language may support none, some, or all of the special methods defined here. A language's compiler may automatically generate default special methods or a programmer may be allowed to optionally define special methods. Most special methods cannot be directly called, but rather the compiler generates code to call them at appropriate times. Static methods are meant to be relevant to all the instances of a class rather than to any specific instance. They are similar to static variables in that sense. An example would be a static method to sum the values of all the variables of every instance of a class. For example, if there were a Product class it might have a static method to compute the average price of all products. A static method can be invoked even if no instances of the class exist yet. Static methods are called "static" because they are resolved at compile time based on the class they are called on and not dynamically as in the case with instance methods, which are resolved polymorphically based on the runtime type of the object. In Java, a commonly used static method is Math.max(). This static method has no owning object and does not run on an instance. It receives all information from its arguments. Copy-assignment operators define actions to be performed by the compiler when a class object is assigned to a class object of the same type. Operator methods define or redefine operator symbols and define the operations to be performed with the symbol and the associated method parameters. C++ example: Member functions in C++ Some procedural languages were extended with object-oriented capabilities to leverage the large skill sets and legacy code for those languages but still provide the benefits of object-oriented development. Perhaps the most well-known example is C++, an object-oriented extension of the C programming language. Due to the design requirements to add the object-oriented paradigm on to an existing procedural language, message passing in C++ has some unique capabilities and terminologies. For example, in C++ a method is known as a member function. C++ also has the concept of virtual functions which are member functions that can be overridden in derived classes and allow for dynamic dispatch. Virtual functions are the means by which a C++ class can achieve polymorphic behavior. Non-virtual member functions, or regular methods, are those that do not participate in polymorphism. C++ Example: See also Notes References |
======================================== |
[SOURCE: https://en.wikipedia.org/wiki/Meta_Platforms#References] | [TOKENS: 8626] |
Contents Meta Platforms Meta Platforms, Inc. (doing business as Meta) is an American multinational technology company headquartered in Menlo Park, California. Meta owns and operates several prominent social media platforms and communication services, including Facebook, Instagram, WhatsApp, Messenger, Threads and Manus. The company also operates an advertising network for its own sites and third parties; as of 2023[update], advertising accounted for 97.8 percent of its total revenue. Meta has been described as a part of Big Tech, which refers to the largest six tech companies in the United States, Alphabet (Google), Amazon, Apple, Meta (Facebook), Microsoft, and Nvidia, which are also the largest companies in the world by market capitalization. The company was originally established in 2004 as TheFacebook, Inc., and was renamed Facebook, Inc. in 2005. In 2021, it rebranded as Meta Platforms, Inc. to reflect a strategic shift toward developing the metaverse—an interconnected digital ecosystem spanning virtual and augmented reality technologies. In 2023, Meta was ranked 31st on the Forbes Global 2000 list of the world's largest public companies. As of 2022, it was the world's third-largest spender on research and development, with R&D expenses totaling US$35.3 billion. History Facebook filed for an initial public offering (IPO) on January 1, 2012. The preliminary prospectus stated that the company sought to raise $5 billion, had 845 million monthly active users, and a website accruing 2.7 billion likes and comments daily. After the IPO, Zuckerberg would retain 22% of the total shares and 57% of the total voting power in Facebook. Underwriters valued the shares at $38 each, valuing the company at $104 billion, the largest valuation yet for a newly public company. On May 16, one day before the IPO, Facebook announced it would sell 25% more shares than originally planned due to high demand. The IPO raised $16 billion, making it the third-largest in US history (slightly ahead of AT&T Mobility and behind only General Motors and Visa). The stock price left the company with a higher market capitalization than all but a few U.S. corporations—surpassing heavyweights such as Amazon, McDonald's, Disney, and Kraft Foods—and made Zuckerberg's stock worth $19 billion. The New York Times stated that the offering overcame questions about Facebook's difficulties in attracting advertisers to transform the company into a "must-own stock". Jimmy Lee of JPMorgan Chase described it as "the next great blue-chip". Writers at TechCrunch, on the other hand, expressed skepticism, stating, "That's a big multiple to live up to, and Facebook will likely need to add bold new revenue streams to justify the mammoth valuation." Trading in the stock, which began on May 18, was delayed that day due to technical problems with the Nasdaq exchange. The stock struggled to stay above the IPO price for most of the day, forcing underwriters to buy back shares to support the price. At the closing bell, shares were valued at $38.23, only $0.23 above the IPO price and down $3.82 from the opening bell value. The opening was widely described by the financial press as a disappointment. The stock set a new record for trading volume of an IPO. On May 25, 2012, the stock ended its first full week of trading at $31.91, a 16.5% decline. On May 22, 2012, regulators from Wall Street's Financial Industry Regulatory Authority announced that they had begun to investigate whether banks underwriting Facebook had improperly shared information only with select clients rather than the general public. Massachusetts Secretary of State William F. Galvin subpoenaed Morgan Stanley over the same issue. The allegations sparked "fury" among some investors and led to the immediate filing of several lawsuits, one of them a class action suit claiming more than $2.5 billion in losses due to the IPO. Bloomberg estimated that retail investors may have lost approximately $630 million on Facebook stock since its debut. S&P Global Ratings added Facebook to its S&P 500 index on December 21, 2013. On May 2, 2014, Zuckerberg announced that the company would be changing its internal motto from "Move fast and break things" to "Move fast with stable infrastructure". The earlier motto had been described as Zuckerberg's "prime directive to his developers and team" in a 2009 interview in Business Insider, in which he also said, "Unless you are breaking stuff, you are not moving fast enough." In November 2016, Facebook announced the Microsoft Windows client of gaming service Facebook Gameroom, formerly Facebook Games Arcade, at the Unity Technologies developers conference. The client allows Facebook users to play "native" games in addition to its web games. The service was closed in June 2021. Lasso was a short-video sharing app from Facebook similar to TikTok that was launched on iOS and Android in 2018 and was aimed at teenagers. On July 2, 2020, Facebook announced that Lasso would be shutting down on July 10. In 2018, the Oculus lead Jason Rubin sent his 50-page vision document titled "The Metaverse" to Facebook's leadership. In the document, Rubin acknowledged that Facebook's virtual reality business had not caught on as expected, despite the hundreds of millions of dollars spent on content for early adopters. He also urged the company to execute fast and invest heavily in the vision, to shut out HTC, Apple, Google and other competitors in the VR space. Regarding other players' participation in the metaverse vision, he called for the company to build the "metaverse" to prevent their competitors from "being in the VR business in a meaningful way at all". In May 2019, Facebook founded Libra Networks, reportedly to develop their own stablecoin cryptocurrency. Later, it was reported that Libra was being supported by financial companies such as Visa, Mastercard, PayPal and Uber. The consortium of companies was expected to pool in $10 million each to fund the launch of the cryptocurrency coin named Libra. Depending on when it would receive approval from the Swiss Financial Market Supervisory authority to operate as a payments service, the Libra Association had planned to launch a limited format cryptocurrency in 2021. Libra was renamed Diem, before being shut down and sold in January 2022 after backlash from Swiss government regulators and the public. During the COVID-19 pandemic, the use of online services, including Facebook, grew globally. Zuckerberg predicted this would be a "permanent acceleration" that would continue after the pandemic. Facebook hired aggressively, growing from 48,268 employees in March 2020 to more than 87,000 by September 2022. Following a period of intense scrutiny and damaging whistleblower leaks, news started to emerge on October 21, 2021 about Facebook's plan to rebrand the company and change its name. In the Q3 2021 earnings call on October 25, Mark Zuckerberg discussed the ongoing criticism of the company's social services and the way it operates, and pointed to the pivoting efforts to building the metaverse – without mentioning the rebranding and the name change. The metaverse vision and the name change from Facebook, Inc. to Meta Platforms was introduced at Facebook Connect on October 28, 2021. Based on Facebook's PR campaign, the name change reflects the company's shifting long term focus of building the metaverse, a digital extension of the physical world by social media, virtual reality and augmented reality features. "Meta" had been registered as a trademark in the United States in 2018 (after an initial filing in 2015) for marketing, advertising, and computer services, by a Canadian company that provided big data analysis of scientific literature. This company was acquired in 2017 by the Chan Zuckerberg Initiative (CZI), a foundation established by Zuckerberg and his wife, Priscilla Chan, and became one of their projects. Following the rebranding announcement, CZI announced that it had already decided to deprioritize the earlier Meta project, thus it would be transferring its rights to the name to Meta Platforms, and the previous project would end in 2022. Soon after the rebranding, in early February 2022, Meta reported a greater-than-expected decline in profits in the fourth quarter of 2021. It reported no growth in monthly users, and indicated it expected revenue growth to stall. It also expected measures taken by Apple Inc. to protect user privacy to cost it some $10 billion in advertisement revenue, an amount equal to roughly 8% of its revenue for 2021. In meeting with Meta staff the day after earnings were reported, Zuckerberg blamed competition for user attention, particularly from video-based apps such as TikTok. The 27% reduction in the company's share price which occurred in reaction to the news eliminated some $230 billion of value from Meta's market capitalization. Bloomberg described the decline as "an epic rout that, in its sheer scale, is unlike anything Wall Street or Silicon Valley has ever seen". Zuckerberg's net worth fell by as much as $31 billion. Zuckerberg owns 13% of Meta, and the holding makes up the bulk of his wealth. According to published reports by Bloomberg on March 30, 2022, Meta turned over data such as phone numbers, physical addresses, and IP addresses to hackers posing as law enforcement officials using forged documents. The law enforcement requests sometimes included forged signatures of real or fictional officials. When asked about the allegations, a Meta representative said, "We review every data request for legal sufficiency and use advanced systems and processes to validate law enforcement requests and detect abuse." In June 2022, Sheryl Sandberg, the chief operating officer of 14 years, announced she would step down that year. Zuckerberg said that Javier Olivan would replace Sandberg, though in a “more traditional” role. In March 2022, Meta (except Meta-owned WhatsApp) and Instagram were banned in Russia and added to the Russian list of terrorist and extremist organizations for alleged Russophobia and hate speech (up to genocidal calls) amid the ongoing Russian invasion of Ukraine. Meta appealed against the ban, but it was upheld by a Moscow court in June of the same year. Also in March 2022, Meta and Italian eyewear giant Luxottica released Ray-Ban Stories, a series of smartglasses which could play music and take pictures. Meta and Luxottica parent company EssilorLuxottica declined to disclose sales on the line of products as of September 2022, though Meta has expressed satisfaction with its customer feedback. In July 2022, Meta saw its first year-on-year revenue decline when its total revenue slipped by 1% to $28.8bn. Analysts and journalists accredited the loss to its advertising business, which has been limited by Apple's app tracking transparency feature and the number of people who have opted not to be tracked by Meta apps. Zuckerberg also accredited the decline to increasing competition from TikTok. On October 27, 2022, Meta's market value dropped to $268 billion, a loss of around $700 billion compared to 2021, and its shares fell by 24%. It lost its spot among the top 20 US companies by market cap, despite reaching the top 5 in the previous year. In November 2022, Meta laid off 11,000 employees, 13% of its workforce. Zuckerberg said the decision to aggressively increase Meta's investments had been a mistake, as he had wrongly predicted that the surge in e-commerce would last beyond the COVID-19 pandemic. He also attributed the decline to increased competition, a global economic downturn and "ads signal loss". Plans to lay off a further 10,000 employees began in April 2023. The layoffs were part of a general downturn in the technology industry, alongside layoffs by companies including Google, Amazon, Tesla, Snap, Twitter and Lyft. Starting from 2022, Meta scrambled to catch up to other tech companies in adopting specialized artificial intelligence hardware and software. It had been using less expensive CPUs instead of GPUs for AI work, but that approach turned out to be less efficient. The company gifted the Inter-university Consortium for Political and Social Research $1.3 million to finance the Social Media Archive's aim to make their data available to social science research. In 2023, Ireland's Data Protection Commissioner imposed a record EUR 1.2 billion fine on Meta for transferring data from Europe to the United States without adequate protections for EU citizens.: 250 In March 2023, Meta announced a new round of layoffs that would cut 10,000 employees and close 5,000 open positions to make the company more efficient. Meta revenue surpassed analyst expectations for the first quarter of 2023 after announcing that it was increasing its focus on AI. On July 6, Meta launched a new app, Threads, a competitor to Twitter. Meta announced its artificial intelligence model Llama 2 in July 2023, available for commercial use via partnerships with major cloud providers like Microsoft. It was the first project to be unveiled out of Meta's generative AI group after it was set up in February. It would not charge access or usage but instead operate with an open-source model to allow Meta to ascertain what improvements need to be made. Prior to this announcement, Meta said it had no plans to release Llama 2 for commercial use. An earlier version of Llama was released to academics. In August 2023, Meta announced its permanent removal of news content from Facebook and Instagram in Canada due to the Online News Act, which requires Canadian news outlets to be compensated for content shared on its platform. The Online News Act was in effect by year-end, but Meta will not participate in the regulatory process. In October 2023, Zuckerberg said that AI would be Meta's biggest investment area in 2024. Meta finished 2023 as one of the best-performing technology stocks of the year, with its share price up 150 percent. Its stock reached an all-time high in January 2024, bringing Meta within 2% of achieving $1 trillion market capitalization. In November 2023 Meta Platforms launched an ad-free service in Europe, allowing subscribers to opt-out of personal data being collected for targeted advertising. A group of 28 European organizations, including Max Schrems' advocacy group NOYB, the Irish Council for Civil Liberties, Wikimedia Europe, and the Electronic Privacy Information Center, signed a 2024 letter to the European Data Protection Board (EDPB) expressing concern that this subscriber model would undermine privacy protections, specifically GDPR data protection standards. Meta removed the Facebook and Instagram accounts of Iran's Supreme Leader Ali Khamenei in February 2024, citing repeated violations of its Dangerous Organizations & Individuals policy. As of March, Meta was under investigation by the FDA for alleged use of their social media platforms to sell illegal drugs. On 16 May 2024, the European Commission began an investigation into Meta over concerns related to child safety. In May 2023, Iraqi social media influencer Esaa Ahmed-Adnan encountered a troubling issue when Instagram removed his posts, citing false copyright violations despite his content being original and free from copyrighted material. He discovered that extortionists were behind these takedowns, offering to restore his content for $3,000 or provide ongoing protection for $1,000 per month. This scam, exploiting Meta’s rights management tools, became widespread in the Middle East, revealing a gap in Meta’s enforcement in developing regions. An Iraqi nonprofit Tech4Peace’s founder, Aws al-Saadi helped Ahmed-Adnan and others, but the restoration process was slow, leading to significant financial losses for many victims, including prominent figures like Ammar al-Hakim. This situation highlighted Meta’s challenges in balancing global growth with effective content moderation and protection. On 16 September 2024, Meta announced it had banned Russian state media outlets from its platforms worldwide due to concerns about "foreign interference activity." This decision followed allegations that RT and its employees funneled $10 million through shell companies to secretly fund influence campaigns on various social media channels. Meta's actions were part of a broader effort to counter Russian covert influence operations, which had intensified since the invasion. At its 2024 Connect conference, Meta presented Orion, its first pair of augmented reality glasses. Though Orion was originally intended to be sold to consumers, the manufacturing process turned out to be too complex and expensive. Instead, the company pivoted to producing a small number of the glasses to be used internally. On 4 October 2024, Meta announced about its new AI model called Movie Gen, capable of generating realistic video and audio clips based on user prompts. Meta stated it would not release Movie Gen for open development, preferring to collaborate directly with content creators and integrate it into its products by the following year. The model was built using a combination of licensed and publicly available datasets. On October 31, 2024, ProPublica published an investigation into deceptive political advertisement scams that sometimes use hundreds of hijacked profiles and facebook pages run by organized networks of scammers. The authors cited spotty enforcement by Meta as a major reason for the extent of the issue. In November 2024, TechCrunch reported that Meta were considering building a $10bn global underwater cable spanning 25,000 miles. In the same month, Meta closed down 2 million accounts on Facebook and Instagram that were linked to scam centers in Myanmar, Laos, Cambodia, the Philippines, and the United Arab Emirates doing pig butchering scams. In December 2024, Meta announced that, beginning February 2025, they would require advertisers to run ads about financial services in Australia to verify information about who are the beneficiary and the payer in a bid to regulate scams. On December 4, 2024, Meta announced it will invest US$10 billion for its largest AI data center in northeast Louisiana, powered by natural gas facilities. On the 11th of that month, Meta experienced a global outage, impacting accounts on all of their social media and messaging applications. Outage reports from DownDetector reached 70,000+ and 100,000+ within minutes for Instagram and Facebook, respectively. In January 2025, Meta announced plans to roll back its diversity, equity, and inclusion (DEI) initiatives, citing shifts in the "legal and policy landscape" in the United States following the 2024 presidential election. The decision followed reports that CEO Mark Zuckerberg sought to align the company more closely with the incoming Trump administration, including changes to content moderation policies and executive leadership. The new content moderation policies continued to bar insults about a person's intellect or mental illness, but made an exception to allow calling LGBTQ people mentally ill because they are gay or transgender. Later that month, Meta agreed to pay $25 million to settle a 2021 lawsuit brought by Donald Trump for suspending his social media accounts after the January 6 riots. Changes to Meta's moderation policies were controversial among its oversight board, with a significant divide in opinion between the board's US conservatives and its global members. In June 2025, Meta Platforms Inc. has decided to make a multibillion-dollar investment into artificial intelligence startup Scale AI. The financing could exceed $10 billion in value which would make it one of the largest private company funding events of all time. In October 2025, it was announced that Meta would be laying off 600 employees in the artificial intelligence unit to perform better and simpler. They referred to their AI unit as "bloated" and are seeking to trim down the department. This mass layoff is going to impact Meta’s AI infrastructure units, Fundamental Artificial Intelligence Research unit (FAIR) and other product-related positions. Mergers and acquisitions Meta has acquired multiple companies (often identified as talent acquisitions). One of its first major acquisitions was in April 2012, when it acquired Instagram for approximately US$1 billion in cash and stock. In October 2013, Facebook, Inc. acquired Onavo, an Israeli mobile web analytics company. In February 2014, Facebook, Inc. announced it would buy mobile messaging company WhatsApp for US$19 billion in cash and stock. The acquisition was completed on October 6. Later that year, Facebook bought Oculus VR for $2.3 billion in cash and stock, which released its first consumer virtual reality headset in 2016. In late November 2019, Facebook, Inc. announced the acquisition of the game developer Beat Games, responsible for developing one of that year's most popular VR games, Beat Saber. In Late 2022, after Facebook Inc rebranded to Meta Platforms Inc, Oculus was rebranded to Meta Quest. In May 2020, Facebook, Inc. announced it had acquired Giphy for a reported cash price of $400 million. It will be integrated with the Instagram team. However, in August 2021, UK's Competition and Markets Authority (CMA) stated that Facebook, Inc. might have to sell Giphy, after an investigation found that the deal between the two companies would harm competition in display advertising market. Facebook, Inc. was fined $70 million by CMA for deliberately failing to report all information regarding the acquisition and the ongoing antitrust investigation. In October 2022, the CMA ruled for a second time that Meta be required to divest Giphy, stating that Meta already controls half of the advertising in the UK. Meta agreed to the sale, though it stated that it disagrees with the decision itself. In May 2023, Giphy was divested to Shutterstock for $53 million. In November 2020, Facebook, Inc. announced that it planned to purchase the customer-service platform and chatbot specialist startup Kustomer to promote companies to use their platform for business. It has been reported that Kustomer valued at slightly over $1 billion. The deal was closed in February 2022 after regulatory approval. In September 2022, Meta acquired Lofelt, a Berlin-based haptic tech startup. In December 2025, it was announced Meta had acquired the AI-wearables startup, Limitless. In the same month, they also acquired another AI startup, Manus AI, for $2 billion. Manus announced in December that its platform had achieved $100mm in recurring revenue just 8 months after its launch and Meta said it will scale the platform to many other businesses. In January 2026, it was announced Meta proposed acquisition of Manus was undergoing preliminary scrutiny by Chinese regulators. The examination concerns the cross-border transfer of artificial intelligence technology developed in China. Lobbying In 2020, Facebook, Inc. spent $19.7 million on lobbying, hiring 79 lobbyists. In 2019, it had spent $16.7 million on lobbying and had a team of 71 lobbyists, up from $12.6 million and 51 lobbyists in 2018. Facebook was the largest spender of lobbying money among the Big Tech companies in 2020. The lobbying team includes top congressional aide John Branscome, who was hired in September 2021, to help the company fend off threats from Democratic lawmakers and the Biden administration. In December 2024, Meta donated $1 million to the inauguration fund for then-President-elect Donald Trump. In 2025, Meta was listed among the donors funding the construction of the White House State Ballroom. Partnerships February 2026, Meta announced a long-term partnership with Nvidia. Censorship In August 2024, Mark Zuckerberg sent a letter to Jim Jordan indicating that during the COVID-19 pandemic the Biden administration repeatedly asked Meta to limit certain COVID-19 content, including humor and satire, on Facebook and Instagram. In 2016 Meta hired Jordana Cutler, formerly an employee at the Israeli Embassy to the United States, as its policy chief for Israel and the Jewish Diaspora. In this role, Cutler pushed for the censorship of accounts belonging to Students for Justice in Palestine chapters in the United States. Critics have said that Cutler's position gives the Israeli government an undue influence over Meta policy, and that few countries have such high levels of contact with Meta policymakers. Following the election of Donald Trump in 2025, various sources noted possible censorship related to the Democratic Party on Instagram and other Meta platforms. In February 2025, a Meta rep flagged journalist Gil Duran's article and other "critiques of tech industry figures" as spam or sensitive content, limiting their reach. In March 2025, Meta attempted to block former employee Sarah Wynn-Williams from promoting or further distributing her memoir, Careless People, that includes allegations of unaddressed sexual harassment in the workplace by senior executives. The New York Times reports that the arbitration is among Meta's most forcible attempts to repudiate a former employee's account of workplace dynamics. Publisher Macmillan reacted to the ruling by the Emergency International Arbitral Tribunal by stating that it will ignore its provisions. As of 15 March 2025[update], hardback and digital versions of Careless People were being offered for sale by major online retailers. From October 2025, Meta began removing and restricting access for accounts related to LGBTQ, reproductive health and abortion information pages on its platforms. Martha Dimitratou, executive director of Repro Uncensored, called Meta's shadow-banning of these issues "One of the biggest waves of censorship we are seeing". Disinformation concerns Since its inception, Meta has been accused of being a host for fake news and misinformation. In the wake of the 2016 United States presidential election, Zuckerberg began to take steps to eliminate the prevalence of fake news, as the platform had been criticized for its potential influence on the outcome of the election. The company initially partnered with ABC News, the Associated Press, FactCheck.org, Snopes and PolitiFact for its fact-checking initiative; as of 2018, it had over 40 fact-checking partners across the world, including The Weekly Standard. A May 2017 review by The Guardian found that the platform's fact-checking initiatives of partnering with third-party fact-checkers and publicly flagging fake news were regularly ineffective, and appeared to be having minimal impact in some cases. In 2018, journalists working as fact-checkers for the company criticized the partnership, stating that it had produced minimal results and that the company had ignored their concerns. In 2024 Meta's decision to continue to disseminate a falsified video of US president Joe Biden, even after it had been proven to be fake, attracted criticism and concern. In January 2025, Meta ended its use of third-party fact-checkers in favor of a user-run community notes system similar to the one used on X. While Zuckerberg supported these changes, saying that the amount of censorship on the platform was excessive, the decision received criticism by fact-checking institutions, stating that the changes would make it more difficult for users to identify misinformation. Meta also faced criticism for weakening its policies on hate speech that were designed to protect minorities and LGBTQ+ individuals from bullying and discrimination. While moving its content review teams from California to Texas, Meta changed their hateful conduct policy to eliminate restrictions on anti-LGBT and anti-immigrant hate speech, as well as explicitly allowing users to accuse LGBT people of being mentally ill or abnormal based on their sexual orientation or gender identity. In January 2025, Meta faced significant criticism for its role in removing LGBTQ+ content from its platforms, amid its broader efforts to address anti-LGBTQ+ hate speech. The removal of LGBTQ+ themes was noted as part of the wider crackdown on content deemed to violate its community guidelines. Meta's content moderation policies, which were designed to combat harmful speech and protect users from discrimination, inadvertently led to the removal or restriction of LGBTQ+ content, particularly posts highlighting LGBTQ+ identities, support, or political issues. According to reports, LGBTQ+ posts, including those that simply celebrated pride or advocated for LGBTQ+ rights, were flagged and removed for reasons that some critics argue were vague or inconsistently applied. Many LGBTQ+ activists and users on Meta's platforms expressed concern that such actions stifled visibility and expression, potentially isolating LGBTQ+ individuals and communities, especially in spaces that were historically important for outreach and support. Lawsuits Numerous lawsuits have been filed against the company, both when it was known as Facebook, Inc., and as Meta Platforms. In March 2020, the Office of the Australian Information Commissioner (OAIC) sued Facebook, for significant and persistent infringements of the rule on privacy involving the Cambridge Analytica fiasco. Every violation of the Privacy Act is subject to a theoretical cumulative liability of $1.7 million. The OAIC estimated that a total of 311,127 Australians had been exposed. On December 8, 2020, the U.S. Federal Trade Commission and 46 states (excluding Alabama, Georgia, South Carolina, and South Dakota), the District of Columbia and the territory of Guam, launched Federal Trade Commission v. Facebook as an antitrust lawsuit against Facebook. The lawsuit concerns Facebook's acquisition of two competitors—Instagram and WhatsApp—and the ensuing monopolistic situation. FTC alleges that Facebook holds monopolistic power in the U.S. social networking market and seeks to force the company to divest from Instagram and WhatsApp to break up the conglomerate. William Kovacic, a former chairman of the Federal Trade Commission, argued the case will be difficult to win as it would require the government to create a counterfactual argument of an internet where the Facebook-WhatsApp-Instagram entity did not exist, and prove that harmed competition or consumers. In November 2025, it was ruled that Meta did not violate antitrust laws and holds no monopoly in the market. On December 24, 2021, a court in Russia fined Meta for $27 million after the company declined to remove unspecified banned content. The fine was reportedly tied to the company's annual revenue in the country. In May 2022, a lawsuit was filed in Kenya against Meta and its local outsourcing company Sama. Allegedly, Meta has poor working conditions in Kenya for workers moderating Facebook posts. According to the lawsuit, 260 screeners were declared redundant with confusing reasoning. The lawsuit seeks financial compensation and an order that outsourced moderators be given the same health benefits and pay scale as Meta employees. In June 2022, 8 lawsuits were filed across the U.S. over the allege that excessive exposure to platforms including Facebook and Instagram has led to attempted or actual suicides, eating disorders and sleeplessness, among other issues. The litigation follows a former Facebook employee's testimony in Congress that the company refused to take responsibility. The company noted that tools have been developed for parents to keep track of their children's activity on Instagram and set time limits, in addition to Meta's "Take a break" reminders. In addition, the company is providing resources specific to eating disorders as well as developing AI to prevent children under the age of 13 signing up for Facebook or Instagram. In June 2022, Meta settled a lawsuit with the US Department of Justice. The lawsuit, which was filed in 2019, alleged that the company enabled housing discrimination through targeted advertising, as it allowed homeowners and landlords to run housing ads excluding people based on sex, race, religion, and other characteristics. The U.S. Department of Justice stated that this was in violation of the Fair Housing Act. Meta was handed a penalty of $115,054 and given until December 31, 2022, to shadow the algorithm tool. In January 2023, Meta was fined €390 million for violations of the European Union General Data Protection Regulation. In May 2023, the European Data Protection Board fined Meta a record €1.2 billion for breaching European Union data privacy laws by transferring personal data of Facebook users to servers in the U.S. In July 2024, Meta agreed to pay the state of Texas US$1.4 billion to settle a lawsuit brought by Texas Attorney General Ken Paxton accusing the company of collecting users' biometric data without consent, setting a record for the largest privacy-related settlement ever obtained by a state attorney general. In October 2024, Meta Platforms faced lawsuits in Japan from 30 plaintiffs who claimed they were defrauded by fake investment ads on Facebook and Instagram, featuring false celebrity endorsements. The plaintiffs are seeking approximately $2.8 million in damages. In April 2025, the Kenyan High Court ruled that a US$2.4 billion lawsuit in which three plaintiffs claim that Facebook inflamed civil violence in Ethiopia in 2021 could proceed. In April 2025, Meta was fined €200 million ($230 million) for breaking the Digital Markets Act, by imposing a “consent or pay” system that forces users to either allow their personal data to be used to target advertisements, or pay a subscription fee for advertising-free versions of Facebook and Instagram. In late April 2025, a case was filed against Meta in Ghana over the alleged psychological distress experienced by content moderators employed to take down disturbing social media content including depictions of murders, extreme violence and child sexual abuse. Meta moved the moderation service to the Ghanaian capital of Accra after legal issues in the previous location Kenya. The new moderation company is Teleperformance, a multinational corporation with a history of worker's rights violation. Reports suggests the conditions are worse here than in the previous Kenyan location, with many workers afraid of speaking out due to fear of returning to conflict zones. Workers reported developing mental illnesses, attempted suicides, and low pay. In 26 January 2026, a New Mexico state court case was filed, suggesting that Mark Zuckerberg approved allowing minors to access artificial intelligence chatbot companions that safety staffers warned were capable of sexual interactions. In 2020, the company UReputation, which had been involved in several cases concerning the management of digital armies[clarification needed], filed a lawsuit against Facebook, accusing it of unlawfully transmitting personal data to third parties. Legal actions were initiated in Tunisia, France, and the United States. In 2025, the United States District court for the Northern District of Georgia approved a discovery procedure, allowing UReputation to access documents and evidence held by Meta. Structure Meta's key management consists of: As of October 2022[update], Meta had 83,553 employees worldwide. As of June 2024[update], Meta's board consisted of the following directors; Meta Platforms is mainly owned by institutional investors, who hold around 80% of all shares. Insiders control the majority of voting shares. The three largest individual investors in 2024 were Mark Zuckerberg, Sheryl Sandberg and Christopher K. Cox. The largest shareholders in late 2024/early 2025 were: Roger McNamee, an early Facebook investor and Zuckerberg's former mentor, said Facebook had "the most centralized decision-making structure I have ever encountered in a large company". Facebook co-founder Chris Hughes has stated that chief executive officer Mark Zuckerberg has too much power, that the company is now a monopoly, and that, as a result, it should be split into multiple smaller companies. In an op-ed in The New York Times, Hughes said he was concerned that Zuckerberg had surrounded himself with a team that did not challenge him, and that it is the U.S. government's job to hold him accountable and curb his "unchecked power". He also said that "Mark's power is unprecedented and un-American." Several U.S. politicians agreed with Hughes. European Union Commissioner for Competition Margrethe Vestager stated that splitting Facebook should be done only as "a remedy of the very last resort", and that it would not solve Facebook's underlying problems. Revenue Facebook ranked No. 34 in the 2020 Fortune 500 list of the largest United States corporations by revenue, with almost $86 billion in revenue most of it coming from advertising. One analysis of 2017 data determined that the company earned US$20.21 per user from advertising. According to New York, since its rebranding, Meta has reportedly lost $500 billion as a result of new privacy measures put in place by companies such as Apple and Google which prevents Meta from gathering users' data. In February 2015, Facebook announced it had reached two million active advertisers, with most of the gain coming from small businesses. An active advertiser was defined as an entity that had advertised on the Facebook platform in the last 28 days. In March 2016, Facebook announced it had reached three million active advertisers with more than 70% from outside the United States. Prices for advertising follow a variable pricing model based on auctioning ad placements, and potential engagement levels of the advertisement itself. Similar to other online advertising platforms like Google and Twitter, targeting of advertisements is one of the chief merits of digital advertising compared to traditional media. Marketing on Meta is employed through two methods based on the viewing habits, likes and shares, and purchasing data of the audience, namely targeted audiences and "look alike" audiences. The U.S. IRS challenged the valuation Facebook used when it transferred IP from the U.S. to Facebook Ireland (now Meta Platforms Ireland) in 2010 (which Facebook Ireland then revalued higher before charging out), as it was building its double Irish tax structure. The case is ongoing and Meta faces a potential fine of $3–5bn. The U.S. Tax Cuts and Jobs Act of 2017 changed Facebook's global tax calculations. Meta Platforms Ireland is subject to the U.S. GILTI tax of 10.5% on global intangible profits (i.e. Irish profits). On the basis that Meta Platforms Ireland Limited is paying some tax, the effective minimum US tax for Facebook Ireland will be circa 11%. In contrast, Meta Platforms Inc. would incur a special IP tax rate of 13.125% (the FDII rate) if its Irish business relocated to the U.S. Tax relief in the U.S. (21% vs. Irish at the GILTI rate) and accelerated capital expensing, would make this effective U.S. rate around 12%. The insignificance of the U.S./Irish tax difference was demonstrated when Facebook moved 1.5bn non-EU accounts to the U.S. to limit exposure to GDPR. Facilities Users outside of the U.S. and Canada contract with Meta's Irish subsidiary, Meta Platforms Ireland Limited (formerly Facebook Ireland Limited), allowing Meta to avoid US taxes for all users in Europe, Asia, Australia, Africa and South America. Meta is making use of the Double Irish arrangement which allows it to pay 2–3% corporation tax on all international revenue. In 2010, Facebook opened its fourth office, in Hyderabad, India, which houses online advertising and developer support teams and provides support to users and advertisers. In India, Meta is registered as Facebook India Online Services Pvt Ltd. It also has offices or planned sites in Chittagong, Bangladesh; Dublin, Ireland; and Austin, Texas, among other cities. Facebook opened its London headquarters in 2017 in Fitzrovia in central London. Facebook opened an office in Cambridge, Massachusetts in 2018. The offices were initially home to the "Connectivity Lab", a group focused on bringing Internet access to those who do not have access to the Internet. In April 2019, Facebook opened its Taiwan headquarters in Taipei. In March 2022, Meta opened new regional headquarters in Dubai. In September 2023, it was reported that Meta had paid £149m to British Land to break the lease on Triton Square London office. Meta reportedly had another 18 years left on its lease on the site. As of 2023, Facebook operated 21 data centers. It committed to purchase 100% renewable energy and to reduce its greenhouse gas emissions 75% by 2020. Its data center technologies include Fabric Aggregator, a distributed network system that accommodates larger regions and varied traffic patterns. Reception US Representative Alexandria Ocasio-Cortez responded in a tweet to Zuckerberg's announcement about Meta, saying: "Meta as in 'we are a cancer to democracy metastasizing into a global surveillance and propaganda machine for boosting authoritarian regimes and destroying civil society ... for profit!'" Ex-Facebook employee Frances Haugen and whistleblower behind the Facebook Papers responded to the rebranding efforts by expressing doubts about the company's ability to improve while led by Mark Zuckerberg, and urged the chief executive officer to resign. In November 2021, a video published by Inspired by Iceland went viral, in which a Zuckerberg look-alike promoted the Icelandverse, a place of "enhanced actual reality without silly looking headsets". In a December 2021 interview, SpaceX and Tesla chief executive officer Elon Musk said he could not see a compelling use-case for the VR-driven metaverse, adding: "I don't see someone strapping a frigging screen to their face all day." In January 2022, Louise Eccles of The Sunday Times logged into the metaverse with the intention of making a video guide. She wrote: Initially, my experience with the Oculus went well. I attended work meetings as an avatar and tried an exercise class set in the streets of Paris. The headset enabled me to feel the thrill of carving down mountains on a snowboard and the adrenaline rush of climbing a mountain without ropes. Yet switching to the social apps, where you mingle with strangers also using VR headsets, it was at times predatory and vile. Eccles described being sexually harassed by another user, as well as "accents from all over the world, American, Indian, English, Australian, using racist, sexist, homophobic and transphobic language". She also encountered users as young as 7 years old on the platform, despite Oculus headsets being intended for users over 13. See also References External links 37°29′06″N 122°08′54″W / 37.48500°N 122.14833°W / 37.48500; -122.14833 |
======================================== |
[SOURCE: https://en.wikipedia.org/wiki/Mars#cite_note-Williams_2016-78] | [TOKENS: 11899] |
Contents Mars Mars is the fourth planet from the Sun. It is also known as the "Red Planet", for its orange-red appearance. Mars is a desert-like rocky planet with a tenuous atmosphere that is primarily carbon dioxide (CO2). At the average surface level the atmospheric pressure is a few thousandths of Earth's, atmospheric temperature ranges from −153 to 20 °C (−243 to 68 °F), and cosmic radiation is high. Mars retains some water, in the ground as well as thinly in the atmosphere, forming cirrus clouds, fog, frost, larger polar regions of permafrost and ice caps (with seasonal CO2 snow), but no bodies of liquid surface water. Its surface gravity is roughly a third of Earth's or double that of the Moon. Its diameter, 6,779 km (4,212 mi), is about half the Earth's, or twice the Moon's, and its surface area is the size of all the dry land of Earth. Fine dust is prevalent across the surface and the atmosphere, being picked up and spread at the low Martian gravity even by the weak wind of the tenuous atmosphere. The terrain of Mars roughly follows a north-south divide, the Martian dichotomy, with the northern hemisphere mainly consisting of relatively flat, low lying plains, and the southern hemisphere of cratered highlands. Geologically, the planet is fairly active with marsquakes trembling underneath the ground, but also hosts many enormous volcanoes that are extinct (the tallest is Olympus Mons, 21.9 km or 13.6 mi tall), as well as one of the largest canyons in the Solar System (Valles Marineris, 4,000 km or 2,500 mi long). Mars has two natural satellites that are small and irregular in shape: Phobos and Deimos. With a significant axial tilt of 25 degrees, Mars experiences seasons, like Earth (which has an axial tilt of 23.5 degrees). A Martian solar year is equal to 1.88 Earth years (687 Earth days), a Martian solar day (sol) is equal to 24.6 hours. Mars formed along with the other planets approximately 4.5 billion years ago. During the martian Noachian period (4.5 to 3.5 billion years ago), its surface was marked by meteor impacts, valley formation, erosion, the possible presence of water oceans and the loss of its magnetosphere. The Hesperian period (beginning 3.5 billion years ago and ending 3.3–2.9 billion years ago) was dominated by widespread volcanic activity and flooding that carved immense outflow channels. The Amazonian period, which continues to the present, is the currently dominating and remaining influence on geological processes. Because of Mars's geological history, the possibility of past or present life on Mars remains an area of active scientific investigation, with some possible traces needing further examination. Being visible with the naked eye in Earth's sky as a red wandering star, Mars has been observed throughout history, acquiring diverse associations in different cultures. In 1963 the first flight to Mars took place with Mars 1, but communication was lost en route. The first successful flyby exploration of Mars was conducted in 1965 with Mariner 4. In 1971 Mariner 9 entered orbit around Mars, being the first spacecraft to orbit any body other than the Moon, Sun or Earth; following in the same year were the first uncontrolled impact (Mars 2) and first successful landing (Mars 3) on Mars. Probes have been active on Mars continuously since 1997. At times, more than ten probes have simultaneously operated in orbit or on the surface, more than at any other planet beyond Earth. Mars is an often proposed target for future crewed exploration missions, though no such mission is currently planned. Natural history Scientists have theorized that during the Solar System's formation, Mars was created as the result of a random process of run-away accretion of material from the protoplanetary disk that orbited the Sun. Mars has many distinctive chemical features caused by its position in the Solar System. Elements with comparatively low boiling points, such as chlorine, phosphorus, and sulfur, are much more common on Mars than on Earth; these elements were probably pushed outward by the young Sun's energetic solar wind. After the formation of the planets, the inner Solar System may have been subjected to the so-called Late Heavy Bombardment. About 60% of the surface of Mars shows a record of impacts from that era, whereas much of the remaining surface is probably underlain by immense impact basins caused by those events. However, more recent modeling has disputed the existence of the Late Heavy Bombardment. There is evidence of an enormous impact basin in the Northern Hemisphere of Mars, spanning 10,600 by 8,500 kilometres (6,600 by 5,300 mi), or roughly four times the size of the Moon's South Pole–Aitken basin, which would be the largest impact basin yet discovered if confirmed. It has been hypothesized that the basin was formed when Mars was struck by a Pluto-sized body about four billion years ago. The event, thought to be the cause of the Martian hemispheric dichotomy, created the smooth Borealis basin that covers 40% of the planet. A 2023 study shows evidence, based on the orbital inclination of Deimos (a small moon of Mars), that Mars may once have had a ring system 3.5 billion years to 4 billion years ago. This ring system may have been formed from a moon, 20 times more massive than Phobos, orbiting Mars billions of years ago; and Phobos would be a remnant of that ring. Epochs: The geological history of Mars can be split into many periods, but the following are the three primary periods: Geological activity is still taking place on Mars. The Athabasca Valles is home to sheet-like lava flows created about 200 million years ago. Water flows in the grabens called the Cerberus Fossae occurred less than 20 million years ago, indicating equally recent volcanic intrusions. The Mars Reconnaissance Orbiter has captured images of avalanches. Physical characteristics Mars is approximately half the diameter of Earth or twice that of the Moon, with a surface area only slightly less than the total area of Earth's dry land. Mars is less dense than Earth, having about 15% of Earth's volume and 11% of Earth's mass, resulting in about 38% of Earth's surface gravity. Mars is the only presently known example of a desert planet, a rocky planet with a surface akin to that of Earth's deserts. The red-orange appearance of the Martian surface is caused by iron(III) oxide (nanophase Fe2O3) and the iron(III) oxide-hydroxide mineral goethite. It can look like butterscotch; other common surface colors include golden, brown, tan, and greenish, depending on the minerals present. Like Earth, Mars is differentiated into a dense metallic core overlaid by less dense rocky layers. The outermost layer is the crust, which is on average about 42–56 kilometres (26–35 mi) thick, with a minimum thickness of 6 kilometres (3.7 mi) in Isidis Planitia, and a maximum thickness of 117 kilometres (73 mi) in the southern Tharsis plateau. For comparison, Earth's crust averages 27.3 ± 4.8 km in thickness. The most abundant elements in the Martian crust are silicon, oxygen, iron, magnesium, aluminum, calcium, and potassium. Mars is confirmed to be seismically active; in 2019, it was reported that InSight had detected and recorded over 450 marsquakes and related events. Beneath the crust is a silicate mantle responsible for many of the tectonic and volcanic features on the planet's surface. The upper Martian mantle is a low-velocity zone, where the velocity of seismic waves is lower than surrounding depth intervals. The mantle appears to be rigid down to the depth of about 250 km, giving Mars a very thick lithosphere compared to Earth. Below this the mantle gradually becomes more ductile, and the seismic wave velocity starts to grow again. The Martian mantle does not appear to have a thermally insulating layer analogous to Earth's lower mantle; instead, below 1050 km in depth, it becomes mineralogically similar to Earth's transition zone. At the bottom of the mantle lies a basal liquid silicate layer approximately 150–180 km thick. The Martian mantle appears to be highly heterogenous, with dense fragments up to 4 km across, likely injected deep into the planet by colossal impacts ~4.5 billion years ago; high-frequency waves from eight marsquakes slowed as they passed these localized regions, and modeling indicates the heterogeneities are compositionally distinct debris preserved because Mars lacks plate tectonics and has a sluggishly convecting interior that prevents complete homogenization. Mars's iron and nickel core is at least partially molten, and may have a solid inner core. It is around half of Mars's radius, approximately 1650–1675 km, and is enriched in light elements such as sulfur, oxygen, carbon, and hydrogen. The temperature of the core is estimated to be 2000–2400 K, compared to 5400–6230 K for Earth's solid inner core. In 2025, based on data from the InSight lander, a group of researchers reported the detection of a solid inner core 613 kilometres (381 mi) ± 67 kilometres (42 mi) in radius. Mars is a terrestrial planet with a surface that consists of minerals containing silicon and oxygen, metals, and other elements that typically make up rock. The Martian surface is primarily composed of tholeiitic basalt, although parts are more silica-rich than typical basalt and may be similar to andesitic rocks on Earth, or silica glass. Regions of low albedo suggest concentrations of plagioclase feldspar, with northern low albedo regions displaying higher than normal concentrations of sheet silicates and high-silicon glass. Parts of the southern highlands include detectable amounts of high-calcium pyroxenes. Localized concentrations of hematite and olivine have been found. Much of the surface is deeply covered by finely grained iron(III) oxide dust. The Phoenix lander returned data showing Martian soil to be slightly alkaline and containing elements such as magnesium, sodium, potassium and chlorine. These nutrients are found in soils on Earth, and are necessary for plant growth. Experiments performed by the lander showed that the Martian soil has a basic pH of 7.7, and contains 0.6% perchlorate by weight, concentrations that are toxic to humans. Streaks are common across Mars and new ones appear frequently on steep slopes of craters, troughs, and valleys. The streaks are dark at first and get lighter with age. The streaks can start in a tiny area, then spread out for hundreds of metres. They have been seen to follow the edges of boulders and other obstacles in their path. The commonly accepted hypotheses include that they are dark underlying layers of soil revealed after avalanches of bright dust or dust devils. Several other explanations have been put forward, including those that involve water or even the growth of organisms. Environmental radiation levels on the surface are on average 0.64 millisieverts of radiation per day, and significantly less than the radiation of 1.84 millisieverts per day or 22 millirads per day during the flight to and from Mars. For comparison the radiation levels in low Earth orbit, where Earth's space stations orbit, are around 0.5 millisieverts of radiation per day. Hellas Planitia has the lowest surface radiation at about 0.342 millisieverts per day, featuring lava tubes southwest of Hadriacus Mons with potentially levels as low as 0.064 millisieverts per day, comparable to radiation levels during flights on Earth. Although Mars has no evidence of a structured global magnetic field, observations show that parts of the planet's crust have been magnetized, suggesting that alternating polarity reversals of its dipole field have occurred in the past. This paleomagnetism of magnetically susceptible minerals is similar to the alternating bands found on Earth's ocean floors. One hypothesis, published in 1999 and re-examined in October 2005 (with the help of the Mars Global Surveyor), is that these bands suggest plate tectonic activity on Mars four billion years ago, before the planetary dynamo ceased to function and the planet's magnetic field faded. Geography and features Although better remembered for mapping the Moon, Johann Heinrich von Mädler and Wilhelm Beer were the first areographers. They began by establishing that most of Mars's surface features were permanent and by more precisely determining the planet's rotation period. In 1840, Mädler combined ten years of observations and drew the first map of Mars. Features on Mars are named from a variety of sources. Albedo features are named for classical mythology. Craters larger than roughly 50 km are named for deceased scientists and writers and others who have contributed to the study of Mars. Smaller craters are named for towns and villages of the world with populations of less than 100,000. Large valleys are named for the word "Mars" or "star" in various languages; smaller valleys are named for rivers. Large albedo features retain many of the older names but are often updated to reflect new knowledge of the nature of the features. For example, Nix Olympica (the snows of Olympus) has become Olympus Mons (Mount Olympus). The surface of Mars as seen from Earth is divided into two kinds of areas, with differing albedo. The paler plains covered with dust and sand rich in reddish iron oxides were once thought of as Martian "continents" and given names like Arabia Terra (land of Arabia) or Amazonis Planitia (Amazonian plain). The dark features were thought to be seas, hence their names Mare Erythraeum, Mare Sirenum and Aurorae Sinus. The largest dark feature seen from Earth is Syrtis Major Planum. The permanent northern polar ice cap is named Planum Boreum. The southern cap is called Planum Australe. Mars's equator is defined by its rotation, but the location of its Prime Meridian was specified, as was Earth's (at Greenwich), by choice of an arbitrary point; Mädler and Beer selected a line for their first maps of Mars in 1830. After the spacecraft Mariner 9 provided extensive imagery of Mars in 1972, a small crater (later called Airy-0), located in the Sinus Meridiani ("Middle Bay" or "Meridian Bay"), was chosen by Merton E. Davies, Harold Masursky, and Gérard de Vaucouleurs for the definition of 0.0° longitude to coincide with the original selection. Because Mars has no oceans, and hence no "sea level", a zero-elevation surface had to be selected as a reference level; this is called the areoid of Mars, analogous to the terrestrial geoid. Zero altitude was defined by the height at which there is 610.5 Pa (6.105 mbar) of atmospheric pressure. This pressure corresponds to the triple point of water, and it is about 0.6% of the sea level surface pressure on Earth (0.006 atm). For mapping purposes, the United States Geological Survey divides the surface of Mars into thirty cartographic quadrangles, each named for a classical albedo feature it contains. In April 2023, The New York Times reported an updated global map of Mars based on images from the Hope spacecraft. A related, but much more detailed, global Mars map was released by NASA on 16 April 2023. The vast upland region Tharsis contains several massive volcanoes, which include the shield volcano Olympus Mons. The edifice is over 600 km (370 mi) wide. Because the mountain is so large, with complex structure at its edges, giving a definite height to it is difficult. Its local relief, from the foot of the cliffs which form its northwest margin to its peak, is over 21 km (13 mi), a little over twice the height of Mauna Kea as measured from its base on the ocean floor. The total elevation change from the plains of Amazonis Planitia, over 1,000 km (620 mi) to the northwest, to the summit approaches 26 km (16 mi), roughly three times the height of Mount Everest, which in comparison stands at just over 8.8 kilometres (5.5 mi). Consequently, Olympus Mons is either the tallest or second-tallest mountain in the Solar System; the only known mountain which might be taller is the Rheasilvia peak on the asteroid Vesta, at 20–25 km (12–16 mi). The dichotomy of Martian topography is striking: northern plains flattened by lava flows contrast with the southern highlands, pitted and cratered by ancient impacts. It is possible that, four billion years ago, the Northern Hemisphere of Mars was struck by an object one-tenth to two-thirds the size of Earth's Moon. If this is the case, the Northern Hemisphere of Mars would be the site of an impact crater 10,600 by 8,500 kilometres (6,600 by 5,300 mi) in size, or roughly the area of Europe, Asia, and Australia combined, surpassing Utopia Planitia and the Moon's South Pole–Aitken basin as the largest impact crater in the Solar System. Mars is scarred by 43,000 impact craters with a diameter of 5 kilometres (3.1 mi) or greater. The largest exposed crater is Hellas, which is 2,300 kilometres (1,400 mi) wide and 7,000 metres (23,000 ft) deep, and is a light albedo feature clearly visible from Earth. There are other notable impact features, such as Argyre, which is around 1,800 kilometres (1,100 mi) in diameter, and Isidis, which is around 1,500 kilometres (930 mi) in diameter. Due to the smaller mass and size of Mars, the probability of an object colliding with the planet is about half that of Earth. Mars is located closer to the asteroid belt, so it has an increased chance of being struck by materials from that source. Mars is more likely to be struck by short-period comets, i.e., those that lie within the orbit of Jupiter. Martian craters can[discuss] have a morphology that suggests the ground became wet after the meteor impact. The large canyon, Valles Marineris (Latin for 'Mariner Valleys, also known as Agathodaemon in the old canal maps), has a length of 4,000 kilometres (2,500 mi) and a depth of up to 7 kilometres (4.3 mi). The length of Valles Marineris is equivalent to the length of Europe and extends across one-fifth the circumference of Mars. By comparison, the Grand Canyon on Earth is only 446 kilometres (277 mi) long and nearly 2 kilometres (1.2 mi) deep. Valles Marineris was formed due to the swelling of the Tharsis area, which caused the crust in the area of Valles Marineris to collapse. In 2012, it was proposed that Valles Marineris is not just a graben, but a plate boundary where 150 kilometres (93 mi) of transverse motion has occurred, making Mars a planet with possibly a two-tectonic plate arrangement. Images from the Thermal Emission Imaging System (THEMIS) aboard NASA's Mars Odyssey orbiter have revealed seven possible cave entrances on the flanks of the volcano Arsia Mons. The caves, named after loved ones of their discoverers, are collectively known as the "seven sisters". Cave entrances measure from 100 to 252 metres (328 to 827 ft) wide and they are estimated to be at least 73 to 96 metres (240 to 315 ft) deep. Because light does not reach the floor of most of the caves, they may extend much deeper than these lower estimates and widen below the surface. "Dena" is the only exception; its floor is visible and was measured to be 130 metres (430 ft) deep. The interiors of these caverns may be protected from micrometeoroids, UV radiation, solar flares and high energy particles that bombard the planet's surface. Martian geysers (or CO2 jets) are putative sites of small gas and dust eruptions that occur in the south polar region of Mars during the spring thaw. "Dark dune spots" and "spiders" – or araneiforms – are the two most visible types of features ascribed to these eruptions. Similarly sized dust will settle from the thinner Martian atmosphere sooner than it would on Earth. For example, the dust suspended by the 2001 global dust storms on Mars only remained in the Martian atmosphere for 0.6 years, while the dust from Mount Pinatubo took about two years to settle. However, under current Martian conditions, the mass movements involved are generally much smaller than on Earth. Even the 2001 global dust storms on Mars moved only the equivalent of a very thin dust layer – about 3 μm thick if deposited with uniform thickness between 58° north and south of the equator. Dust deposition at the two rover sites has proceeded at a rate of about the thickness of a grain every 100 sols. Atmosphere Mars lost its magnetosphere 4 billion years ago, possibly because of numerous asteroid strikes, so the solar wind interacts directly with the Martian ionosphere, lowering the atmospheric density by stripping away atoms from the outer layer. Both Mars Global Surveyor and Mars Express have detected ionized atmospheric particles trailing off into space behind Mars, and this atmospheric loss is being studied by the MAVEN orbiter. Compared to Earth, the atmosphere of Mars is quite rarefied. Atmospheric pressure on the surface today ranges from a low of 30 Pa (0.0044 psi) on Olympus Mons to over 1,155 Pa (0.1675 psi) in Hellas Planitia, with a mean pressure at the surface level of 600 Pa (0.087 psi). The highest atmospheric density on Mars is equal to that found 35 kilometres (22 mi) above Earth's surface. The resulting mean surface pressure is only 0.6% of Earth's 101.3 kPa (14.69 psi). The scale height of the atmosphere is about 10.8 kilometres (6.7 mi), which is higher than Earth's 6 kilometres (3.7 mi), because the surface gravity of Mars is only about 38% of Earth's. The atmosphere of Mars consists of about 96% carbon dioxide, 1.93% argon and 1.89% nitrogen along with traces of oxygen and water. The atmosphere is quite dusty, containing particulates about 1.5 μm in diameter which give the Martian sky a tawny color when seen from the surface. It may take on a pink hue due to iron oxide particles suspended in it. Despite repeated detections of methane on Mars, there is no scientific consensus as to its origin. One suggestion is that methane exists on Mars and that its concentration fluctuates seasonally. The existence of methane could be produced by non-biological process such as serpentinization involving water, carbon dioxide, and the mineral olivine, which is known to be common on Mars, or by Martian life. Compared to Earth, its higher concentration of atmospheric CO2 and lower surface pressure may be why sound is attenuated more on Mars, where natural sources are rare apart from the wind. Using acoustic recordings collected by the Perseverance rover, researchers concluded that the speed of sound there is approximately 240 m/s for frequencies below 240 Hz, and 250 m/s for those above. Auroras have been detected on Mars. Because Mars lacks a global magnetic field, the types and distribution of auroras there differ from those on Earth; rather than being mostly restricted to polar regions as is the case on Earth, a Martian aurora can encompass the planet. In September 2017, NASA reported radiation levels on the surface of the planet Mars were temporarily doubled, and were associated with an aurora 25 times brighter than any observed earlier, due to a massive, and unexpected, solar storm in the middle of the month. Mars has seasons, alternating between its northern and southern hemispheres, similar to on Earth. Additionally the orbit of Mars has, compared to Earth's, a large eccentricity and approaches perihelion when it is summer in its southern hemisphere and winter in its northern, and aphelion when it is winter in its southern hemisphere and summer in its northern. As a result, the seasons in its southern hemisphere are more extreme and the seasons in its northern are milder than would otherwise be the case. The summer temperatures in the south can be warmer than the equivalent summer temperatures in the north by up to 30 °C (54 °F). Martian surface temperatures vary from lows of about −110 °C (−166 °F) to highs of up to 35 °C (95 °F) in equatorial summer. The wide range in temperatures is due to the thin atmosphere which cannot store much solar heat, the low atmospheric pressure (about 1% that of the atmosphere of Earth), and the low thermal inertia of Martian soil. The planet is 1.52 times as far from the Sun as Earth, resulting in just 43% of the amount of sunlight. Mars has the largest dust storms in the Solar System, reaching speeds of over 160 km/h (100 mph). These can vary from a storm over a small area, to gigantic storms that cover the entire planet. They tend to occur when Mars is closest to the Sun, and have been shown to increase global temperature. Seasons also produce dry ice covering polar ice caps. Hydrology While Mars contains water in larger amounts, most of it is dust covered water ice at the Martian polar ice caps. The volume of water ice in the south polar ice cap, if melted, would be enough to cover most of the surface of the planet with a depth of 11 metres (36 ft). Water in its liquid form cannot persist on the surface due to Mars's low atmospheric pressure, which is less than 1% that of Earth. Only at the lowest of elevations are the pressure and temperature high enough for liquid water to exist for short periods. Although little water is present in the atmosphere, there is enough to produce clouds of water ice and different cases of snow and frost, often mixed with snow of carbon dioxide dry ice. Landforms visible on Mars strongly suggest that liquid water has existed on the planet's surface. Huge linear swathes of scoured ground, known as outflow channels, cut across the surface in about 25 places. These are thought to be a record of erosion caused by the catastrophic release of water from subsurface aquifers, though some of these structures have been hypothesized to result from the action of glaciers or lava. One of the larger examples, Ma'adim Vallis, is 700 kilometres (430 mi) long, much greater than the Grand Canyon, with a width of 20 kilometres (12 mi) and a depth of 2 kilometres (1.2 mi) in places. It is thought to have been carved by flowing water early in Mars's history. The youngest of these channels is thought to have formed only a few million years ago. Elsewhere, particularly on the oldest areas of the Martian surface, finer-scale, dendritic networks of valleys are spread across significant proportions of the landscape. Features of these valleys and their distribution strongly imply that they were carved by runoff resulting from precipitation in early Mars history. Subsurface water flow and groundwater sapping may play important subsidiary roles in some networks, but precipitation was probably the root cause of the incision in almost all cases. Along craters and canyon walls, there are thousands of features that appear similar to terrestrial gullies. The gullies tend to be in the highlands of the Southern Hemisphere and face the Equator; all are poleward of 30° latitude. A number of authors have suggested that their formation process involves liquid water, probably from melting ice, although others have argued for formation mechanisms involving carbon dioxide frost or the movement of dry dust. No partially degraded gullies have formed by weathering and no superimposed impact craters have been observed, indicating that these are young features, possibly still active. Other geological features, such as deltas and alluvial fans preserved in craters, are further evidence for warmer, wetter conditions at an interval or intervals in earlier Mars history. Such conditions necessarily require the widespread presence of crater lakes across a large proportion of the surface, for which there is independent mineralogical, sedimentological and geomorphological evidence. Further evidence that liquid water once existed on the surface of Mars comes from the detection of specific minerals such as hematite and goethite, both of which sometimes form in the presence of water. The chemical signature of water vapor on Mars was first unequivocally demonstrated in 1963 by spectroscopy using an Earth-based telescope. In 2004, Opportunity detected the mineral jarosite. This forms only in the presence of acidic water, showing that water once existed on Mars. The Spirit rover found concentrated deposits of silica in 2007 that indicated wet conditions in the past, and in December 2011, the mineral gypsum, which also forms in the presence of water, was found on the surface by NASA's Mars rover Opportunity. It is estimated that the amount of water in the upper mantle of Mars, represented by hydroxyl ions contained within Martian minerals, is equal to or greater than that of Earth at 50–300 parts per million of water, which is enough to cover the entire planet to a depth of 200–1,000 metres (660–3,280 ft). On 18 March 2013, NASA reported evidence from instruments on the Curiosity rover of mineral hydration, likely hydrated calcium sulfate, in several rock samples including the broken fragments of "Tintina" rock and "Sutton Inlier" rock as well as in veins and nodules in other rocks like "Knorr" rock and "Wernicke" rock. Analysis using the rover's DAN instrument provided evidence of subsurface water, amounting to as much as 4% water content, down to a depth of 60 centimetres (24 in), during the rover's traverse from the Bradbury Landing site to the Yellowknife Bay area in the Glenelg terrain. In September 2015, NASA announced that they had found strong evidence of hydrated brine flows in recurring slope lineae, based on spectrometer readings of the darkened areas of slopes. These streaks flow downhill in Martian summer, when the temperature is above −23 °C, and freeze at lower temperatures. These observations supported earlier hypotheses, based on timing of formation and their rate of growth, that these dark streaks resulted from water flowing just below the surface. However, later work suggested that the lineae may be dry, granular flows instead, with at most a limited role for water in initiating the process. A definitive conclusion about the presence, extent, and role of liquid water on the Martian surface remains elusive. Researchers suspect much of the low northern plains of the planet were covered with an ocean hundreds of meters deep, though this theory remains controversial. In March 2015, scientists stated that such an ocean might have been the size of Earth's Arctic Ocean. This finding was derived from the ratio of protium to deuterium in the modern Martian atmosphere compared to that ratio on Earth. The amount of Martian deuterium (D/H = 9.3 ± 1.7 10−4) is five to seven times the amount on Earth (D/H = 1.56 10−4), suggesting that ancient Mars had significantly higher levels of water. Results from the Curiosity rover had previously found a high ratio of deuterium in Gale Crater, though not significantly high enough to suggest the former presence of an ocean. Other scientists caution that these results have not been confirmed, and point out that Martian climate models have not yet shown that the planet was warm enough in the past to support bodies of liquid water. Near the northern polar cap is the 81.4 kilometres (50.6 mi) wide Korolev Crater, which the Mars Express orbiter found to be filled with approximately 2,200 cubic kilometres (530 cu mi) of water ice. In November 2016, NASA reported finding a large amount of underground ice in the Utopia Planitia region. The volume of water detected has been estimated to be equivalent to the volume of water in Lake Superior (which is 12,100 cubic kilometers). During observations from 2018 through 2021, the ExoMars Trace Gas Orbiter spotted indications of water, probably subsurface ice, in the Valles Marineris canyon system. Orbital motion Mars's average distance from the Sun is roughly 230 million km (143 million mi), and its orbital period is 687 (Earth) days. The solar day (or sol) on Mars is only slightly longer than an Earth day: 24 hours, 39 minutes, and 35.244 seconds. A Martian year is equal to 1.8809 Earth years, or 1 year, 320 days, and 18.2 hours. The gravitational potential difference and thus the delta-v needed to transfer between Mars and Earth is the second lowest for Earth. The axial tilt of Mars is 25.19° relative to its orbital plane, which is similar to the axial tilt of Earth. As a result, Mars has seasons like Earth, though on Mars they are nearly twice as long because its orbital period is that much longer. In the present day, the orientation of the north pole of Mars is close to the star Deneb. Mars has a relatively pronounced orbital eccentricity of about 0.09; of the seven other planets in the Solar System, only Mercury has a larger orbital eccentricity. It is known that in the past, Mars has had a much more circular orbit. At one point, 1.35 million Earth years ago, Mars had an eccentricity of roughly 0.002, much less than that of Earth today. Mars's cycle of eccentricity is 96,000 Earth years compared to Earth's cycle of 100,000 years. Mars has its closest approach to Earth (opposition) in a synodic period of 779.94 days. It should not be confused with Mars conjunction, where the Earth and Mars are at opposite sides of the Solar System and form a straight line crossing the Sun. The average time between the successive oppositions of Mars, its synodic period, is 780 days; but the number of days between successive oppositions can range from 764 to 812. The distance at close approach varies between about 54 and 103 million km (34 and 64 million mi) due to the planets' elliptical orbits, which causes comparable variation in angular size. At their furthest Mars and Earth can be as far as 401 million km (249 million mi) apart. Mars comes into opposition from Earth every 2.1 years. The planets come into opposition near Mars's perihelion in 2003, 2018 and 2035, with the 2020 and 2033 events being particularly close to perihelic opposition. The mean apparent magnitude of Mars is +0.71 with a standard deviation of 1.05. Because the orbit of Mars is eccentric, the magnitude at opposition from the Sun can range from about −3.0 to −1.4. The minimum brightness is magnitude +1.86 when the planet is near aphelion and in conjunction with the Sun. At its brightest, Mars (along with Jupiter) is second only to Venus in apparent brightness. Mars usually appears distinctly yellow, orange, or red. When farthest away from Earth, it is more than seven times farther away than when it is closest. Mars is usually close enough for particularly good viewing once or twice at 15-year or 17-year intervals. Optical ground-based telescopes are typically limited to resolving features about 300 kilometres (190 mi) across when Earth and Mars are closest because of Earth's atmosphere. As Mars approaches opposition, it begins a period of retrograde motion, which means it will appear to move backwards in a looping curve with respect to the background stars. This retrograde motion lasts for about 72 days, and Mars reaches its peak apparent brightness in the middle of this interval. Moons Mars has two relatively small (compared to Earth's) natural moons, Phobos (about 22 km (14 mi) in diameter) and Deimos (about 12 km (7.5 mi) in diameter), which orbit at 9,376 km (5,826 mi) and 23,460 km (14,580 mi) around the planet. The origin of both moons is unclear, although a popular theory states that they were asteroids captured into Martian orbit. Both satellites were discovered in 1877 by Asaph Hall and were named after the characters Phobos (the deity of panic and fear) and Deimos (the deity of terror and dread), twins from Greek mythology who accompanied their father Ares, god of war, into battle. Mars was the Roman equivalent to Ares. In modern Greek, the planet retains its ancient name Ares (Aris: Άρης). From the surface of Mars, the motions of Phobos and Deimos appear different from that of the Earth's satellite, the Moon. Phobos rises in the west, sets in the east, and rises again in just 11 hours. Deimos, being only just outside synchronous orbit – where the orbital period would match the planet's period of rotation – rises as expected in the east, but slowly. Because the orbit of Phobos is below a synchronous altitude, tidal forces from Mars are gradually lowering its orbit. In about 50 million years, it could either crash into Mars's surface or break up into a ring structure around the planet. The origin of the two satellites is not well understood. Their low albedo and carbonaceous chondrite composition have been regarded as similar to asteroids, supporting a capture theory. The unstable orbit of Phobos would seem to point toward a relatively recent capture. But both have circular orbits near the equator, which is unusual for captured objects, and the required capture dynamics are complex. Accretion early in the history of Mars is plausible, but would not account for a composition resembling asteroids rather than Mars itself, if that is confirmed. Mars may have yet-undiscovered moons, smaller than 50 to 100 metres (160 to 330 ft) in diameter, and a dust ring is predicted to exist between Phobos and Deimos. A third possibility for their origin as satellites of Mars is the involvement of a third body or a type of impact disruption. More-recent lines of evidence for Phobos having a highly porous interior, and suggesting a composition containing mainly phyllosilicates and other minerals known from Mars, point toward an origin of Phobos from material ejected by an impact on Mars that reaccreted in Martian orbit, similar to the prevailing theory for the origin of Earth's satellite. Although the visible and near-infrared (VNIR) spectra of the moons of Mars resemble those of outer-belt asteroids, the thermal infrared spectra of Phobos are reported to be inconsistent with chondrites of any class. It is also possible that Phobos and Deimos were fragments of an older moon, formed by debris from a large impact on Mars, and then destroyed by a more recent impact upon the satellite. More recently, a study conducted by a team of researchers from multiple countries suggests that a lost moon, at least fifteen times the size of Phobos, may have existed in the past. By analyzing rocks which point to tidal processes on the planet, it is possible that these tides may have been regulated by a past moon. Human observations and exploration The history of observations of Mars is marked by oppositions of Mars when the planet is closest to Earth and hence is most easily visible, which occur every couple of years. Even more notable are the perihelic oppositions of Mars, which are distinguished because Mars is close to perihelion, making it even closer to Earth. The ancient Sumerians named Mars Nergal, the god of war and plague. During Sumerian times, Nergal was a minor deity of little significance, but, during later times, his main cult center was the city of Nineveh. In Mesopotamian texts, Mars is referred to as the "star of judgement of the fate of the dead". The existence of Mars as a wandering object in the night sky was also recorded by the ancient Egyptian astronomers and, by 1534 BCE, they were familiar with the retrograde motion of the planet. By the period of the Neo-Babylonian Empire, the Babylonian astronomers were making regular records of the positions of the planets and systematic observations of their behavior. For Mars, they knew that the planet made 37 synodic periods, or 42 circuits of the zodiac, every 79 years. They invented arithmetic methods for making minor corrections to the predicted positions of the planets. In Ancient Greece, the planet was known as Πυρόεις. Commonly, the Greek name for the planet now referred to as Mars, was Ares. It was the Romans who named the planet Mars, for their god of war, often represented by the sword and shield of the planet's namesake. In the fourth century BCE, Aristotle noted that Mars disappeared behind the Moon during an occultation, indicating that the planet was farther away. Ptolemy, a Greek living in Alexandria, attempted to address the problem of the orbital motion of Mars. Ptolemy's model and his collective work on astronomy was presented in the multi-volume collection later called the Almagest (from the Arabic for "greatest"), which became the authoritative treatise on Western astronomy for the next fourteen centuries. Literature from ancient China confirms that Mars was known by Chinese astronomers by no later than the fourth century BCE. In the East Asian cultures, Mars is traditionally referred to as the "fire star" (火星) based on the Wuxing system. In 1609 Johannes Kepler published a 10 year study of Martian orbit, using the diurnal parallax of Mars, measured by Tycho Brahe, to make a preliminary calculation of the relative distance to the planet. From Brahe's observations of Mars, Kepler deduced that the planet orbited the Sun not in a circle, but in an ellipse. Moreover, Kepler showed that Mars sped up as it approached the Sun and slowed down as it moved farther away, in a manner that later physicists would explain as a consequence of the conservation of angular momentum.: 433–437 In 1610 the first use of a telescope for astronomical observation, including Mars, was performed by Italian astronomer Galileo Galilei. With the telescope the diurnal parallax of Mars was again measured in an effort to determine the Sun-Earth distance. This was first performed by Giovanni Domenico Cassini in 1672. The early parallax measurements were hampered by the quality of the instruments. The only occultation of Mars by Venus observed was that of 13 October 1590, seen by Michael Maestlin at Heidelberg. By the 19th century, the resolution of telescopes reached a level sufficient for surface features to be identified. On 5 September 1877, a perihelic opposition to Mars occurred. The Italian astronomer Giovanni Schiaparelli used a 22-centimetre (8.7 in) telescope in Milan to help produce the first detailed map of Mars. These maps notably contained features he called canali, which, with the possible exception of the natural canyon Valles Marineris, were later shown to be an optical illusion. These canali were supposedly long, straight lines on the surface of Mars, to which he gave names of famous rivers on Earth. His term, which means "channels" or "grooves", was popularly mistranslated in English as "canals". Influenced by the observations, the orientalist Percival Lowell founded an observatory which had 30- and 45-centimetre (12- and 18-in) telescopes. The observatory was used for the exploration of Mars during the last good opportunity in 1894, and the following less favorable oppositions. He published several books on Mars and life on the planet, which had a great influence on the public. The canali were independently observed by other astronomers, like Henri Joseph Perrotin and Louis Thollon in Nice, using one of the largest telescopes of that time. The seasonal changes (consisting of the diminishing of the polar caps and the dark areas formed during Martian summers) in combination with the canals led to speculation about life on Mars, and it was a long-held belief that Mars contained vast seas and vegetation. As bigger telescopes were used, fewer long, straight canali were observed. During observations in 1909 by Antoniadi with an 84-centimetre (33 in) telescope, irregular patterns were observed, but no canali were seen. The first spacecraft from Earth to visit Mars was Mars 1 of the Soviet Union, which flew by in 1963, but contact was lost en route. NASA's Mariner 4 followed and became the first spacecraft to successfully transmit from Mars; launched on 28 November 1964, it made its closest approach to the planet on 15 July 1965. Mariner 4 detected the weak Martian radiation belt, measured at about 0.1% that of Earth, and captured the first images of another planet from deep space. Once spacecraft visited the planet during the 1960s and 1970s, many previous concepts of Mars were radically broken. After the results of the Viking life-detection experiments, the hypothesis of a dead planet was generally accepted. The data from Mariner 9 and Viking allowed better maps of Mars to be made. Until 1997 and after Viking 1 shut down in 1982, Mars was only visited by three unsuccessful probes, two flying past without contact (Phobos 1, 1988; Mars Observer, 1993), and one (Phobos 2 1989) malfunctioning in orbit before reaching its destination Phobos. In 1997 Mars Pathfinder became the first successful rover mission beyond the Moon and started together with Mars Global Surveyor (operated until late 2006) an uninterrupted active robotic presence at Mars that has lasted until today. It produced complete, extremely detailed maps of the Martian topography, magnetic field and surface minerals. Starting with these missions a range of new improved crewless spacecraft, including orbiters, landers, and rovers, have been sent to Mars, with successful missions by the NASA (United States), Jaxa (Japan), ESA, United Kingdom, ISRO (India), Roscosmos (Russia), the United Arab Emirates, and CNSA (China) to study the planet's surface, climate, and geology, uncovering the different elements of the history and dynamic of the hydrosphere of Mars and possible traces of ancient life. As of 2023[update], Mars is host to ten functioning spacecraft. Eight are in orbit: 2001 Mars Odyssey, Mars Express, Mars Reconnaissance Orbiter, MAVEN, ExoMars Trace Gas Orbiter, the Hope orbiter, and the Tianwen-1 orbiter. Another two are on the surface: the Mars Science Laboratory Curiosity rover and the Perseverance rover. Collected maps are available online at websites including Google Mars. NASA provides two online tools: Mars Trek, which provides visualizations of the planet using data from 50 years of exploration, and Experience Curiosity, which simulates traveling on Mars in 3-D with Curiosity. Planned missions to Mars include: As of February 2024[update], debris from these types of missions has reached over seven tons. Most of it consists of crashed and inactive spacecraft as well as discarded components. In April 2024, NASA selected several companies to begin studies on providing commercial services to further enable robotic science on Mars. Key areas include establishing telecommunications, payload delivery and surface imaging. Habitability and habitation During the late 19th century, it was widely accepted in the astronomical community that Mars had life-supporting qualities, including the presence of oxygen and water. However, in 1894 W. W. Campbell at Lick Observatory observed the planet and found that "if water vapor or oxygen occur in the atmosphere of Mars it is in quantities too small to be detected by spectroscopes then available". That observation contradicted many of the measurements of the time and was not widely accepted. Campbell and V. M. Slipher repeated the study in 1909 using better instruments, but with the same results. It was not until the findings were confirmed by W. S. Adams in 1925 that the myth of the Earth-like habitability of Mars was finally broken. However, even in the 1960s, articles were published on Martian biology, putting aside explanations other than life for the seasonal changes on Mars. The current understanding of planetary habitability – the ability of a world to develop environmental conditions favorable to the emergence of life – favors planets that have liquid water on their surface. Most often this requires the orbit of a planet to lie within the habitable zone, which for the Sun is estimated to extend from within the orbit of Earth to about that of Mars. During perihelion, Mars dips inside this region, but Mars's thin (low-pressure) atmosphere prevents liquid water from existing over large regions for extended periods. The past flow of liquid water demonstrates the planet's potential for habitability. Recent evidence has suggested that any water on the Martian surface may have been too salty and acidic to support regular terrestrial life. The environmental conditions on Mars are a challenge to sustaining organic life: the planet has little heat transfer across its surface, it has poor insulation against bombardment by the solar wind due to the absence of a magnetosphere and has insufficient atmospheric pressure to retain water in a liquid form (water instead sublimes to a gaseous state). Mars is nearly, or perhaps totally, geologically dead; the end of volcanic activity has apparently stopped the recycling of chemicals and minerals between the surface and interior of the planet. Evidence suggests that the planet was once significantly more habitable than it is today, but whether living organisms ever existed there remains unknown. The Viking probes of the mid-1970s carried experiments designed to detect microorganisms in Martian soil at their respective landing sites and had positive results, including a temporary increase in CO2 production on exposure to water and nutrients. This sign of life was later disputed by scientists, resulting in a continuing debate, with NASA scientist Gilbert Levin asserting that Viking may have found life. A 2014 analysis of Martian meteorite EETA79001 found chlorate, perchlorate, and nitrate ions in sufficiently high concentrations to suggest that they are widespread on Mars. UV and X-ray radiation would turn chlorate and perchlorate ions into other, highly reactive oxychlorines, indicating that any organic molecules would have to be buried under the surface to survive. Small quantities of methane and formaldehyde detected by Mars orbiters are both claimed to be possible evidence for life, as these chemical compounds would quickly break down in the Martian atmosphere. Alternatively, these compounds may instead be replenished by volcanic or other geological means, such as serpentinite. Impact glass, formed by the impact of meteors, which on Earth can preserve signs of life, has also been found on the surface of the impact craters on Mars. Likewise, the glass in impact craters on Mars could have preserved signs of life, if life existed at the site. The Cheyava Falls rock discovered on Mars in June 2024 has been designated by NASA as a "potential biosignature" and was core sampled by the Perseverance rover for possible return to Earth and further examination. Although highly intriguing, no definitive final determination on a biological or abiotic origin of this rock can be made with the data currently available. Several plans for a human mission to Mars have been proposed, but none have come to fruition. The NASA Authorization Act of 2017 directed NASA to study the feasibility of a crewed Mars mission in the early 2030s; the resulting report concluded that this would be unfeasible. In addition, in 2021, China was planning to send a crewed Mars mission in 2033. Privately held companies such as SpaceX have also proposed plans to send humans to Mars, with the eventual goal to settle on the planet. As of 2024, SpaceX has proceeded with the development of the Starship launch vehicle with the goal of Mars colonization. In plans shared with the company in April 2024, Elon Musk envisions the beginning of a Mars colony within the next twenty years. This would be enabled by the planned mass manufacturing of Starship and initially sustained by resupply from Earth, and in situ resource utilization on Mars, until the Mars colony reaches full self sustainability. Any future human mission to Mars will likely take place within the optimal Mars launch window, which occurs every 26 months. The moon Phobos has been proposed as an anchor point for a space elevator. Besides national space agencies and space companies, groups such as the Mars Society and The Planetary Society advocate for human missions to Mars. In culture Mars is named after the Roman god of war (Greek Ares), but was also associated with the demi-god Heracles (Roman Hercules) by ancient Greek astronomers, as detailed by Aristotle. This association between Mars and war dates back at least to Babylonian astronomy, in which the planet was named for the god Nergal, deity of war and destruction. It persisted into modern times, as exemplified by Gustav Holst's orchestral suite The Planets, whose famous first movement labels Mars "The Bringer of War". The planet's symbol, a circle with a spear pointing out to the upper right, is also used as a symbol for the male gender. The symbol dates from at least the 11th century, though a possible predecessor has been found in the Greek Oxyrhynchus Papyri. The idea that Mars was populated by intelligent Martians became widespread in the late 19th century. Schiaparelli's "canali" observations combined with Percival Lowell's books on the subject put forward the standard notion of a planet that was a drying, cooling, dying world with ancient civilizations constructing irrigation works. Many other observations and proclamations by notable personalities added to what has been termed "Mars Fever". In the present day, high-resolution mapping of the surface of Mars has revealed no artifacts of habitation, but pseudoscientific speculation about intelligent life on Mars still continues. Reminiscent of the canali observations, these speculations are based on small scale features perceived in the spacecraft images, such as "pyramids" and the "Face on Mars". In his book Cosmos, planetary astronomer Carl Sagan wrote: "Mars has become a kind of mythic arena onto which we have projected our Earthly hopes and fears." The depiction of Mars in fiction has been stimulated by its dramatic red color and by nineteenth-century scientific speculations that its surface conditions might support not just life but intelligent life. This gave way to many science fiction stories involving these concepts, such as H. G. Wells's The War of the Worlds, in which Martians seek to escape their dying planet by invading Earth; Ray Bradbury's The Martian Chronicles, in which human explorers accidentally destroy a Martian civilization; as well as Edgar Rice Burroughs's series Barsoom, C. S. Lewis's novel Out of the Silent Planet (1938), and a number of Robert A. Heinlein stories before the mid-sixties. Since then, depictions of Martians have also extended to animation. A comic figure of an intelligent Martian, Marvin the Martian, appeared in Haredevil Hare (1948) as a character in the Looney Tunes animated cartoons of Warner Brothers, and has continued as part of popular culture to the present. After the Mariner and Viking spacecraft had returned pictures of Mars as a lifeless and canal-less world, these ideas about Mars were abandoned; for many science-fiction authors, the new discoveries initially seemed like a constraint, but eventually the post-Viking knowledge of Mars became itself a source of inspiration for works like Kim Stanley Robinson's Mars trilogy. See also Notes References Further reading External links Solar System → Local Interstellar Cloud → Local Bubble → Gould Belt → Orion Arm → Milky Way → Milky Way subgroup → Local Group → Local Sheet → Local Volume → Virgo Supercluster → Laniakea Supercluster → Pisces–Cetus Supercluster Complex → Local Hole → Observable universe → UniverseEach arrow (→) may be read as "within" or "part of". |
======================================== |
[SOURCE: https://en.wikipedia.org/wiki/Break_statement] | [TOKENS: 4979] |
Contents Control flow In software, control flow (or flow of control) describes how execution progresses from one command to the next. In many contexts, such as machine code and an imperative programming language, control progresses sequentially (to the command located immediately after the currently executing command) except when a command transfers control to another point – in which case the command is classified as a control flow command. Depending on context, other terms are used instead of command. For example, in machine code, the typical term is instruction and in an imperative language, the typical term is statement. Although an imperative language encodes control flow explicitly, languages of other programming paradigms are less focused on control flow. A declarative language specifies desired results without prescribing an order of operations. A functional language uses both language constructs and functions to control flow even though they are usually not called control flow statements. For a central processing unit instruction set, a control flow instruction often alters the program counter and is either an unconditional branch (a.k.a. jump) or a conditional branch. An alternative approach is predication which conditionally enables instructions instead of branching. An asynchronous control flow transfer such as an interrupt or a signal alters the normal flow of control to a handler before returning control to where it was interrupted. One way to attack software is to redirect the flow of execution. A variety of control-flow integrity techniques, including stack canaries, buffer overflow protection, shadow stacks, and vtable pointer verification, are used to defend against these attacks. Structure Control flow is closely related to code structure. Control flows along lines defined by structure and the execution rules of a language. This general concept of structure is not be confused with structured programming which limits structure to sequencing, selection and iteration based on block organization. Sequential execution is the most basic structure. Although not all code is sequential in nature, imperative code is. A label identifies a position in source code. Some control flow statements reference a label so that control jumps to the labeled line. Other than marking a position, a label has no other effect. Some languages limit a label to a number which is sometimes called a line number, although that implies the inherent index of the line, not a label. None-the-less, such numeric labels are typically required to increment from top to bottom in a file even if not be sequential. For example, in BASIC: In many languages, a label is an alphanumeric identifier, usually appearing at the start of a line and immediately followed by a colon. For example, the following C code defines a label Success on line 3 which identifies a jump target point at the first statement that follows it (line 4). Most languages provide for organizing sequences of code as a block. When used with a control statement, the beginning of a block provides a jump target. For example, in the following C code (which uses curly braces to delimit a block), control jumps from line 1 to 4 if done is false. Control Many control commands have been devised for programming languages. This section describes notable constructs; it is organized by functionality. A function provides for control flow in that when called, execution jumps to the start of the function's code and when it completes, control returns the calling point. In the following C code, control jumps from line 6 to 2 in order to call function foo(). Then, after completing the function body (printing "Hi"), control returns to after the call, line 7. A branch command moves the point of execution from the point in the code that contains the command to the point that the command specifies. A jump command unconditionally branches control to another point in the code, and is the most basic form of controlling the flow of code. In a high-level language, this is often provided as a goto statement. Although the keyword may be upper or lower case or one or two words depending on the language, it is like: goto label. When control reaches a goto statement, control then jumps to the statement that follows the indicated label. The goto statement has been considered harmful by many computer scientists, including Dijkstra. A conditional statement jumps control based on the value of a Boolean expression. Common variations include: The following Pascal code shows a simple if-then-else. The syntax is similar in Ada: In C: In bash: In Python: In Lisp: A multiway branch jumps control based on matching values. There is usually a provision for a default action if no match is found. A switch statement can allow compiler optimizations, such as lookup tables. In dynamic languages, the cases may not be limited to constant expressions, and might extend to pattern matching, as in the shell script example on the right, where the *) implements the default case as a glob matching any string. Case logic can also be implemented in functional form, as in SQL's decode statement. The following Pascal code shows a relatively simple switch statement. Pascal uses the case keyword instead of switch. In Ada: In C: In Bash: In Lisp: In Fortran: A loop is a sequence of statements, loop body, which is executed a number of times based on runtime state. The body is executed once for each item of a collection (definite iteration), until a condition is met (indefinite iteration), or infinitely. A loop inside the loop body is called a nested loop. Early exit from a loop may be supported via a break statement. In a functional programming language, such as Haskell and Scheme, both recursive and iterative processes are expressed with tail recursive procedures instead of looping constructs that are syntactic. A relatively simple yet useful loop iterates over a range of numeric values. A simple form starts at an integer value, ends at a larger integer value and iterates for each integer value between. Often, the increment can be any integer value (even negative, to loop from a larger to a smaller value). Example in BASIC: Example in Pascal: Example in Fortran: In many programming languages, only integers can be used at all or reliably. As a floating-point number is represented imprecisely due to hardware constraints, the following loop might iterate 9 or 10 times, depending on various factors such as rounding error, hardware, compiler. Furthermore, if the increment of X occurs by repeated addition, accumulated rounding errors may mean that the value of X in each iteration can differ quite significantly from the commonly expected sequence of 0.1, 0.2, 0.3, ..., 1.0. Some loop constructs iterate until a condition is true. Some variations test the condition at the start of the loop; others test at the end. If the test is at the start, the body may be skipped completely. At the end, the body is always executed at least once. Example in Visual Basic: Example in Pascal: Example in C family of pre-test: Example in C family of post-test: Although using the for keyword, the three-part C-style loop is a condition-based construct, not a numeric-based one. The second part, the condition, is evaluated before each loop, so the loop is pre-test. The first part is a place to initialize state, and the third part is for incrementing for the next iteration, but both aspects can be performed elsewhere. The following C code implements the logic of a numeric loop that iterates for i from 0 to n-1. Some loop constructs enumerate the items of a collection; iterating for each item. Example in Smalltalk: Example in Pascal: Example in Raku: Example in TCL: Example in PHP: Example in Java: Example in C#: Example in PowerShell where 'foreach' is an alias of 'ForEach-Object': Example in Fortran: Scala has for-expressions, which generalise collection-controlled loops, and also support other uses, such as asynchronous programming. Haskell has do-expressions and comprehensions, which together provide similar function to for-expressions in Scala. In computer programming, an infinite loop (or endless loop) is a sequence of instructions that, as written, will continue endlessly, unless an external intervention occurs, such as turning off power via a switch or pulling a plug. It may be intentional. There is no general algorithm to determine whether a computer program contains an infinite loop or not; this is the halting problem. Common loop structures sometimes result in duplicated code, either repeated statements or repeated conditions. This arises for various reasons and has various proposed solutions to eliminate or minimize code duplication. Other than the traditional unstructured solution of a goto statement, general structured solutions include having a conditional (if statement) inside the loop (possibly duplicating the condition but not the statements) or wrapping repeated logic in a function (so there is a duplicated function call, but the statements are not duplicated). A common case is where the start of the loop is always executed, but the end may be skipped on the last iteration. This was dubbed by Dijkstra a loop which is performed "n and a half times", and is now called the loop-and-a-half problem. Common cases include reading data in the first part, checking for end of data, and then processing the data in the second part; or processing, checking for end, and then preparing for the next iteration. In these cases, the first part of the loop is executed n {\displaystyle n} times, but the second part is only executed n − 1 {\displaystyle n-1} times. This problem has been recognized at least since 1967 by Knuth, with Wirth suggesting solving it via early loop exit. Since the 1990s this has been the most commonly taught solution, using a break statement, as in: A subtlety of this solution is that the condition is the opposite of a usual while condition: rewriting while condition ... repeat with an exit in the middle requires reversing the condition: loop ... if not condition exit ... repeat. The loop with test in the middle control structure explicitly supports the loop-an-a-half use case, without reversing the condition. A loop construct provides for structured completion criteria that either results in another iteration or continuing execution after the loop statement. But, various unstructured control flow constructs are supported by many languages. Early exit jumps control to after the loop body For example, when searching a list, can stop looping when the item is found. Some programming languages provide a statement such as break (most languages), Exit (Visual Basic), or last (Perl). In the following Ada code, the loop exits when X is 0. A more idiomatic style uses exit when: Python supports conditional execution of code depending on whether a loop was exited early (with a break statement) or not by using an else-clause with the loop. In the following Python code, the else clause is linked to the for statement, and not the inner if statement. Both Python's for and while loops support such an else clause, which is executed only if early exit of the loop has not occurred. Some languages support breaking out of nested loops; in theory circles, these are called multi-level breaks. One common use example is searching a multi-dimensional table. This can be done either via multilevel breaks (break out of N levels), as in bash and PHP, or via labeled breaks (break out and continue at given label), as in Ada, Go, Java, Rust and Perl. Alternatives to multilevel breaks include single breaks, together with a state variable which is tested to break out another level; exceptions, which are caught at the level being broken out to; placing the nested loops in a function and using return to effect termination of the entire nested loop; or using a label and a goto statement. Neither C nor C++ currently have multilevel break or named loops, and the usual alternative is to use a goto to implement a labeled break. However, the inclusion of this feature has been proposed, and was added to C2Y., following the Java syntax. Python does not have a multilevel break or continue – this was proposed in PEP 3136, and rejected on the basis that the added complexity was not worth the rare legitimate use. The notion of multi-level breaks is of some interest in theoretical computer science, because it gives rise to what is today called the Kosaraju hierarchy. In 1973 S. Rao Kosaraju refined the structured program theorem by proving that it is possible to avoid adding additional variables in structured programming, as long as arbitrary-depth, multi-level breaks from loops are allowed. Furthermore, Kosaraju proved that a strict hierarchy of programs exists: for every integer n, there exists a program containing a multi-level break of depth n that cannot be rewritten as a program with multi-level breaks of depth less than n without introducing added variables. In his 2004 textbook, David Watt uses Tennent's notion of sequencer to explain the similarity between multi-level breaks and return statements. Watt notes that a class of sequencers known as escape sequencers, defined as "sequencer that terminates execution of a textually enclosing command or procedure", encompasses both breaks from loops (including multi-level breaks) and return statements. As commonly implemented, however, return sequencers may also carry a (return) value, whereas the break sequencer as implemented in contemporary languages usually cannot. The following structure was proposed by Dahl in 1972: The construction here can be thought of as a do loop with the while check in the middle, which allows clear loop-and-a-half logic. Further, by omitting individual components, this single construction can replace several constructions in most programming languages. If xxx1 is omitted, we get a loop with the test at the top (a traditional while loop). If xxx2 is omitted, we get a loop with the test at the bottom, equivalent to a do while loop in many languages. If while is omitted, we get an infinite loop. This construction also allows keeping the same polarity of the condition even when in the middle, unlike early exit, which requires reversing the polarity (adding a not), functioning as until instead of while. This structure is not widely supported, with most languages instead using if ... break for conditional early exit. This is supported by some languages, such as Forth, where the syntax is BEGIN ... WHILE ... REPEAT, and the shell script languages Bourne shell (sh) and bash, where the syntax is while ... do ... done or until ... do ... done, as: The shell syntax works because the while (or until) loop accepts a list of commands as a condition, formally: The value (exit status) of the list of test-commands is the value of the last command, and these can be separated by newlines, resulting in the idiomatic form above. Similar constructions are possible in C and C++ with the comma operator, and other languages with similar constructs, which allow shoehorning a list of statements into the while condition: While legal, this is marginal, and it is primarily used, if at all, only for short modify-then-test cases, as in: Loop variants and loop invariants are used to express correctness of loops. In practical terms, a loop variant is an integer expression which has an initial non-negative value. The variant's value must decrease during each loop iteration but must never become negative during the correct execution of the loop. Loop variants are used to guarantee that loops will terminate. A loop invariant is an assertion which must be true before the first loop iteration and remain true after each iteration. This implies that when a loop terminates correctly, both the exit condition and the loop invariant are satisfied. Loop invariants are used to monitor specific properties of a loop during successive iterations. Some programming languages, such as Eiffel contain native support for loop variants and invariants. In other cases, support is an add-on, such as the Java Modeling Language's specification for loop statements in Java. Some Lisp dialects provide an extensive sublanguage for describing Loops. An early example can be found in Conversional Lisp of Interlisp. Common Lisp provides a Loop macro which implements such a sublanguage. Many programming languages, especially those favoring more dynamic styles of programming, offer constructs for non-local control flow which cause execution to jump from the current execution point to a predeclared point. Notable examples follow. The earliest Fortran compilers supported statements for handling exceptional conditions including IF ACCUMULATOR OVERFLOW, IF QUOTIENT OVERFLOW, and IF DIVIDE CHECK. In the interest of machine independence, they were not included in FORTRAN IV and the Fortran 66 Standard. However, since Fortran 2003 it is possible to test for numerical issues via calls to functions in the IEEE_EXCEPTIONS module. PL/I has some 22 standard conditions (e.g., ZERODIVIDE SUBSCRIPTRANGE ENDFILE) which can be raised and which can be intercepted by: ON condition action; Programmers can also define and use their own named conditions. Like the unstructured if, only one statement can be specified so in many cases a GOTO is needed to decide where flow of control should resume. Unfortunately, some implementations had a substantial overhead in both space and time (especially SUBSCRIPTRANGE), so many programmers tried to avoid using conditions. A typical example of syntax: Many modern languages natively support exception handling. Generally, exceptional control flow starts with an exception object being thrown (a.k.a. raised). Control then proceeds to the inner-most exception handler for the call stack. If the handler handles the exception, then flow control reverts to normal. Otherwise, control proceeds outward to containing handlers until one handles the exception or the program reaches the outermost scope and exits. As control flows to progressively outer handlers, aspects that would normally occur such as popping the call stack are handled automatically. The following C++ code demonstrates structured exception handling. If an exception propagates from the execution of doSomething() and the exception object type matches one of the types specified in a catch clause, then that clause is executed. For example, if an exception of type SomeException is propagated by doSomething(), then control jumps from line 2 to 4 and the message "Caught SomeException" is printed and then control jumps to after the try statement, line 8. If an exception of any other type is propagated, then control jumps from line 2 to 6. If no exception, then control jumps from 2 to 8. Many languages use the C++ keywords (throw, try and catch), but some languages use other keywords. For example, Ada uses exception to introduce an exception handler and when instead of catch. AppleScript incorporates placeholders in the syntax to extract information about the exception as shown in the following AppleScript code. In many languages (including Object Pascal, D, Java, C#, and Python) a finally clause at the end of a try statement is executed at the end of the try statement, whether an exception propagates from the rest of the try or not. The following C# code ensures that the stream stream is closed. Since this pattern is common, C# provides the using statement to ensure cleanup. In the following code, even if ProcessStuff() propagates an exception, the stream object is released. Python's with statement and Ruby's block argument to File.open are used to similar effect. In computer science, a continuation is an abstract representation of the control state of a computer program. A continuation implements (reifies) the program control state, i.e. the continuation is a data structure that represents the computational process at a given point in the process's execution; the created data structure can be accessed by the programming language, instead of being hidden in the runtime environment. Continuations are useful for encoding other control mechanisms in programming languages such as exceptions, generators, coroutines, and so on. The "current continuation" or "continuation of the computation step" is the continuation that, from the perspective of running code, would be derived from the current point in a program's execution. The term continuations can also be used to refer to first-class continuations, which are constructs that give a programming language the ability to save the execution state at any point and return to that point at a later point in the program, possibly multiple times. In computer science, a generator is a routine that can be used to control the iteration behaviour of a loop. All generators are also iterators. A generator is very similar to a function that returns an array, in that a generator has parameters, can be called, and generates a sequence of values. However, instead of building an array containing all the values and returning them all at once, a generator yields the values one at a time, which requires less memory and allows the caller to get started processing the first few values immediately. In short, a generator looks like a function but behaves like an iterator. Generators can be implemented in terms of more expressive control flow constructs, such as coroutines or first-class continuations. Generators, also known as semicoroutines, are a special case of (and weaker than) coroutines, in that they always yield control back to the caller (when passing a value back), rather than specifying a coroutine to jump to; see comparison of coroutines with generators. Coroutines are computer program components that allow execution to be suspended and resumed, generalizing subroutines for cooperative multitasking. Coroutines are well-suited for implementing familiar program components such as cooperative tasks, exceptions, event loops, iterators, infinite lists and pipes. They have been described as "functions whose execution you can pause". Melvin Conway coined the term coroutine in 1958 when he applied it to the construction of an assembly program. The first published explanation of the coroutine appeared later, in 1963. In computer programming, COMEFROM is a control flow statement that causes control flow to jump to the statement after it when control reaches the point specified by the COMEFROM argument. The statement is intended to be the opposite of goto and is considered to be more a joke than serious computer science. Often the specified jump point is identified as a label. For example, COMEFROM x specifies that when control reaches the label x, then control continues at the statement after the COMEFROM. A major difference with goto is that goto depends on the local structure of the code, while COMEFROM depends on the global structure. A goto statement transfers control when control reaches the statement, but COMEFROM requires the processor (i.e. interpreter) to scan for COMEFROM statements so that when control reaches any of the specified points, the processor can make the jump. The resulting logic tends to be difficult to understand since there is no indication near a jump point that control will in fact jump. One must study the entire program to see if any COMEFROM statements reference that point. The semantics of a COMEFROM statement varies by programming language. In some languages, the jump occurs before the statement at the specified point is executed and in others the jump occurs after. Depending on the language, multiple COMEFROM statements that reference the same point may be invalid, non-deterministic, executed in some order, or induce parallel or otherwise concurrent processing as seen in Threaded Intercal.[citation needed] COMEFROM was initially seen in lists of joke assembly language instructions (as 'CMFRM'). It was elaborated upon in a Datamation article by R. Lawrence Clark in 1973, written in response to Edsger Dijkstra's letter Go To Statement Considered Harmful. COMEFROM was eventually implemented in the C-INTERCAL variant of the esoteric programming language INTERCAL along with the even more obscure 'computed COMEFROM'. There were also Fortran proposals for 'assigned COME FROM' and a 'DONT' statement (to complement the existing 'DO' loop). Zahn's construct was proposed in 1974, and discussed in Knuth (1974). A modified version is presented here. exitwhen is used to specify the events which may occur within xxx, their occurrence is indicated by using the name of the event as a statement. When some event does occur, the relevant action is carried out, and then control passes just after endexit. This construction provides a very clear separation between determining that some situation applies, and the action to be taken for that situation. exitwhen is conceptually similar to exception handling, and exceptions or similar constructs are used for this purpose in many languages. The following simple example involves searching a two-dimensional table for a particular item. See also References External links |
======================================== |
[SOURCE: https://en.wikipedia.org/wiki/NOP_(code)] | [TOKENS: 2189] |
Contents NOP (code) In computer science, a NOP, no-op, or NOOP (pronounced "no op"; short for no operation) is a machine language instruction and its assembly language mnemonic, programming language statement, or computer protocol command that does nothing. Machine language instructions Some computer instruction sets include an instruction whose purpose is to not change the state of any of the programmer-accessible registers, status flags, or memory. It often takes a well-defined number of clock cycles to execute. In other instruction sets, there is no explicit NOP instruction, but the assembly language mnemonic NOP represents an instruction which acts as a NOP; e.g., on the SPARC, sethi 0, %g0. A NOP must not access memory, as that could cause a memory fault or page fault. A NOP is most commonly used for timing purposes, to force memory alignment, to prevent hazards, to occupy a branch delay slot, to render void an existing instruction such as a jump, as a target of an execute instruction, or as a place-holder to be replaced by active instructions later on in program development (or to replace removed instructions when reorganizing would be problematic or time-consuming). In some cases, a NOP can have minor side effects; for example, on the Motorola 68000 series of processors, the NOP opcode causes a synchronization of the pipeline. Listed below are the NOP instruction for some CPU architectures: 0x0F 0x1F 0x00 0x0F 0x1F 0x40 0x00 0x0F 0x1F 0x44 0x00 0x00 0x66 0x0F 0x1F 0x44 0x00 0x00 0x0F 0x1F 0x80 0x00 0x00 0x00 0x00 0x0F 0x1F 0x84 0x00 0x00 0x00 0x00 0x00 0x66 0x0F 0x1F 0x84 0x00 0x00 0x00 0x00 0x00 0x0F 0x1F is a two-byte NOP opcode that takes a ModR/M operand upon which no memory is accessed and no registers are written. ModR/M and operands added to this are: 0x00 is [EAX] 0x40 0x00 is [EAX + 00H] 0x44 0x00 0x00 is [EAX + EAX*1 + 00H] 0x80 0x00 0x00 0x00 0x00 is [EAX + 00000000H] 0x84 0x00 0x00 0x00 0x00 0x00 is [EAX + EAX*1 + 00000000H] All five NOP forms include a 6-bit qp field (bits 5:0) and a 21-bit immediate field (bit 36 + bits 25:6). These fields may be set to any value with no effect on the operation of the instruction. (The encodings listed here result from setting these fields to all-0s – which is common but not required). The NOP.x form of the instruction additionally consumes a second 41-bit instruction slot – the contents of this slot is considered to be providing 41 additional immediate-bits, for a total immediate-size of 62 bits. In the case of both the NOP and NOPR instructions, the first 0 in the second byte is the "mask" value, the condition to test such as equal, not equal, high, low, etc. If the mask is 0, no branch occurs. In the case of the NOPR instruction, the second value in the second byte is the register to branch on. If register 0 is chosen, no branch occurs regardless of the mask value. Thus, if either of the two values in the second byte is 0, the branch will not happen. If the first 0 in the second byte is 0, the value of the second value in the second byte is irrelevant on most processors; however, on the IBM System/360 Model 91, if that value refers to register 15, the instruction will wait for all previously-decoded instructions to complete before continuing. In the case of the NOP instruction, the second value in the second byte is the "base" register of a combined base register, displacement register and offset address. If the base register is also 0, the branch is not taken regardless of the value of the displacement register or displacement address. Suggested opcode for 68020 and later 68k processors if a NOP without pipeline synchronization is desired. ('n' may take any 4-bit value.): 3-21 : 4-189 Under the Power ISA, many apparent no-op instruction encodings have significant side-effects – therefore, no-op encodings other than ori r0,r0,0 should be carefully avoided unless these side-effects are specifically intended. For example: From a hardware design point of view, unmapped areas of a bus are often designed to return zeroes; since the NOP slide behavior is often desirable, it gives a bias to coding it with the all-zeroes opcode. Code A function or a sequence of programming language statements is a NOP or null statement if it has no effect. Null statements may be required by the syntax of some languages in certain contexts. In Ada, the null statement serves as a NOP. As the syntax forbids that control statements or functions be empty, the null statement must be used to specify that no action is required. (Thus, if the programmer forgets to write a sequence of statements, the program will fail to compile.) The simplest NOP statement in C is the null statement, which is just a semi-colon in a context requiring a statement. Most C compilers generate no code for null statements, which has historical and performance reasons. An empty block (compound statement) is also a NOP, and may be more legible, but will still have no code generated for it by the compiler. In some cases, such as the body of a function, a block must be used, but this can be empty. In C, statements cannot be empty—simple statements must end with a ; (semicolon) while compound statements are enclosed in {} (braces), which does not itself need a following semicolon. Thus in contexts where a statement is grammatically required, some such null statement can be used. The null statement is useless by itself, but it can have a syntactic use in a wider context, e.g., within the context of a loop: alternatively, or more tersely: The last form might generate a warning with some compilers or compiler options, as a semicolon placed after a parenthesis at the end of a line usually indicates the end of a function call expression. The above code continues calling the function getchar() until it returns a \n (newline) character, essentially fast-forwarding the current reading location of standard input to the beginning of next line. In Fortran, the CONTINUE statement is used in some contexts such as the last statement in a DO loop, although it can be used anywhere, and does not have any functionality. The JavaScript language does not have a built-in NOP statement. Many implementations are possible: Alternatives, in situations where a function is required, are: The AngularJS framework provides angular.noop function that performs no operations. The jQuery library provides a function jQuery.noop(), which does nothing. The Lodash library provides a function _.noop(), which returns undefined and does nothing. As with C, the ; used by itself can be used as a null statement in Pascal. In fact, due to the specification of the language, in a BEGIN / END block, the semicolon is optional before the END statement, thus a semicolon used there is superfluous. Also, a block consisting of BEGIN END; may be used as a placeholder to indicate no action, even if placed inside another BEGIN / END block. The Python programming language has a pass statement which has no effect when executed and thus serves as a NOP. It is primarily used to ensure correct syntax due to Python's indentation-sensitive syntax; for example the syntax for definition of a class requires an indented block with the class logic, which has to be expressed as pass when it should be empty. The ':' [colon] command is a shell builtin that has similar effect to a "NOP" (a do-nothing operation). It is not technically an NOP, as it changes the special parameter $? (exit status of last command) to 0. It may be considered a synonym for the shell builtin 'true', and its exit status is true (0). The TeX typographical system's macro language has the \relax command. It does nothing by itself, but may be used to prevent the immediately preceding command from parsing any subsequent tokens. NOP protocol commands Many computer protocols, such as telnet, include a NOP command that a client can issue to request a response from the server without requesting any other actions. Such a command can be used to ensure the connection is still alive or that the server is responsive. A NOOP command is part of the following protocols (this is a partial list): Unlike the other protocols listed, the IMAP4 NOOP command has a specific purpose—it allows the server to send any pending notifications to the client. While most telnet or FTP servers respond to a NOOP command with "OK" or "+OK", some programmers have added quirky responses to the client. For example, the ftpd daemon of MINIX responds to NOOP with the message: Cracking NOPs are often involved when cracking software that checks for serial numbers, specific hardware or software requirements, presence or absence of hardware dongles, etc. in the form of a NOP slide. This process is accomplished by altering functions and subroutines to bypass security checks and instead simply return the expected value being checked for. Because most of the instructions in the security check routine will be unused, these would be replaced with NOPs, thus removing the software's security functionality without altering the positioning of everything which follows in the binary. Security exploits The NOP opcode can be used to form a NOP slide, which allows code to execute when the exact value of the instruction pointer is indeterminate (e.g., when a buffer overflow causes a function's return address on the stack to be overwritten). See also References |
======================================== |
[SOURCE: https://en.wikipedia.org/wiki/Living_Interplanetary_Flight_Experiment] | [TOKENS: 769] |
Contents Living Interplanetary Flight Experiment The Living Interplanetary Flight Experiment (LIFE or Phobos LIFE) was an interplanetary mission developed by the Planetary Society. It consisted of sending selected microorganisms on a three-year interplanetary round-trip in a small capsule aboard the Russian Fobos-Grunt spacecraft in 2011, which was a failed sample-return mission to the Martian moon Phobos. The Fobos-Grunt mission failed to leave Earth orbit and was destroyed. The goal was to test whether selected organisms can survive an undetermined number of years in deep space by flying them through interplanetary space. The experiment would have tested one aspect of panspermia, the hypothesis that life could survive space travel, if protected inside rocks blasted by impact off one planet to land on another. Precursor Prior to the Phobos LIFE experiment, a precursor LIFE prototype was successfully flown in 2011 aboard the final flight of Space Shuttle Endeavour, STS-134. Known as the Shuttle-LIFE (also LIFE) experiment. The experiment The project includes representatives of all three domains of life: bacteria, eukaryota and archaea. The capsule was transporting 10 types of organisms in 30 self-contained samples, i.e., each in triplicate. In addition, one or more natural native soil samples were flown in their own self-contained capsule. The Phobos-Soil sample return mission was the only attempted biological science mission that would have returned to Earth from deep space, far beyond the protection of Earth's magnetic field; sending biological samples through deep space is therefore a much better test of interplanetary survivability than sending the samples on a typical Earth-orbiting flight. The project was being done in collaboration with the Russian Space Research Institute, the Institute for Biomedical Problems of the Russian Academy of Sciences, the Moscow State University, the American Type Culture Collection (ATCC), and the Institute for Aerospace Medicine in Germany. Three fundamental guidelines governed the selection of the organisms: First, the organisms selected represent the three domains of life – eukaryote, bacteria and archaea. Second, the organisms are very well studied (e.g., having their genome sequenced and studied in many other experiments) to make it possible to accurately assess the effects of the long exposure to space. If they had already been studied in space conditions so much the better, since it would enable researchers to pinpoint precisely how organisms were affected by the years-long exposure to the interplanetary environment. Finally, a strong preference was given to organisms that appear to stand the best chance of surviving the journey. These are extremophiles, organisms that thrive in conditions that would kill the vast majority of Earthly creatures. The 10 'passenger' organisms selected are listed below: Bacteria Archaea Eukaryote The mass of the Bio-Module on board the Fobos-Grunt spacecraft was 100 grams or less. The design is a short cylinder. The bio-module provided 30 small tubes (3 millimeters in diameter) for individual microbe samples. It also accommodated a native sample of bacteria – derived from a permafrost region on Earth – within a cavity 26 mm in diameter. Mission failure The module passed stress tests including a shake test with vibrations at frequencies to 1,100 Hz and an impact test of 4,000 g, designed to simulate the potential impact of the capsule on Earth. The LIFE experiment was launched on November 8, 2011 on board the Fobos-Grunt. However, the spacecraft failed to depart Earth orbit due to a programming error, and fell back to Earth in the Pacific Ocean. The module was not recovered. The team is seeking out future exploratory opportunities. See also References External links |
======================================== |
[SOURCE: https://en.wikipedia.org/wiki/Mary_(programming_language)] | [TOKENS: 405] |
Contents Mary (programming language) Mary is a programming language designed and implemented by Mark Rain at RUNIT in Trondheim, Norway during the 1970s. It borrowed many features from ALGOL 68 but was designed for systems programming (machine-oriented programming), with a subset of operations being reserved for higher-level usage. An unusual feature of its syntax was that expressions were constructed using the conventional infix operators, but all of them had the same precedence and evaluation went from left to right unless there were brackets. Assignment had the destination on the right and assignment was considered just another operator. Similar to C, several language features appear to have existed to allow producing reasonably well optimised code, despite a quite primitive code generator in the compiler. These included operators similar to the += et alter in C and explicit register declarations for variables. Notable features: Compilers were made for Kongsberg Våpenfabrikk's SM-4 and Norsk Data Nord-10/ND-100 mini-computers. The original Mary compiler was written in NU ALGOL, ran on the Univac-1100 series and was used to bootstrap a native compiler for ND-100/Sintran-III. RUNIT implemented a CHILL compiler written in Mary which ran on ND-100 and had Intel 8086 and 80286 targets. When this compiler was ported to the VAX platform, a common backend for Mary and CHILL was implemented. Later, backends for i386 and SPARC were available. Since the Mary compiler was implemented in Mary, it was possible to run the compiler on all these platforms. An improved version, Mary/2, was developed using a new compiler in the United States. Mary is no longer maintained. Example See also References This article is based on material taken from the Free On-line Dictionary of Computing prior to 1 November 2008 and incorporated under the "relicensing" terms of the GFDL, version 1.3 or later. Further reading |
======================================== |
[SOURCE: https://en.wikipedia.org/w/index.php?title=Non-player_character&action=edit§ion=6] | [TOKENS: 1431] |
Editing Non-player character (section) Copy and paste: – — ° ′ ″ ≈ ≠ ≤ ≥ ± − × ÷ ← → · § Cite your sources: <ref></ref> {{}} {{{}}} | [] [[]] [[Category:]] #REDIRECT [[]] <s></s> <sup></sup> <sub></sub> <code></code> <pre></pre> <blockquote></blockquote> <ref></ref> <ref name="" /> {{Reflist}} <references /> <includeonly></includeonly> <noinclude></noinclude> {{DEFAULTSORT:}} <nowiki></nowiki> <!-- --> <span class="plainlinks"></span> Symbols: ~ | ¡ ¿ † ‡ ↔ ↑ ↓ • ¶ # ∞ ‹› «» ¤ ₳ ฿ ₵ ¢ ₡ ₢ $ ₫ ₯ € ₠ ₣ ƒ ₴ ₭ ₤ ℳ ₥ ₦ ₧ ₰ £ ៛ ₨ ₪ ৳ ₮ ₩ ¥ ♠ ♣ ♥ ♦ 𝄫 ♭ ♮ ♯ 𝄪 © ¼ ½ ¾ Latin: A a Á á À à  â Ä ä Ǎ ǎ Ă ă Ā ā à ã Å å Ą ą Æ æ Ǣ ǣ B b C c Ć ć Ċ ċ Ĉ ĉ Č č Ç ç D d Ď ď Đ đ Ḍ ḍ Ð ð E e É é È è Ė ė Ê ê Ë ë Ě ě Ĕ ĕ Ē ē Ẽ ẽ Ę ę Ẹ ẹ Ɛ ɛ Ǝ ǝ Ə ə F f G g Ġ ġ Ĝ ĝ Ğ ğ Ģ ģ H h Ĥ ĥ Ħ ħ Ḥ ḥ I i İ ı Í í Ì ì Î î Ï ï Ǐ ǐ Ĭ ĭ Ī ī Ĩ ĩ Į į Ị ị J j Ĵ ĵ K k Ķ ķ L l Ĺ ĺ Ŀ ŀ Ľ ľ Ļ ļ Ł ł Ḷ ḷ Ḹ ḹ M m Ṃ ṃ N n Ń ń Ň ň Ñ ñ Ņ ņ Ṇ ṇ Ŋ ŋ O o Ó ó Ò ò Ô ô Ö ö Ǒ ǒ Ŏ ŏ Ō ō Õ õ Ǫ ǫ Ọ ọ Ő ő Ø ø Œ œ Ɔ ɔ P p Q q R r Ŕ ŕ Ř ř Ŗ ŗ Ṛ ṛ Ṝ ṝ S s Ś ś Ŝ ŝ Š š Ş ş Ș ș Ṣ ṣ ß T t Ť ť Ţ ţ Ț ț Ṭ ṭ Þ þ U u Ú ú Ù ù Û û Ü ü Ǔ ǔ Ŭ ŭ Ū ū Ũ ũ Ů ů Ų ų Ụ ụ Ű ű Ǘ ǘ Ǜ ǜ Ǚ ǚ Ǖ ǖ V v W w Ŵ ŵ X x Y y Ý ý Ŷ ŷ Ÿ ÿ Ỹ ỹ Ȳ ȳ Z z Ź ź Ż ż Ž ž ß Ð ð Þ þ Ŋ ŋ Ə ə Greek: Ά ά Έ έ Ή ή Ί ί Ό ό Ύ ύ Ώ ώ Α α Β β Γ γ Δ δ Ε ε Ζ ζ Η η Θ θ Ι ι Κ κ Λ λ Μ μ Ν ν Ξ ξ Ο ο Π π Ρ ρ Σ σ ς Τ τ Υ υ Φ φ Χ χ Ψ ψ Ω ω {{Polytonic|}} Cyrillic: А а Б б В в Г г Ґ ґ Ѓ ѓ Д д Ђ ђ Е е Ё ё Є є Ж ж З з Ѕ ѕ И и І і Ї ї Й й Ј ј К к Ќ ќ Л л Љ љ М м Н н Њ њ О о П п Р р С с Т т Ћ ћ У у Ў ў Ф ф Х х Ц ц Ч ч Џ џ Ш ш Щ щ Ъ ъ Ы ы Ь ь Э э Ю ю Я я ́ IPA: t̪ d̪ ʈ ɖ ɟ ɡ ɢ ʡ ʔ ɸ β θ ð ʃ ʒ ɕ ʑ ʂ ʐ ç ʝ ɣ χ ʁ ħ ʕ ʜ ʢ ɦ ɱ ɳ ɲ ŋ ɴ ʋ ɹ ɻ ɰ ʙ ⱱ ʀ ɾ ɽ ɫ ɬ ɮ ɺ ɭ ʎ ʟ ɥ ʍ ɧ ʼ ɓ ɗ ʄ ɠ ʛ ʘ ǀ ǃ ǂ ǁ ɨ ʉ ɯ ɪ ʏ ʊ ø ɘ ɵ ɤ ə ɚ ɛ œ ɜ ɝ ɞ ʌ ɔ æ ɐ ɶ ɑ ɒ ʰ ʱ ʷ ʲ ˠ ˤ ⁿ ˡ ˈ ˌ ː ˑ ̪ {{IPA|}} This page is a member of 11 hidden categories (help): |
======================================== |
[SOURCE: https://en.wikipedia.org/wiki/File:Internet_users_per_100_inhabitants_ITU.svg] | [TOKENS: 121] |
File:Internet users per 100 inhabitants ITU.svg Summary Data represented in this chart, references for each year's data, and the script to generate the graph are on Github Licensing File history Click on a date/time to view the file as it appeared at that time. File usage The following 8 pages use this file: Global file usage The following other wikis use this file: Metadata This file contains additional information, probably added from the digital camera or scanner used to create or digitize it. If the file has been modified from its original state, some details may not fully reflect the modified file. |
======================================== |
[SOURCE: https://en.wikipedia.org/wiki/Microsoft_Macro_Assembler] | [TOKENS: 1213] |
Contents Microsoft Macro Assembler Microsoft Macro Assembler (MASM) is an x86 assembler that uses the Intel syntax for MS-DOS and Microsoft Windows. Beginning with MASM 8.0, there are two versions of the assembler: One for 16-bit & 32-bit assembly sources, and another (ML64) for 64-bit sources only. MASM is maintained by Microsoft, but since version 6.12 it has not been sold as a separate product. It is instead supplied with various Microsoft SDKs and C compilers. Recent versions of MASM are included with Microsoft Visual Studio. Notable applications compiled using MASM are RollerCoaster Tycoon which was 99% written in assembly language and built with MASM. History The earliest versions of MASM date back to 1981. They were sold either as the generic "Microsoft Macro Assembler" for all x86 machines or as the OEM version specifically for IBM PCs. By Version 4.0, the IBM release was dropped. Up to Version 3.0, MASM was also bundled with a smaller companion assembler, ASM.EXE. This was intended for PCs with only 64k of memory and lacked some features of the full MASM, such as the ability to use code macros. MS-DOS versions up to 4.x included Microsoft's LINK utility, which was designed to convert intermediate OBJ files generated by MASM and other compilers; however, as users who did not program had no use of the utility, it was moved to their compiler packages. Version 4.0, released October 1985, added support for 286 instructions. Version 5.0, released August 1987, supported 386 instructions, and also shorthand mnemonics for segment descriptors (.code, .data, etc.), but it could still only generate real mode executables. Through version 5.0, MASM was available as an MS-DOS application only. Versions 5.1 and 6.0 were available as both MS-DOS and OS/2 applications. Version 6.0, released in 1991, added parameter passing with "invoke" and some other high level-like constructs, in addition to the already existing high level-like records, among other things. Both 6.0 and 6.0B were able to be run on an 8086 processor but could generate flat 32-bit 386 code. In 1992, 6.1 was released, which added support for the COFF object format used by Windows NT, and removed support for OS/2. 6.1 was built as a bi-modal binary before the Win32 API was finalized, and is incompatible with running on Windows NT due to missing exports. In 1993 full support for protected mode 32-bit applications and the Pentium instruction set was added. The 6.11 MASM binary at that time (1993) was shipped as a "bi-modal" (win32, i.e. PE) DOS-extended binary (using the Phar Lap TNT DOS extender). However, the setup.exe is an MZ executable so won't run under 64-bit versions of Windows, and the bi-modal ml.exe is compressed, and the decomp.exe is an NE executable, so also won't run under 64-bit versions of Windows, so you effectively need access to 32-bit Windows (or MSDOS) in order to install it. Version 6.11 is the last version of MASM that will run under MS-DOS. There were a series of patches available, up to 6.11d, that need 32-bit Windows to run, but the patched ml.exe still has the Phar Lap dos extender so can still be run under MSDOS. By the end of 1997, MASM fully supported Windows 95 and included some AMD-specific instructions. In 1999, Intel released macros for SIMD and MMX instructions, which were shortly thereafter supported natively by MASM. With the 6.15 release in 2000, Microsoft discontinued support for MASM as a separate product, instead subsuming it into the Visual Studio toolset. Though it was still compatible with Windows 98, current versions of Visual Studio were not. Support for 64-bit processors was not added until the release of Visual Studio 2005, with MASM 8.0. After 25 June 2015, there are at least three different MASMs with the version number 14.00.23026. In Microsoft Visual Studio 2015 Enterprise Edition, there is one "amd64_x86" ml and two ml64s, "x86_amd64" and "amd64". They run on different platforms targeting different platforms: Object module formats supported by MASM Early versions of MASM generated object modules using the OMF format, which was used to create binaries for MS-DOS or OS/2. Since version 6.1, MASM is able to produce object modules in the Portable Executable (PE/COFF) format. PE/COFF is compatible with recent Microsoft C compilers, and object modules produced by either MASM or the C compiler can be routinely intermixed and linked into Win32 and Win64 binaries. Assemblers compatible with MASM Some other assemblers can assemble most code written for MASM, with the exception of more complex macros. Mixed language programming support Documentation for 1987's version 5.1 included support for "Microsoft BASIC, C, FORTRAN, Pascal." Reception In a review of three assemblers, Michael Blaszczak of BYTE in February 1989 found that MASM 5.1 was the slowest and complained the most about code. He concluded that "MASM takes some getting used to, but it gets the job done" despite "more than its fair share of frustrating quirks and oddities". See also References External links |
======================================== |
[SOURCE: https://en.wikipedia.org/wiki/Extraterrestrial_life#Context] | [TOKENS: 11349] |
Contents Extraterrestrial life Extraterrestrial life, or alien life (colloquially aliens), is life that originates from another world rather than on Earth. No extraterrestrial life has yet been scientifically or conclusively detected. Such life might range from simple forms such as prokaryotes to intelligent beings, possibly bringing forth civilizations that might be far more, or far less, advanced than humans. The Drake equation speculates about the existence of sapient life elsewhere in the universe. The science of extraterrestrial life is known as astrobiology. Speculation about inhabited worlds beyond Earth dates back to antiquity. Early Christian writers, including Augustine, discussed ideas from thinkers like Democritus and Epicurus about countless worlds in the vast universe. Pre-modern writers typically assumed extraterrestrial "worlds" were inhabited by living beings. William Vorilong, in the 15th century, acknowledged the possibility Jesus could have visited extraterrestrial worlds to redeem their inhabitants.: 26 In 1440, Nicholas of Cusa suggested Earth is a "brilliant star"; he theorized that all celestial bodies, even the Sun, could host life. Descartes wrote that there were no means to prove the stars were not inhabited by "intelligent creatures", but their existence was a matter of speculation.: 67 In comparison to the life-abundant Earth, the vast majority of intrasolar and extrasolar planets and moons have harsh surface conditions and disparate atmospheric chemistry, or lack an atmosphere. However, there are many extreme and chemically harsh ecosystems on Earth that do support forms of life and are often hypothesized to be the origin of life on Earth. Examples include life surrounding hydrothermal vents, acidic hot springs, and volcanic lakes, as well as halophiles and the deep biosphere. Since the mid-20th century, researchers have searched for extraterrestrial life and intelligence. Solar system studies focus on Venus, Mars, Europa, and Titan, while exoplanet discoveries now total 6,022 confirmed planets in 4,490 systems as of October 2025. Depending on the category of search, methods range from analysis of telescope and specimen data to radios used to detect and transmit interstellar communication. Interstellar travel remains largely hypothetical, with only the Voyager 1 and Voyager 2 probes confirmed to have entered the interstellar medium. The concept of extraterrestrial life, especially intelligent life, has greatly influenced culture and fiction. A key debate centers on contacting extraterrestrial intelligence: some advocate active attempts, while others warn it could be risky, given human history of exploiting other societies. Context Initially, after the Big Bang, the universe was too hot to allow life. It is estimated that the temperature of the universe was around 10 billion Kelvin at the one-second mark. Roughly 15 million years later, it cooled to temperate levels, though the elements of organic life were yet nonexistent. The only freely available elements at that point were hydrogen and helium. Carbon and oxygen (and later, water) would not appear until 50 million years later, created through stellar fusion. At that point, the difficulty for life to appear was not the temperature, but the scarcity of free heavy elements. Planetary systems emerged, and the first organic compounds may have formed in the protoplanetary disk of dust grains that would eventually create rocky planets like Earth. Although Earth was in a molten state after its birth and may have burned any organics that fell on it, it would have been more receptive once it cooled down. Once the right conditions on Earth were met, life started by a chemical process known as abiogenesis. Alternatively, life may have formed less frequently, then spread—by meteoroids, for example—between habitable planets in a process called panspermia. During most of its stellar evolution, stars combine hydrogen nuclei to make helium nuclei by stellar fusion, and the comparatively lighter weight of helium allows the star to release the extra energy. The process continues until the star uses all of its available fuel, with the speed of consumption being related to the size of the star. During its last stages, stars start combining helium nuclei to form carbon nuclei. The larger stars can further combine carbon nuclei to create oxygen and silicon, oxygen into neon and sulfur, and so on until iron. Ultimately, the star blows much of its content back into the stellar medium, where it would join clouds that would eventually become new generations of stars and planets. Many of those materials are the raw components of life on Earth. As this process takes place in all the universe, said materials are ubiquitous in the cosmos and not a rarity from the Solar System. Earth is a planet in the Solar System, a planetary system formed by a star at the center, the Sun, and the objects that orbit it: other planets, moons, asteroids, and comets. The sun is part of the Milky Way, a galaxy. The Milky Way is part of the Local Group, a galaxy group that is in turn part of the Laniakea Supercluster. The universe is composed of all similar structures in existence. The immense distances between celestial objects are a difficulty for studying extraterrestrial life. So far, humans have only set foot on the Moon and sent robotic probes to other planets and moons in the Solar System. Although probes can withstand conditions that may be lethal to humans, the distances cause time delays: the New Horizons took nine years after launch to reach Pluto. No probe has ever reached extrasolar planetary systems. The Voyager 2 left the Solar System at a speed of 50,000 kilometers per hour; if it headed towards the Alpha Centauri system, the closest one to Earth at 4.4 light years, it would reach it in 100,000 years. Under current technology, such systems can only be studied by telescopes, which have limitations. It is estimated that dark matter has a larger amount of combined matter than stars and gas clouds, but as it plays no role in the stellar evolution of stars and planets, it is usually not taken into account by astrobiology. There is an area around a star, the circumstellar habitable zone or "Goldilocks zone", wherein water may be at the right temperature to exist in liquid form at a planetary surface. This area is neither too close to the star, where water would become steam, nor too far away, where water would be frozen as ice. However, although useful as an approximation, planetary habitability is complex and defined by several factors. Being in the habitable zone is not enough for a planet to be habitable, not even to actually have such liquid water. Venus is located in the solar system's habitable zone, but does not have liquid water because of the conditions of its atmosphere. Jovian planets or gas giants are not considered habitable even if they orbit close enough to their stars as hot Jupiters, due to crushing atmospheric pressures. The actual distances for the habitable zones vary according to the type of star, and even the solar activity of each specific star influences the local habitability. The type of star also defines the time the habitable zone will exist, as its presence and limits will change along with the star's stellar evolution. The Big Bang occurred 13.8 billion years ago, the Solar System was formed 4.6 billion years ago, and the first hominids appeared 6 million years ago. Life on other planets may have started, evolved, given birth to extraterrestrial intelligences, and perhaps even faced a planetary extinction event millions or billions of years ago. When considered from a cosmic perspective, the brief times of existence of Earth's species may suggest that extraterrestrial life may be equally fleeting under such a scale. During a period of about 7 million years, from about 10 to 17 million years after the Big Bang, the background temperature was between 373 and 273 K (100 and 0 °C; 212 and 32 °F), allowing the possibility of liquid water if any planets existed. Avi Loeb (2014) speculated that primitive life might in principle have appeared during this window, which he called "the Habitable Epoch of the Early Universe". Life on Earth is quite ubiquitous across the planet and has adapted over time to almost all the available environments in it, extremophiles and the deep biosphere thrive at even the most hostile ones. As a result, it is inferred that life in other celestial bodies may be equally adaptive. However, the origin of life is unrelated to its ease of adaptation and may have stricter requirements. A celestial body may not have any life on it, even if it were habitable. Likelihood of existence Life in the cosmos beyond Earth has been observed. The hypothesis of ubiquitous extraterrestrial life relies on three main ideas. The first one, the size of the universe, allows for plenty of planets to have a similar habitability to Earth, and the age of the universe gives enough time for a long process analog to the history of Earth to happen there. The second is that the substances that make life, such as carbon and water, are ubiquitous in the universe. The third is that the physical laws are universal, which means that the forces that would facilitate or prevent the existence of life would be the same ones as on Earth. According to this argument, made by scientists such as Carl Sagan and Stephen Hawking, it would be improbable for life not to exist somewhere else other than Earth. This argument is embodied in the Copernican principle, which states that Earth does not occupy a unique position in the Universe, and the mediocrity principle, which states that there is nothing special about life on Earth. Other authors consider instead that life in the cosmos, or at least multicellular life, may actually be rare. The Rare Earth hypothesis maintains that life on Earth is possible because of a series of factors that range from the location in the galaxy and the configuration of the Solar System to local characteristics of the planet, and that it is unlikely that another planet simultaneously meets all such requirements. The proponents of this hypothesis consider that very little evidence suggests the existence of extraterrestrial life and that, at this point, it is just a desired result and not a reasonable scientific explanation for any gathered data. In 1961, astronomer and astrophysicist Frank Drake devised the Drake equation as a way to stimulate scientific dialogue at a meeting on the search for extraterrestrial intelligence (SETI). The Drake equation is a probabilistic argument used to estimate the number of active, communicative extraterrestrial civilizations in the Milky Way galaxy. The Drake equation is:: xix where: and Drake's proposed estimates are as follows, but numbers on the right side of the equation are agreed as speculative and open to substitution: 10,000 = 5 ⋅ 0.5 ⋅ 2 ⋅ 1 ⋅ 0.2 ⋅ 1 ⋅ 10,000 {\displaystyle 10{,}000=5\cdot 0.5\cdot 2\cdot 1\cdot 0.2\cdot 1\cdot 10{,}000} [better source needed] The Drake equation has proved controversial since, although it is written as a math equation, none of its values were known at the time. Although some values may eventually be measured, others are based on social sciences and are not knowable by their very nature. This does not allow one to make noteworthy conclusions from the equation. Based on observations from the Hubble Space Telescope, there are nearly 2 trillion galaxies in the observable universe. It is estimated that at least ten percent of all Sun-like stars have a system of planets. In other words, there are 6.25×1018 stars with planets orbiting them in the observable universe. Even if it is assumed that only one out of a billion of these stars has planets supporting life, there would be some 6.25 billion life-supporting planetary systems in the observable universe. A 2013 study based on results from the Kepler spacecraft estimated that the Milky Way contains at least as many planets as it does stars, resulting in 100–400 billion exoplanets. The Nebular hypothesis that explains the formation of the Solar System and other planetary systems would suggest that those can have several configurations, and not all of them may have rocky planets within the habitable zone. The apparent contradiction between high estimates of the probability of the existence of extraterrestrial civilisations and the lack of evidence for such civilisations is known as the Fermi paradox. Dennis W. Sciama claimed that life's existence in the universe depends on various fundamental constants. Zhi-Wei Wang and Samuel L. Braunstein suggest that a random universe capable of supporting life is likely to be just barely able to do so, giving a potential explanation to the Fermi paradox. Biochemical basis If extraterrestrial life exists, it could range from simple microorganisms and multicellular organisms similar to animals or plants, to complex alien intelligences akin to humans. When scientists talk about extraterrestrial life, they consider all those types. Although it is possible that extraterrestrial life may have other configurations, scientists use the hierarchy of lifeforms from Earth for simplicity, as it is the only one known to exist. The first basic requirement for life is an environment with non-equilibrium thermodynamics, which means that the thermodynamic equilibrium must be broken by a source of energy. The traditional sources of energy in the cosmos are the stars, such as for life on Earth, which depends on the energy of the sun. However, there are other alternative energy sources, such as volcanoes, plate tectonics, and hydrothermal vents. There are ecosystems on Earth in deep areas of the ocean that do not receive sunlight, and take energy from black smokers instead. Magnetic fields and radioactivity have also been proposed as sources of energy, although they would be less efficient ones. Life on Earth requires water in a liquid state as a solvent in which biochemical reactions take place. It is highly unlikely that an abiogenesis process can start within a gaseous or solid medium: the atom speeds, either too fast or too slow, make it difficult for specific ones to meet and start chemical reactions. A liquid medium also allows the transport of nutrients and substances required for metabolism. Sufficient quantities of carbon and other elements, along with water, might enable the formation of living organisms on terrestrial planets with a chemical make-up and temperature range similar to that of Earth. Life based on ammonia rather than water has been suggested as an alternative, though this solvent appears less suitable than water. It is also conceivable that there are forms of life whose solvent is a liquid hydrocarbon, such as methane, ethane or propane. Another unknown aspect of potential extraterrestrial life would be the chemical elements that would compose it. Life on Earth is largely composed of carbon, but there could be other hypothetical types of biochemistry. A replacement for carbon would need to be able to create complex molecules, store information required for evolution, and be freely available in the medium. To create DNA, RNA, or a close analog, such an element should be able to bind its atoms with many others, creating complex and stable molecules. It should be able to create at least three covalent bonds: two for making long strings and at least a third to add new links and allow for diverse information. Only nine elements meet this requirement: boron, nitrogen, phosphorus, arsenic, antimony (three bonds), carbon, silicon, germanium and tin (four bonds). As for abundance, carbon, nitrogen, and silicon are the most abundant ones in the universe, far more than the others. On Earth's crust the most abundant of those elements is silicon, in the Hydrosphere it is carbon and in the atmosphere, it is carbon and nitrogen. Silicon, however, has disadvantages over carbon. The molecules formed with silicon atoms are less stable, and more vulnerable to acids, oxygen, and light. An ecosystem of silicon-based lifeforms would require very low temperatures, high atmospheric pressure, an atmosphere devoid of oxygen, and a solvent other than water. The low temperatures required would add an extra problem, the difficulty to kickstart a process of abiogenesis to create life in the first place. Norman Horowitz, head of the Jet Propulsion Laboratory bioscience section for the Mariner and Viking missions from 1965 to 1976 considered that the great versatility of the carbon atom makes it the element most likely to provide solutions, even exotic solutions, to the problems of survival of life on other planets. However, he also considered that the conditions found on Mars were incompatible with carbon based life. Even if extraterrestrial life is based on carbon and uses water as a solvent, like Earth life, it may still have a radically different biochemistry. Life is generally considered to be a product of natural selection. It has been proposed that to undergo natural selection a living entity must have the capacity to replicate itself, the capacity to avoid damage/decay, and the capacity to acquire and process resources in support of the first two capacities. Life on Earth may have started with an RNA world and later evolved to its current form, where some of the RNA tasks were transferred to DNA and proteins. Extraterrestrial life may still be stuck using RNA, or evolve into other configurations. It is unclear if our biochemistry is the most efficient one that could be generated, or which elements would follow a similar pattern. However, it is likely that, even if cells had a different composition to those from Earth, they would still have a cell membrane. Life on Earth jumped from prokaryotes to eukaryotes and from unicellular organisms to multicellular organisms through evolution. So far no alternative process to achieve such a result has been conceived, even if hypothetical. Evolution requires life to be divided into individual organisms, and no alternative organisation has been satisfactorily proposed either. At the basic level, membranes define the limit of a cell, between it and its environment, while remaining partially open to exchange energy and resources with it. The evolution from simple cells to eukaryotes, and from them to multicellular lifeforms, is not guaranteed. The Cambrian explosion took place thousands of millions of years after the origin of life, and its causes are not fully known yet. On the other hand, the jump to multicellularity took place several times, which suggests that it could be a case of convergent evolution, and so likely to take place on other planets as well. Palaeontologist Simon Conway Morris considers that convergent evolution would lead to kingdoms similar to our plants and animals, and that many features are likely to develop in alien animals as well, such as bilateral symmetry, limbs, digestive systems and heads with sensory organs. Scientists from the University of Oxford analysed it from the perspective of evolutionary theory and wrote in a study in the International Journal of Astrobiology that aliens may be similar to humans. The planetary context would also have an influence: a planet with higher gravity would have smaller animals, and other types of stars can lead to non-green photosynthesizers. The amount of energy available would also affect biodiversity, as an ecosystem sustained by black smokers or hydrothermal vents would have less energy available than those sustained by a star's light and heat, and so its lifeforms would not grow beyond a certain complexity. There is also research in assessing the capacity of life for developing intelligence. It has been suggested that this capacity arises with the number of potential niches a planet contains, and that the complexity of life itself is reflected in the information density of planetary environments, which in turn can be computed from its niches. It is common knowledge that the conditions on other planets in the solar system, in addition to the many galaxies outside of the Milky Way galaxy, are very harsh and seem to be too extreme to harbor any life. The environmental conditions on these planets can have intense UV radiation paired with extreme temperatures, lack of water, and much more that can lead to conditions that don't seem to favor the creation or maintenance of extraterrestrial life. However, there has been much historical evidence that some of the earliest and most basic forms of life on Earth originated in some extreme environments that seem unlikely to have harbored life at least at one point in Earth's history. Fossil evidence as well as many historical theories backed up by years of research and studies have marked environments like hydrothermal vents or acidic hot springs as some of the first places that life could have originated on Earth. These environments can be considered extreme when compared to the typical ecosystems that the majority of life on Earth now inhabit, as hydrothermal vents are scorching hot due to the magma escaping from the Earth's mantle and meeting the much colder oceanic water. Even in today's world, there can be a diverse population of bacteria found inhabiting the area surrounding these hydrothermal vents which can suggest that some form of life can be supported even in the harshest of environments like the other planets in the solar system. The aspects of these harsh environments that make them ideal for the origin of life on Earth, as well as the possibility of creation of life on other planets, is the chemical reactions forming spontaneously. For example, the hydrothermal vents found on the ocean floor are known to support many chemosynthetic processes which allow organisms to utilize energy through reduced chemical compounds that fix carbon. In return, these reactions will allow for organisms to live in relatively low oxygenated environments while maintaining enough energy to support themselves. The early Earth environment was reducing and therefore, these carbon fixing compounds were necessary for the survival and possible origin of life on Earth. With the little amount of information that scientists have found regarding the atmosphere on other planets in the Milky Way galaxy and beyond, the atmospheres are most likely reducing or with very low oxygen levels, especially when compared with Earth's atmosphere. If there were the necessary elements and ions on these planets, the same carbon fixing, reduced chemical compounds occurring around hydrothermal vents could also occur on these planets' surfaces and possibly result in the origin of extraterrestrial life. Planetary habitability in the Solar System The Solar System has a wide variety of planets, dwarf planets, and moons, and each one is studied for its potential to host life. Each one has its own specific conditions that may benefit or harm life. So far, the only lifeforms found are those from Earth. No extraterrestrial intelligence other than humans exists or has ever existed within the Solar System. Astrobiologist Mary Voytek points out that it would be unlikely to find large ecosystems, as they would have already been detected by now. The inner Solar System is likely devoid of life. However, Venus is still of interest to astrobiologists, as it is a terrestrial planet that was likely similar to Earth in its early stages and developed in a different way. There is a greenhouse effect, the surface is the hottest in the Solar System, sulfuric acid clouds, all surface liquid water is lost, and it has a thick carbon-dioxide atmosphere with huge pressure. Comparing both helps to understand the precise differences that lead to beneficial or harmful conditions for life. And despite the conditions against life on Venus, there are suspicions that microbial life-forms may still survive in high-altitude clouds. Mars is a cold and almost airless desert, inhospitable to life. However, recent studies revealed that water on Mars used to be quite abundant, forming rivers, lakes, and perhaps even oceans. Mars may have been habitable back then, and life on Mars may have been possible. But when the planetary core ceased to generate a magnetic field, solar winds removed the atmosphere and the planet became vulnerable to solar radiation. Ancient life-forms may still have left fossilised remains, and microbes may still survive deep underground. As mentioned, the gas giants and ice giants are unlikely to contain life. The most distant solar system bodies, found in the Kuiper Belt and outwards, are locked in permanent deep-freeze, but cannot be ruled out completely. Although the giant planets themselves are highly unlikely to have life, there is much hope to find it on moons orbiting these planets. Europa, from the Jovian system, has a subsurface ocean below a thick layer of ice. Ganymede and Callisto also have subsurface oceans, but life is less likely in them because water is sandwiched between layers of solid ice. Europa would have contact between the ocean and the rocky surface, which helps the chemical reactions. It may be difficult to dig so deep in order to study those oceans, though. Enceladus, a tiny moon of Saturn with another subsurface ocean, may not need to be dug, as it releases water to space in eruption columns. The space probe Cassini flew inside one of these, but could not make a full study because NASA did not expect this phenomenon and did not equip the probe to study ocean water. Still, Cassini detected complex organic molecules, salts, evidence of hydrothermal activity, hydrogen, and methane. Titan is the only celestial body in the Solar System besides Earth that has liquid bodies on the surface. It has rivers, lakes, and rain of hydrocarbons, methane, and ethane, and even a cycle similar to Earth's water cycle. This special context encourages speculations about lifeforms with different biochemistry, but the cold temperatures would make such chemistry take place at a very slow pace. Water is rock-solid on the surface, but Titan does have a subsurface water ocean like several other moons. However, it is of such a great depth that it would be very difficult to access it for study. Scientific search The science that searches and studies life in the universe, both on Earth and elsewhere, is called astrobiology. With the study of Earth's life, the only known form of life, astrobiology seeks to study how life starts and evolves and the requirements for its continuous existence. This helps to determine what to look for when searching for life in other celestial bodies. This is a complex area of study, and uses the combined perspectives of several scientific disciplines, such as astronomy, biology, chemistry, geology, oceanography, and atmospheric sciences. The scientific search for extraterrestrial life is being carried out both directly and indirectly. As of September 2017[update], 3,667 exoplanets in 2,747 systems have been identified, and other planets and moons in the Solar System hold the potential for hosting primitive life such as microorganisms. As of 8 February 2021, an updated status of studies considering the possible detection of lifeforms on Venus (via phosphine) and Mars (via methane) was reported. Scientists search for biosignatures within the Solar System by studying planetary surfaces and examining meteorites. Some claim to have identified evidence that microbial life has existed on Mars. In 1996, a controversial report stated that structures resembling nanobacteria were discovered in a meteorite, ALH84001, formed of rock ejected from Mars. Although all the unusual properties of the meteorite were eventually explained as the result of inorganic processes, the controversy over its discovery laid the groundwork for the development of astrobiology. An experiment on the two Viking Mars landers reported gas emissions from heated Martian soil samples that some scientists argue are consistent with the presence of living microorganisms. Lack of corroborating evidence from other experiments on the same samples suggests that a non-biological reaction is a more likely hypothesis. In February 2005 NASA scientists reported they may have found some evidence of extraterrestrial life on Mars. The two scientists, Carol Stoker and Larry Lemke of NASA's Ames Research Center, based their claim on methane signatures found in Mars's atmosphere resembling the methane production of some forms of primitive life on Earth, as well as on their own study of primitive life near the Rio Tinto river in Spain. NASA officials soon distanced NASA from the scientists' claims, and Stoker herself backed off from her initial assertions. In November 2011, NASA launched the Mars Science Laboratory that landed the Curiosity rover on Mars. It is designed to assess the past and present habitability on Mars using a variety of scientific instruments. The rover landed on Mars at Gale Crater in August 2012. A group of scientists at Cornell University started a catalog of microorganisms, with the way each one reacts to sunlight. The goal is to help with the search for similar organisms in exoplanets, as the starlight reflected by planets rich in such organisms would have a specific spectrum, unlike that of starlight reflected from lifeless planets. If Earth was studied from afar with this system, it would reveal a shade of green, as a result of the abundance of plants with photosynthesis. In August 2011, NASA studied meteorites found on Antarctica, finding adenine, guanine, hypoxanthine, and xanthine. Adenine and guanine are components of DNA, and the others are used in other biological processes. The studies ruled out pollution of the meteorites on Earth, as those components would not be freely available the way they were found in the samples. This discovery suggests that several organic molecules that serve as building blocks of life may be generated within asteroids and comets. In October 2011, scientists reported that cosmic dust contains complex organic compounds ("amorphous organic solids with a mixed aromatic-aliphatic structure") that could be created naturally, and rapidly, by stars. It is still unclear if those compounds played a role in the creation of life on Earth, but Sun Kwok, of the University of Hong Kong, thinks so. "If this is the case, life on Earth may have had an easier time getting started as these organics can serve as basic ingredients for life." In August 2012, and in a world first, astronomers at Copenhagen University reported the detection of a specific sugar molecule, glycolaldehyde, in a distant star system. The molecule was found around the protostellar binary IRAS 16293-2422, which is located 400 light years from Earth. Glycolaldehyde is needed to form ribonucleic acid, or RNA, which is similar in function to DNA. This finding suggests that complex organic molecules may form in stellar systems prior to the formation of planets, eventually arriving on young planets early in their formation. In December 2023, astronomers reported the first time discovery, in the plumes of Enceladus, moon of the planet Saturn, of hydrogen cyanide, a possible chemical essential for life as we know it, as well as other organic molecules, some of which are yet to be better identified and understood. According to the researchers, "these [newly discovered] compounds could potentially support extant microbial communities or drive complex organic synthesis leading to the origin of life." Although most searches are focused on the biology of extraterrestrial life, an extraterrestrial intelligence capable enough to develop a civilization may be detectable by other means as well. Technology may generate technosignatures, effects on the native planet that may not be caused by natural causes. There are three main types of techno-signatures considered: interstellar communications, effects on the atmosphere, and planetary-sized structures such as Dyson spheres. Organizations such as the SETI Institute search the cosmos for potential forms of communication. They started with radio waves, and now search for laser pulses as well. The challenge for this search is that there are natural sources of such signals as well, such as gamma-ray bursts and supernovae, and the difference between a natural signal and an artificial one would be in its specific patterns. Astronomers intend to use artificial intelligence for this, as it can manage large amounts of data and is devoid of biases and preconceptions. Besides, even if there is an advanced extraterrestrial civilization, there is no guarantee that it is transmitting radio communications in the direction of Earth. The length of time required for a signal to travel across space means that a potential answer may arrive decades or centuries after the initial message. The atmosphere of Earth is rich in nitrogen dioxide as a result of air pollution, which can be detectable. The natural abundance of carbon, which is also relatively reactive, makes it likely to be a basic component of the development of a potential extraterrestrial technological civilization, as it is on Earth. Fossil fuels may likely be generated and used on such worlds as well. The abundance of chlorofluorocarbons in the atmosphere can also be a clear technosignature, considering their role in ozone depletion. Light pollution may be another technosignature, as multiple lights on the night side of a rocky planet can be a sign of advanced technological development. However, modern telescopes are not strong enough to study exoplanets with the required level of detail to perceive it. The Kardashev scale proposes that a civilization may eventually start consuming energy directly from its local star. This would require giant structures built next to it, called Dyson spheres. Those speculative structures would cause an excess infrared radiation, that telescopes may notice. The infrared radiation is typical of young stars, surrounded by dusty protoplanetary disks that will eventually form planets. An older star such as the Sun would have no natural reason to have excess infrared radiation. The presence of heavy elements in a star's light-spectrum is another potential biosignature; such elements would (in theory) be found if the star were being used as an incinerator/repository for nuclear waste products. Some astronomers search for extrasolar planets that may be conducive to life, narrowing the search to terrestrial planets within the habitable zones of their stars. Since 1992, over four thousand exoplanets have been discovered (6,128 planets in 4,584 planetary systems including 1,017 multiple planetary systems as of 30 October 2025). The extrasolar planets so far discovered range in size from that of terrestrial planets similar to Earth's size to that of gas giants larger than Jupiter. The number of observed exoplanets is expected to increase greatly in the coming years.[better source needed] The Kepler space telescope has also detected a few thousand candidate planets, of which about 11% may be false positives. There is at least one planet on average per star. About 1 in 5 Sun-like stars[a] have an "Earth-sized"[b] planet in the habitable zone,[c] with the nearest expected to be within 12 light-years distance from Earth. Assuming 200 billion stars in the Milky Way,[d] that would be 11 billion potentially habitable Earth-sized planets in the Milky Way, rising to 40 billion if red dwarfs are included. The rogue planets in the Milky Way possibly number in the trillions. The nearest known exoplanet is Proxima Centauri b, located 4.2 light-years (1.3 pc) from Earth in the southern constellation of Centaurus. As of March 2014[update], the least massive exoplanet known is PSR B1257+12 A, which is about twice the mass of the Moon. The most massive planet listed on the NASA Exoplanet Archive is DENIS-P J082303.1−491201 b, about 29 times the mass of Jupiter, although according to most definitions of a planet, it is too massive to be a planet and may be a brown dwarf instead. Almost all of the planets detected so far are within the Milky Way, but there have also been a few possible detections of extragalactic planets. The study of planetary habitability also considers a wide range of other factors in determining the suitability of a planet for hosting life. One sign that a planet probably already contains life is the presence of an atmosphere with significant amounts of oxygen, since that gas is highly reactive and generally would not last long without constant replenishment. This replenishment occurs on Earth through photosynthetic organisms. One way to analyse the atmosphere of an exoplanet is through spectrography when it transits its star, though this might only be feasible with dim stars like white dwarfs. History and cultural impact The modern concept of extraterrestrial life is based on assumptions that were not commonplace during the early days of astronomy. The first explanations for the celestial objects seen in the night sky were based on mythology. Scholars from Ancient Greece were the first to consider that the universe is inherently understandable and rejected explanations based on supernatural incomprehensible forces, such as the myth of the Sun being pulled across the sky in the chariot of Apollo. They had not developed the scientific method yet and based their ideas on pure thought and speculation, but they developed precursor ideas to it, such as that explanations had to be discarded if they contradict observable facts. The discussions of those Greek scholars established many of the pillars that would eventually lead to the idea of extraterrestrial life, such as Earth being round and not flat. The cosmos was first structured in a geocentric model that considered that the sun and all other celestial bodies revolve around Earth. However, they did not consider them as worlds. In Greek understanding, the world was composed by both Earth and the celestial objects with noticeable movements. Anaximander thought that the cosmos was made from apeiron, a substance that created the world, and that the world would eventually return to the cosmos. Eventually two groups emerged, the atomists that thought that matter at both Earth and the cosmos was equally made of small atoms of the classical elements (earth, water, fire and air), and the Aristotelians who thought that those elements were exclusive of Earth and that the cosmos was made of a fifth one, the aether. Atomist Epicurus thought that the processes that created the world, its animals and plants should have created other worlds elsewhere, along with their own animals and plants. Aristotle thought instead that all the earth element naturally fell towards the center of the universe, and that would make it impossible for other planets to exist elsewhere. Under that reasoning, Earth was not only in the center, it was also the only planet in the universe. Cosmic pluralism, the plurality of worlds, or simply pluralism, describes the philosophical belief in numerous "worlds" in addition to Earth, which might harbor extraterrestrial life. The earliest recorded assertion of extraterrestrial human life is found in ancient scriptures of Jainism. There are multiple "worlds" mentioned in Jain scriptures that support human life. These include, among others, Bharat Kshetra, Mahavideh Kshetra, Airavat Kshetra, and Hari kshetra. Medieval Muslim writers like Fakhr al-Din al-Razi and Muhammad al-Baqir supported cosmic pluralism on the basis of the Qur'an. Chaucer's poem The House of Fame engaged in medieval thought experiments that postulated the plurality of worlds. However, those ideas about other worlds were different from the current knowledge about the structure of the universe, and did not postulate the existence of planetary systems other than the Solar System. When those authors talk about other worlds, they talk about places located at the center of their own systems, and with their own stellar vaults and cosmos surrounding them. The Greek ideas and the disputes between atomists and Aristotelians outlived the fall of the Greek empire. The Great Library of Alexandria compiled information about it, part of which was translated by Islamic scholars and thus survived the end of the Library. Baghdad combined the knowledge of the Greeks, the Indians, the Chinese and its own scholars, and the knowledge expanded through the Byzantine Empire. From there it eventually returned to Europe by the time of the Middle Ages. However, as the Greek atomist doctrine held that the world was created by random movements of atoms, with no need for a creator deity, it became associated with atheism, and the dispute intertwined with religious ones. Still, the Church did not react to those topics in a homogeneous way, and there were stricter and more permissive views within the church itself. The first known mention of the term 'panspermia' was in the writings of the 5th-century BC Greek philosopher Anaxagoras. He proposed the idea that life exists everywhere. By the time of the late Middle Ages there were many known inaccuracies in the geocentric model, but it was kept in use because naked eye observations provided limited data. Nicolaus Copernicus started the Copernican Revolution by proposing that the planets revolve around the sun rather than Earth. His proposal had little acceptance at first because, as he kept the assumption that orbits were perfect circles, his model led to as many inaccuracies as the geocentric one. Tycho Brahe improved the available data with naked-eye observatories, which worked with highly complex sextants and quadrants. Tycho could not make sense of his observations, but Johannes Kepler did: orbits were not perfect circles, but ellipses. This knowledge benefited the Copernican model, which worked now almost perfectly. The invention of the telescope a short time later, perfected by Galileo Galilei, clarified the final doubts, and the paradigm shift was completed. Under this new understanding, the notion of extraterrestrial life became feasible: if Earth is but just a planet orbiting around a star, there may be planets similar to Earth elsewhere. The astronomical study of distant bodies also proved that physical laws are the same elsewhere in the universe as on Earth, with nothing making the planet truly special. The new ideas were met with resistance from the Catholic church. Galileo was tried for the heliocentric model, which was considered heretical, and forced to recant it. The best-known early-modern proponent of ideas of extraterrestrial life was the Italian philosopher Giordano Bruno, who argued in the 16th century for an infinite universe in which every star is surrounded by its own planetary system. Bruno wrote that other worlds "have no less virtue nor a nature different to that of our earth" and, like Earth, "contain animals and inhabitants". Bruno's belief in the plurality of worlds was one of the charges leveled against him by the Venetian Holy Inquisition, which tried and executed him. The heliocentric model was further strengthened by the postulation of the theory of gravity by Sir Isaac Newton. This theory provided the mathematics that explains the motions of all things in the universe, including planetary orbits. By this point, the geocentric model was definitely discarded. By this time, the use of the scientific method had become a standard, and new discoveries were expected to provide evidence and rigorous mathematical explanations. Science also took a deeper interest in the mechanics of natural phenomena, trying to explain not just the way nature works but also the reasons for working that way. There was very little actual discussion about extraterrestrial life before this point, as the Aristotelian ideas remained influential while geocentrism was still accepted. When it was finally proved wrong, it not only meant that Earth was not the center of the universe, but also that the lights seen in the sky were not just lights, but physical objects. The notion that life may exist in them as well soon became an ongoing topic of discussion, although one with no practical ways to investigate. The possibility of extraterrestrials remained a widespread speculation as scientific discovery accelerated. William Herschel, the discoverer of Uranus, was one of many 18th–19th-century astronomers who believed that the Solar System is populated by alien life. Other scholars of the period who championed "cosmic pluralism" included Immanuel Kant and Benjamin Franklin. At the height of the Enlightenment, even the Sun and Moon were considered candidates for extraterrestrial inhabitants. Speculation about life on Mars increased in the late 19th century, following telescopic observation of apparent Martian canals – which soon, however, turned out to be optical illusions. Despite this, in 1895, American astronomer Percival Lowell published his book Mars, followed by Mars and its Canals in 1906, proposing that the canals were the work of a long-gone civilisation. Spectroscopic analysis of Mars's atmosphere began in earnest in 1894, when U.S. astronomer William Wallace Campbell showed that neither water nor oxygen was present in the Martian atmosphere. By 1909 better telescopes and the best perihelic opposition of Mars since 1877 conclusively put an end to the canal hypothesis. As a consequence of the belief in the spontaneous generation there was little thought about the conditions of each celestial body: it was simply assumed that life would thrive anywhere. This theory was disproved by Louis Pasteur in the 19th century. Popular belief in thriving alien civilisations elsewhere in the solar system still remained strong until Mariner 4 and Mariner 9 provided close images of Mars, which debunked forever the idea of the existence of Martians and decreased the previous expectations of finding alien life in general. The end of the spontaneous generation belief forced investigation into the origin of life. Although abiogenesis is the more accepted theory, a number of authors reclaimed the term "panspermia" and proposed that life was brought to Earth from elsewhere. Some of those authors are Jöns Jacob Berzelius (1834), Kelvin (1871), Hermann von Helmholtz (1879) and, somewhat later, by Svante Arrhenius (1903). The science fiction genre, although not so named during the time, developed during the late 19th century. The expansion of the genre of extraterrestrials in fiction influenced the popular perception over the real-life topic, making people eager to jump to conclusions about the discovery of aliens. Science marched at a slower pace, some discoveries fueled expectations and others dashed excessive hopes. For example, with the advent of telescopes, most structures seen on the Moon or Mars were immediately attributed to Selenites or Martians, and later ones (such as more powerful telescopes) revealed that all such discoveries were natural features. A famous case is the Cydonia region of Mars, first imaged by the Viking 1 orbiter. The low-resolution photos showed a rock formation that resembled a human face, but later spacecraft took photos in higher detail that showed that there was nothing special about the site. The search and study of extraterrestrial life became a science of its own, astrobiology. Also known as exobiology, this discipline is studied by the NASA, the ESA, the INAF, and others. Astrobiology studies life from Earth as well, but with a cosmic perspective. For example, abiogenesis is of interest to astrobiology, not because of the origin of life on Earth, but for the chances of a similar process taking place in other celestial bodies. Many aspects of life, from its definition to its chemistry, are analyzed as either likely to be similar in all forms of life across the cosmos or only native to Earth. Astrobiology, however, remains constrained by the current lack of extraterrestrial life-forms to study, as all life on Earth comes from the same ancestor, and it is hard to infer general characteristics from a group with a single example to analyse. The 20th century came with great technological advances, speculations about future hypothetical technologies, and an increased basic knowledge of science by the general population thanks to science divulgation through the mass media. The public interest in extraterrestrial life and the lack of discoveries by mainstream science led to the emergence of pseudosciences that provided affirmative, if questionable, answers to the existence of aliens. Ufology claims that many unidentified flying objects (UFOs) would be spaceships from alien species, and ancient astronauts hypothesis claim that aliens would have visited Earth in antiquity and prehistoric times but people would have failed to understand it by then. Most UFOs or UFO sightings can be readily explained as sightings of Earth-based aircraft (including top-secret aircraft), known astronomical objects or weather phenomenons, or as hoaxes. Looking beyond the pseudosciences, Lewis White Beck strove to elevate the level of public discourse on the topic of extraterrestrial life by tracing the evolution of philosophical thought over the centuries from ancient times into the modern era. His review of the contributions made by Lucretius, Plutarch, Aristotle, Copernicus, Immanuel Kant, John Wilkins, Charles Darwin and Karl Marx demonstrated that even in modern times, humanity could be profoundly influenced in its search for extraterrestrial life by subtle and comforting archetypal ideas which are largely derived from firmly held religious, philosophical and existential belief systems. On a positive note, however, Beck further argued that even if the search for extraterrestrial life proves to be unsuccessful, the endeavor itself could have beneficial consequences by assisting humanity in its attempt to actualize superior ways of living here on Earth. By the 21st century, it was accepted that multicellular life in the Solar System can only exist on Earth, but the interest in extraterrestrial life increased regardless. This is a result of the advances in several sciences. The knowledge of planetary habitability allows to consider on scientific terms the likelihood of finding life at each specific celestial body, as it is known which features are beneficial and harmful for life. Astronomy and telescopes also improved to the point exoplanets can be confirmed and even studied, increasing the number of search places. Life may still exist elsewhere in the Solar System in unicellular form, but the advances in spacecraft allow to send robots to study samples in situ, with tools of growing complexity and reliability. Although no extraterrestrial life has been found and life may still be just a rarity from Earth, there are scientific reasons to suspect that it can exist elsewhere, and technological advances that may detect it if it does. Many scientists are optimistic about the chances of finding alien life. In the words of SETI's Frank Drake, "All we know for sure is that the sky is not littered with powerful microwave transmitters". Drake noted that it is entirely possible that advanced technology results in communication being carried out in some way other than conventional radio transmission. At the same time, the data returned by space probes, and giant strides in detection methods, have allowed science to begin delineating habitability criteria on other worlds, and to confirm that at least other planets are plentiful, though aliens remain a question mark. The Wow! signal, detected in 1977 by a SETI project, remains a subject of speculative debate. On the other hand, other scientists are pessimistic. Jacques Monod wrote that "Man knows at last that he is alone in the indifferent immensity of the universe, whence which he has emerged by chance". In 2000, geologist and paleontologist Peter Ward and astrobiologist Donald Brownlee published a book entitled Rare Earth: Why Complex Life is Uncommon in the Universe.[better source needed] In it, they discussed the Rare Earth hypothesis, in which they claim that Earth-like life is rare in the universe, whereas microbial life is common. Ward and Brownlee are open to the idea of evolution on other planets that is not based on essential Earth-like characteristics such as DNA and carbon. As for the possible risks, theoretical physicist Stephen Hawking warned in 2010 that humans should not try to contact alien life forms. He warned that aliens might pillage Earth for resources. "If aliens visit us, the outcome would be much as when Columbus landed in America, which didn't turn out well for the Native Americans", he said. Jared Diamond had earlier expressed similar concerns. On 20 July 2015, Hawking and Russian billionaire Yuri Milner, along with the SETI Institute, announced a well-funded effort, called the Breakthrough Initiatives, to expand efforts to search for extraterrestrial life. The group contracted the services of the 100-meter Robert C. Byrd Green Bank Telescope in West Virginia in the United States and the 64-meter Parkes Telescope in New South Wales, Australia. On 13 February 2015, scientists (including Geoffrey Marcy, Seth Shostak, Frank Drake and David Brin) at a convention of the American Association for the Advancement of Science, discussed Active SETI and whether transmitting a message to possible intelligent extraterrestrials in the Cosmos was a good idea; one result was a statement, signed by many, that a "worldwide scientific, political and humanitarian discussion must occur before any message is sent". Government responses The 1967 Outer Space Treaty and the 1979 Moon Agreement define rules of planetary protection against potentially hazardous extraterrestrial life. COSPAR also provides guidelines for planetary protection. A committee of the United Nations Office for Outer Space Affairs had in 1977 discussed for a year strategies for interacting with extraterrestrial life or intelligence. The discussion ended without any conclusions. As of 2010, the UN lacks response mechanisms for the case of an extraterrestrial contact. One of the NASA divisions is the Office of Safety and Mission Assurance (OSMA), also known as the Planetary Protection Office. A part of its mission is to "rigorously preclude backward contamination of Earth by extraterrestrial life." In 2016, the Chinese Government released a white paper detailing its space program. According to the document, one of the research objectives of the program is the search for extraterrestrial life. It is also one of the objectives of the Chinese Five-hundred-meter Aperture Spherical Telescope (FAST) program. In 2020, Dmitry Rogozin, the head of the Russian space agency, said the search for extraterrestrial life is one of the main goals of deep space research. He also acknowledged the possibility of existence of primitive life on other planets of the Solar System. The French space agency has an office for the study of "non-identified aero spatial phenomena". The agency is maintaining a publicly accessible database of such phenomena, with over 1600 detailed entries. According to the head of the office, the vast majority of entries have a mundane explanation; but for 25% of entries, their extraterrestrial origin can neither be confirmed nor denied. In 2020, chairman of the Israel Space Agency Isaac Ben-Israel stated that the probability of detecting life in outer space is "quite large". But he disagrees with his former colleague Haim Eshed who stated that there are contacts between an advanced alien civilisation and some of Earth's governments. In fiction Although the idea of extraterrestrial peoples became feasible once astronomy developed enough to understand the nature of planets, they were not thought of as being any different from humans. Having no scientific explanation for the origin of mankind and its relation to other species, there was no reason to expect them to be any other way. This was changed by the 1859 book On the Origin of Species by Charles Darwin, which proposed the theory of evolution. Now with the notion that evolution on other planets may take other directions, science fiction authors created bizarre aliens, clearly distinct from humans. A usual way to do that was to add body features from other animals, such as insects or octopuses. Costuming and special effects feasibility alongside budget considerations forced films and TV series to tone down the fantasy, but these limitations lessened since the 1990s with the advent of computer-generated imagery (CGI), and later on as CGI became more effective and less expensive. Real-life events sometimes captivate people's imagination and this influences the works of fiction. For example, during the Barney and Betty Hill incident, the first recorded claim of an alien abduction, the couple reported that they were abducted and experimented on by aliens with oversized heads, big eyes, pale grey skin, and small noses, a description that eventually became the grey alien archetype once used in works of fiction. See also Notes References Further reading External links |
======================================== |
[SOURCE: https://en.wikipedia.org/wiki/Walter_Brattain] | [TOKENS: 2532] |
Contents Walter Brattain Walter Houser Brattain (/ˈbrætn/ BRAT-n; February 10, 1902 – October 13, 1987) was an American solid-state physicist who shared the 1956 Nobel Prize in Physics with John Bardeen and William Shockley for their invention of the point-contact transistor. Brattain devoted much of his life to research on surface states. Early life and education Walter Houser Brattain was born on February 10, 1902, in Amoy (now Xiamen), China, to American parents, Ross R. Brattain and Ottilie Houser. His father was of Scottish descent, while his mother's parents were both immigrants from Stuttgart, Germany. Ross was a teacher at the Ting-Wen Institute,: 11 a private school for Chinese boys. Ottilie was a gifted mathematician. Both were graduates of Whitman College.: 71 Ottilie and baby Walter returned to the United States in 1903, and Ross followed shortly afterward.: 12 The family lived for several years in Spokane, Washington, then settled on a cattle ranch near Tonasket, Washington, in 1911.: 12 : 71 Brattain attended high school in Washington, spending one year at Queen Anne High School, two years at Tonasket High School, and one year at Moran School for Boys. He then attended Whitman College, where he studied under Benjamin H. Brown (physics) and Walter A. Bratton (mathematics). He received his B.S. in 1924 with a double major in Physics and Mathematics. Brattain and his classmates Walker Bleakney, Vladimir Rojansky, and E. John Workman would all go on to have distinguished careers, later becoming known as "the four horsemen of physics".: 71 Brattain's brother Robert, who followed him at Whitman College, also became a physicist.: 71 Brattain obtained an M.A. from the University of Oregon in 1926 and a Ph.D. from the University of Minnesota in 1929. At Minnesota, he had the opportunity to study the new field of quantum mechanics under John Van Vleck. His doctoral thesis, written under John T. Tate, was titled Efficiency of Excitation by Electron Impact and Anomalous Scattering in Mercury Vapor.: 72 Career and research From 1928 to 1929, Brattain worked for the National Bureau of Standards in Washington, D.C., where he helped to develop piezoelectric frequency standards. In August 1929, he joined Joseph A. Becker at Bell Telephone Laboratories as a research physicist. The two men worked on the heat-induced flow of charge carriers in copper oxide rectifiers.: 72 Brattain was able to attend a lecture by Arnold Sommerfeld. Some of their subsequent experiments on thermionic emission provided experimental validation for the Sommerfeld theory. They also did work on the surface state and work function of tungsten and the adsorption of thorium atoms.: 74 Through his studies of rectification and photo-effects on the semiconductor surfaces of cuprous oxide and silicon, Brattain discovered the photo-effect at the free surface of a semiconductor. This work was considered by the Nobel Committee to be one of his chief contributions to solid-state physics. At the time, the telephone industry was heavily dependent on the use of vacuum tubes to control electron flow and amplify current. Vacuum tubes were neither reliable nor efficient, and Bell Labs wanted to develop an alternative technology. As early as the 1930s Brattain worked with William Shockley on the idea of a semiconductor amplifier that used copper oxide, an early and unsuccessful attempt at creating a field-effect transistor. Other researchers at Bell and elsewhere were also experimenting with semiconductors, using materials such as germanium and silicon, but the pre-war research effort was somewhat haphazard and lacked strong theoretical grounding. During World War II, both Brattain and Shockley were separately involved in research on magnetic detection of submarines with the National Defense Research Committee at Columbia University. Brattain's group developed magnetometers sensitive enough to detect anomalies in the Earth's magnetic field caused by submarines.: 104 As a result of this work, in 1944, Brattain patented a design for a magnetometer head. In 1945, Bell Labs reorganized and created a group specifically to do fundamental research in solid-state physics, relating to communications technologies. Creation of the sub-department was authorized by the vice-president for research, Mervin Kelly. An interdisciplinary group, it was co-led by Shockley and Stanley O. Morgan.: 76 The new group was soon joined by John Bardeen. Bardeen was a close friend of Brattain's brother Robert, who had introduced John and Walter in the 1930s. They often played bridge and golf together.: 77 Bardeen was a quantum physicist, Brattain a gifted experimenter in materials science, and Shockley, the leader of their team, was an expert in solid-state physics. According to theories of the time, Shockley's field-effect transistor, a cylinder coated thinly with silicon and mounted close to a metal plate, should have worked. He ordered Brattain and Bardeen to find out why it wouldn't. During November and December, the two men carried out a variety of experiments, attempting to determine why Shockley's device wouldn't amplify. Bardeen was a brilliant theorist; Brattain, equally importantly, "had an intuitive feel for what you could do in semiconductors".: 40 Bardeen theorized that the failure to conduct might be the result of local variations in the surface state which trapped the charge carriers.: 467–468 Brattain and Bardeen eventually managed to create a small level of amplification by pushing a gold metal point into the silicon, and surrounding it with distilled water. Replacing silicon with germanium enhanced the amplification, but only for low frequency currents. On December 16, Brattain devised a method of placing two gold leaf contacts close together on a germanium surface. Brattain reported: "Using this double point contact, contact was made to a germanium surface that had been anodized to 90 volts, electrolyte washed off in H2O and then had some gold spots evaporated on it. The gold contacts were pressed down on the bare surface. Both gold contacts to the surface rectified nicely... One point was used as a grid and the other point as a plate. The bias (D.C.) on the grid had to be positive to get amplification." As described by Bardeen, "The initial experiments with the gold spot suggested immediately that holes were being introduced into the germanium block, increasing the concentration of holes near the surface. The names emitter and collector were chosen to describe this phenomenon. The only question was how the charge of the added holes was compensated. Our first thought was that the charge was compensated by surface states. Shockley later suggested that the charge was compensated by electrons in the bulk and suggested the junction transistor geometry... Later experiments carried out by Brattain and me showed that very likely both occur in the point-contact transistor.": 470 On December 23, 1947, Brattain, Bardeen, and Shockley demonstrated the first working transistor to their colleagues at Bell Labs. Amplifying small electrical signals and supporting the processing of digital information, the transistor is "the key enabler of modern electronics." Convinced by the 1947 demonstration that a major breakthrough was being made, Bell Labs focused intensively on what it now called the Surface States Project. Initially, strict secrecy was observed. Carefully restricted internal conferences within Bell Labs shared information about the work of Brattain, Bardeen, Shockley and others who were engaged in related research.: 471 Patents were registered, recording the invention of the point-contact transistor by Bardeen and Brattain. There was considerable anxiety over whether Ralph Bray and Seymour Benzer, studying resistance in germanium at Purdue University, might make a similar discovery and publish before Bell Labs.: 38–39 On June 30, 1948, Bell Labs held a press conference to publicly announce their discovery. They also adopted an open policy in which new knowledge was freely shared with other institutions. By doing so, they avoided classification of the work as a military secret, and made possible widespread research and development of transistor technology. Bell Labs organized several symposia, open to university, industry and military participants, which were attended by hundreds of scientists in September 1951, April 1952, and 1956. Representatives from international as well as domestic companies attended.: 471–472, 475–476 Shockley believed and stated that he should have received all the credit for the invention of the transistor. He actively excluded Bardeen and Brattain from new areas of research, in particular the junction transistor, which Shockley patented. Shockley's theory of the junction transistor was an "impressive achievement", pointing the way to future solid-state electronics, but it would be several years before its construction would become practically possible.: 43–44 Brattain transferred to another research group within Bell Labs, working with C. G. B. Garrett, and P. J. Boddy. He continued to study the surface properties of solids and the "transistor effect," so as to better understand the various factors underlying semiconductor behavior.: 79–81 Describing it as "an intolerable situation," Bardeen left Bell Labs in 1951 to go to the University of Illinois, where he eventually won a second Nobel Prize for his theory of superconductivity. Shockley left Bell Labs in 1953 and went on to form Shockley Semiconductor Laboratory. In 1956, the three men were jointly awarded the Nobel Prize in Physics by King Gustaf VI Adolf of Sweden "for their researches on semiconductors and their discovery of the transistor effect." Bardeen and Brattain were included for the discovery of the point-contact transistor; Shockley for the development of the junction transistor. Walter Brattain is credited as having said, when told of the award, "I certainly appreciate the honor. It is a great satisfaction to have done something in life and to have been recognized for it in this way. However, much of my good fortune comes from being in the right place, at the right time, and having the right sort of people to work with." Each of the three gave a lecture. Brattain spoke on Surface Properties of Semiconductors, Bardeen on Semiconductor Research Leading to the Point Contact Transistor, and Shockley on Transistor Technology Evokes New Physics. Brattain later collaborated with P. J. Boddy and P. N. Sawyer on several papers on electrochemical processes in living matter.: 80 He became interested in blood clotting after his son required heart surgery. He also collaborated with Whitman chemistry professor David Frasco, using phospholipid bilayers as a model to study the surface of living cells and their absorption processes. Brattain taught at Harvard University as a visiting lecturer in 1952 and at Whitman College as a visiting lecturer in 1962 and 1963, and a visiting professor beginning in 1963. Upon formally retiring from Bell Labs in 1967, he continued to teach at Whitman, becoming an adjunct professor in 1972. He retired from teaching in 1976 but continued to be a consultant at Whitman. At Whitman, the Walter Brattain Scholarships are awarded on a merit basis to "entering students who have achieved high academic excellence in their college preparatory work." All applicants for admission are considered for the scholarship, which is potentially renewable for four years. Personal life and death Brattain married twice. His first wife was chemist Keren Gilmore. They were married in 1935 and had a son, William, in 1943. Keren died on April 10, 1957. The following year, Brattain married Emma Jane (Kirsch) Miller, a mother of three children. Brattain died of Alzheimer's disease on October 13, 1987, in Seattle at the age of 85. He is buried in City Cemetery in Pomeroy, Washington. Recognition Notes References External links |
======================================== |
[SOURCE: https://en.wikipedia.org/wiki/History_of_the_Jews_in_Finland] | [TOKENS: 3678] |
Contents History of the Jews in Finland The history of the Jews in Finland goes back to the late 18th century. Many of the first Jews to arrive were nineteenth-century Russian soldiers known as cantonists who stayed in Finland after their military service ended. Jews were granted full civil rights as Finnish citizens in 1918, making Finland one of the last European countries to do so. During World War II, Finland's position as a co-belligerent with Nazi Germany against the Soviet Union presented a challenge for Finnish Jews who fought in the Finnish army alongside the Nazis. Finnish Jews were protected by Finnish authorities despite German pressure, but eight Austrian Jewish refugees were deported to Germany in 1942, a move that was considered a national scandal. Finnish government apologized for the extradition in 2000. The two synagogues in active use today in Finland were built by Jewish congregations in Helsinki and Turku in 1906 and 1912, respectively. The Vyborg Synagogue (built 1910–1911) was destroyed by Russian air bombings on 30 November 1939, the first day of the Winter War. Today, Finland is home to around 1,800 Jews, of which 1,400 live in the Greater Helsinki area and 200 in Turku. Finnish and Swedish are the most common mother tongues of Jews in Finland, and many also speak Yiddish, German, Russian or Hebrew. Since data collection began in 2008, incidents of antisemitism have been on the rise in Finland. The number of incidents are likely under-reported, as Finland does not have a systematic method for recording specific forms of hate speech that incite violence or hatred. History until 1809 Before Finland was annexed by Russia in 1809, it was a part of the Swedish realm. Swedish laws prohibited Jews from settling in the kingdom, a restriction that stemmed from the 1593 Uppsala Synod, which allowed only Lutheranism to be practiced in Sweden. The 1686 Church Law went even further, requiring Jews and other non-Christians to convert before being allowed to settle. Economic needs occasionally made room for exemptions, and the 1782 Judereglementet laws allowed wealthy Jews to settle in three Swedish cities – all of which fell outside the boundaries of modern-day Finland. In 1806, King Gustav IV Adolf further tightened policy by banning Jewish immigration altogether. As a result, no Jews officially resided in Finland before the 19th century, although some, like Aaron Isaac, conducted business in the province. An exception to this general rule was Old Finland, the southeastern parts of Finland ceded to Russia in the 18th century. Under Russian administration, Jews were allowed to settle legally in these areas. The earliest family to arrive was the Weikaim family, who arrived in Hamina (Fredrikshamn) from Daugavpils (then part of the Russian Empire) in 1799, and moved to Viipuri (Vyborg) in 1815. The Jews in Finland formed a small community involved in trades like tinsmithing and construction. Some of the early settlers converted to Christianity to facilitate their residence. Grand Duchy of Finland In 1809, Finland was ceded to the Russian Empire as an autonomous Grand Duchy. The Russian emperor confirmed the continued application of Swedish laws, including the ban on Jewish settlement. Ambiguities arose due to the presence of Jews already living legally in Old Finland. Some provincial governors in rest of Finland also used their discretion to grant settlement permits to Jewish individuals. Despite the legal difficulties, Russian Jews established themselves in Finland as tradesmen and craftsmen during the period of autonomy from 1809 to 1917, although their number remained small. In 1872, they numbered about 700 individuals, half of whom lived in Helsinki and the other half in Turku (Åbo) and Viipuri. The Jews who inhabited Finland were mostly former soldiers from the Imperial Russian army. These cantonists were forced into the Russian army in childhood and were required to serve at least 25 years. After completing their service, some chose to remain in the regions where they had been stationed. In 1858, a limited exemption was granted to retired Jewish soldiers and their families to allow them to remain in Finland. Finnish authorities often interpreted Jewish regulations more restrictively than their Russian counterparts, emphasizing Swedish-era legislation in order to underscore Finland's legal autonomy. When Jewish civil rights were expanded in Russia under Alexander II, Finland retained more conservative policies. In 1886, the Finnish Senate replaced the 1858 settlement decree with a system of six-month residence permits. The move paralleled Russian efforts under Alexander III and Nicholas I to restrict Jewish mobility and rights. Regulations from 1869 also restricted their right to work, and Jews typically earned a living selling second-hand clothes. In 1888, the Finnish authorities took an even harsher step by expelling Jewish families from certain regions. The first expulsion order targeted 12 Jewish families in Turku, and 34 families in Viipuri were also ordered to leave. This did not drastically change the number of Jews residing in Finland. Some of the expelled families relocated to the United States, while others moved to Palestine. Public debate over Jewish rights intensified in the late 19th century. Liberal voices, especially in the Swedish-language press, supported equal rights, while Finnish-language outlets often expressed concern over a potential influx of impoverished Eastern Jews. In 1872, Leo Mechelin proposed granting Jews full civil rights in the Diet of Finland. His initiative was rejected, particularly by the clerical estate, which feared mass immigration and cultural disruption. Jews were granted full rights as Finnish citizens after Finland had declared independence in 1917. The law came into force in January 1918, making Finland one of the last countries in Europe to grant Jews equal citizenship. Rabbi Naftali Zvi Amsterdam, one of the foremost disciples of Rabbi Yisrael Salanter and the Mussar Movement, served as chief rabbi of Helsinki under Rabbi Yisrael's instruction from 1867 to 1875. Jewish youths in Helsinki founded the sports association IK Stjärnan (later Makkabi Helsinki) in 1906, making it the oldest still-operating Jewish sports club in the world with an uninterrupted history. World War II Finland's involvement in World War II began during the Winter War (30 November 1939 – 13 March 1940), the Soviet Union's invasion of Finland. Finnish Jews evacuated Finnish Karelia alongside other locals. The Vyborg Synagogue was destroyed by air bombings within the first few days of the war. Finland resumed fighting the Soviet Union in the Continuation War (1941 – 1944), whose onset was timed to coincide with Germany's launch of Operation Barbarossa. This resulted in Finland fighting alongside Nazi Germany. 327 Finnish Jews fought for Finland during the war, including 242 rank-and-file soldiers, 52 non-commissioned officers, 18 officers, and 15 medical officers. 21 Jews served in the women's auxiliary Lotta Svärd. In total, 15 Finnish Jews were killed in action in the Winter War, and eight were killed in the Continuation War. As Finland's wartime operations were supported by substantial numbers of German forces, the Finnish front had a field synagogue operating in the presence of Nazi troops. Jewish soldiers were granted leave on Saturdays and Jewish holidays. Finnish Jewish soldiers later participated in the Lapland War against Germany.[citation needed] In November 1942, eight Jewish Austrian refugees (along with 19 others) were deported to Nazi Germany after the head of the Finnish police agreed to turn them over. Seven of the Jews were murdered immediately. According to author Martin Gilbert, these eight were: Georg Kollman; Frans Olof Kollman; Frans Kollman's mother; Hans Eduard Szubilski; Henrich Huppert; Kurt Huppert; Hans Robert Martin Korn, who had been a volunteer in the Winter War; and an unknown individual. When Finnish media reported the news, it caused a national scandal, and ministers resigned in protest. After protests by Lutheran ministers, an Archbishop, and the Social Democratic Party, no more foreign Jewish refugees were deported from Finland. In 2000, Finnish Prime Minister Paavo Lipponen issued an official apology for the extradition of the eight Jewish refugees. Approximately 500 Jewish refugees arrived in Finland during World War II, although about 350 moved on to other countries, including about 160 who were transferred to neutral Sweden for safety reasons on the direct orders of Finnish Army commander Marshal Carl Gustaf Emil Mannerheim. About 40 of the remaining Jewish refugees were forced into compulsory labor service in Salla in Lapland in March 1942. The refugees were moved to Kemijärvi in June and eventually to Suursaari Island in the Gulf of Finland. Although Heinrich Himmler visited Finland twice to try to persuade the authorities to hand over the Jewish population, he was unsuccessful. In 1942, an exchange of Soviet prisoners of war (POWs) took place between Finland and Germany. Approximately 2,600–2,800 Soviet POWs of various nationalities then held by Finland were exchanged for 2,100 Soviet POWs of Baltic Finnic nationalities (Finnish, Karelian, Ingrian, or Estonian) held by Germany, who might have volunteered in the Finnish army. About 2,000 of the POWs handed over by Finland joined the Wehrmacht. Among the rest, there were about 500 people (mainly Soviet political officers) who were considered politically dangerous in Finland. This latter group most likely perished in concentration camps or were executed following guidelines set by the Commissar Order. 47 Jews appear on the list of those extradited, although religion was not a determining factor in extradition. Jews with Finnish citizenship were protected during the war. Late in the conflict, Germany's ambassador to Helsinki Wipert von Blücher concluded in a report to Hitler that Finns would not endanger their citizens of Jewish origin in any situation. Three Finnish Jews were offered the Iron Cross for their wartime service: Leo Skurnik, Salomon Klass, and Dina Poljakoff. Major Leo Skurnik, a district medical officer in the Finnish Army, organized an evacuation of a German field hospital when it came under Soviet shelling. More than 600 patients, including SS soldiers, were evacuated. Captain Salomon Klass, also of the Finnish Army, led a Finnish unit that rescued a German company from encirclement by the Soviets. Dina Poljakoff, a member of Lotta Svärd, the Finnish women's auxiliary service, was a nursing assistant who helped tend to German wounded and came to be greatly admired by her patients. All three refused the award.[dead link] The then-President of Finland, Marshal Mannerheim, attended the memorial service for fallen Finnish Jews at the Helsinki Synagogue on 6 December 1944. After World War II During the 1948 Arab–Israeli War, about 28 Finnish Jews, mostly Finnish Army veterans, fought for the State of Israel. After Israel's establishment, Finland had a high rate of immigration to Israel (known as "aliyah"), which led to a shrinking Jewish population. The community was partly revitalized when some Soviet Jews immigrated to Finland following the collapse of the Soviet Union. As of 2020, the number of Jews in Finland was approximately 1,800, of whom 1,400 lived in Helsinki, about 200 in Turku, and about 50 in Tampere. Jews are well integrated into Finnish society and are represented in nearly all sectors. Most Finnish Jews are corporate employees or self-employed professionals. Most Finnish Jews speak Finnish or Swedish as their mother tongue. Yiddish, German, Russian, and Hebrew are also spoken in the community. The Jews, like Finland's other traditional minorities as well as immigrant groups, are represented on the Advisory Board for Ethnic Relations. There are two synagogues still standing in Finland: one in Helsinki and one in Turku. Helsinki also has a Jewish day school, which serves about 110 students (many of whom are the children of Israelis working in Finland); and a Chabad Lubavitch rabbi is based in the city. Tampere previously had an organized Jewish community, but it stopped functioning in 1981. The other two cities continue to run their community organizations. There are also some Reform Jewish movements in Finland today. Historically, antisemitic hate crimes have been rare, and the Jewish community has been relatively safe.[citation needed] However, there have been some antisemitic crimes reported in the last decade;[timeframe?] the most common types include defamation, verbal threats, and damage to property. In 2011, Ben Zyskowicz, the first Finnish Jewish parliamentarian, was assaulted by a man shouting antisemitic slurs. Four years later, a few campaign advertisements containing Zyskowicz's picture were sprayed with swastikas in Helsinki. In 2023, Zyskowicz was attacked by a man who shouted insults about NATO, Jews and immigrants. In 2015 the Fundamental Rights Agency published its annual overview of data on antisemitism available in the European Union, including information from a report by the Police College of Finland. The semi-frequent report covers religiously motivated hate crimes, including antisemitic crimes. The most recently-documented data is from 2013, when most of the incidents (six out of eleven) concerned verbal threats or harassments. In May 2024, the European Jewish Congress prepared a report titled “Experiences and Views of Antisemitism in Finland – A Report on Discrimination and Hate Crime Targeting Jews" to investigate the rising levels of antisemitism in Finland. The survey respondents consisted of persons over the age of 16 who live in Finland and identify as Jewish. The report was prepared by researchers at the Polin Institute in collaboration with Åbo Akademi University and the Finnish Ministry of Justice. According to the report, over 80% of respondents believed that antisemitism has increased in the past 5 years, while over 70% of respondents stated that Finnish people blame Jewish people for the actions of the Israeli government. Holocaust denial started in Finland almost immediately after the war, with many Finns who had been involved in the far-right and Nazi movements publishing articles questioning the Holocaust. Prominent early Finnish Holocaust deniers include professor C. A. J. Gadolin, CEO Carl-Gustaf Herlitz, architech Carl O. Nordling and ambassador Teo Snellman. In early 1970s, a Finnish translation of a pamphlet denying the Holocaust written by Vera Oredsson was distributed in Finland. Pekka Siitoin's Patriotic Popular Front started distributing a Finnish translation of Richard Harwood's Did Six Million Really Die? in 1976. In the late 1980 and early 1990s, the newspaper Uusi Suunta published by the National Radical Party [fi] wrote how the supposed Jewish owned media maligned fascism with Holocaust "sob stories" that "have never been proven" to undermine the nationalist movement. One issue of Uusi Suunta stated that "The horrors of the concentration camps, that are not true by half, have been tied to the nationalist movement regardless of nationality." In the 1990s the party secretary of the National Democratic Party Olavi Koskela said that "the unbelievable Holocaust lies" enable the Jewish rulership over multicultural, multiracial and multilingual society. Finland - Fatherland's newspaper Kansallinen Rintama made similar arguments. Antisemitism has experienced a resurgence after the Cold War, both in the internet and real life. Medical Licentiate Vesa-Ilkka Laurio wrote a blog frequently denying the Holocaust and criticizing democracy from a Christian fundamentalist perspective. Swedenborgian Nova Hierosolyma society's Uusi Jerusalem website also publishes Holocaust denial material, an article written by Erkki Kivilohkare claiming only 100,000 Jews died in the Holocaust. Another Swedenborgian fundamentalist Markku Juutinen has also denied holocaust in the Kumouksen ääni magazine. In 2013, the newspaper Magneettimedia and its editor-in-chief, department store tycoon Juha Kärkkäinen, were convicted of agitation against a population group in connection with antisemitic content published in the newspaper. Originally launched as a customer and advertising paper, Magneettimedia was distributed to hundreds of thousands of households through the mail and the Kärkkäinen department store chain. It combined alternative health content and conspiracy theories with increasingly explicit antisemitic material. The newspaper published articles denying the Holocaust and articles such as "Zionist terrorism" and "CNN, Goldman Sachs and Zionist Control" translated from David Duke. Following the conviction, Juha Kärkkäinen withdrew from the publication, and the newspaper continued as an online publication under the control of the association Pohjoinen Perinne, whose leadership has been linked to the neo-Nazi Nordic Resistance Movement. Members of the movement have also distributed Holocaust-denial material in schools. Other alternative media sites, including MV-media and Verkkomedia, are known for publishing articles denying the Holocaust. Far-right Finns Party General Secretary Olli Immonen has also multiple times shared a blog criticizing the "Holocaust-religion". The Finns Party leader Jussi Halla-aho also referred to Holocaust as "holohoax". Pseudonymous Thomas Dalton who is a prolific author of Holocaust denial books and has republished On the Jews and Their Lies and The Protocols of the Elders of Zion is suspected of being a researcher in University of Helsinki according to Demokraatti. See also References Further reading External links |
======================================== |
[SOURCE: https://en.wikipedia.org/wiki/Arthropoda] | [TOKENS: 9070] |
Contents Arthropod For fossil groups, see text Arthropods (/ˈɑːrθrəˌpɒd/ AR-thrə-pod) are invertebrates in the phylum Arthropoda. They possess an exoskeleton with a cuticle made of chitin, often mineralised with calcium carbonate, a body with differentiated (metameric) segments, and paired jointed appendages. In order to keep growing, they must go through stages of moulting, a process by which they shed their exoskeleton to reveal a new one. They form an extremely diverse group of up to ten million species. Haemolymph is the analogue of blood for most arthropods. An arthropod has an open circulatory system, with a body cavity called a haemocoel through which haemolymph circulates to the interior organs. Like their exteriors, the internal organs of arthropods are generally built of repeated segments. They have ladder-like nervous systems, with paired ventral nerve cords running through all segments and forming paired ganglia in each segment. Their heads are formed by fusion of varying numbers of segments, and their brains are formed by fusion of the ganglia of these segments and encircle the esophagus. The respiratory and excretory systems of arthropods vary, depending as much on their environment as on the subphylum to which they belong. Arthropods use combinations of compound eyes and pigment-pit ocelli for vision. In most species, the ocelli can only detect the direction from which light is coming, and the compound eyes are the main source of information; however, in spiders, the main eyes are ocelli that can form images and, in a few cases, can swivel to track prey. Arthropods also have a wide range of chemical and mechanical sensors, mostly based on modifications of the many bristles known as setae that project through their cuticles. Similarly, their reproduction and development are varied; all terrestrial species use internal fertilization, but this is sometimes by indirect transfer of the sperm via an appendage or the ground, rather than by direct injection. Aquatic species use either internal or external fertilization. Almost all arthropods lay eggs, with many species giving birth to live young after the eggs have hatched inside the mother. Still, a few are genuinely viviparous, such as aphids. Arthropod hatchlings vary from miniature adults to grubs and caterpillars that lack jointed limbs and eventually undergo a total metamorphosis to produce the adult form. The level of maternal care for hatchlings varies from nonexistent to the prolonged care provided by social insects. The evolutionary ancestry of arthropods dates back to the Cambrian period. The group is generally regarded as monophyletic, and many analyses support the placement of arthropods with cycloneuralians (or their constituent clades) in a superphylum Ecdysozoa. Overall, however, the basal relationships of animals are not yet well resolved. Likewise, the relationships between various arthropod groups are still actively debated. Today, arthropods contribute to the human food supply both directly as food and more importantly, indirectly as pollinators of crops. Some species are known to spread severe disease to humans, livestock, and crops. Etymology The word arthropod comes from the Greek ἄρθρον árthron 'joint', and πούς poús (gen. ποδός podós) 'foot' or 'leg', which together mean "jointed leg", with the word "arthropodes" initially used in anatomical descriptions by Barthélemy Charles Joseph Dumortier published in 1832. The designation "Arthropoda" appears to have been first used in 1843 by the German zoologist Johann Ludwig Christian Gravenhorst (1777–1857). The origin of the name has been the subject of considerable confusion, with credit often given erroneously to Pierre André Latreille or Karl Theodor Ernst von Siebold instead, among various others. Terrestrial arthropods are often called bugs.[Note 1] The term is also occasionally extended to colloquial names for freshwater or marine crustaceans (e.g., Balmain bug, Moreton Bay bug, mudbug) and used by physicians and bacteriologists for disease-causing germs (e.g., superbugs), but entomologists reserve this term for a narrow category of "true bugs", insects of the order Hemiptera. Description Arthropods are invertebrates with segmented bodies and jointed limbs. The exoskeleton or cuticle consists of chitin, a polymer of N-Acetylglucosamine. The cuticle of many crustaceans, beetle mites, the clades Penetini and Archaeoglenini inside the beetle subfamily Phrenapatinae, and millipedes (except for bristly millipedes) is also biomineralized with calcium carbonate. Calcification of the endosternite, an internal structure used for muscle attachments, also occurs in some opiliones, and the pupal cuticle of the fly Bactrocera dorsalis contains calcium phosphate. Arthropoda is the largest animal phylum, with the estimates of the number of arthropod species varying from 1,170,000 to 5~10 million and accounting for over 80 percent of all known living animal species. One arthropod sub-group, the insects, includes more described species than any other taxonomic class. The total number of species remains difficult to determine, as estimates rely on census counts at specific locations, scaled up and projected onto other regions, then totalled – allowing for double-counting – to cover the whole world. Modeling assumptions are involved at each stage, introducing uncertainty. A study in 1992 estimated that there were 500,000 species of animals and plants in Costa Rica alone, of which 365,000 were arthropods. They are important members of marine, freshwater, land and air ecosystems and one of only two major animal groups that have adapted to life in dry environments; the other is amniotes, whose living members are reptiles, birds and mammals. Both the smallest and largest arthropods are crustaceans. The smallest belong to the class Tantulocarida, some of which are less than 100 micrometres (0.0039 in) long. The largest are species in the class Malacostraca, with the legs of the Japanese spider crab potentially spanning up to 4 metres (13 ft) and the American lobster reaching weights over 20 kg (44 lbs). The embryos of all arthropods are segmented, consisting of a series of repeated modules. The last common ancestor of living arthropods probably consisted of a series of undifferentiated segments, each with a pair of appendages that functioned as limbs. However, all known living and fossil arthropods have grouped segments into tagmata in which segments and their limbs are specialized in various ways. The three-part appearance of many insect bodies and the two-part appearance of spiders is a result of this grouping. There are no external signs of segmentation in mites. Arthropods also have two body elements that are not part of this serially repeated pattern of segments, an ocular somite at the front, where the mouth and eyes originated, and a telson at the rear, behind the anus. Originally, it seems that each appendage-bearing segment had two separate pairs of appendages: an upper, unsegmented exite and a lower, segmented endopod. These would later fuse into a single pair of biramous appendages united by a basal segment (protopod or basipod), with the upper branch acting as a gill while the lower branch was used for locomotion. The appendages of most crustaceans and some extinct taxa such as trilobites have another segmented branch known as exopods, but whether these structures have a single origin remain controversial. In some segments of all known arthropods, the appendages have been modified, for example to form gills, mouth-parts, antennae for collecting information, or claws for grasping; arthropods are "like Swiss Army knives, each equipped with a unique set of specialized tools." In many arthropods, appendages have vanished from some regions of the body; it is particularly common for abdominal appendages to have disappeared or be highly modified. The most conspicuous specialization of segments is in the head. The four major groups of arthropods – Chelicerata (sea spiders, horseshoe crabs and arachnids), Myriapoda (symphylans, pauropods, millipedes and centipedes), Pancrustacea (oligostracans, copepods, malacostracans, branchiopods, hexapods, etc.), and the extinct Trilobita – have heads formed of various combinations of segments, with appendages that are missing or specialized in different ways. Despite myriapods and hexapods both having similar head combinations, hexapods are deeply nested within crustacea while myriapods are not, so these traits are believed to have evolved separately. In addition, some extinct arthropods, such as Marrella, belong to none of these groups, as their heads are formed by their own particular combinations of segments and specialized appendages. Working out the evolutionary stages by which all these different combinations could have appeared is so difficult that it has long been known as "The arthropod head problem". In 1960, R. E. Snodgrass even hoped it would not be solved, as he found trying to work out solutions to be fun.[Note 2] Arthropod exoskeletons are made of cuticle, a non-cellular material secreted by the epidermis. Their cuticles vary in the details of their structure, but generally consist of three main layers: the epicuticle, a thin outer waxy coat that moisture-proofs the other layers and gives them some protection; the exocuticle, which consists of chitin and chemically hardened proteins; and the endocuticle, which consists of chitin and unhardened proteins. The exocuticle and endocuticle together are known as the procuticle. Each body segment and limb section is encased in hardened cuticle. The joints between body segments and between limb sections are covered by flexible cuticle. The exoskeletons of most aquatic crustaceans are biomineralized with calcium carbonate extracted from the water. Some terrestrial crustaceans have developed means of storing the mineral, since on land they cannot rely on a steady supply of dissolved calcium carbonate. Biomineralization generally affects the exocuticle and the outer part of the endocuticle. Two recent hypotheses about the evolution of biomineralization in arthropods and other groups of animals propose that it provides tougher defensive armor, and that it allows animals to grow larger and stronger by providing more rigid skeletons; and in either case a mineral-organic composite exoskeleton is cheaper to build than an all-organic one of comparable strength. The cuticle may have setae (bristles) growing from special cells in the epidermis. Setae are as varied in form and function as appendages. For example, they are often used as sensors to detect air or water currents, or contact with objects; aquatic arthropods use feather-like setae to increase the surface area of swimming appendages and to filter food particles out of water; aquatic insects, which are air-breathers, use thick felt-like coats of setae to trap air, extending the time they can spend under water; heavy, rigid setae serve as defensive spines. Although all arthropods use muscles attached to the inside of the exoskeleton to flex their limbs, some still use hydraulic pressure to extend them, a system inherited from their pre-arthropod ancestors; for example, all spiders extend their legs hydraulically and can generate pressures up to eight times their resting level. The exoskeleton cannot stretch and thus restricts growth. Arthropods, therefore, replace their exoskeletons by undergoing ecdysis (moulting), or shedding the old exoskeleton, the exuviae, after growing a new one that is not yet hardened. Moulting cycles run nearly continuously until an arthropod reaches full size. The developmental stages between each moult (ecdysis) until sexual maturity is reached is called an instar. Differences between instars can often be seen in altered body proportions, colors, patterns, changes in the number of body segments or head width. After moulting, i.e. shedding their exoskeleton, the juvenile arthropods continue in their life cycle until they either pupate or moult again. In the initial phase of moulting, the animal stops feeding and its epidermis releases moulting fluid, a mixture of enzymes that digests the endocuticle and thus detaches the old cuticle. This phase begins when the epidermis has secreted a new epicuticle to protect it from the enzymes, and the epidermis secretes the new exocuticle while the old cuticle is detaching. When this stage is complete, the animal's body swells by taking in a large quantity of water or air, causing the old cuticle to split along predefined weaknesses where the old exocuticle was thinnest. It commonly takes several minutes for the animal to struggle out of the old cuticle. At this point, the new one is wrinkled and so soft that the animal cannot support itself and finds it very difficult to move, and the new endocuticle has not yet formed. The animal continues to pump itself up to stretch the new cuticle as much as possible, then hardens the new exocuticle and eliminates the excess air or water. By the end of this phase, the new endocuticle has formed. Many arthropods then eat the discarded cuticle to reclaim its materials. Because arthropods are unprotected and nearly immobilized until the new cuticle has hardened, they are in danger both of being trapped in the old cuticle and of being attacked by predators. Moulting may be responsible for 80 to 90% of all arthropod deaths. Arthropod bodies are also segmented internally, and the nervous, muscular, circulatory, and excretory systems have repeated components. Arthropods come from a lineage of animals that have a coelom, a membrane-lined cavity between the gut and the body wall that accommodates the internal organs. The strong, segmented limbs of arthropods eliminate the need for one of the coelom's main ancestral functions, as a hydrostatic skeleton, which muscles compress in order to change the animal's shape and thus enable it to move. Hence the coelom of the arthropod is reduced to small areas around the reproductive and excretory systems. Its place is largely taken by a hemocoel, a cavity that runs most of the length of the body and through which blood flows. Arthropods have open circulatory systems. Most have a few short, open-ended arteries. In chelicerates and crustaceans, the blood carries oxygen to the tissues, while hexapods use a separate system of tracheae. Many crustaceans and a few chelicerates and tracheates use respiratory pigments to assist oxygen transport. The most common respiratory pigment in arthropods is copper-based hemocyanin; this is used by many crustaceans and a few centipedes. A few crustaceans and insects use iron-based hemoglobin, the respiratory pigment used by vertebrates. As with other invertebrates, the respiratory pigments of those arthropods that have them are generally dissolved in the blood and rarely enclosed in corpuscles as they are in vertebrates. The heart is a muscular tube that runs just under the back and for most of the length of the hemocoel. It contracts in ripples that run from rear to front, pushing blood forwards. Sections not being squeezed by the heart muscle are expanded either by elastic ligaments or by small muscles, in either case connecting the heart to the body wall. Along the heart run a series of paired ostia, non-return valves that allow blood to enter the heart but prevent it from leaving before it reaches the front. Arthropods have a wide variety of respiratory systems. Small species often do not have any, since their high ratio of surface area to volume enables simple diffusion through the body surface to supply enough oxygen. Crustacea usually have gills that are modified appendages. Many arachnids have book lungs. Tracheae, systems of branching tunnels that run from the openings in the body walls, deliver oxygen directly to individual cells in many insects, myriapods and arachnids. Living arthropods have paired main nerve cords running along their bodies below the gut, and in each segment the cords form a pair of ganglia from which sensory and motor nerves run to other parts of the segment. Although the pairs of ganglia in each segment often appear physically fused, they are connected by commissures (relatively large bundles of nerves), which give arthropod nervous systems a characteristic ladder-like appearance. The brain is in the head, encircling and mainly above the esophagus. It consists of the fused ganglia of the acron and one or two of the foremost segments that form the head – a total of three pairs of ganglia in most arthropods, but only two in chelicerates, which do not have antennae or the ganglion connected to them. The ganglia of other head segments are often close to the brain and function as part of it. In insects, these other head ganglia combine into a pair of subesophageal ganglia, under and behind the esophagus. Spiders take this process a step further, as all the segmental ganglia are incorporated into the subesophageal ganglia, which occupy most of the space in the cephalothorax (front "super-segment"). There are two different types of arthropod excretory systems. In aquatic arthropods, the end-product of biochemical reactions that metabolise nitrogen is ammonia, which is so toxic that it needs to be diluted as much as possible with water. The ammonia is then eliminated via any permeable membrane, mainly through the gills. All crustaceans use this system, and its high consumption of water may be responsible for the relative lack of success of crustaceans as land animals. Various groups of terrestrial arthropods have independently developed a different system: the end-product of nitrogen metabolism is uric acid, which can be excreted as dry material; the Malpighian tubule system filters the uric acid and other nitrogenous waste out of the blood in the hemocoel, and dumps these materials into the hindgut, from which they are expelled as feces. Most aquatic arthropods and some terrestrial ones also have organs called nephridia ("little kidneys"), which extract other wastes for excretion as urine. The stiff cuticles of arthropods would block out information about the outside world, except that they are penetrated by many sensors or connections from sensors to the nervous system. In fact, arthropods have modified their cuticles into elaborate arrays of sensors. Various touch sensors, mostly setae, respond to different levels of force, from strong contact to very weak air currents. Chemical sensors provide equivalents of taste and smell, often by means of setae. Pressure sensors often take the form of membranes that function as eardrums, but are connected directly to nerves rather than to auditory ossicles. The antennae of most hexapods include sensor packages that monitor humidity, moisture and temperature. Most arthropods lack balance and acceleration sensors, and rely on their eyes to tell them which way is up. The self-righting behavior of cockroaches is triggered when pressure sensors on the underside of the feet report no pressure. However, many malacostracan crustaceans have statocysts, which provide the same sort of information as the balance and motion sensors of the vertebrate inner ear. The proprioceptors of arthropods, sensors that report the force exerted by muscles and the degree of bending in the body and joints, are well understood. However, little is known about what other internal sensors arthropods may have. Arthropods may have sophisticated visual systems that include one or more usually both of compound eyes and pigment-cup ocelli, ("little eyes"). In most cases, ocelli are only capable of detecting the direction from which light is coming, using the shadow cast by the walls of the cup, although the main eyes of spiders are pigment-cup ocelli that are capable of forming images, and those of jumping spiders can rotate to track prey. Compound eyes consist of fifteen to several thousand independent ommatidia, columns that are usually hexagonal in cross section. Each ommatidium is an independent sensor, with its own light-sensitive cells and often with its own lens and cornea. Compound eyes have a wide field of view, and can detect fast movement and, in some cases, the polarization of light. On the other hand, the relatively large size of ommatidia makes the images rather coarse, and compound eyes are shorter-sighted than those of birds and mammals – although this is not a severe disadvantage, as objects and events within 20 cm (8 in) are most important to most arthropods. Several arthropods have color vision, and that of some insects has been studied in detail; for example, the ommatidia of bees contain receptors for both green and ultra-violet. Reproduction and development A few arthropods, such as barnacles, are hermaphroditic, that is, each can have the organs of both sexes. However, individuals of most species remain of one sex their entire lives. A few species of insects and crustaceans can reproduce by parthenogenesis, especially if conditions favor a "population explosion". However, most arthropods rely on sexual reproduction, and parthenogenetic species often revert to sexual reproduction when conditions become less favorable. The ability to undergo meiosis is widespread among arthropods including both those that reproduce sexually and those that reproduce parthenogenetically. Although meiosis is a major characteristic of arthropods, understanding of its fundamental adaptive benefit has long been regarded as an unresolved problem, that appears to have remained unsettled. Aquatic arthropods may breed by external fertilization, as for example horseshoe crabs do, or by internal fertilization, where the ova remain in the female's body and the sperm must somehow be inserted. All known terrestrial arthropods use internal fertilization. Opiliones (harvestmen), millipedes, and some crustaceans use modified appendages such as gonopods or penises to transfer the sperm directly to the female. However, most male terrestrial arthropods produce spermatophores, waterproof packets of sperm, which the females take into their bodies. A few such species rely on females to find spermatophores that have already been deposited on the ground, but in most cases males only deposit spermatophores when complex courtship rituals look likely to be successful. Most arthropods lay eggs, but scorpions are ovoviviparous: they produce live young after the eggs have hatched inside the mother, and are noted for prolonged maternal care. Newly born arthropods have diverse forms, and insects alone cover the range of extremes. Some hatch as apparently miniature adults (direct development), and in some cases, such as silverfish, the hatchlings do not feed and may be helpless until after their first moult. Many insects (Holometabola) hatch as grubs or caterpillars, which do not have segmented limbs or hardened cuticles, and metamorphose into adult forms by entering an inactive phase in which the larval tissues are broken down and re-used to build the adult body. Dragonfly larvae have the typical cuticles and jointed limbs of arthropods but are flightless water-breathers with extendable jaws. Crustaceans commonly hatch as tiny nauplius larvae that have only three segments and pairs of appendages. Evolutionary history Based on the distribution of shared plesiomorphic features in extant and fossil taxa, the last common ancestor of all arthropods is inferred to have been as a modular organism with each module covered by its own sclerite (armor plate) and bearing a pair of biramous limbs. However, whether the ancestral limb was uniramous or biramous is far from a settled debate. This Ur-arthropod had a ventral mouth, pre-oral antennae and dorsal eyes at the front of the body. It was assumed to have been a non-discriminatory sediment feeder, processing whatever sediment came its way for food, but fossil findings hint that the last common ancestor of both arthropods and Priapulida shared the same specialized mouth apparatus: a circular mouth with rings of teeth used for capturing animal prey. It has been proposed that the Ediacaran animals Parvancorina and Spriggina, from around 555 million years ago, were arthropods, but later study shows that their affinities of being origin of arthropods are not reliable. Small arthropods with bivalve-like shells have been found in Early Cambrian fossil beds dating 541 to 539 million years ago in China and Australia. The earliest Cambrian trilobite fossils are about 520 million years old, but the class was already quite diverse and worldwide, suggesting that they had been around for quite some time. In the Maotianshan shales, which date back to 518 million years ago, arthropods such as Kylinxia and Erratus have been found that seem to represent transitional fossils between stem (e.g. Radiodonta such as Anomalocaris) and true arthropods. Re-examination in the 1970s of the Burgess Shale fossils from about 505 million years ago identified many arthropods, some of which could not be assigned to any of the well-known groups, and thus intensified the debate about the Cambrian explosion. A fossil of Marrella from the Burgess Shale has provided the earliest clear evidence of moulting. The earliest fossil of likely pancrustacean larvae date from about 514 million years ago in the Cambrian, followed by unique taxa like Yicaris and Wujicaris. The purported pancrustacean/crustacean affinity of some cambrian arthropods (e.g. Phosphatocopina, Bradoriida and Hymenocarine taxa like waptiids) were disputed by subsequent studies, as they might branch before the mandibulate crown-group. Within the pancrustacean crown-group, only Malacostraca, Branchiopoda and Pentastomida have Cambrian fossil records. Crustacean fossils are common from the Ordovician period onwards. They have remained almost entirely aquatic, possibly because they never developed excretory systems that conserve water. Arthropods provide the earliest identifiable fossils of land animals, from about 419 million years ago in the Late Silurian, and terrestrial tracks from about 450 million years ago appear to have been made by arthropods. Arthropods possessed attributes that were easy coopted for life on land; their existing jointed exoskeletons provided protection against desiccation, support against gravity and a means of locomotion that was not dependent on water. Around the same time the aquatic, scorpion-like eurypterids became the largest ever arthropods, some as long as 2.5 m (8 ft 2 in). The oldest known arachnid is the trigonotarbid Palaeotarbus jerami, from about 420 million years ago in the Silurian period.[Note 3] Attercopus fimbriunguis, from 386 million years ago in the Devonian period, bears the earliest known silk-producing spigots, but its lack of spinnerets means it was not one of the true spiders, which first appear in the Late Carboniferous over 299 million years ago. The Jurassic and Cretaceous periods provide a large number of fossil spiders, including representatives of many modern families. The oldest known scorpion is Dolichophonus, dated back to 436 million years ago. Lots of Silurian and Devonian scorpions were previously thought to be gill-breathing, hence the idea that scorpions were primitively aquatic and evolved air-breathing book lungs later on. However subsequent studies reveal most of them lacking reliable evidence for an aquatic lifestyle, while exceptional aquatic taxa (e.g. Waeringoscorpio) most likely derived from terrestrial scorpion ancestors. The oldest fossil record of hexapod is obscure, as most of the candidates are poorly preserved and their hexapod affinities had been disputed. An iconic example is the Devonian Rhyniognatha hirsti, dated at 396 to 407 million years ago, its mandibles are thought to be a type found only in winged insects, which suggests that the earliest insects appeared in the Silurian period. However, later study shows that Rhyniognatha most likely represent a myriapod, not even a hexapod. The unequivocal oldest known hexapod is the springtail Rhyniella, from about 410 million years ago in the Devonian period, and the palaeodictyopteran Delitzschala bitterfeldensis, from about 325 million years ago in the Carboniferous period, respectively. The Mazon Creek lagerstätten from the Late Carboniferous, about 300 million years ago, include about 200 species, some gigantic by modern standards, and indicate that insects had occupied their main modern ecological niches as herbivores, detritivores and insectivores. Social termites and ants first appear in the Early Cretaceous, and advanced social bees have been found in Late Cretaceous rocks but did not become abundant until the Middle Cenozoic. From 1952 to 1977, zoologist Sidnie Manton and others argued that arthropods are polyphyletic, in other words, that they do not share a common ancestor that was itself an arthropod. Instead, they proposed that three separate groups of "arthropods" evolved separately from common worm-like ancestors: the chelicerates, including spiders and scorpions; the crustaceans; and the uniramia, consisting of onychophorans, myriapods and hexapods. These arguments usually bypassed trilobites, as the evolutionary relationships of this class were unclear. Proponents of polyphyly argued the following: that the similarities between these groups are the results of convergent evolution, as natural consequences of having rigid, segmented exoskeletons; that the three groups use different chemical means of hardening the cuticle; that there were significant differences in the construction of their compound eyes; that it is hard to see how such different configurations of segments and appendages in the head could have evolved from the same ancestor; and that crustaceans have biramous limbs with separate gill and leg branches, while the other two groups have uniramous limbs in which the single branch serves as a leg. includes Aysheaia and Peripatus includes Hallucigenia and Microdictyon includes modern tardigrades aswell as extinct animals likeKerygmachela and Opabinia Anomalocaris includes living groups andextinct forms such as trilobites Further analysis and discoveries in the 1990s reversed this view, and led to acceptance that arthropods are monophyletic, in other words they are inferred to share a common ancestor that was itself an arthropod. For example, Graham Budd's analyses of Kerygmachela in 1993 and of Opabinia in 1996 convinced him that these animals were similar to onychophorans and to various Early Cambrian "lobopods", and he presented an "evolutionary family tree" that showed these as "aunts" and "cousins" of all arthropods. These changes made the scope of the term "arthropod" unclear, and Claus Nielsen proposed that the wider group should be labelled "Panarthropoda" ("all the arthropods") while the animals with jointed limbs and hardened cuticles should be called "Euarthropoda" ("true arthropods"). A contrary view was presented in 2003, when Jan Bergström and Hou Xian-guang argued that, if arthropods were a "sister-group" to any of the anomalocarids, they must have lost and then re-evolved features that were well-developed in the anomalocarids. The earliest known arthropods ate mud in order to extract food particles from it, and possessed variable numbers of segments with unspecialized appendages that functioned as both gills and legs. Anomalocarids were, by the standards of the time, huge and sophisticated predators with specialized mouths and grasping appendages, fixed numbers of segments some of which were specialized, tail fins, and gills that were very different from those of arthropods. In 2006, they suggested that arthropods were more closely related to lobopods and tardigrades than to anomalocarids. In 2014, it was found that tardigrades were more closely related to arthropods than velvet worms. Spiralia (annelids, molluscs, brachiopods, chaetognatha, etc.) Nematoida (nematodes and close relatives) Scalidophora (priapulids and Kinorhyncha, and Loricifera) Onychophorans Tardigrades Chelicerates †Euthycarcinoids Myriapods Crustaceans Hexapods Relationships of Ecdysozoa to each other and to annelids, etc.,[failed verification] including euthycarcinoids Higher up the "family tree", the Annelida have traditionally been considered the closest relatives of the Panarthropoda, since both groups have segmented bodies, and the combination of these groups was labelled Articulata. There had been competing proposals that arthropods were closely related to other groups such as nematodes, priapulids and tardigrades, but these remained minority views because it was difficult to specify in detail the relationships between these groups. In the 1990s, molecular phylogenetic analyses of DNA sequences produced a coherent scheme showing arthropods as members of a superphylum labelled Ecdysozoa ("animals that moult"), which contained nematodes, priapulids and tardigrades but excluded annelids. This was backed up by studies of the anatomy and development of these animals, which showed that many of the features that supported the Articulata hypothesis showed significant differences between annelids and the earliest Panarthropods in their details, and some were hardly present at all in arthropods. This hypothesis groups annelids with molluscs and brachiopods in another superphylum, Lophotrochozoa. If the Ecdysozoa hypothesis is correct, then segmentation of arthropods and annelids either has evolved convergently or has been inherited from a much older ancestor and subsequently lost in several other lineages, such as the non-arthropod members of the Ecdysozoa. giant lobopodians † gilled lobopodians † Radiodonta † Chelicerata Megacheira † Artiopoda † Isoxyida † Mandibulata Aside from the four major living groups (crustaceans, chelicerates, myriapods and hexapods), a number of fossil forms, mostly from the early Cambrian period, are difficult to place taxonomically, either from lack of obvious affinity to any of the main groups or from clear affinity to several of them. Marrella was the first one to be recognized as significantly different from the well-known groups. Modern interpretations of the basal, extinct stem-group of Arthropoda recognised the following groups, from most basal to most crownward: The Deuteropoda is a recently established clade uniting the crown-group (living) arthropods with these possible "upper stem-group" fossils taxa. The clade is defined by important changes to the structure of the head region such as the appearance of a differentiated deutocerebral appendage pair, which excludes more basal taxa like radiodonts and "gilled lobopodians". Controversies remain about the positions of various extinct arthropod groups. Some studies recover Megacheira as closely related to chelicerates, while others recover them as outside the group containing Chelicerate and Mandibulata as stem-group euarthropods. The placement of the Artiopoda (which contains the extinct trilobites and similar forms) is also a frequent subject of dispute. The main hypotheses position them in the clade Arachnomorpha with the Chelicerates. However, one of the newer hypotheses is that the chelicerae have originated from the same pair of appendages that evolved into antennae in the ancestors of Mandibulata, which would place trilobites, which had antennae, closer to Mandibulata than Chelicerata, in the clade Antennulata. The fuxianhuiids, usually suggested to be stem-group arthropods, have been suggested to be Mandibulates in some recent studies. The Hymenocarina, a group of bivalved arthropods, previously thought to have been stem-group members of the group, have been demonstrated to be mandibulates based on the presence of mandibles. List of arthropod groups and genera († denotes extinct taxa) The phylum Arthropoda is typically subdivided into four subphyla, of which one is extinct: The phylogeny of the major extant arthropod groups has been an area of considerable interest and dispute. Recent studies strongly suggest that Crustacea, as traditionally defined, is paraphyletic, with Hexapoda having evolved from within it, so that Crustacea and Hexapoda form a clade, Pancrustacea. The position of Myriapoda, Chelicerata and Pancrustacea remains unclear as of April 2012[update]. In some studies, Myriapoda is grouped with Chelicerata (forming Myriochelata); in other studies, Myriapoda is grouped with Pancrustacea (forming Mandibulata), or Myriapoda may be sister to Chelicerata plus Pancrustacea. The following cladogram shows the internal relationships between all the living classes of arthropods as of the late 2010s, as well as the estimated timing for some of the clades: Pycnogonida Xiphosura Arachnida Chilopoda Symphyla Pauropoda Diplopoda Ostracoda Mystacocarida Ichthyostraca Copepoda Malacostraca Tantulocarida Thecostraca Cephalocarida Branchiopoda Remipedia Collembola Protura Diplura Insecta Interaction with humans Crustaceans such as crabs, lobsters, crayfish, shrimp, and prawns have long been part of human cuisine, and are now raised commercially. Insects and their grubs are at least as nutritious as meat, and are eaten both raw and cooked in many cultures, though not most European, Hindu, and Islamic cultures. Cooked tarantulas are considered a delicacy in Cambodia, and by the Piaroa Indians of southern Venezuela, after the highly irritant hairs – the spider's main defense system – are removed. Humans also unintentionally eat arthropods in other foods, and food safety regulations lay down acceptable contamination levels for different kinds of food material.[Note 4][Note 5] The intentional cultivation of arthropods and other small animals for human food, referred to as minilivestock, is now emerging in animal husbandry as an ecologically sound concept. Commercial butterfly breeding provides Lepidoptera stock to butterfly conservatories, educational exhibits, schools, research facilities, and cultural events. However, the greatest contribution of arthropods to human food supply is by pollination: A 2008 study examined the 100 crops that FAO lists as grown for food, and estimated pollination's economic value as €153 billion, or 9.5 per cent of the value of world agricultural production used for human food in 2005. Besides pollinating, bees produce honey, which is the basis of a rapidly growing industry and international trade. The red dye cochineal, produced from a Central American species of insect, was economically important to the Aztecs and Mayans. While the region was under Spanish control, it became Mexico's second most-lucrative export, and is now regaining some of the ground it lost to synthetic competitors. Shellac, a resin secreted by a species of insect native to southern Asia, was historically used in great quantities for many applications in which it has mostly been replaced by synthetic resins, but it is still used in woodworking and as a food additive. The blood of horseshoe crabs contains a clotting agent, limulus amebocyte lysate, which is now used to test that antibiotics and kidney machines are free of dangerous bacteria, and to detect spinal meningitis. Forensic entomology uses evidence provided by arthropods to establish the time and sometimes the place of death of a human, and in some cases the cause. Recently insects have also gained attention as potential sources of drugs and other medicinal substances. The relative simplicity of the arthropods' body plan, allowing them to move on a variety of surfaces both on land and in water, have made them useful as models for robotics. The redundancy provided by segments allows arthropods and biomimetic robots to move normally, even with damaged or lost appendages. Although arthropods are the most numerous phylum on Earth, and thousands of arthropod species are venomous, they inflict relatively few serious bites and stings on humans. Far more serious are the effects on humans of diseases like malaria carried by blood-sucking insects. Other blood-sucking insects infect livestock with diseases that kill many animals and greatly reduce the usefulness of others. Ticks can cause tick paralysis and several parasite-borne diseases in humans. A few of the closely related mites also infest humans, causing intense itching, and others cause allergic diseases, including hay fever, asthma, and eczema. Many species of arthropods, principally insects but also mites, are agricultural and forest pests. The mite Varroa destructor has become the largest single problem faced by beekeepers worldwide. Efforts to control arthropod pests by large-scale use of pesticides have caused long-term effects on human health and on biodiversity. Increasing arthropod resistance to pesticides has led to the development of integrated pest management using a wide range of measures including biological control. Predatory mites may be useful in controlling some mite pests. See also Notes References External links |
======================================== |
[SOURCE: https://en.wikipedia.org/wiki/Mars_Geyser_Hopper] | [TOKENS: 1844] |
Contents Mars Geyser Hopper The Mars Geyser Hopper (MGH) was proposed in 2012 as a NASA design reference mission for a Discovery-class spacecraft concept that would investigate the springtime carbon dioxide Martian geysers found in regions around the south pole of Mars. The power technology that MGH proposed to use was the Advanced Stirling radioisotope generator (ASRG). NASA finished the ASRG design and made one test unit of the device but the program was concluded by the mid-2010s. Neither InSight nor any of the next Discovery's semi-finalists use the ASRG or an RTG due to high demand and limited supply of the type of plutonium it relies on. Background The Discovery program was started in the 1990s following discussions at NASA for a new program, and has achieved such missions as Genesis, Deep Impact and Kepler among others; this is the program this mission was designed for selection in, at least initially. One of the first unmanned robotic spacecraft to do a hop was Surveyor 6 lunar lander, which successfully soft landed on the Moon 1967 and conducted a post-landing hop. Another possibly for a hopper mission may be Saturn's moon Enceladus. Hoppers are noted for their ability to potentially visit different landing sites. Another hopper-type mission was the Comet Hopper, which won a Discovery semi-finalist award to study a hopping mission to the Comet 46P/Wirtanen. There was some speculation in 2012 that the Geyser Hopper mission could be flown after the InSight Mars lander. Mission overview The mission was projected to cost US$350 million and to meet a cost cap of no more than US$425 million, not including the launch cost. In order to reduce the cost and minimize risk, the spacecraft concept is based on a previous spacecraft design, the Mars Phoenix lander, which has a demonstrated flight heritage that incorporates soft landing capability and incorporates a restartable rocket propulsion system, suitable to be repurposed for this mission requirements. The spacecraft would land at a target landing area near the south pole of Mars, where geysers exist over a stretch of several hundred kilometers with densities of at least one geyser every 1 to 2 kilometres (0.62 to 1.24 mi) and have the ability to "hop" at least twice from its landed location after a summertime landing to reposition itself close to a geyser site, and wait through the winter until the first sunlight of spring to witness first-hand the Martian geyser phenomenon and investigate the debris pattern and channel. Martian geysers are unlike any terrestrial geological phenomenon. The shapes and unusual spider appearance of these features have stimulated a variety of scientific hypotheses about their origin, ranging from differences in frosting reflectance, to explanations involving biological processes. However, all current geophysical models assume some sort of geyser-like activity. Their characteristics and formation process are still a matter of debate. The seasonal frosting and defrosting of CO2 ice results in the appearance of a number of features, such dark dune spots with spider-like rilles or channels below the ice, where spider-like radial channels are carved between the ground and ice, giving it an appearance of spider webs, then, pressure accumulating in their interior ejects gas and dark basaltic sand or dust, which is deposited on the ice surface and thus, forming dark dune spots. This process is rapid, observed happening in the space of a few days, weeks or months, a growth rate rather unusual in geology – especially for Mars. The primary mission duration, starting from launch, is 30 months, comprising 8 months of interplanetary cruise followed by a primary mission of 22 months (one Mars year) on the surface. The spacecraft will enter the atmosphere, and make a rocket-powered soft landing in a region of the south pole where geysers are known to form. This landing will take place during the polar summer, when the surface is free of ice. The predicted landing ellipse is 20 by 50 kilometres (12 mi × 31 mi) and hence the landing will be targeted to a region, and not to a specific geyser location. During the first post-landing phase, it will conduct science operations to characterize the landing site, to understand the surface geology of the area during the ice-free summer period. The spacecraft will then stow its science instruments and re-ignite the engines for a first hop of a distance of up to 2 kilometers (1.2 mi). This hop is designed to place the lander in a location where it can directly probe the geyser region, examining the surface at a spot where a geyser had been. Once again, the spacecraft will stow its instruments and activate the engines for a second hop, a distance of ~100 meters (330 ft). This hop will place the lander onto the winter-over site, a spot chosen to be a relatively high elevation where the lander can get a good view of the surroundings, close to but not located on the site of a known geyser, and outside the fall-out pattern of the expected debris plume. The spacecraft will characterize the local area during the remaining sunlight, and then go into "winter-over mode". The lander will continue to transmit engineering status data and meteorological reports during the winter, but will not conduct major science operations. On the arrival of polar spring, the lander will study the geyser phenomenon from the location selected for optimum viewing. Automated geyser detection on board the spacecraft will scan the environment, although the routine imagery will be buffered on the spacecraft, images will not be relayed to Earth until the spacecraft detects a geyser. This triggers high-speed, high-resolution imagery, including LIDAR characterization of particle motion and infrared spectroscopy. Simultaneously, the science instruments will do chemical analysis of any fallout particles spewed onto the surface of the lander. Geysers erupt at a rate of about one a day during peak springtime season. If more than one is detected simultaneously, the spacecraft algorithm will focus on the nearest or "best". The lander will continue this primary geyser science for a period of about 90 days. Tens of geyser observations are expected over the spring/summer season. Extended mission operations, if desired, would continue the observation through a full Martian year and into the second Martian summer. The hopper concept could also be used for exploration missions other than the polar geyser observation mission discussed here. The ability to make multiple rocket-powered hops from an initial landing location to a science region of interest would be valuable across a large range of terrain on Mars, as well as elsewhere in the Solar System, and would demonstrate a new form of rover with the ability to traverse far more rugged terrain than any previous missions, a mission concept that would be applicable to exploration of many planets and moons. Spacecraft The geyser phenomenon occurs following an extended period of complete darkness, and the geysers themselves occur at the beginning of polar spring, when temperatures are in the range of −150 °C (−238 °F), and the sun angle is only a few degrees above the horizon. The extreme environment, low Sun angles during the geyser occurrence, and the fact that it would be desirable to emplace the probe well before the occurrence of the geysers, during a period of no sunlight, makes this a difficult environment for the use of solar arrays as the primary power source. Thus, this is an attractive mission for use of the Advanced Stirling Radioisotope Generator (ASRG) with a mass of 126 kilograms (278 lb) including a Li-ion battery for use during Entry/Descent/Landing (EDL) as well as during the hops when there is a short duration requirement for additional power. However, the ASRG development was cancelled by NASA in 2013. Hopping propulsion is based on the Phoenix landing system, using integrated hydrazine monopropellant blow-down system with 15 Aerojet MR-107N thrusters with Isp 230 sec for landing and hopping. RCS is four pairs of Aerojet MR-103D thrusters at 215 sec Isp, and one Aerojet MR-102 thruster at 220 sec Isp. The system will be fueled with 191 kg of propellant. The lander will communicate through X-band direct to Earth on cruise deck for transit; it will then use UHF antenna. Imaging and all data relaying would be coordinated with the Mars Reconnaissance Orbiter operations team. The science instruments include stereo cameras (MastCam) to view the geyser events and a robotic arm (from Phoenix) to dig beneath the soil surface and gather soil samples for chemical analysis on the Hopper. A light detection and ranging instrument (LIDAR), a landing camera and a thermal spectrometer for remote geological analysis as well as weather sensing are included. See also References External links |
======================================== |
[SOURCE: https://en.wikipedia.org/wiki/Special:BookSources/978-0-7486-7710-8] | [TOKENS: 380] |
Contents Book sources This page allows users to search multiple sources for a book given a 10- or 13-digit International Standard Book Number. Spaces and dashes in the ISBN do not matter. This page links to catalogs of libraries, booksellers, and other book sources where you will be able to search for the book by its International Standard Book Number (ISBN). Online text Google Books and other retail sources below may be helpful if you want to verify citations in Wikipedia articles, because they often let you search an online version of the book for specific words or phrases, or you can browse through the book (although for copyright reasons the entire book is usually not available). At the Open Library (part of the Internet Archive) you can borrow and read entire books online. Online databases Subscription eBook databases Libraries Alabama Alaska California Colorado Connecticut Delaware Florida Georgia Illinois Indiana Iowa Kansas Kentucky Massachusetts Michigan Minnesota Missouri Nebraska New Jersey New Mexico New York North Carolina Ohio Oklahoma Oregon Pennsylvania Rhode Island South Carolina South Dakota Tennessee Texas Utah Washington state Wisconsin Bookselling and swapping Find your book on a site that compiles results from other online sites: These sites allow you to search the catalogs of many individual booksellers: Non-English book sources If the book you are looking for is in a language other than English, you might find it helpful to look at the equivalent pages on other Wikipedias, linked below – they are more likely to have sources appropriate for that language. Find other editions The WorldCat xISBN tool for finding other editions is no longer available. However, there is often a "view all editions" link on the results page from an ISBN search. Google books often lists other editions of a book and related books under the "about this book" link. You can convert between 10 and 13 digit ISBNs with these tools: Find on Wikipedia See also Get free access to research! Research tools and services Outreach Get involved |
======================================== |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.