text stringlengths 0 473k |
|---|
[SOURCE: https://en.wikipedia.org/wiki/Tianwen-3] | [TOKENS: 559] |
Contents Tianwen-3 Tianwen-3 (Chinese: 天问三号) is a planned Mars sample-return mission by China which would send two spacecraft (an orbiter/Earth-returner and a lander/ascent-vehicle) via two separate launches to Mars. Together, the two spacecraft will seek to obtain samples of Martian rocks and soil and then return the cached samples to Earth. The mission is scheduled to launch in 2028. The primary goal of the mission is to search for signatures of life on Mars. Overview In summer 2022 during a deep space exploration technology forum held at Nanjing University, Sun Zezhou, chief designer of the Tianwen-1 mission, detailed plans for the mission based on a two-launch architecture. The mission constitutes part of the Tianwen series of space missions. The current mission architecture envisions two launches around 2028 by the Long March 5 carrier rocket. One launch will send an orbiter/return-vehicle. A second launch will send a lander/ascent-vehicle. Once the lander arrives on the Martian surface it will collect surface samples, via a drill on the lander and possibly an autonomous mobile robot with multiple legs. After several months on the Martian surface and after storing the samples collected by the lander and mobile robot, the ascent vehicle will launch from atop the lander and rendezvous with the waiting orbiter. The ascent vehicle will transfer the collected samples to the orbiter/return-vehicle, which will depart for Earth. The samples will be returned to Earth via an atmospheric reentry vehicle. The main goal of the mission is search for life signatures. In 2024, scientists identified 51 potential landing sites in line with the missions current engineering constraints (altitudes less than −3 km, latitudes from 17°N to 30°N, slopes less than 8°, and 'rock abundance' less than 10%): 1 in Amazonis Planitia, 12 in Utopia Planitia, and 38 in Chryse Planitia or Arabia Terra. International Collaboration On 11 March 2025, CNSA opened the Tianwen-3 Mars sample-return mission to potential international partners. International teams can propose "piggyback payloads requiring support from the Tianwen-3 spacecraft or independent scientific instruments" for inclusion on the spacecraft. 15 kilograms of mass has been allocated on the Tianwen-3 Earth-return spacecraft for potential international scientific payloads and another 5 kilograms are available for partner payloads on the Mars orbiter. Interested parties may submit proposals by 30 June 2025 with a final selection targeted for October 2025. Flight hardware for the selected payloads are to be delivered in 2027. See also References External links |
======================================== |
[SOURCE: https://en.wikipedia.org/wiki/Database_management_system] | [TOKENS: 8348] |
Contents Database In computing, a database is an organized collection of data or a type of data store based on the use of a database management system (DBMS), the software that interacts with end users, applications, and the database itself to capture and analyze the data. The DBMS additionally encompasses the core facilities provided to administer the database. The sum total of the database, the DBMS and the associated applications can be referred to as a database system. Often the term "database" is also used loosely to refer to any of the DBMS, the database system or an application associated with the database. Before digital storage and retrieval of data became widespread, index cards were used for data storage in a wide range of applications and environments: in the home to record and store recipes, shopping lists, contact information and other organizational data; in business to record presentation notes, project research and notes, and contact information; in schools as flash cards or other visual aids; and in academic research to hold data such as bibliographical citations or notes in a card file. Professional book indexers used index cards in the creation of book indexes until they were replaced by indexing software in the 1980s and 1990s. Small databases can be stored on a file system, while large databases are hosted on computer clusters or cloud storage. The design of databases spans formal techniques and practical considerations, including data modeling, efficient data representation and storage, query languages, security and privacy of sensitive data, and distributed computing issues, including supporting concurrent access and fault tolerance. Computer scientists may classify database management systems according to the database models that they support. Relational databases became dominant in the 1980s. These model data as rows and columns in a series of tables, and the vast majority use SQL for writing and querying data. In the 2000s, non-relational databases became popular, collectively referred to as NoSQL, because they use different query languages. Terminology and overview Formally, a "database" refers to a set of related data accessed through the use of a "database management system" (DBMS), which is an integrated set of computer software that allows users to interact with one or more databases and provides access to all of the data contained in the database (although restrictions may exist that limit access to particular data). The DBMS provides various functions that allow entry, storage and retrieval of large quantities of information and provides ways to manage how that information is organized. Because of the close relationship between them, the term "database" is often used casually to refer to both a database and the DBMS used to manipulate it. Outside the world of professional information technology, the term database is often used to refer to any collection of related data (such as a spreadsheet or a card index) as size and usage requirements typically necessitate use of a database management system. Existing DBMSs provide various functions that allow management of a database and its data which can be classified into four main functional groups: Both a database and its DBMS conform to the principles of a particular database model. "Database system" refers collectively to the database model, database management system, and database. Physically, database servers are dedicated computers that hold the actual databases and run only the DBMS and related software. Database servers are usually multiprocessor computers, with generous memory and RAID disk arrays used for stable storage. Hardware database accelerators, connected to one or more servers via a high-speed channel, are also used in large-volume transaction processing environments. DBMSs are found at the heart of most database applications. DBMSs may be built around a custom multitasking kernel with built-in networking support, but modern DBMSs typically rely on a standard operating system to provide these functions.[citation needed] Since DBMSs comprise a significant market, computer and storage vendors often take into account DBMS requirements in their own development plans. Databases and DBMSs can be categorized according to the database model(s) that they support (such as relational or XML), the type(s) of computer they run on (from a server cluster to a mobile phone), the query language(s) used to access the database (such as SQL or XQuery), and their internal engineering, which affects performance, scalability, resilience, and security. History The sizes, capabilities, and performance of databases and their respective DBMSs have grown in orders of magnitude. These performance increases were enabled by the technology progress in the areas of processors, computer memory, computer storage, and computer networks. The concept of a database was made possible by the emergence of direct access storage media such as magnetic disks, which became widely available in the mid-1960s; earlier systems relied on sequential storage of data on magnetic tape. The subsequent development of database technology can be divided into three eras based on data model or structure: navigational, SQL/relational, and post-relational. The two main early navigational data models were the hierarchical model and the CODASYL model (network model). These were characterized by the use of pointers (often physical disk addresses) to follow relationships from one record to another. The relational model, first proposed in 1970 by Edgar F. Codd, departed from this tradition by insisting that applications should search for data by content, rather than by following links. The relational model employs sets of ledger-style tables, each used for a different type of entity. Only in the mid-1980s did computing hardware become powerful enough to allow the wide deployment of relational systems (DBMSs plus applications). By the early 1990s, however, relational systems dominated in all large-scale data processing applications, and as of 2018[update] they remain dominant: IBM Db2, Oracle, MySQL, and Microsoft SQL Server are the most searched DBMS. The dominant database language, standardized SQL for the relational model, has influenced database languages for other data models.[citation needed] Object databases were developed in the 1980s to overcome the inconvenience of object–relational impedance mismatch, which led to the coining of the term "post-relational" and also the development of hybrid object–relational databases. The next generation of post-relational databases in the late 2000s became known as NoSQL databases, introducing fast key–value stores and document-oriented databases. A competing "next generation" known as NewSQL databases attempted new implementations that retained the relational/SQL model while aiming to match the high performance of NoSQL compared to commercially available relational DBMSs. The introduction of the term database coincided with the availability of direct-access storage (disks and drums) from the mid-1960s onwards. The term represented a contrast with the tape-based systems of the past, allowing shared interactive use rather than daily batch processing. The Oxford English Dictionary cites a 1962 report by the System Development Corporation of California as the first to use the term "data-base" in a specific technical sense. As computers grew in speed and capability, a number of general-purpose database systems emerged; by the mid-1960s a number of such systems had come into commercial use. Interest in a standard began to grow, and Charles Bachman, author of one such product, the Integrated Data Store (IDS), founded the Database Task Group within CODASYL, the group responsible for the creation and standardization of COBOL. In 1971, the Database Task Group delivered their standard, which generally became known as the CODASYL approach, and soon a number of commercial products based on this approach entered the market. The CODASYL approach offered applications the ability to navigate around a linked data set which was formed into a large network. Applications could find records by one of three methods: Later systems added B-trees to provide alternate access paths. Many CODASYL databases also added a declarative query language for end users (as distinct from the navigational API). However, CODASYL databases were complex and required significant training and effort to produce useful applications. IBM also had its own DBMS in 1966, known as Information Management System (IMS). IMS was a development of software written for the Apollo program on the System/360. IMS was generally similar in concept to CODASYL, but used a strict hierarchy for its model of data navigation instead of CODASYL's network model. Both concepts later became known as navigational databases due to the way data was accessed: the term was popularized by Bachman's 1973 Turing Award presentation The Programmer as Navigator. IMS is classified by IBM as a hierarchical database. IDMS and Cincom Systems' TOTAL databases are classified as network databases. IMS remains in use as of 2014[update]. Edgar F. Codd worked at IBM in San Jose, California, in an office primarily involved in the development of hard disk systems. He was unhappy with the navigational model of the CODASYL approach, notably the lack of a "search" facility. In 1970, he wrote a number of papers that outlined a new approach to database construction that eventually culminated in the groundbreaking A Relational Model of Data for Large Shared Data Banks. The paper described a new system for storing and working with large databases. Instead of records being stored in some sort of linked list of free-form records as in CODASYL, Codd's idea was to organize the data as a number of "tables", each table being used for a different type of entity. Each table would contain a fixed number of columns containing the attributes of the entity. One or more columns of each table were designated as a primary key by which the rows of the table could be uniquely identified; cross-references between tables always used these primary keys, rather than disk addresses, and queries would join tables based on these key relationships, using a set of operations based on the mathematical system of relational calculus (from which the model takes its name). Splitting the data into a set of normalized tables (or relations) aimed to ensure that each "fact" was only stored once, thus simplifying update operations. Virtual tables called views could present the data in different ways for different users, but views could not be directly updated. Codd used mathematical terms to define the model: relations, tuples, and domains rather than tables, rows, and columns. The terminology that is now familiar came from early implementations. Codd would later criticize the tendency for practical implementations to depart from the mathematical foundations on which the model was based. The use of primary keys (user-oriented identifiers) to represent cross-table relationships, rather than disk addresses, had two primary motivations. From an engineering perspective, it enabled tables to be relocated and resized without expensive database reorganization. But Codd was more interested in the difference in semantics: the use of explicit identifiers made it easier to define update operations with clean mathematical definitions, and it also enabled query operations to be defined in terms of the established discipline of first-order predicate calculus; because these operations have clean mathematical properties, it becomes possible to rewrite queries in provably correct ways, which is the basis of query optimization. There is no loss of expressiveness compared with the hierarchic or network models, though the connections between tables are no longer so explicit. In the hierarchic and network models, records were allowed to have a complex internal structure. For example, the salary history of an employee might be represented as a "repeating group" within the employee record. In the relational model, the process of normalization led to such internal structures being replaced by data held in multiple tables, connected only by logical keys. For instance, a common use of a database system is to track information about users, their name, login information, various addresses and phone numbers. In the navigational approach, all of this data would be placed in a single variable-length record. In the relational approach, the data would be normalized into a user table, an address table and a phone number table (for instance). Records would be created in these optional tables only if the address or phone numbers were actually provided. As well as identifying rows/records using logical identifiers rather than disk addresses, Codd changed the way in which applications assembled data from multiple records. Rather than requiring applications to gather data one record at a time by navigating the links, they would use a declarative query language that expressed what data was required, rather than the access path by which it should be found. Finding an efficient access path to the data became the responsibility of the database management system, rather than the application programmer. This process, called query optimization, depended on the fact that queries were expressed in terms of mathematical logic. Codd's paper inspired teams at various universities to research the subject, including one at University of California, Berkeley led by Eugene Wong and Michael Stonebraker, who started INGRES using funding that had already been allocated for a geographical database project and student programmers to produce code. Beginning in 1973, INGRES delivered its first test products which were generally ready for widespread use in 1979. INGRES was similar to System R in a number of ways, including the use of a "language" for data access, known as QUEL. Over time, INGRES moved to the emerging SQL standard. IBM itself did one test implementation of the relational model, PRTV, and a production one, Business System 12, both now discontinued. Honeywell wrote MRDS for Multics, and now there are two new implementations: Alphora Dataphor and Rel. Most other DBMS implementations usually called relational are actually SQL DBMSs. In 1970, the University of Michigan began development of the MICRO Information Management System based on D.L. Childs' Set-Theoretic Data model. The university in 1974 hosted a debate between Codd and Bachman which Bruce Lindsay of IBM later described as "throwing lightning bolts at each other!". MICRO was used to manage very large data sets by the US Department of Labor, the U.S. Environmental Protection Agency, and researchers from the University of Alberta, the University of Michigan, and Wayne State University. It ran on IBM mainframe computers using the Michigan Terminal System. The system remained in production until 1998. In the 1970s and 1980s, attempts were made to build database systems with integrated hardware and software. The underlying philosophy was that such integration would provide higher performance at a lower cost. Examples were IBM System/38, the early offering of Teradata, and the Britton Lee, Inc. database machine. Another approach to hardware support for database management was ICL's CAFS accelerator, a hardware disk controller with programmable search capabilities. In the long term, these efforts were generally unsuccessful because specialized database machines could not keep pace with the rapid development and progress of general-purpose computers. Thus most database systems nowadays are software systems running on general-purpose hardware, using general-purpose computer data storage. However, this idea is still pursued in certain applications by some companies like Netezza and Oracle (Exadata). IBM formed a team led by Codd that started working on a prototype system, System R despite opposition from others at the company. The first version was ready in 1974/5, and work then started on multi-table systems in which the data could be split so that all of the data for a record (some of which is optional) did not have to be stored in a single large "chunk". Subsequent multi-user versions were tested by customers in 1978 and 1979, by which time a standardized query language – SQL[citation needed] – had been added. Codd's ideas were establishing themselves as both workable and superior to CODASYL, pushing IBM to develop a true production version of System R, known as SQL/DS, and, later, Database 2 (IBM Db2). Larry Ellison's Oracle Database (or more simply, Oracle) started from a different chain, based on IBM's papers on System R. Though Oracle V1 implementations were completed in 1978, it was not until Oracle Version 2 when Ellison beat IBM to market in 1979. Stonebraker went on to apply the lessons from INGRES to develop a new database, Postgres, which is now known as PostgreSQL. PostgreSQL is often used for global mission-critical applications (the .org and .info domain name registries use it as their primary data store, as do many large companies and financial institutions). In Sweden, Codd's paper was also read and Mimer SQL was developed in the mid-1970s at Uppsala University. In 1984, this project was consolidated into an independent enterprise. Another data model, the entity–relationship model, emerged in 1976 and gained popularity for database design as it emphasized a more familiar description than the earlier relational model. Later on, entity–relationship constructs were retrofitted as a data modeling construct for the relational model, and the difference between the two has become irrelevant.[citation needed] Besides IBM and various software companies such as Sybase and Informix Corporation, most large computer hardware vendors by the 1980s had their own database systems such as DEC's VAX Rdb/VMS. The decade ushered in the age of desktop computing. The new computers empowered their users with spreadsheets like Lotus 1-2-3 and database software like dBASE. The dBASE product was lightweight and easy for any computer user to understand out of the box. C. Wayne Ratliff, the creator of dBASE, stated: "dBASE was different from programs like BASIC, C, FORTRAN, and COBOL in that a lot of the dirty work had already been done. The data manipulation is done by dBASE instead of by the user, so the user can concentrate on what he is doing, rather than having to mess with the dirty details of opening, reading, and closing files, and managing space allocation." dBASE was one of the top selling software titles in the 1980s and early 1990s. By the start of the decade databases had become a billion-dollar industry in about ten years. The 1990s, along with a rise in object-oriented programming, saw a growth in how data in various databases were handled. Programmers and designers began to treat the data in their databases as objects. That is to say that if a person's data were in a database, that person's attributes, such as their address, phone number, and age, were now considered to belong to that person instead of being extraneous data. This allows for relations between data to be related to objects and their attributes and not to individual fields. The term "object–relational impedance mismatch" described the inconvenience of translating between programmed objects and database tables. Object databases and object–relational databases attempt to solve this problem by providing an object-oriented language (sometimes as extensions to SQL) that programmers can use as alternative to purely relational SQL. On the programming side, libraries known as object–relational mappings (ORMs) attempt to solve the same problem. Database sales grew rapidly during the dotcom bubble and, after its end, the rise of ecommerce. The popularity of open source databases such as MySQL has grown since 2000, to the extent that Ken Jacobs of Oracle said in 2005 that perhaps "these guys are doing to us what we did to IBM". XML databases are a type of structured document-oriented database that allows querying based on XML document attributes. XML databases are mostly used in applications where the data is conveniently viewed as a collection of documents, with a structure that can vary from the very flexible to the highly rigid: examples include scientific articles, patents, tax filings, and personnel records. NoSQL databases are often very fast, do not require fixed table schemas, avoid join operations by storing denormalized data, and are designed to scale horizontally. In recent years, there has been a strong demand for massively distributed databases with high partition tolerance, but according to the CAP theorem, it is impossible for a distributed system to simultaneously provide consistency, availability, and partition tolerance guarantees. A distributed system can satisfy any two of these guarantees at the same time, but not all three. For that reason, many NoSQL databases are using what is called eventual consistency to provide both availability and partition tolerance guarantees with a reduced level of data consistency. NewSQL is a class of modern relational databases that aims to provide the same scalable performance of NoSQL systems for online transaction processing (read-write) workloads while still using SQL and maintaining the ACID guarantees of a traditional database system. Use cases Databases are used to support internal operations of organizations and to underpin online interactions with customers and suppliers (see Enterprise software). Databases are used to hold administrative information and more specialized data, such as engineering data or economic models. Examples include computerized library systems, flight reservation systems, computerized parts inventory systems, and many content management systems that store websites as collections of webpages in a database. Classification One way to classify databases involves the type of their contents, for example: bibliographic, document-text, statistical, or multimedia objects. Another way is by their application area, for example: accounting, music compositions, movies, banking, manufacturing, or insurance. A third way is by some technical aspect, such as the database structure or interface type. This section lists a few of the adjectives used to characterize different kinds of databases. Database management system Connolly and Begg define database management system (DBMS) as a "software system that enables users to define, create, maintain and control access to the database." Examples of DBMS's include MySQL, MariaDB, PostgreSQL, Microsoft SQL Server, Oracle Database, and Microsoft Access. The DBMS acronym is sometimes extended to indicate the underlying database model, with RDBMS for the relational, OODBMS for the object (oriented) and ORDBMS for the object–relational model. Other extensions can indicate some other characteristics, such as DDBMS for a distributed database management systems. The functionality provided by a DBMS can vary enormously. The core functionality is the storage, retrieval and update of data. Codd proposed the following functions and services a fully-fledged general purpose DBMS should provide: It is also generally to be expected the DBMS will provide a set of utilities for such purposes as may be necessary to administer the database effectively, including import, export, monitoring, defragmentation and analysis utilities. The core part of the DBMS interacting between the database and the application interface sometimes referred to as the database engine. Often DBMSs will have configuration parameters that can be statically and dynamically tuned, for example the maximum amount of main memory on a server the database can use. The trend is to minimize the amount of manual configuration, and for cases such as embedded databases the need to target zero-administration is paramount. The large major enterprise DBMSs have tended to increase in size and functionality and have involved up to thousands of human years of development effort throughout their lifetime.[a] Early multi-user DBMS typically only allowed for the application to reside on the same computer with access via terminals or terminal emulation software. The client–server architecture was a development where the application resided on a client desktop and the database on a server allowing the processing to be distributed. This evolved into a multitier architecture incorporating application servers and web servers with the end user interface via a web browser with the database only directly connected to the adjacent tier. A general-purpose DBMS will provide public application programming interfaces (API) and optionally a processor for database languages such as SQL to allow applications to be written to interact with and manipulate the database. A special purpose DBMS may use a private API and be specifically customized and linked to a single application. For example, an email system performs many of the functions of a general-purpose DBMS such as message insertion, message deletion, attachment handling, blocklist lookup, associating messages an email address and so forth however these functions are limited to what is required to handle email. Application External interaction with the database will be via an application program that interfaces with the DBMS. This can range from a database tool that allows users to execute SQL queries textually or graphically, to a website that happens to use a database to store and search information. A programmer will code interactions to the database (sometimes referred to as a datasource) via an application program interface (API) or via a database language. The particular API or language chosen will need to be supported by DBMS, possibly indirectly via a preprocessor or a bridging API. Some API's aim to be database independent, ODBC being a commonly known example. Other common API's include JDBC and ADO.NET. Database languages Database languages are special-purpose languages, which allow one or more of the following tasks, sometimes distinguished as sublanguages: Database languages are specific to a particular data model. Notable examples include: A database language may also incorporate features like: Storage Database storage is the container of the physical materialization of a database. It comprises the internal (physical) level in the database architecture. It also contains all the information needed (e.g., metadata, "data about the data", and internal data structures) to reconstruct the conceptual level and external level from the internal level when needed. Databases as digital objects contain three layers of information which must be stored: the data, the structure, and the semantics. Proper storage of all three layers is needed for future preservation and longevity of the database. Putting data into permanent storage is generally the responsibility of the database engine a.k.a. "storage engine". Though typically accessed by a DBMS through the underlying operating system (and often using the operating systems' file systems as intermediates for storage layout), storage properties and configuration settings are extremely important for the efficient operation of the DBMS, and thus are closely maintained by database administrators. A DBMS, while in operation, always has its database residing in several types of storage (e.g., memory and external storage). The database data and the additional needed information, possibly in very large amounts, are coded into bits. Data typically reside in the storage in structures that look completely different from the way the data look at the conceptual and external levels, but in ways that attempt to optimize (the best possible) these levels' reconstruction when needed by users and programs, as well as for computing additional types of needed information from the data (e.g., when querying the database). Some DBMSs support specifying which character encoding was used to store data, so multiple encodings can be used in the same database. Various low-level database storage structures are used by the storage engine to serialize the data model so it can be written to the medium of choice. Techniques such as indexing may be used to improve performance. Conventional storage is row-oriented, but there are also column-oriented and correlation databases. Often storage redundancy is employed to increase performance. A common example is storing materialized views, which consist of frequently needed external views or query results. Storing such views saves the expensive computing them each time they are needed. The downsides of materialized views are the overhead incurred when updating them to keep them synchronized with their original updated database data, and the cost of storage redundancy. Occasionally a database employs storage redundancy by database objects replication (with one or more copies) to increase data availability (both to improve performance of simultaneous multiple end-user accesses to the same database object, and to provide resiliency in a case of partial failure of a distributed database). Updates of a replicated object need to be synchronized across the object copies. In many cases, the entire database is replicated. With data virtualization, the data used remains in its original locations and real-time access is established to allow analytics across multiple sources. This can aid in resolving some technical difficulties such as compatibility problems when combining data from various platforms, lowering the risk of error caused by faulty data, and guaranteeing that the newest data is used. Furthermore, avoiding the creation of a new database containing personal information can make it easier to comply with privacy regulations. However, with data virtualization, the connection to all necessary data sources must be operational as there is no local copy of the data, which is one of the main drawbacks of the approach. Security Database security deals with all various aspects of protecting the database content, its owners, and its users. It ranges from protection from intentional unauthorized database uses to unintentional database accesses by unauthorized entities (e.g., a person or a computer program). Database access control deals with controlling who (a person or a certain computer program) are allowed to access what information in the database. The information may comprise specific database objects (e.g., record types, specific records, data structures), certain computations over certain objects (e.g., query types, or specific queries), or using specific access paths to the former (e.g., using specific indexes or other data structures to access information). Database access controls are set by special authorized (by the database owner) personnel that uses dedicated protected security DBMS interfaces. This may be managed directly on an individual basis, or by the assignment of individuals and privileges to groups, or (in the most elaborate models) through the assignment of individuals and groups to roles which are then granted entitlements. Data security prevents unauthorized users from viewing or updating the database. Using passwords, users are allowed access to the entire database or subsets of it called "subschemas". For example, an employee database can contain all the data about an individual employee, but one group of users may be authorized to view only payroll data, while others are allowed access to only work history and medical data. If the DBMS provides a way to interactively enter and update the database, as well as interrogate it, this capability allows for managing personal databases. Data security in general deals with protecting specific chunks of data, both physically (i.e., from corruption, or destruction, or removal; e.g., see physical security), or the interpretation of them, or parts of them to meaningful information (e.g., by looking at the strings of bits that they comprise, concluding specific valid credit-card numbers; e.g., see data encryption). Change and access logging records who accessed which attributes, what was changed, and when it was changed. Logging services allow for a forensic database audit later by keeping a record of access occurrences and changes. Sometimes application-level code is used to record changes rather than leaving this in the database. Monitoring can be set up to attempt to detect security breaches. Therefore, organizations must take database security seriously because of the many benefits it provides. Organizations will be safeguarded from security breaches and hacking activities like firewall intrusion, virus spread, and ransom ware. This helps in protecting the company's essential information, which cannot be shared with outsiders at any cause. Transactions and concurrency Database transactions can be used to introduce some level of fault tolerance and data integrity after recovery from a crash. A database transaction is a unit of work, typically encapsulating a number of operations over a database (e.g., reading a database object, writing, acquiring or releasing a lock, etc.), an abstraction supported in database and also other systems. Each transaction has well defined boundaries in terms of which program/code executions are included in that transaction (determined by the transaction's programmer via special transaction commands). The acronym ACID describes some ideal properties of a database transaction: atomicity, consistency, isolation, and durability. Migration A database built with one DBMS is not portable to another DBMS (i.e., the other DBMS cannot run it). However, in some situations, it is desirable to migrate a database from one DBMS to another. The reasons are primarily economical (different DBMSs may have different total costs of ownership or TCOs), functional, and operational (different DBMSs may have different capabilities). The migration involves the database's transformation from one DBMS type to another. The transformation should maintain (if possible) the database related application (i.e., all related application programs) intact. Thus, the database's conceptual and external architectural levels should be maintained in the transformation. It may be desired that also some aspects of the architecture internal level are maintained. A complex or large database migration may be a complicated and costly (one-time) project by itself, which should be factored into the decision to migrate. This is in spite of the fact that tools may exist to help migration between specific DBMSs. Typically, a DBMS vendor provides tools to help import databases from other popular DBMSs. Building, maintaining, and tuning After designing a database for an application, the next stage is building the database. Typically, an appropriate general-purpose DBMS can be selected to be used for this purpose. A DBMS provides the needed user interfaces to be used by database administrators to define the needed application's data structures within the DBMS's respective data model. Other user interfaces are used to select needed DBMS parameters (like security related, storage allocation parameters, etc.). When the database is ready (all its data structures and other needed components are defined), it is typically populated with initial application's data (database initialization, which is typically a distinct project; in many cases using specialized DBMS interfaces that support bulk insertion) before making it operational. In some cases, the database becomes operational while empty of application data, and data are accumulated during its operation. After the database is created, initialized and populated it needs to be maintained. Various database parameters may need changing and the database may need to be tuned (tuning) for better performance; application's data structures may be changed or added, new related application programs may be written to add to the application's functionality, etc. Backup and restore Sometimes it is desired to bring a database back to a previous state (for many reasons, e.g., cases when the database is found corrupted due to a software error, or if it has been updated with erroneous data). To achieve this, a backup operation is done occasionally or continuously, where each desired database state (i.e., the values of its data and their embedding in database's data structures) is kept within dedicated backup files (many techniques exist to do this effectively). When it is decided by a database administrator to bring the database back to this state (e.g., by specifying this state by a desired point in time when the database was in this state), these files are used to restore that state. Static analysis Static analysis techniques for software verification can be applied also in the scenario of query languages. In particular, the *Abstract interpretation framework has been extended to the field of query languages for relational databases as a way to support sound approximation techniques. The semantics of query languages can be tuned according to suitable abstractions of the concrete domain of data. The abstraction of relational database systems has many interesting applications, in particular, for security purposes, such as fine-grained access control, watermarking, etc. Miscellaneous features Other DBMS features might include: Increasingly, there are calls for a single system that incorporates all of these core functionalities into the same build, test, and deployment framework for database management and source control. Borrowing from other developments in the software industry, some market such offerings as "DevOps for database". Design and modeling The first task of a database designer is to produce a conceptual data model that reflects the structure of the information to be held in the database. A common approach to this is to develop an entity–relationship model, often with the aid of drawing tools. Another popular approach is the Unified Modeling Language. A successful data model will accurately reflect the possible state of the external world being modeled: for example, if people can have more than one phone number, it will allow this information to be captured. Designing a good conceptual data model requires a good understanding of the application domain; it typically involves asking deep questions about the things of interest to an organization, like "can a customer also be a supplier?", or "if a product is sold with two different forms of packaging, are those the same product or different products?", or "if a plane flies from New York to Dubai via Frankfurt, is that one flight or two (or maybe even three)?". The answers to these questions establish definitions of the terminology used for entities (customers, products, flights, flight segments) and their relationships and attributes. Producing the conceptual data model sometimes involves input from business processes, or the analysis of workflow in the organization. This can help to establish what information is needed in the database, and what can be left out. For example, it can help when deciding whether the database needs to hold historic data as well as current data. Having produced a conceptual data model that users are happy with, the next stage is to translate this into a schema that implements the relevant data structures within the database. This process is often called logical database design, and the output is a logical data model expressed in the form of a schema. Whereas the conceptual data model is (in theory at least) independent of the choice of database technology, the logical data model will be expressed in terms of a particular database model supported by the chosen DBMS. (The terms data model and database model are often used interchangeably, but in this article we use data model for the design of a specific database, and database model for the modeling notation used to express that design). The most popular database model for general-purpose databases is the relational model, or more precisely, the relational model as represented by the SQL language. The process of creating a logical database design using this model uses a methodical approach known as normalization. The goal of normalization is to ensure that each elementary "fact" is only recorded in one place, so that insertions, updates, and deletions automatically maintain consistency. The final stage of database design is to make the decisions that affect performance, scalability, recovery, security, and the like, which depend on the particular DBMS. This is often called physical database design, and the output is the physical data model. A key goal during this stage is data independence, meaning that the decisions made for performance optimization purposes should be invisible to end-users and applications. There are two types of data independence: Physical data independence and logical data independence. Physical design is driven mainly by performance requirements, and requires a good knowledge of the expected workload and access patterns, and a deep understanding of the features offered by the chosen DBMS. Another aspect of physical database design is security. It involves both defining access control to database objects as well as defining security levels and methods for the data itself. A database model is a type of data model that determines the logical structure of a database and fundamentally determines in which manner data can be stored, organized, and manipulated. The most popular example of a database model is the relational model (or the SQL approximation of relational), which uses a table-based format. Common logical data models for databases include: An object–relational database combines the two related structures. Physical data models include: Other models include: Specialized models are optimized for particular types of data: A database management system provides three views of the database data: While there is typically only one conceptual and internal view of the data, there can be any number of different external views. This allows users to see database information in a more business-related way rather than from a technical, processing viewpoint. For example, a financial department of a company needs the payment details of all employees as part of the company's expenses, but does not need details about employees that are in the interest of the human resources department. Thus different departments need different views of the company's database. The three-level database architecture relates to the concept of data independence which was one of the major initial driving forces of the relational model. The idea is that changes made at a certain level do not affect the view at a higher level. For example, changes in the internal level do not affect application programs written using conceptual level interfaces, which reduces the impact of making physical changes to improve performance. The conceptual view provides a level of indirection between internal and external. On the one hand it provides a common view of the database, independent of different external view structures, and on the other hand it abstracts away details of how the data are stored or managed (internal level). In principle every level, and even every external view, can be presented by a different data model. In practice usually a given DBMS uses the same data model for both the external and the conceptual levels (e.g., relational model). The internal level, which is hidden inside the DBMS and depends on its implementation, requires a different level of detail and uses its own types of data structure types. Research Database technology has been an active research topic since the 1960s, both in academia and in the research and development groups of companies (for example IBM Research). Research activity includes theory and development of prototypes. Notable research topics have included models, the atomic transaction concept, related concurrency control techniques, query languages and query optimization methods, RAID, and more. The database research area has several dedicated academic journals (for example, ACM Transactions on Database Systems-TODS, Data and Knowledge Engineering-DKE) and annual conferences (e.g., ACM SIGMOD, ACM PODS, VLDB, IEEE ICDE). See also Notes References Sources Further reading External links |
======================================== |
[SOURCE: https://en.wikipedia.org/wiki/Julis,_Gaza] | [TOKENS: 1545] |
Contents Julis, Gaza Julis (Arabic: جولس) was a Palestinian Arab village in the Gaza Subdistrict, located 26.5 kilometers (16.5 mi) northeast of Gaza on a slight elevation along the southern coastal plain. In 1945, there were 1,030 inhabitants in the village. It was depopulated during the 1948 Arab-Israeli War. History Julis was built on an archaeological site whose ancient name is unknown. Potsherds from the Mamluk era have been found in the village. The village had a Maqam (shrine) which was constructed with ancient materials. Julis was inhabited in the 15th century. In 1472–1473 CE, Sultan Qaitbay endowed it for the benefit of his Jerusalem madrasa. A branch of the Jaradat tribe, originating from Sa'ir, settled in Julis. Julis was incorporated into the Ottoman Empire in 1517 with the rest of Palestine. In first Ottoman tax register of 1526/7 the village was unpopulated. By 1596 CE, however, the village had been refounded as part of the nahiya of Gaza, in the Liwa of Gaza. It consisted of 204 persons (37 households), all Muslims. They paid a fixed tax-rate of 33,3 % on agricultural products, including wheat, barley, fruit, beehives, vineyards and goats; a total of 10,400 akçe. 6,5/24 of the revenue went to a Waqf. Marom and Taxel have shown that during the seventeenth to eighteenth centuries, nomadic economic and security pressures led to settlement abandonment around Majdal ‘Asqalān, and the southern coastal plain in general. The population of abandoned villages moved to surviving settlements, while the lands of abandoned settlements continued to be cultivated by neighboring villages. Thus, Julis absorbed the lands of Bardagha, mentioned separately as an inhabited village in the Ottoman tax registers of the 16th century. In 1838, Julis was noted as a village in the District of Gaza. In 1863 Victor Guérin found the village to be located on a hillock and containing five hundred inhabitants. It had a oualy, dedicated to Scheik Mohammed, was internally decorated with two fragments. Guérin further noted that "Several marble columns are laid across the mouth of the well, with furniture arranged around it." An Ottoman village list of about 1870 indicated 101 houses and a population of 307, though the population count included only men. In 1882, the PEF's Survey of Western Palestine described Julis as mostly built of adobe brick structures, and it had a well to the south, and a pool surrounded by gardens to the northeast. In the 1922 census of Palestine, conducted by the British Mandate authorities, Jules had a population of 481 Muslims, increasing in the 1931 census to 682, still all Muslims, in 165 houses. The village was laid out in a square, sandwiched between the two highways and bounded at one end by the traffic circle where they intersected. Its adobe and cement houses were constructed close together. The village had a mosque, and a shrine dedicated to Shaykh Khayr. According to local tradition, Khayr was a Muslim soldier killed fighting against the Crusaders. Village shops were scattered along the highway and in 1937 a school was opened; it had an enrollment of 86 students in the mid-1940s. Underground water was abundant in Julis and was used for domestic methods. During World War II, the British authorities built a highway that passed through Julis parallel and feeding traffic to the coastal highway. The road also intersected at the village with the highway leading from al-Majdal (Ashkelon) to the Jerusalem–Jaffa highway. This gave the village importance as a transportation center. The British also constructed a military camp in Julis to control the junction. In the 1945 statistics Julis had a population of 1,030 Muslims, with a total of 13,584 dunams of land, according to an official land and population survey. Of this, 1,360 dunams were used for citrus and bananas, 931 for plantations and irrigable land, 10,803 for cereals, while 30 dunams were built-up land. In addition to agriculture, residents practiced animal husbandry which formed was an important source of income for the town. In 1943, they owned 278 heads of cattle, 346 sheep over a year old, 138 goats over a year old, 35 camels, 5 horses, 14 mules, 114 donkeys, 1010 fowls, and 776 pigeons. On May 27–28, 1948, the Givati Brigade's First Battalion captured a military barracks in Julis during Operation Barak, but failed to gain control of the village itself. Egyptian forces attempted to recapture it almost immediately. According to the History of the Haganah, "The defenders of the place [Givati forces] blocked enemy units which tried... to infiltrate the barracks from the direction of the village of Julis." The Haganah account says that Julis was captured on June 11, as the Givati's Third Battalion mounted a number of operations to occupy a number of villages before the first truce of the war took place. However, in Gamal Abdel Nasser's memoirs, he recalled the maneuvers having taken place soon after the truce came into effect. At the end of the truce, Julis became one of the many main positions the Egyptians failed to recapture. The Egyptian Army's Sixth Battalion which Nasser was chief of staff of, was ordered to take back the position. In later years, Nasser was very critical of the operation's planning, writing "Once again we were a facing a battle for which we had no preparation. We had no information about the enemy at Julis." In the few hours before his unit was to move towards Julis, Nasser organized a quick reconnaissance of the position. During the course of the battle, his commanding officer ordered him to participate in the actual fighting, leaving his unit without direction or coordination. After getting hold of a few aerial photographs of the village, Nasser convinced his commander that "even if we had succeeded in entering Julis... it would have turned into a cemetery for our forces." He argued that Julis was indefensible without the barracks which overlooked it. On July 10, after many Egyptian casualties, the battle was called off. According to the Haganah, the Givati units repulsed an Egyptian attack in which no Israeli soldier was injured. A close colleague of Nasser, Isma'il Mohieddin was killed during the battle. Following the war the area was incorporated into the State of Israel and the moshav of Hodaya was established on village lands southwest of the village site in 1949. According to Palestinian historian Walid Khalidi, "Only a few houses remain. Most of them are made of cement, and have simple architectural features: flat roofs and rectangular doors and windows. One has two storeys and another has an 'illiyya. (A guest room on the top floor.) One house, in the southwestern section of the site, is occupied by Jewish residents." References Bibliography External links |
======================================== |
[SOURCE: https://en.wikipedia.org/wiki/Ultra-Orthodox_Jews] | [TOKENS: 13145] |
Contents Haredi Judaism Haredi Judaism (Hebrew: יהדות חֲרֵדִית, romanized: Yahadut Ḥaredit, IPA: [χaʁeˈdi]) is a branch of Orthodox Judaism that is characterized by its strict interpretation of religious sources and its accepted halakha (Jewish law) and traditions, in opposition to more accommodating values and practices. Its members are often referred to as ultra-Orthodox in English, a term considered pejorative by many of its adherents, who prefer the term strictly Orthodox or Haredi (plural: Haredim). Haredim regard themselves as the most authentic custodians of Jewish religious law and tradition which, in their opinion, is binding and unchangeable. Many consider all other contemporary expressions of Judaism, including Modern Orthodoxy, as "deviations from God's laws", although other movements of Judaism would disagree. Some scholars have suggested that Haredi Judaism is a reaction to societal changes, including political emancipation, the Haskalah movement derived from the Enlightenment, acculturation, secularization, religious reform in all its forms from mild to extreme, and the rise of the Jewish national movements. In contrast to Modern Orthodox Jews, Haredim segregate themselves from other parts of society, although some Haredi communities encourage young people to get a professional degree or establish a business. Furthermore, some Haredi groups, like Chabad-Lubavitch, encourage outreach to non-observant and unaffiliated Jews. As of 2020, there were about 2.1 million Haredim globally, representing 14% of the world's Jewish population. Haredim primarily live in Israel (17% of Israeli Jews and 14% of Israel's total population), North America (12% of American Jews), and Western Europe (most notably Antwerp and Stamford Hill in London). Absence of intermarriage, coupled with both a high birth and retention rate, spur rapid growth of the Haredi population, which is on pace to more than double every 20 years. Their numbers have been further boosted since the 1970s by secular Jews adopting a Haredi lifestyle as part of the baal teshuva movement; however, this has been somewhat offset by those leaving. Terminology The term Haredi is a Modern Hebrew adjective derived from the Biblical verb hared, which appears in the Book of Isaiah (66:2; its plural haredim appears in Isaiah 66:5) and is translated as "[one who] trembles" at the word of God. The word connotes an awe-inspired fear to perform the will of God; it is used to distinguish them from other Orthodox Jews (similar to the names used by Christian Quakers and Shakers to describe their relationship to God). The term most commonly used by outsiders, for example most American news organizations, is ultra-Orthodox Judaism. Hillel Halkin suggests the origins of the term may date to the 1950s, a period in which Haredi survivors of the Holocaust first began arriving in America. However, Isaac Leeser (1806–1868) was described in 1916 as "ultra-Orthodox". The word Haredi is often used in the Jewish diaspora in place of the term ultra-Orthodox, which many view as inaccurate or offensive, it being seen as a derogatory term suggesting extremism; English-language alternatives that have been proposed include fervently Orthodox, strictly Orthodox, or traditional Orthodox. Others, however, dispute the characterization of the term as pejorative. Ari L. Goldman, a professor at Columbia University, notes that the term simply serves a practical purpose to distinguish a specific part of the Orthodox community, and is not meant as pejorative. Others, such as Samuel Heilman, criticized terms such as ultra-Orthodox and traditional Orthodox, arguing that they misidentify Haredi Jews as more authentically Orthodox than others, as opposed to adopting customs and practises that reflect their desire to separate from the outside world. The community has sometimes been characterized as traditional Orthodox, in contradistinction to the Modern Orthodox, the other major branch of Orthodox Judaism, and not to be confused with the movement represented by the Union for Traditional Judaism, which originated in Conservative Judaism. Haredi Jews also use other terms to refer to themselves. Common Yiddish words include Yidn (Jews), erlekhe Yidn (virtuous Jews), ben Torah (son of the Torah), frum (pious), and heimish (home-like; i.e., "our crowd"). In Israel, Haredi Jews are sometimes also called by the derogatory slang words dos (plural dosim), that mimics the traditional Ashkenazi Hebrew pronunciation of the Hebrew word datiyim (religious), and more rarely, sh'chorim (blacks), a reference to the black clothes they typically wear; a related informal term used in English is black hat. Population Due to its imprecise definition, lack of data collection, and rapid change over time, estimating the global Haredi population is difficult. The true number of Haredim may be significantly underestimated due to their reluctance to participate in surveys and censuses. In 1992, out of a total of 1,500,000 Orthodox Jews worldwide, about 550,000 were Haredi (half of them in Israel). One estimate given in 2011 stated that there were approximately 1.3 million Haredi Jews globally. Studies have shown a very high growth rate, with a large young population. Haredi population grew to 2.1 million in 2020 and is expected to double by 2040. The vast majority of Haredi Jews are Ashkenazi. However, some 20% of the Haredi population are thought to belong to the Sephardic Haredi stream. In recent decades, Haredi society has grown due to the addition of a religious population that identifies with the Shas movement. The percentage of people leaving the Haredi population has been estimated between 6% and 18%. Israel has the largest Haredi population. In 1948, there were about 35,000 to 45,000 Haredi Jews in Israel. By 1980, Haredim made up 4% of the Israeli population. Haredim made up 9.9% of the Israeli population in 2009, with 750,000 out of 7,552,100; by 2014, that figure had risen to 11.1%, with 910,500 Haredim out of a total Israeli population of 8,183,400. According to a December 2017 study conducted by the Israeli Democracy Institute, the number of Haredi Jews in Israel exceeded 1 million in 2017, making up 12% of the population in Israel. In 2019, Haredim reached a population of almost 1,126,000; over the next year, it reached 1,175,000 (12.6% of total population). By the end of 2023, it reached almost 1,335,000, or 13.6% of total population; at the end of 2024, it passed over 1,392,000, thus representing 13.9% of the population, and by the end of 2025 it numbered over 1,452,000 or 14.3% of the total population of Israel. The number of Haredi Jews in Israel continues to rise rapidly, with their current population growth rate being 4% per year. The number of children per woman is 7.2, and the share of Haredim among those under the age of 20 was 16.3% in 2009 (29% of Jews). By 2030, the Haredi Jewish community is projected to make up 16% of the total population, and by 2065, a third of the Israeli population, including non-Jews. By then, one in two Israeli children would be Haredi. It is also projected that the number of Haredim in 2059 may be between 2.73 and 5.84 million, of an estimated total number of Israeli Jews between 6.09 and 9.95 million. The largest Israeli Haredi concentrations are in Jerusalem, Bnei Brak, Modi'in Illit, Beitar Illit, Beit Shemesh, Kiryat Ye'arim, Ashdod, Rekhasim, Safed, and El'ad. Two Haredi cities, Kasif and Harish, are planned.[citation needed] The United States has the second largest Haredi population, which has a growth rate on pace to double every 20 years. In 2000, there were 360,000 Haredi Jews in the US (7.2 per cent of the approximately 5 million Jews in the U.S.); by 2006, demographers estimate the number had grown to 468,000 (30% increase), or 9.4 percent of all U.S. Jews. In 2013, it was estimated that there were 530,000 total ultra-Orthodox Jews in the United States, or 10% of all American Jews. By 2011, 61% of all Jewish children in Eight-County New York City metropolitan area were Orthodox, with Haredim making up 49%. In 2020, it was estimated that there were approximately 700,000 total ultra-Orthodox Jews in the United States, or 12% of all American Jews. This number is expected to grow significantly in the coming years, due to high Haredi birth rates in America. Most American Haredi Jews live in the greater New York metropolitan area. The largest centers of Haredi and Hasidic life in New York are found in Brooklyn. The New York City borough of Queens is home to a growing Haredi population, mainly affiliated with the Yeshiva Chofetz Chaim and Yeshivas Ohr HaChaim in Kew Gardens Hills and Yeshiva Shaar Hatorah in Kew Gardens. Many of the students attend Queens College. There are major yeshivas and communities of Haredi Jews in Far Rockaway, such as Yeshiva of Far Rockaway and a number of others. Hasidic shtibelach exist in these communities as well, mostly catering to Haredi Jews who follow Hasidic customs, while living a Litvish or Modern Orthodox cultural lifestyle, although small Hasidic enclaves do exist, such as in the Bayswater section of Far Rockaway. One of the oldest Haredi communities in New York is on the Lower East Side, home to the Mesivtha Tifereth Jerusalem. Washington Heights, in northern Manhattan, is the historical home to German Jews, with Khal Adath Jeshurun and Yeshiva Rabbi Samson Raphael Hirsch. The presence of Yeshiva University attracts young people, many of whom remain in the area after graduation. The Yeshiva Sh'or Yoshuv, together with many synagogues in the Lawrence neighborhood and other Five Towns neighborhoods, such as Woodmere and Cedarhurst, have attracted many Haredi Jews. The Hudson Valley, north of New York City, has the most rapidly growing Haredi communities, such as the Hasidic communities in Kiryas Joel of Satmar Hasidim, and New Square of the Skver. A vast community of Haredi Jews lives in the Monsey, New York, area. There are significant Haredi communities in Lakewood (New Jersey), home to the largest non-Hasidic Lithuanian yeshiva in America, Beth Medrash Govoha. There are also sizable communities in Teaneck, Englewood, Mahwah, Passaic and Edison, where a branch of the Rabbi Jacob Joseph Yeshiva opened in 1982. There is also a community of Syrian Jews favorable to the Haredim in their midst in Deal, New Jersey. The Haredi community of New Haven has close to 150 families and a number of thriving Haredi educational institutions. Waterbury, Connecticut has a growing Haredi community, in Waterbury proper, and in the neighboring areas of Blueridge and Naugatuck. Baltimore, Maryland, has a large Haredi population. The major yeshiva is Yeshivas Ner Yisroel, founded in 1933, with thousands of alumni and their families. Ner Yisroel is also a Maryland state-accredited college, and has agreements with Johns Hopkins University, Towson University, Loyola College in Maryland, University of Baltimore, and University of Maryland, Baltimore County, allowing undergraduate students to take night courses at these colleges and universities in a variety of academic fields. The agreement also allows the students to receive academic credits for their religious studies. Silver Spring, Maryland, and its environs has a growing Haredi community, mostly of highly educated and skilled professionals working for the United States government in various capacities, most living in Kemp Mill, White Oak, and Woodside, and many of its children attend the Yeshiva of Greater Washington and Yeshivas Ner Yisroel in Baltimore. Aventura, Sunny Isles Beach, Golden Beach, Surfside and Bal Harbour are home to a large and growing Haredi population. The community is long-established in the area, with several synagogues including The Shul of Bal Harbour, Young Israel of Bal Harbour, Aventura Chabad, Beit Rambam, Safra Synagogue of Aventura, and Chabad of Sunny Isles; mikvehs, Jewish schools and kosher restaurants. The community has recently grown much further, due to many Orthodox Jews from New York moving to Florida during the COVID-19 pandemic. North of Miami, the communities of Boca Raton, Lauderhill, Boynton Beach, and Hollywood have significant Haredi populations. Los Angeles has many Haredi Jews, most living in the Pico-Robertson and Fairfax (Fairfax Avenue-La Brea Avenue) areas. Chicago is home to the Haredi Telshe Yeshiva, with many other Haredim living in the city. Haredim in Philadelphia primarily live in Bala Cynwyd, and the community is centered around Aish HaTorah and the Philadelphia Community Kollel. In Pittsburgh a small yeshiva opened in 1945. Today there are approximately 200 Chabad families living in the Squirrel Hill neighborhood. Kingston has a young growing Chabad Haredi community which has been growing steadily over the past 20 years since the first families moved there when a yeshiva was opened. Denver has a large Haredi population of Ashkenazi origin, dating back to the early 1920s. The Haredi Denver West Side Jewish Community adheres to Litvak Jewish traditions (Lithuanian), and has several congregations located within their communities. Boston and Brookline, Massachusetts, have the largest Haredi populations in New England. One of the oldest Haredi Lithuanian yeshivas, Telshe Yeshiva, transplanted itself to Cleveland in 1941. Beachwood, Ohio has a large and growing Haredi community, and is a heavily Jewish suburb of Cleveland. The haredi community is centered around the Beachwood Kehilla and Green Road Synagogue, has a mikvah and a Jewish day school. In 1998, the Haredi population in the Jewish community of the United Kingdom was estimated at 27,000 (13% of affiliated Jews). The largest communities are located in London, particularly Stamford Hill, Golders Green, Hendon, Edgware; in Salford and Prestwich in Greater Manchester; and in Gateshead. A 2007 study asserted that three out of four British Jewish births were Haredi, who then accounted for 17% of British Jews (45,500 out of around 275,000). Another study in 2010 established that there were 9,049 Haredi households in the UK, which would account for a population of nearly 53,400, or 20% of the community. The Board of Deputies of British Jews has predicted that the Haredi community will become the largest group in Anglo-Jewry within the next three decades: In comparison with the national average of 2.4 children per family, Haredi families have an average of 5.9 children, and consequently, the population distribution is heavily biased to the under-20-year-olds. By 2006, membership of Haredi synagogues had doubled since 1990. British Haredi fertility rate has also been estimated to be as high as 6.9 children per woman. An investigation by The Independent in 2014 reported that more than 1,000 children in Haredi communities were attending illegal schools where secular knowledge is banned, and they learn only religious texts, meaning they leave school with no qualifications and often unable to speak any English. The 2018 Survey by the Jewish Policy Research (JPR) and the Board of Deputies of British Jews showed that the high birth rate in the Haredi and Orthodox community reversed the decline in the Jewish population in Britain. In 2020, it was estimated that there were approximately 76,000 total ultra-Orthodox Jews in the United Kingdom, or 25% of all British Jews, a significant increase from 1998 and 2010. About 25,000 Haredim live in the Jewish community of France, mostly people of Sephardic, Maghrebi Jewish descent. Important communities are located in Paris (19th arrondissement), Strasbourg, and Lyon. Other important communities, mostly of Ashkenazi Jews, are the Antwerp community in Belgium, as well as in the Swiss communities of Zürich and Basel, and in the Dutch community in Amsterdam. There is also a Haredi community in Vienna, in the Jewish community of Austria. Other countries with significant Haredi populations include: Canada, with a total number of 30,000 Haredim, with large Haredi centres in Montreal and Toronto; South Africa, primarily in Johannesburg; and an estimated 7,500 Haredim in Australia, centred in Melbourne. Haredi communities also exist in Argentina, especially in Buenos Aires, and in Brazil, primarily in São Paulo. A Haredi city is under construction (2021) in Mexico near Ixtapan de la Sal. Decades after The Holocaust, Haredim are growing again in Budapest, opening several new synagogues and two mikvehs in the city over the past couple of years. History Throughout Jewish history, Judaism has always faced internal and external challenges to its beliefs and practices which have emerged over time and produced counter-responses. According to its adherents, Haredi Judaism is a continuation of Rabbinic Judaism, and the immediate forebears of contemporary Haredi Jews were the Jewish religious traditionalists of Central and Eastern Europe who fought against secular modernization's influence which reduced Jewish religious observance. Indeed, adherents of Haredi Judaism, just like Rabbinic Jews, see their beliefs as part of an unbroken tradition which dates back to the revelation at Sinai. However, most historians of Orthodoxy consider Haredi Judaism, in its most modern incarnation, to date back to the beginning of the 20th century. For centuries, before Jewish emancipation, European Jews were forced to live in ghettos where Jewish culture and religious observance were preserved. Change began in the wake of the Age of Enlightenment, when some European liberals sought to include the Jewish population in the emerging empires and nation states. The influence of the Haskalah movement (Jewish Enlightenment) was also evident. Supporters of the Haskalah held that Judaism must change, in keeping with the social changes around them. Other Jews insisted on strict adherence to halakha (Jewish law and custom). In Germany, the opponents of Reform rallied to Samson Raphael Hirsch, who led a secession from German Jewish communal organizations to form a strictly Orthodox movement, with its own network of synagogues and religious schools. His approach was to accept the tools of modern scholarship and apply them in defence of Orthodox Judaism. In the Polish–Lithuanian Commonwealth (including areas traditionally considered Lithuanian), Jews true to traditional values gathered under the banner of Agudas Shlumei Emunei Yisroel. Moses Sofer was opposed to any philosophical, social, or practical change to customary Orthodox practice. Thus, he did not allow any secular studies to be added to the curriculum of his Pressburg Yeshiva. Sofer's student Moshe Schick, together with Sofer's sons Shimon and Samuel Benjamin, took an active role in arguing against the Reform movement. Others, such as Hillel Lichtenstein, advocated an even more stringent position for Orthodoxy. A major historic event was the meltdown after the Universal Israelite Congress of 1868–1869 in Pest, Hungary. In an attempt to unify all streams of Judaism under one constitution, the Orthodox offered the Shulchan Aruch as the ruling Code of law and observance. This was dismissed by the reformists, leading many Orthodox rabbis to resign from the Congress and form their own social and political groups. Hungarian Jewry split into two major institutionally sectarian groups: Orthodox, and Neolog. However, some communities refused to join either of the groups, calling themselves "Status Quo".[citation needed] Schick demonstrated support in 1877 for the separatist policies of Samson Raphael Hirsch in Germany. Schick's own son was enrolled in the Hildesheimer Rabbinical Seminary, headed by Azriel Hildesheimer, which taught secular studies. Hirsch, however, did not reciprocate, and expressed astonishment at Schick's halakhic contortions in condemning even those Status Quo communities that clearly adhered to halakha. Lichtenstein opposed Hildesheimer, and his son Hirsh Hildesheimer, as they made use of the German language in sermons from the pulpit and seemed to lean in the direction of Zionism. Shimon Sofer was somewhat more lenient than Lichtenstein on the use of German in sermons, allowing the practice as needed for the sake of keeping cordial relations with the various governments. Likewise, he allowed extra-curricular studies of the gymnasium for students whose rabbinical positions would be recognized by the governments, stipulating the necessity to prove the strict adherence to the God-fearing standards per individual case. In 1912, the World Agudath Israel was founded, to differentiate itself from the Torah Nationalist Mizrachi and secular Zionist organizations. It was dominated by the Hasidic rebbes and Lithuanian rabbis and roshei yeshiva (deans). The organization nominated rabbis who subsequently were elected as representatives in the Polish legislature Sejm, such as Meir Shapiro and Yitzhak-Meir Levin. Not all Hasidic factions joined the Agudath Israel, remaining independent instead, such as Machzikei Hadat of Galicia. In 1919, Yosef Chaim Sonnenfeld and Yitzchok Yerucham Diskin founded the Edah HaChareidis as part of Agudath Israel in then-Mandate Palestine. In 1924, Agudath Israel obtained 75 percent of the votes in the Kehilla elections. The Orthodox community polled some 16,000 of a total 90,000 at the Knesseth Israel in 1929. But Sonnenfeld lobbied Sir John Chancellor, the High Commissioner, for separate representation in the Palestine Communities Ordinance from that of the Knesseth Israel. He explained that the Agudas Israel community would cooperate with the Vaad Leumi and the National Jewish Council in matters pertaining to the municipality, but sought to protect its religious convictions independently. The community petitioned the Permanent Mandates Commission of the League of Nations on this issue. The one community principle was victorious, despite their opposition, but this is seen as the creation of the Haredi community in Israel, separate from the other Orthodox and Zionist movements. In 1932, Sonnenfeld was succeeded by Yosef Tzvi Dushinsky, a disciple of the Shevet Sofer, one of the grandchildren of Moses Sofer. Dushinsky promised to build up a strong Jewish Orthodoxy at peace with the other Jewish communities and the non-Jews. In general, the present-day Haredi population originate from two distinct post-Holocaust waves. The vast majority of Hasidic and Litvak communities were destroyed during the Holocaust. Although Hasidic customs have largely been preserved, the customs of Lithuanian Jewry, including its unique Hebrew pronunciation, have been almost lost. Litvish customs are still preserved primarily by the few older Jews who were born in Lithuania prior to the Holocaust. In the decade or so after 1945, there was a strong drive to revive and maintain these lifestyles by some notable Haredi leaders. The Chazon Ish was particularly prominent in the early days of the State of Israel. Aharon Kotler established many of the Haredi schools and yeshivas in the United States and Israel; and Joel Teitelbaum had a significant impact on revitalizing Hasidic Jewry, as well as many of the Jews who fled Hungary during the 1956 revolution who became followers of his Satmar dynasty, and became the largest Hasidic group in the world. These Jews typically have maintained a connection only with other religious family members. As such, those growing up in such families have little or no contact with non-Haredi Jews. The second wave began in the 1970s associated with the religious revival of the so-called baal teshuva movement, although most of the newly religious become Orthodox, and not necessarily fully Haredi.[citation needed] The formation and spread of the Sephardic Haredi lifestyle movement also began in the 1980s by Ovadia Yosef, alongside the establishment of the Shas party in 1984. This led many Sephardi Jews to adopt the clothing and culture of the Lithuanian Haredi Judaism, though it had no historical basis in their own tradition.[citation needed] Many yeshivas were also established specifically for new adopters of the Haredi way of life.[citation needed] The original Haredi population has been instrumental in the expansion of their lifestyle, though criticisms have been made of discrimination towards the later adopters of the Haredi lifestyle in shidduchim (matchmaking) and the school system. Practices and beliefs The Haredim represent the conservative or pietistic form of Jewish fundamentalism, distinct from the radical fundamentalism of Gush Emunim, and emphasising withdrawal from, and disdain for, the secular world, and the creation of an alternative world which insulates the Torah and the life it prescribes from outside influences. Haredi Judaism is not an institutionally cohesive or homogeneous group, but comprises a diversity of spiritual and cultural orientations, generally divided into a broad range of Hasidic courts and Litvishe-Yeshivish streams from Eastern Europe, and Oriental Sephardic Haredi Jews. These groups often differ significantly from one another in their specific ideologies and lifestyles, as well as the degree of stringency in religious practice, rigidity of religious philosophy, and isolation from the general culture that they maintain.[citation needed] Some Haredis encourage outreach to less observant and unaffiliated Jews and hilonim (secular Israeli Jews). Some scholars, including some secular and Reform Jews, describe the Haredim as "radical fundamentalists". Efforts to keep clear of external influence is a core characteristic of Haredi Judaism. Historically, new mediums of communication such as books, newspapers and magazines, and later tapes, CDs and television, were dealt with by either transforming and controlling the content, or choosing to have rabbinic leadership censor it selectively or altogether. In the modern digital era, difficulty in censoring the Internet and conversely, the Internet's importance, resulted in a decades long and ongoing struggle of comprehension, adaption, and regulation on the part of rabbinical leadership and community activists. These beliefs and practices, which have been interpreted as "isolationist", can bring them into conflict with authorities. In 2018, a Haredi school in the United Kingdom was rated as "inadequate" by the Office for Standards in Education, after repeated complaints were raised about the censoring of textbooks and exam papers which contained mentions of homosexuality, examples of women socializing with men, pictures showing women's shoulders and legs, or information that contradicts a creationist worldview. Haredi life, like Orthodox Jewish life in general, is very family-centered and ordered. Boys and girls attend separate schools, and proceed to higher Torah study, in a yeshiva or seminary, respectively, starting anywhere between the ages of 13 and 18. A significant proportion of young men remain in yeshiva until their marriage (often arranged). After marriage, many Haredi men continue their Torah studies in a kollel. Studying in secular institutions is often discouraged, although educational facilities for vocational training in a Haredi framework do exist. In the United States and Europe, the majority of Haredi males are active in the workforce. For various reasons, in Israel, a majority (56%) of their male members do not work, though some of those are part of the unofficial workforce. Haredi families (and Orthodox Jewish families in general) are usually much larger than non-Orthodox families, with an average of seven children per family, but it is not unheard of for families to have twelve or more children. About 80% of female Haredi Jews in Israel work. Haredi Jews are typically opposed to the viewing of television and films, and the reading of secular newspapers and books. There has been a strong campaign against the Internet, and Internet-enabled mobile phones without filters have also been banned by leading rabbis. In May 2012, 40,000 Haredim gathered at Citi Field, a baseball park in New York City, to discuss the dangers of unfiltered Internet. The event was organized by the Ichud HaKehillos LeTohar HaMachane. The Internet has been allowed for business purposes, so long as filters are installed. In some instances, forms of recreation which conform to Jewish law are treated as antithetical to Haredi Judaism. In 2013, the Rabbinical Court of the Ashkenazi Community in the Haredi settlement of Beitar Illit ruled against Zumba (a type of dance fitness) classes, although they were held with a female instructor and all-female participants. The Court said in part: "Both in form and manner, the activity [Zumba] is entirely at odds with both the ways of the Torah and the holiness of Israel, as are the songs associated to it." Jewish Chicago has lauded the Haredim for their lifestyle, arguing that it has low crime and drugs, and a strong sense of family and community. With Haredi Judaism having a heavy emphasis on marriage, especially while young, some members rely on the shidduch (matchmaking) system. They employ a schadhan (a professional matchmaker) to support them in their search for a spouse. While there is no current statistical data showing how many people use the services of a schadhan, it is estimated that the vast majority of Haredi couples were paired by one. However, with the broader societal shift to online dating, matchmaking in Orthodox and Haredi Judaism has started making inroads online. Vastly different from the most popular online dating services, apps like Shidduch pair couples based upon shared values and life goals. To do this, users fill-out a digital resume. The app was made possible by a partnership between its developers and the Orthodox Union — the same group responsible for kosher food certification ("Circle-U"). The standard mode of dress for males of the Lithuanian stream is a black or navy suit and a white shirt. Headgear includes black Fedora or Homburg hats, with black skull caps. Pre-war Lithuanian yeshiva students also wore light coloured suits, along with beige or grey hats, and prior to the 1990s, it was common for Americans of the Lithuanian stream to wear coloured shirts throughout the week, reserving white shirts for Shabbos. Beards are common among Haredi and many other Orthodox Jewish men, and Hasidic men will almost never be clean-shaven. Women adhere to the laws of modest dress, and wear long skirts and sleeves, high necklines, and, if married, some form of hair covering. Haredi women never wear trousers, although most do wear pajama-trousers within the home at night. Over the years, it has become popular among some Haredi women to wear sheitels (wigs), that are thought to be more attractive than their own natural hair (drawing criticism from some more conservative Haredi rabbis). Mainstream Sephardi Haredi rabbi Ovadia Yosef forbade the wearing of wigs altogether. Haredi women often dress more freely and casually within the home, as long as the body remains covered in accordance with the halakha. More modernized Haredi women are somewhat more lenient in matters of their dress, and some follow the latest trends and fashions, while conforming to halakha. Non-Lithuanian Hasidic men and women differ from the Lithuanian stream by having a much more specific dress code, the most obvious difference for men being the full-length suit jacket (rekel) on weekdays, and the fur hat (shtreimel) and silk caftan (bekishe) on the Sabbath. Haredi neighborhoods have been said by some to be safer, with less violent crime, although this is a generalization, and even that may apply to only specific communities, rather than all. In Israel, the entrances to some of the most extreme Haredi neighborhoods are fitted with signs that ask for modest clothing to be worn. Some areas are known to have "modesty patrols", and people dressed in ways perceived as immodest may suffer harassment, and advertisements featuring scantily dressed models may be targeted for vandalism. These concerns are also addressed through public lobbying and legal avenues. During the week-long Rio Carnival in Rio de Janeiro, many of the city's 7,000 Orthodox Jews feel compelled to leave the town, due to the immodest exposure of participants. In 2001, Haredi campaigners in Jerusalem succeeded in persuading the Egged bus company to get all their advertisements approved by a special committee. By 2011, Egged had gradually removed all bus adverts that featured women, in response to their continuous defacement. A court order that stated such action was discriminatory led to Egged's decision not to feature people at all (neither male nor female). Depictions of certain other creatures, such as space aliens, were also banned, in order not to offend Haredi sensibilities. Haredi Jews also campaign against other types of advertising that promote activities they deem offensive or inappropriate. Due to halakha, i.e., activities that Orthodox Jews believe are prohibited on Shabbat, most state-run buses in Israel do not run on Saturdays, regardless of whether riders are Orthodox, or even whether they are Jewish. In a similar vein, Haredi Jews in Israel have demanded that the roads in their neighborhoods be closed on Saturdays, vehicular traffic being viewed as an "intolerable provocation" upon their religious lifestyle (see Driving on Shabbat in Jewish law). In most cases, the authorities granted permission after Haredi petitioning and demonstrations, some of them including fierce clashes between Haredi Jews and secular counter-demonstrators, and violence against police and motorists. While Jewish modesty law requires gender separation under various circumstances, observers have contended that there is a growing trend among some groups of Hasidic Haredi Jews to extend its observance to the public arena. In the Hasidic village of Kiryas Joel, New York, an entrance sign asks visitors to "maintain sex separation in all public areas", and the bus stops have separate waiting areas for men and women. In New Square, another Hasidic enclave, men and women are expected to walk on opposite sides of the road. In Israel, Jerusalem residents of Mea Shearim were banned from erecting a street barrier dividing men and women during the week-long Sukkot festival's nightly parties; and street signs requesting that women avoid certain pavements in Beit Shemesh have been repeatedly removed by the municipality. Since 1973, buses catering to Haredi Jews running from Rockland County and Brooklyn into Manhattan have had separate areas for men and women, allowing passengers to conduct on-board prayer services. Although the lines are privately operated, they serve the general public, and in 2011, the set-up was challenged on grounds of discrimination, and the arrangement was deemed illegal. During 2010–2012, there was much public debate in Israel surrounding the existence of segregated Haredi Mehadrin bus lines (whose policy calls for both men and women to stay in their respective areas: men in the front of the bus, and women in the rear of the bus) following an altercation that occurred after a woman refused to move to the rear of the bus to sit among the women. A subsequent court ruling stated that while voluntary segregation should be allowed, forced separation is unlawful. Israeli national airline El Al has agreed to provide gender-separated flights in consideration of Haredi requirements. Education in the Haredi community is strictly segregated by sex. Yeshiva education for boys is primarily focused on the study of Jewish scriptures, such as the Torah and Talmud (non-Hasidic yeshivas in the United States teach secular studies in the afternoon); girls obtain studies both in Jewish religious education as well as broader secular subjects. In 1930s Poland, the Agudath Israel movement published its own Yiddish-language paper, Dos Yiddishe Tagblatt. In 1950, the Agudah started printing Hamodia, a Hebrew-language Israeli daily. Haredi publications tend to shield their readership from objectionable material, and perceive themselves as a "counterculture", desisting from advertising secular entertainment and events. The editorial policy of a Haredi newspaper is determined by a rabbinical board, and every edition is checked by a rabbinical censor. A strict policy of modesty is characteristic of the Haredi press in recent years, and pictures of women are usually not printed. In 2009, the Israeli daily Yated Ne'eman doctored an Israeli cabinet photograph replacing two female ministers with images of men, and in 2013, the Bakehilah magazine pixelated the faces of women appearing in a photograph of the Warsaw Ghetto Uprising. The mainstream Haredi political Shas party also refrains from publishing female images. Among Haredi publishers which have not adopted this policy is ArtScroll, which does publish pictures of women in their books. No coverage is given to serious crime, violence, sex, or drugs, and little coverage is given to non-Orthodox streams of Judaism. Inclusion of "immoral" content is avoided, and when publication of such stories is a necessity, they are often written ambiguously. The Haredi press generally takes an ambivalent stance towards Zionism and gives more coverage to issues that concern the Haredi community, such as the drafting of girls and yeshiva students into the army, autopsies, and Shabbat observance. In Israel, it portrays the secular world as "spitefully anti-Semitic", and describes secular youth as "mindless, immoral, drugged, and unspeakably lewd". Such attacks have led to Haredi editors being warned about libelous provocations. While the Haredi press is extensive and varied in Israel, only around half the Haredi population reads newspapers. Around 10% read secular newspapers, while 40% do not read any newspaper at all. According to a 2007 survey, 27% read the weekend Friday edition of Hamodia, and 26% the Yated Ne'eman. In 2006, the most-read Haredi magazine in Israel was the Mishpacha weekly, which sold 110,000 copies. Other popular Haredi publications include Ami Magazine and The Flatbush Jewish Journal. Haredi leaders have at times suggested a ban on the internet and any internet-capable device, their reasoning being that the immense amount of information can be corrupting, and the ability to use the internet with no observation from the community can lead to individuation. Some Haredi businessmen utilize the internet throughout the week, but they still observe Shabbat in every aspect by not accepting or processing orders from Friday evening to Saturday evening. They utilize the internet under strict filters and guidelines. The Kosher cell phone was introduced to the Jewish public with the sole ability to call other phones. It was unable to utilize the internet, text other phones, and had no camera feature. In fact, a kosher phone plan was created, with decreased rates for kosher-to-kosher calls, to encourage community. News hotlines are an important source of news in the Haredi world. Since many Haredi Jews do not listen to the radio or have access to the internet, even if they read newspapers, they are left with little or no access to breaking news. News hotlines were formed to fill this gap, and many have expanded to additional fields over time. Currently, many news lines provide rabbinic lectures, entertainment, business advice, and similar services, in addition to their primary function of reporting the news. Many Hasidic sects maintain their own hotlines, where relevant internal news is reported and the group's perspective can be advocated for. In the Israeli Haredi community, there are dozens of prominent hotlines, in both Yiddish and Hebrew. Some Haredi hotlines have played significant public roles. In Israel From the founding of Zionism in the 1890s, Haredi Jews leaders voiced objections to its secular orientation. After the establishment of the State of Israel, some Haredi Jews observed the Israeli Independence Day as a day of mourning and referred to Israeli state-holidays as byimey edeyhem ("idolatrous holidays"). The chief political division among Haredi Jews has been in their approach to the State of Israel. After Israeli independence, different Haredi movements took varying positions on it. Only a minority of Haredi Jews consider themselves to be Zionists. Haredim who do not consider themselves Zionists fall into two-camps: non-Zionist, and anti-Zionist. Non-Zionist Haredim, who comprise the majority, do not object to the State of Israel as an independent Jewish state, and many even consider it to be positive, but they do not believe that it has any religious significance. Anti-Zionist Haredim, who are a minority, but are more publicly visible than the non-Zionist majority, believe that any Jewish independence prior to the coming of the Messiah is a sin. The ideologically non-Zionist United Torah Judaism alliance comprising Agudat Yisrael and Degel HaTorah (and the umbrella organizations World Agudath Israel and Agudath Israel of America) represents a moderate and pragmatic stance of cooperation with the State of Israel, and participation in the political system. UTJ has been a participant in numerous coalition governments, seeking to influence state and society in a more religious direction and maintain welfare and religious funding policies. In general, their position is supportive of Israel. Haredim who are stridently anti-Zionist are under the umbrella of Edah HaChareidis, who reject participation in politics and state funding of its affiliated institutions, in contradistinction to Agudah-affiliated institutions. Neturei Karta is a very small activist organization of anti-Zionist Haredim, whose controversial activities have been strongly condemned, including by other anti-Zionist Haredim. Haredi support is often required to form coalition governments in the Knesset. In recent years, some rebbes affiliated with Agudath Israel, such as the Sadigura rebbe Avrohom Yaakov Friedman, have taken stances closer to the Israeli right wing on security, settlements and withdrawal from the Gaza Strip. Shas represents Sephardi and Mizrahi Haredim, and, while having many points in common with Ashkenazi Haredim, differs from them by its more enthusiastic support for the State of Israel and the IDF. The Sikirim group, an anti-Zionist group composed of Haredi Jews, is considered a radical organization by Israelis. The purpose of marriage in the Haredi (and Orthodox) viewpoint is for the purpose of companionship, as well as for the purpose of having children. There is a high rate of marriage in the Haredi community. 83% are married, compared to the non-Haredi community in Israel of 63%. Marriage is viewed as holy, and as the natural home for a man and a woman to truly love each other. In 2016, the divorce rate in Israel was 5% among the Haredi population, compared to the general population rate of 14%. In 2016, Haaretz claimed that divorces among Haredim are increasing in Israel. In 2017, some predominantly Haredi cities reported the highest growth rates in divorce in the Israel, in the context of generally falling rates of divorce, and in 2018, some predominantly Haredi cities reported drops in divorce, in the context of generally rising rates of divorce. When the divorce is linked to one spouse leaving the community, the one who chooses to leave is often shunned from his or her communities and forced to abandon their children, as most courts prefer keeping children in an established status quo. Haredim primarily educate their children in their own private schools, starting with chederim for pre-school to primary school ages, to yeshivos for boys from secondary school ages, and in seminaries, often called Bais Yaakovs, for girls of secondary school ages. Only Jewish religiously observant students are admitted, and parents must agree to abide by the rules of the school to keep their children enrolled. Yeshivas are headed by rosh yeshivas (deans) and principals. Many Hasidic schools in Israel, Europe, and North America teach few (or no) secular subjects, while some of the Litvish (Lithuanian style) schools in Israel follow educational policies to the Hasidic school. In the U.S., most teach secular subjects to boys and girls, as part of a dual curriculum of secular subjects (generally called "English") and Torah subjects. Yeshivas teach mostly Talmud and Rabbinic literature, while the girls' schools teach Jewish Law, Midrash, and Tanach (Hebrew Bible). Between 2007 and 2017, the number of Haredim studying in higher education had risen from 1,000 to 10,800. In 2007, the Kemach Foundation was established to become an investor in the sector's social and economic development, and provide opportunities for employment. Through the philanthropy of Leo Noé of London, later joined by the Wolfson family of New York and Elie Horn from Brazil, Kemach has facilitated academic and vocational training. With a $22 million budget, including government funding, Kemach provides individualized career assessment, academic or vocational scholarships, and job placement for the entire Haredi population in Israel. The Foundation is managed by specialists who, coming from the Haredi sector themselves, are familiar with the community's needs and sensitivities. By April 2014, more than 17,800 Haredim have received the services of Kemach, and more than 7,500 have received, or continue to receive, monthly scholarships to fund their academic or vocational studies. From 500 graduates, the net benefits to the government would be 80.8 million NIS if they work for one year, 572.3 million NIS if they work for 5 years, and 2.8 billion NIS (discounted) if they work for 30 years. The Council for Higher Education announced in 2012 that it was investing NIS 180 million over the following five years to establish appropriate frameworks for the education of Haredim, focusing on specific professions. The largest Haredi campus in Israel is The Haredi Campus - The Academic College Ono. In the midst of a controversy surrounding the limited secular education in some Haredi yeshivas, New York City mayor Eric Adams held up the Haredi yeshiva model as a model to emulate, arguing that "We need to ask, 'What are we doing wrong in our schools?' And learn what you are doing in the yeshivas to improve education." Tucker Carlson, in an interview with a former yeshiva student, observed that the yeshiva system, with its emphasis on asking questions, "seems like a great education". Upon the establishment of the State of Israel in 1948, universal conscription was instituted for all able-bodied Jewish males. However, military-aged Haredi men were exempted from service in the Israel Defense Forces (IDF) under the Torato Umanuto arrangement, which officially granted deferred entry into the IDF for yeshiva students, but in practice allowed young Haredi men to serve for a significantly reduced period of time or bypass military service altogether. At that time, the Haredi population was very low and only 400 individuals were affected. However, the Haredi population rapidly grew. In 2018, the Israel Democracy Institute estimated that the Haredim comprised 12% of Israel's total population and 15% of its Jewish population. Haredim are also younger than the general population. Their absence from the IDF attracts significant resentment from secular Israelis. The most common criticisms of the exemption policy are: Over the years, as many as 1,000 Haredi Jews have volunteered to serve in a Haredi Jewish unit of the IDF known as the Netzah Yehuda Battalion, or Nahal Haredi. The vast majority of Haredi men, however, continue to receive deferments from military service. Haredim usually reject the practice of IDF service and contend that: The Torato Umanuto arrangement was enshrined in the Tal Law that came into force in 2002. The High Court of Justice later ruled that it could not be extended in its current form beyond August 2012. A replacement was expected. The IDF was, however, experiencing a shortage of personnel, and there were pressures to reduce the scope of the Torato Umanuto exemption. In March 2014, Israel's parliament approved legislation to end exemptions from military service for Haredi seminary students. The bill was passed by 65 votes to one, and an amendment allowing civilian national service by 67 to one. In June 2024, the Supreme Court of Israel declared any continued exemption of IDF conscription unlawful. The army began drafting 3,000 Haredi men the following month. There has been much uproar in Haredi society following actions towards Haredi conscription. While some Haredim see this as a great social and economic opportunity, others (including leading rabbis among them) strongly oppose this move. Among the extreme Haredim, there have been some more severe reactions. Several Haredi leaders have threatened that Haredi populations would leave the country if forced to enlist. Others have fueled public incitement against secular and National-Religious Jews, and specifically against politicians Yair Lapid and Naftali Bennett, who support and promote Haredi enlistment. Some Haredim have taken to threatening their fellows who agree to enlist, to the point of physically attacking some of them. The Shahar program, also known as Shiluv Haredim (Ultra-Orthodox integration), allows Haredi men aged 22 to 26 to serve in the army for about a year and a half. At the beginning of their service, they study mathematics and English, which are often not well covered in Haredi boy schools. The program is partly aimed at encouraging Haredi participation in the workforce after military service. However, not all beneficiaries seem to be Haredim. As of 2013[update], figures from the Central Bureau of Statistics on employment rates place Haredi women at 73%, close to the 80% for the non-Haredi Jewish women's national figure; while the number of working Haredi men has increased to 56%, it is still far below the 90% of non-Haredi Jewish men nationwide. As of 2021[update], most Haredi boys instead go to yeshivas and then continue to study at yeshiva after getting married. The Trajtenberg Committee, charged in 2011 with drafting proposals for economic and social change, called, among other things, for increasing employment among the Haredi population. Its proposals included encouraging military or national service and offering college prep courses for volunteers, creating more employment centers targeting Haredim and experimental matriculation prep courses after yeshiva hours. The committee also called for increasing the number of Haredi students receiving technical training through the Industry, Trade, and Labor Ministry and forcing Haredi schools to carry out standardized testing, as is done at other public schools. It is estimated that half as many of the Haredi community are in employment as the rest of population. This has led to increasing financial deprivation, and 50% of children within the community live below the poverty line. This puts strain on each family, the community, and often the Israeli economy. The demographic trend indicates the community will constitute an increasing percentage of the population, and consequently, Israel faces an economic challenge in the years ahead due to fewer people in the labor force. A report commissioned by the Treasury found that the Israeli economy may lose more than six billion shekels annually as a result of low Haredi participation in the workforce. The OECD in a 2010 report stated that, "Haredi families are frequently jobless, or are one-earner families in low-paid employment. Poverty rates are around 60% for Haredim." As of 2017, according to an Israeli finance ministry study, the Haredi participation rate in the labour force is 51%, compared to 89% for the rest of Israeli Jews. A 2018 study by Oren Heller, a National Insurance Institute of Israel senior economic researcher, has found that while upward mobility among Haredim is significantly greater than the national average, unlike it, this tends not to translate into significantly higher pay. Haredi families living in Israel benefited from government-subsidized child care when the father studied Torah and the mother worked at least 24 hours per week. However, after Israeli Finance Minister Avigdor Liberman introduced a new policy in 2021, families in which the father is a full-time yeshiva student are no longer eligible for a daycare subsidy. Under this policy, fathers must also work at least part-time in order for the family to qualify for the subsidy. The move was denounced by Haredi leaders. A 2025 Israel Democracy Institute study found that although Haredim made up 14% of Israel's working-age population in 2023, they generated only 4% of national tax revenue. As a result, the average non-Haredi worker is projected to pay an extra 3,540 shekels in taxes in 2025. Only 23% of Haredim pay income tax, compared to 62% of non-Haredi Jewish men and 46% of women. Employment among Haredi men declined to 54% in 2024, while rates for Haredi women rose to 81% in 2023—just 2% below non-Haredi women. Due to a lack of secular education, many Haredi men are poorly equipped for the labor market, leading to lower household incomes. Despite contributing less in taxes, Haredi households consume more state services, receiving transportation and municipal tax discounts, housing aid, and other benefits; the Kohelet Policy Forum reported that 80% of Haredi households are net receivers of public funds. The IDI called this imbalance unsustainable. The Haredim in general are materially poorer than most other Israelis, but still represent an important market sector due to their bloc purchasing habits. For this reason, some companies and organizations in Israel refrain from including women or other images deemed immodest in their advertisements to avoid Haredi consumer boycotts. More than 50 percent of Haredim live below the poverty line, compared with 15 percent of the rest of the population. Their families are also larger, with Haredi women having an average of 6.7 children, while the average Jewish Israeli woman has 3 children. Families with many children often receive economic support through governmental child allowances, government assistance in housing, as well as specific funds by their own community institutions. In recent years, there has been a process of reconciliation and an attempt to merge Haredi Jews with Israeli society, although employment discrimination is widespread. Haredi Jews such as satirist Kobi Arieli, publicist Sehara Blau, and politician Israel Eichler write regularly for leading Israeli newspapers. Another important factor in the reconciliation process has been the activities of ZAKA, a Haredi organization known for providing emergency medical attention at the scene of suicide bombings, and Yad Sarah, the largest national volunteer organization in Israel established in 1977 by former Haredi mayor of Jerusalem, Uri Lupolianski. It is estimated that Yad Sarah saves the country's economy an estimated $320 million in hospital fees and long-term care costs each year. Present leadership and organizations Notwithstanding the authority of Chief Rabbis of Israel (Ashkenazi: David Lau, Sephardi: Yitzhak Yosef), or the wide acknowledgement of specific rabbis in Israel (for example, Rabbi Gershon Edelstein of the non-Hasidic Lithuanian Jews, and Yaakov Aryeh Alter, who heads the Ger Hasidic dynasty, the largest Hasidic group in Israel), Haredi and Hasidic factions generally align with the independent authority of their respective group leaders. Other representative associations may be linked to specific Haredi and Hasidic groups. For example: Haredi political parties in Israel include: Past leaders of Haredi Jewry Leaders of Haredi Jewry in America included: Leaders of Haredi Jewry in Israel included Controversies People who decide to leave Haredi communities are sometimes shunned and pressured or forced to abandon their children. Cases of pedophilia, sexual violence, assaults, and abuses against women and children occur in roughly the same rates in Haredi communities as in the general population; however, they are rarely discussed or reported to the authorities, and frequently downplayed by members of the communities. To receive a religious divorce, a Jewish woman needs her husband's consent in the form of a get (Jewish divorce document). Without this consent, any future offspring of the wife would be considered mamzerim (bastards/impure). If the circumstances truly warrant a divorce, and the husband is unwilling, a dayan (rabbinic judge) has the prerogative of instituting community shunning measures to "coerce him until he agrees", with physical force reserved only for the rarest of cases. The New York divorce coercion gang was a Haredi Jewish group that kidnapped, and in some cases tortured, Jewish men in the New York metropolitan area to force them to grant their wives gittin (religious divorces). The Federal Bureau of Investigation (FBI) broke up the group after conducting a sting operation against the gang in October 2013. The sting resulted in the prosecution of four men, three of whom were convicted in late 2015. In January 2023, the Times of Israel reported that Haredi citizens in Israel pay just 2% of the country's total income tax revenues, despite making up 13.9% of the nation's population. Furthermore, the article's author described their communities as an "epicenter of poverty", with over 60% of Haredi households classified as "poor" on the government's socio-economic index, with that figure remaining nearly constant in every Haredi community. While this disparity has been present in Israel for decades, it has garnered more attention since December 2022 for numerous reasons. First, Haredi families have the highest fertility rate in Israel, at 6.6 births per woman. In comparison, the average fertility rate in Israel is much lower, at 2.9 per woman. Current projections estimate that the Haredi population will double by 2036, and they will comprise 16% of the total population by 2030. The second aspect of the controversy surrounds their political connections to Israel's Religious Zionist alliance. Historically, they have remained politically uninvolved, but since the 1990s, they have continuously engaged more. Today, members of Israel's ultra-Orthodox community have long enjoyed benefits: exemption from army service for Torah students, government stipends for those choosing full-time religious study over work, and separate schools that receive state funds, even though their curriculums often do not fully teach government-mandated subjects. Today, many Israeli Haredi men do not work, preferring to study the Torah full-time, thus resulting in their high poverty rate. Hundreds of thousands of ultra-Orthodox men assembled in Jerusalem in October 2025 to demonstrate against conscription into the Israel Defence Forces (IDF). Hebrew media sources have highlighted that, in recent years, this issue has been the sole matter that has brought together all sects and factions within the ultra-Orthodox community. As per Israel's Channel 12, the last protest that similarly unified the Haredi community was another anti-conscription rally approximately a decade earlier. In media A Life Apart: Hasidism in America is a documentary film produced and directed by Menachem Daum and Oren Rudavsky, which aimed to portray the Hasidic Haredi world in more positive terms, stressing the close family ties as well as their rich traditions. Shtisel is an Israeli television drama series about a Haredi family in Jerusalem which has led to more favorable feelings about Haredi Jews. See also References Bibliography External links |
======================================== |
[SOURCE: https://en.wikipedia.org/wiki/PlayStation_(console)#cite_ref-father_23-6] | [TOKENS: 10728] |
Contents PlayStation (console) The PlayStation[a] (codenamed PSX, abbreviated as PS, and retroactively PS1 or PS one) is a home video game console developed and marketed by Sony Computer Entertainment. It was released in Japan on 3 December 1994, followed by North America on 9 September 1995, Europe on 29 September 1995, and other regions following thereafter. As a fifth-generation console, the PlayStation primarily competed with the Nintendo 64 and the Sega Saturn. Sony began developing the PlayStation after a failed venture with Nintendo to create a CD-ROM peripheral for the Super Nintendo Entertainment System in the early 1990s. The console was primarily designed by Ken Kutaragi and Sony Computer Entertainment in Japan, while additional development was outsourced in the United Kingdom. An emphasis on 3D polygon graphics was placed at the forefront of the console's design. PlayStation game production was designed to be streamlined and inclusive, enticing the support of many third party developers. The console proved popular for its extensive game library, popular franchises, low retail price, and aggressive youth marketing which advertised it as the preferable console for adolescents and adults. Critically acclaimed games that defined the console include Gran Turismo, Crash Bandicoot, Spyro the Dragon, Tomb Raider, Resident Evil, Metal Gear Solid, Tekken 3, and Final Fantasy VII. Sony ceased production of the PlayStation on 23 March 2006—over eleven years after it had been released, and in the same year the PlayStation 3 debuted. More than 4,000 PlayStation games were released, with cumulative sales of 962 million units. The PlayStation signaled Sony's rise to power in the video game industry. It received acclaim and sold strongly; in less than a decade, it became the first computer entertainment platform to ship over 100 million units. Its use of compact discs heralded the game industry's transition from cartridges. The PlayStation's success led to a line of successors, beginning with the PlayStation 2 in 2000. In the same year, Sony released a smaller and cheaper model, the PS one. History The PlayStation was conceived by Ken Kutaragi, a Sony executive who managed a hardware engineering division and was later dubbed "the Father of the PlayStation". Kutaragi's interest in working with video games stemmed from seeing his daughter play games on Nintendo's Famicom. Kutaragi convinced Nintendo to use his SPC-700 sound processor in the Super Nintendo Entertainment System (SNES) through a demonstration of the processor's capabilities. His willingness to work with Nintendo was derived from both his admiration of the Famicom and conviction in video game consoles becoming the main home-use entertainment systems. Although Kutaragi was nearly fired because he worked with Nintendo without Sony's knowledge, president Norio Ohga recognised the potential in Kutaragi's chip and decided to keep him as a protégé. The inception of the PlayStation dates back to a 1988 joint venture between Nintendo and Sony. Nintendo had produced floppy disk technology to complement cartridges in the form of the Family Computer Disk System, and wanted to continue this complementary storage strategy for the SNES. Since Sony was already contracted to produce the SPC-700 sound processor for the SNES, Nintendo contracted Sony to develop a CD-ROM add-on, tentatively titled the "Play Station" or "SNES-CD". The PlayStation name had already been trademarked by Yamaha, but Nobuyuki Idei liked it so much that he agreed to acquire it for an undisclosed sum rather than search for an alternative. Sony was keen to obtain a foothold in the rapidly expanding video game market. Having been the primary manufacturer of the MSX home computer format, Sony had wanted to use their experience in consumer electronics to produce their own video game hardware. Although the initial agreement between Nintendo and Sony was about producing a CD-ROM drive add-on, Sony had also planned to develop a SNES-compatible Sony-branded console. This iteration was intended to be more of a home entertainment system, playing both SNES cartridges and a new CD format named the "Super Disc", which Sony would design. Under the agreement, Sony would retain sole international rights to every Super Disc game, giving them a large degree of control despite Nintendo's leading position in the video game market. Furthermore, Sony would also be the sole benefactor of licensing related to music and film software that it had been aggressively pursuing as a secondary application. The Play Station was to be announced at the 1991 Consumer Electronics Show (CES) in Las Vegas. However, Nintendo president Hiroshi Yamauchi was wary of Sony's increasing leverage at this point and deemed the original 1988 contract unacceptable upon realising it essentially handed Sony control over all games written on the SNES CD-ROM format. Although Nintendo was dominant in the video game market, Sony possessed a superior research and development department. Wanting to protect Nintendo's existing licensing structure, Yamauchi cancelled all plans for the joint Nintendo–Sony SNES CD attachment without telling Sony. He sent Nintendo of America president Minoru Arakawa (his son-in-law) and chairman Howard Lincoln to Amsterdam to form a more favourable contract with Dutch conglomerate Philips, Sony's rival. This contract would give Nintendo total control over their licences on all Philips-produced machines. Kutaragi and Nobuyuki Idei, Sony's director of public relations at the time, learned of Nintendo's actions two days before the CES was due to begin. Kutaragi telephoned numerous contacts, including Philips, to no avail. On the first day of the CES, Sony announced their partnership with Nintendo and their new console, the Play Station. At 9 am on the next day, in what has been called "the greatest ever betrayal" in the industry, Howard Lincoln stepped onto the stage and revealed that Nintendo was now allied with Philips and would abandon their work with Sony. Incensed by Nintendo's renouncement, Ohga and Kutaragi decided that Sony would develop their own console. Nintendo's contract-breaking was met with consternation in the Japanese business community, as they had broken an "unwritten law" of native companies not turning against each other in favour of foreign ones. Sony's American branch considered allying with Sega to produce a CD-ROM-based machine called the Sega Multimedia Entertainment System, but the Sega board of directors in Tokyo vetoed the idea when Sega of America CEO Tom Kalinske presented them the proposal. Kalinske recalled them saying: "That's a stupid idea, Sony doesn't know how to make hardware. They don't know how to make software either. Why would we want to do this?" Sony halted their research, but decided to develop what it had developed with Nintendo and Sega into a console based on the SNES. Despite the tumultuous events at the 1991 CES, negotiations between Nintendo and Sony were still ongoing. A deal was proposed: the Play Station would still have a port for SNES games, on the condition that it would still use Kutaragi's audio chip and that Nintendo would own the rights and receive the bulk of the profits. Roughly two hundred prototype machines were created, and some software entered development. Many within Sony were still opposed to their involvement in the video game industry, with some resenting Kutaragi for jeopardising the company. Kutaragi remained adamant that Sony not retreat from the growing industry and that a deal with Nintendo would never work. Knowing that they had to take decisive action, Sony severed all ties with Nintendo on 4 May 1992. To determine the fate of the PlayStation project, Ohga chaired a meeting in June 1992, consisting of Kutaragi and several senior Sony board members. Kutaragi unveiled a proprietary CD-ROM-based system he had been secretly working on which played games with immersive 3D graphics. Kutaragi was confident that his LSI chip could accommodate one million logic gates, which exceeded the capabilities of Sony's semiconductor division at the time. Despite gaining Ohga's enthusiasm, there remained opposition from a majority present at the meeting. Older Sony executives also opposed it, who saw Nintendo and Sega as "toy" manufacturers. The opposers felt the game industry was too culturally offbeat and asserted that Sony should remain a central player in the audiovisual industry, where companies were familiar with one another and could conduct "civili[s]ed" business negotiations. After Kutaragi reminded him of the humiliation he suffered from Nintendo, Ohga retained the project and became one of Kutaragi's most staunch supporters. Ohga shifted Kutaragi and nine of his team from Sony's main headquarters to Sony Music Entertainment Japan (SMEJ), a subsidiary of the main Sony group, so as to retain the project and maintain relationships with Philips for the MMCD development project. The involvement of SMEJ proved crucial to the PlayStation's early development as the process of manufacturing games on CD-ROM format was similar to that used for audio CDs, with which Sony's music division had considerable experience. While at SMEJ, Kutaragi worked with Epic/Sony Records founder Shigeo Maruyama and Akira Sato; both later became vice-presidents of the division that ran the PlayStation business. Sony Computer Entertainment (SCE) was jointly established by Sony and SMEJ to handle the company's ventures into the video game industry. On 27 October 1993, Sony publicly announced that it was entering the game console market with the PlayStation. According to Maruyama, there was uncertainty over whether the console should primarily focus on 2D, sprite-based graphics or 3D polygon graphics. After Sony witnessed the success of Sega's Virtua Fighter (1993) in Japanese arcades, the direction of the PlayStation became "instantly clear" and 3D polygon graphics became the console's primary focus. SCE president Teruhisa Tokunaka expressed gratitude for Sega's timely release of Virtua Fighter as it proved "just at the right time" that making games with 3D imagery was possible. Maruyama claimed that Sony further wanted to emphasise the new console's ability to utilise redbook audio from the CD-ROM format in its games alongside high quality visuals and gameplay. Wishing to distance the project from the failed enterprise with Nintendo, Sony initially branded the PlayStation the "PlayStation X" (PSX). Sony formed their European division and North American division, known as Sony Computer Entertainment Europe (SCEE) and Sony Computer Entertainment America (SCEA), in January and May 1995. The divisions planned to market the new console under the alternative branding "PSX" following the negative feedback regarding "PlayStation" in focus group studies. Early advertising prior to the console's launch in North America referenced PSX, but the term was scrapped before launch. The console was not marketed with Sony's name in contrast to Nintendo's consoles. According to Phil Harrison, much of Sony's upper management feared that the Sony brand would be tarnished if associated with the console, which they considered a "toy". Since Sony had no experience in game development, it had to rely on the support of third-party game developers. This was in contrast to Sega and Nintendo, which had versatile and well-equipped in-house software divisions for their arcade games and could easily port successful games to their home consoles. Recent consoles like the Atari Jaguar and 3DO suffered low sales due to a lack of developer support, prompting Sony to redouble their efforts in gaining the endorsement of arcade-savvy developers. A team from Epic Sony visited more than a hundred companies throughout Japan in May 1993 in hopes of attracting game creators with the PlayStation's technological appeal. Sony found that many disliked Nintendo's practices, such as favouring their own games over others. Through a series of negotiations, Sony acquired initial support from Namco, Konami, and Williams Entertainment, as well as 250 other development teams in Japan alone. Namco in particular was interested in developing for PlayStation since Namco rivalled Sega in the arcade market. Attaining these companies secured influential games such as Ridge Racer (1993) and Mortal Kombat 3 (1995), Ridge Racer being one of the most popular arcade games at the time, and it was already confirmed behind closed doors that it would be the PlayStation's first game by December 1993, despite Namco being a longstanding Nintendo developer. Namco's research managing director Shegeichi Nakamura met with Kutaragi in 1993 to discuss the preliminary PlayStation specifications, with Namco subsequently basing the Namco System 11 arcade board on PlayStation hardware and developing Tekken to compete with Virtua Fighter. The System 11 launched in arcades several months before the PlayStation's release, with the arcade release of Tekken in September 1994. Despite securing the support of various Japanese studios, Sony had no developers of their own by the time the PlayStation was in development. This changed in 1993 when Sony acquired the Liverpudlian company Psygnosis (later renamed SCE Liverpool) for US$48 million, securing their first in-house development team. The acquisition meant that Sony could have more launch games ready for the PlayStation's release in Europe and North America. Ian Hetherington, Psygnosis' co-founder, was disappointed after receiving early builds of the PlayStation and recalled that the console "was not fit for purpose" until his team got involved with it. Hetherington frequently clashed with Sony executives over broader ideas; at one point it was suggested that a television with a built-in PlayStation be produced. In the months leading up to the PlayStation's launch, Psygnosis had around 500 full-time staff working on games and assisting with software development. The purchase of Psygnosis marked another turning point for the PlayStation as it played a vital role in creating the console's development kits. While Sony had provided MIPS R4000-based Sony NEWS workstations for PlayStation development, Psygnosis employees disliked the thought of developing on these expensive workstations and asked Bristol-based SN Systems to create an alternative PC-based development system. Andy Beveridge and Martin Day, owners of SN Systems, had previously supplied development hardware for other consoles such as the Mega Drive, Atari ST, and the SNES. When Psygnosis arranged an audience for SN Systems with Sony's Japanese executives at the January 1994 CES in Las Vegas, Beveridge and Day presented their prototype of the condensed development kit, which could run on an ordinary personal computer with two extension boards. Impressed, Sony decided to abandon their plans for a workstation-based development system in favour of SN Systems's, thus securing a cheaper and more efficient method for designing software. An order of over 600 systems followed, and SN Systems supplied Sony with additional software such as an assembler, linker, and a debugger. SN Systems produced development kits for future PlayStation systems, including the PlayStation 2 and was bought out by Sony in 2005. Sony strived to make game production as streamlined and inclusive as possible, in contrast to the relatively isolated approach of Sega and Nintendo. Phil Harrison, representative director of SCEE, believed that Sony's emphasis on developer assistance reduced most time-consuming aspects of development. As well as providing programming libraries, SCE headquarters in London, California, and Tokyo housed technical support teams that could work closely with third-party developers if needed. Sony did not favour their own over non-Sony products, unlike Nintendo; Peter Molyneux of Bullfrog Productions admired Sony's open-handed approach to software developers and lauded their decision to use PCs as a development platform, remarking that "[it was] like being released from jail in terms of the freedom you have". Another strategy that helped attract software developers was the PlayStation's use of the CD-ROM format instead of traditional cartridges. Nintendo cartridges were expensive to manufacture, and the company controlled all production, prioritising their own games, while inexpensive compact disc manufacturing occurred at dozens of locations around the world. The PlayStation's architecture and interconnectability with PCs was beneficial to many software developers. The use of the programming language C proved useful, as it safeguarded future compatibility of the machine should developers decide to make further hardware revisions. Despite the inherent flexibility, some developers found themselves restricted due to the console's lack of RAM. While working on beta builds of the PlayStation, Molyneux observed that its MIPS processor was not "quite as bullish" compared to that of a fast PC and said that it took his team two weeks to port their PC code to the PlayStation development kits and another fortnight to achieve a four-fold speed increase. An engineer from Ocean Software, one of Europe's largest game developers at the time, thought that allocating RAM was a challenging aspect given the 3.5 megabyte restriction. Kutaragi said that while it would have been easy to double the amount of RAM for the PlayStation, the development team refrained from doing so to keep the retail cost down. Kutaragi saw the biggest challenge in developing the system to be balancing the conflicting goals of high performance, low cost, and being easy to program for, and felt he and his team were successful in this regard. Its technical specifications were finalised in 1993 and its design during 1994. The PlayStation name and its final design were confirmed during a press conference on May 10, 1994, although the price and release dates had not been disclosed yet. Sony released the PlayStation in Japan on 3 December 1994, a week after the release of the Sega Saturn, at a price of ¥39,800. Sales in Japan began with a "stunning" success with long queues in shops. Ohga later recalled that he realised how important PlayStation had become for Sony when friends and relatives begged for consoles for their children. PlayStation sold 100,000 units on the first day and two million units within six months, although the Saturn outsold the PlayStation in the first few weeks due to the success of Virtua Fighter. By the end of 1994, 300,000 PlayStation units were sold in Japan compared to 500,000 Saturn units. A grey market emerged for PlayStations shipped from Japan to North America and Europe, with buyers of such consoles paying up to £700. "When September 1995 arrived and Sony's Playstation roared out of the gate, things immediately felt different than [sic] they did with the Saturn launch earlier that year. Sega dropped the Saturn $100 to match the Playstation's $299 debut price, but sales weren't even close—Playstations flew out the door as fast as we could get them in stock. Before the release in North America, Sega and Sony presented their consoles at the first Electronic Entertainment Expo (E3) in Los Angeles on 11 May 1995. At their keynote presentation, Sega of America CEO Tom Kalinske revealed that their Saturn console would be released immediately to select retailers at a price of $399. Next came Sony's turn: Olaf Olafsson, the head of SCEA, summoned Steve Race, the head of development, to the conference stage, who said "$299" and left the audience with a round of applause. The attention to the Sony conference was further bolstered by the surprise appearance of Michael Jackson and the showcase of highly anticipated games, including Wipeout (1995), Ridge Racer and Tekken (1994). In addition, Sony announced that no games would be bundled with the console. Although the Saturn had released early in the United States to gain an advantage over the PlayStation, the surprise launch upset many retailers who were not informed in time, harming sales. Some retailers such as KB Toys responded by dropping the Saturn entirely. The PlayStation went on sale in North America on 9 September 1995. It sold more units within two days than the Saturn had in five months, with almost all of the initial shipment of 100,000 units sold in advance and shops across the country running out of consoles and accessories. The well-received Ridge Racer contributed to the PlayStation's early success, — with some critics considering it superior to Sega's arcade counterpart Daytona USA (1994) — as did Battle Arena Toshinden (1995). There were over 100,000 pre-orders placed and 17 games available on the market by the time of the PlayStation's American launch, in comparison to the Saturn's six launch games. The PlayStation released in Europe on 29 September 1995 and in Australia on 15 November 1995. By November it had already outsold the Saturn by three to one in the United Kingdom, where Sony had allocated a £20 million marketing budget during the Christmas season compared to Sega's £4 million. Sony found early success in the United Kingdom by securing listings with independent shop owners as well as prominent High Street chains such as Comet and Argos. Within its first year, the PlayStation secured over 20% of the entire American video game market. From September to the end of 1995, sales in the United States amounted to 800,000 units, giving the PlayStation a commanding lead over the other fifth-generation consoles,[b] though the SNES and Mega Drive from the fourth generation still outsold it. Sony reported that the attach rate of sold games and consoles was four to one. To meet increasing demand, Sony chartered jumbo jets and ramped up production in Europe and North America. By early 1996, the PlayStation had grossed $2 billion (equivalent to $4.106 billion 2025) from worldwide hardware and software sales. By late 1996, sales in Europe totalled 2.2 million units, including 700,000 in the UK. Approximately 400 PlayStation games were in development, compared to around 200 games being developed for the Saturn and 60 for the Nintendo 64. In India, the PlayStation was launched in test market during 1999–2000 across Sony showrooms, selling 100 units. Sony finally launched the console (PS One model) countrywide on 24 January 2002 with the price of Rs 7,990 and 26 games available from start. PlayStation was also doing well in markets where it was never officially released. For example, in Brazil, due to the registration of the trademark by a third company, the console could not be released, which was why the market was taken over by the officially distributed Sega Saturn during the first period, but as the Sega console withdraws, PlayStation imports and large piracy increased. In another market, China, the most popular 32-bit console was Sega Saturn, but after leaving the market, PlayStation grown with a base of 300,000 users until January 2000, although Sony China did not have plans to release it. The PlayStation was backed by a successful marketing campaign, allowing Sony to gain an early foothold in Europe and North America. Initially, PlayStation demographics were skewed towards adults, but the audience broadened after the first price drop. While the Saturn was positioned towards 18- to 34-year-olds, the PlayStation was initially marketed exclusively towards teenagers. Executives from both Sony and Sega reasoned that because younger players typically looked up to older, more experienced players, advertising targeted at teens and adults would draw them in too. Additionally, Sony found that adults reacted best to advertising aimed at teenagers; Lee Clow surmised that people who started to grow into adulthood regressed and became "17 again" when they played video games. The console was marketed with advertising slogans stylised as "LIVE IN YUR WRLD. PLY IN URS" (Live in Your World. Play in Ours.) and "U R NOT E" (red E). The four geometric shapes were derived from the symbols for the four buttons on the controller. Clow thought that by invoking such provocative statements, gamers would respond to the contrary and say "'Bullshit. Let me show you how ready I am.'" As the console's appeal enlarged, Sony's marketing efforts broadened from their earlier focus on mature players to specifically target younger children as well. Shortly after the PlayStation's release in Europe, Sony tasked marketing manager Geoff Glendenning with assessing the desires of a new target audience. Sceptical over Nintendo and Sega's reliance on television campaigns, Glendenning theorised that young adults transitioning from fourth-generation consoles would feel neglected by marketing directed at children and teenagers. Recognising the influence early 1990s underground clubbing and rave culture had on young people, especially in the United Kingdom, Glendenning felt that the culture had become mainstream enough to help cultivate PlayStation's emerging identity. Sony partnered with prominent nightclub owners such as Ministry of Sound and festival promoters to organise dedicated PlayStation areas where demonstrations of select games could be tested. Sheffield-based graphic design studio The Designers Republic was contracted by Sony to produce promotional materials aimed at a fashionable, club-going audience. Psygnosis' Wipeout in particular became associated with nightclub culture as it was widely featured in venues. By 1997, there were 52 nightclubs in the United Kingdom with dedicated PlayStation rooms. Glendenning recalled that he had discreetly used at least £100,000 a year in slush fund money to invest in impromptu marketing. In 1996, Sony expanded their CD production facilities in the United States due to the high demand for PlayStation games, increasing their monthly output from 4 million discs to 6.5 million discs. This was necessary because PlayStation sales were running at twice the rate of Saturn sales, and its lead dramatically increased when both consoles dropped in price to $199 that year. The PlayStation also outsold the Saturn at a similar ratio in Europe during 1996, with 2.2 million consoles sold in the region by the end of the year. Sales figures for PlayStation hardware and software only increased following the launch of the Nintendo 64. Tokunaka speculated that the Nintendo 64 launch had actually helped PlayStation sales by raising public awareness of the gaming market through Nintendo's added marketing efforts. Despite this, the PlayStation took longer to achieve dominance in Japan. Tokunaka said that, even after the PlayStation and Saturn had been on the market for nearly two years, the competition between them was still "very close", and neither console had led in sales for any meaningful length of time. By 1998, Sega, encouraged by their declining market share and significant financial losses, launched the Dreamcast as a last-ditch attempt to stay in the industry. Although its launch was successful, the technically superior 128-bit console was unable to subdue Sony's dominance in the industry. Sony still held 60% of the overall video game market share in North America at the end of 1999. Sega's initial confidence in their new console was undermined when Japanese sales were lower than expected, with disgruntled Japanese consumers reportedly returning their Dreamcasts in exchange for PlayStation software. On 2 March 1999, Sony officially revealed details of the PlayStation 2, which Kutaragi announced would feature a graphics processor designed to push more raw polygons than any console in history, effectively rivalling most supercomputers. The PlayStation continued to sell strongly at the turn of the new millennium: in June 2000, Sony released the PSOne, a smaller, redesigned variant which went on to outsell all other consoles in that year, including the PlayStation 2. In 2005, PlayStation became the first console to ship 100 million units with the PlayStation 2 later achieving this faster than its predecessor. The combined successes of both PlayStation consoles led to Sega retiring the Dreamcast in 2001, and abandoning the console business entirely. The PlayStation was eventually discontinued on 23 March 2006—over eleven years after its release, and less than a year before the debut of the PlayStation 3. Hardware The main microprocessor is a R3000 CPU made by LSI Logic operating at a clock rate of 33.8688 MHz and 30 MIPS. This 32-bit CPU relies heavily on the "cop2" 3D and matrix math coprocessor on the same die to provide the necessary speed to render complex 3D graphics. The role of the separate GPU chip is to draw 2D polygons and apply shading and textures to them: the rasterisation stage of the graphics pipeline. Sony's custom 16-bit sound chip supports ADPCM sources with up to 24 sound channels and offers a sampling rate of up to 44.1 kHz and music sequencing. It features 2 MB of main RAM, with an additional 1 MB of video RAM. The PlayStation has a maximum colour depth of 16.7 million true colours with 32 levels of transparency and unlimited colour look-up tables. The PlayStation can output composite, S-Video or RGB video signals through its AV Multi connector (with older models also having RCA connectors for composite), displaying resolutions from 256×224 to 640×480 pixels. Different games can use different resolutions. Earlier models also had proprietary parallel and serial ports that could be used to connect accessories or multiple consoles together; these were later removed due to a lack of usage. The PlayStation uses a proprietary video compression unit, MDEC, which is integrated into the CPU and allows for the presentation of full motion video at a higher quality than other consoles of its generation. Unusual for the time, the PlayStation lacks a dedicated 2D graphics processor; 2D elements are instead calculated as polygons by the Geometry Transfer Engine (GTE) so that they can be processed and displayed on screen by the GPU. While running, the GPU can also generate a total of 4,000 sprites and 180,000 polygons per second, in addition to 360,000 per second flat-shaded. The PlayStation went through a number of variants during its production run. Externally, the most notable change was the gradual reduction in the number of external connectors from the rear of the unit. This started with the original Japanese launch units; the SCPH-1000, released on 3 December 1994, was the only model that had an S-Video port, as it was removed from the next model. Subsequent models saw a reduction in number of parallel ports, with the final version only retaining one serial port. Sony marketed a development kit for amateur developers known as the Net Yaroze (meaning "Let's do it together" in Japanese). It was launched in June 1996 in Japan, and following public interest, was released the next year in other countries. The Net Yaroze allowed hobbyists to create their own games and upload them via an online forum run by Sony. The console was only available to buy through an ordering service and with the necessary documentation and software to program PlayStation games and applications through C programming compilers. On 7 July 2000, Sony released the PS One (stylised as "PS one" or "PSone"), a smaller, redesigned version of the original PlayStation. It was the highest-selling console through the end of the year, outselling all other consoles—including the PlayStation 2. In 2002, Sony released a 5-inch (130 mm) LCD screen add-on for the PS One, referred to as the "Combo pack". It also included a car cigarette lighter adaptor adding an extra layer of portability. Production of the LCD "Combo Pack" ceased in 2004, when the popularity of the PlayStation began to wane in markets outside Japan. A total of 28.15 million PS One units had been sold by the time it was discontinued in March 2006. Three iterations of the PlayStation's controller were released over the console's lifespan. The first controller, the PlayStation controller, was released alongside the PlayStation in December 1994. It features four individual directional buttons (as opposed to a conventional D-pad), a pair of shoulder buttons on both sides, Start and Select buttons in the centre, and four face buttons consisting of simple geometric shapes: a green triangle, red circle, blue cross, and a pink square (, , , ). Rather than depicting traditionally used letters or numbers onto its buttons, the PlayStation controller established a trademark which would be incorporated heavily into the PlayStation brand. Teiyu Goto, the designer of the original PlayStation controller, said that the circle and cross represent "yes" and "no", respectively (though this layout is reversed in Western versions); the triangle symbolises a point of view and the square is equated to a sheet of paper to be used to access menus. The European and North American models of the original PlayStation controllers are roughly 10% larger than its Japanese variant, to account for the fact the average person in those regions has larger hands than the average Japanese person. Sony's first analogue gamepad, the PlayStation Analog Joystick (often erroneously referred to as the "Sony Flightstick"), was first released in Japan in April 1996. Featuring two parallel joysticks, it uses potentiometer technology previously used on consoles such as the Vectrex; instead of relying on binary eight-way switches, the controller detects minute angular changes through the entire range of motion. The stick also features a thumb-operated digital hat switch on the right joystick, corresponding to the traditional D-pad, and used for instances when simple digital movements were necessary. The Analog Joystick sold poorly in Japan due to its high cost and cumbersome size. The increasing popularity of 3D games prompted Sony to add analogue sticks to its controller design to give users more freedom over their movements in virtual 3D environments. The first official analogue controller, the Dual Analog Controller, was revealed to the public in a small glass booth at the 1996 PlayStation Expo in Japan, and released in April 1997 to coincide with the Japanese releases of analogue-capable games Tobal 2 and Bushido Blade. In addition to the two analogue sticks (which also introduced two new buttons mapped to clicking in the analogue sticks), the Dual Analog controller features an "Analog" button and LED beneath the "Start" and "Select" buttons which toggles analogue functionality on or off. The controller also features rumble support, though Sony decided that haptic feedback would be removed from all overseas iterations before the United States release. A Sony spokesman stated that the feature was removed for "manufacturing reasons", although rumours circulated that Nintendo had attempted to legally block the release of the controller outside Japan due to similarities with the Nintendo 64 controller's Rumble Pak. However, a Nintendo spokesman denied that Nintendo took legal action. Next Generation's Chris Charla theorised that Sony dropped vibration feedback to keep the price of the controller down. In November 1997, Sony introduced the DualShock controller. Its name derives from its use of two (dual) vibration motors (shock). Unlike its predecessor, its analogue sticks feature textured rubber grips, longer handles, slightly different shoulder buttons and has rumble feedback included as standard on all versions. The DualShock later replaced its predecessors as the default controller. Sony released a series of peripherals to add extra layers of functionality to the PlayStation. Such peripherals include memory cards, the PlayStation Mouse, the PlayStation Link Cable, the Multiplayer Adapter (a four-player multitap), the Memory Drive (a disk drive for 3.5-inch floppy disks), the GunCon (a light gun), and the Glasstron (a monoscopic head-mounted display). Released exclusively in Japan, the PocketStation is a memory card peripheral which acts as a miniature personal digital assistant. The device features a monochrome liquid crystal display (LCD), infrared communication capability, a real-time clock, built-in flash memory, and sound capability. Sharing similarities with the Dreamcast's VMU peripheral, the PocketStation was typically distributed with certain PlayStation games, enhancing them with added features. The PocketStation proved popular in Japan, selling over five million units. Sony planned to release the peripheral outside Japan but the release was cancelled, despite receiving promotion in Europe and North America. In addition to playing games, most PlayStation models are equipped to play CD-Audio. The Asian model SCPH-5903 can also play Video CDs. Like most CD players, the PlayStation can play songs in a programmed order, shuffle the playback order of the disc and repeat one song or the entire disc. Later PlayStation models use a music visualisation function called SoundScope. This function, as well as a memory card manager, is accessed by starting the console without either inserting a game or closing the CD tray, thereby accessing a graphical user interface (GUI) for the PlayStation BIOS. The GUI for the PS One and PlayStation differ depending on the firmware version: the original PlayStation GUI had a dark blue background with rainbow graffiti used as buttons, while the early PAL PlayStation and PS One GUI had a grey blocked background with two icons in the middle. PlayStation emulation is versatile and can be run on numerous modern devices. Bleem! was a commercial emulator which was released for IBM-compatible PCs and the Dreamcast in 1999. It was notable for being aggressively marketed during the PlayStation's lifetime, and was the centre of multiple controversial lawsuits filed by Sony. Bleem! was programmed in assembly language, which allowed it to emulate PlayStation games with improved visual fidelity, enhanced resolutions, and filtered textures that was not possible on original hardware. Sony sued Bleem! two days after its release, citing copyright infringement and accusing the company of engaging in unfair competition and patent infringement by allowing use of PlayStation BIOSs on a Sega console. Bleem! were subsequently forced to shut down in November 2001. Sony was aware that using CDs for game distribution could have left games vulnerable to piracy, due to the growing popularity of CD-R and optical disc drives with burning capability. To preclude illegal copying, a proprietary process for PlayStation disc manufacturing was developed that, in conjunction with an augmented optical drive in Tiger H/E assembly, prevented burned copies of games from booting on an unmodified console. Specifically, all genuine PlayStation discs were printed with a small section of deliberate irregular data, which the PlayStation's optical pick-up was capable of detecting and decoding. Consoles would not boot game discs without a specific wobble frequency contained in the data of the disc pregap sector (the same system was also used to encode discs' regional lockouts). This signal was within Red Book CD tolerances, so PlayStation discs' actual content could still be read by a conventional disc drive; however, the disc drive could not detect the wobble frequency (therefore duplicating the discs omitting it), since the laser pick-up system of any optical disc drive would interpret this wobble as an oscillation of the disc surface and compensate for it in the reading process. Early PlayStations, particularly early 1000 models, experience skipping full-motion video or physical "ticking" noises from the unit. The problems stem from poorly placed vents leading to overheating in some environments, causing the plastic mouldings inside the console to warp slightly and create knock-on effects with the laser assembly. The solution is to sit the console on a surface which dissipates heat efficiently in a well vented area or raise the unit up slightly from its resting surface. Sony representatives also recommended unplugging the PlayStation when it is not in use, as the system draws in a small amount of power (and therefore heat) even when turned off. The first batch of PlayStations use a KSM-440AAM laser unit, whose case and movable parts are all built out of plastic. Over time, the plastic lens sled rail wears out—usually unevenly—due to friction. The placement of the laser unit close to the power supply accelerates wear, due to the additional heat, which makes the plastic more vulnerable to friction. Eventually, one side of the lens sled will become so worn that the laser can tilt, no longer pointing directly at the CD; after this, games will no longer load due to data read errors. Sony fixed the problem by making the sled out of die-cast metal and placing the laser unit further away from the power supply on later PlayStation models. Due to an engineering oversight, the PlayStation does not produce a proper signal on several older models of televisions, causing the display to flicker or bounce around the screen. Sony decided not to change the console design, since only a small percentage of PlayStation owners used such televisions, and instead gave consumers the option of sending their PlayStation unit to a Sony service centre to have an official modchip installed, allowing play on older televisions. Game library The PlayStation featured a diverse game library which grew to appeal to all types of players. Critically acclaimed PlayStation games included Final Fantasy VII (1997), Crash Bandicoot (1996), Spyro the Dragon (1998), Metal Gear Solid (1998), all of which became established franchises. Final Fantasy VII is credited with allowing role-playing games to gain mass-market appeal outside Japan, and is considered one of the most influential and greatest video games ever made. The PlayStation's bestselling game is Gran Turismo (1997), which sold 10.85 million units. After the PlayStation's discontinuation in 2006, the cumulative software shipment was 962 million units. Following its 1994 launch in Japan, early games included Ridge Racer, Crime Crackers, King's Field, Motor Toon Grand Prix, Toh Shin Den (i.e. Battle Arena Toshinden), and Kileak: The Blood. The first two games available at its later North American launch were Jumping Flash! (1995) and Ridge Racer, with Jumping Flash! heralded as an ancestor for 3D graphics in console gaming. Wipeout, Air Combat, Twisted Metal, Warhawk and Destruction Derby were among the popular first-year games, and the first to be reissued as part of Sony's Greatest Hits or Platinum range. At the time of the PlayStation's first Christmas season, Psygnosis had produced around 70% of its launch catalogue; their breakthrough racing game Wipeout was acclaimed for its techno soundtrack and helped raise awareness of Britain's underground music community. Eidos Interactive's action-adventure game Tomb Raider contributed substantially to the success of the console in 1996, with its main protagonist Lara Croft becoming an early gaming icon and garnering unprecedented media promotion. Licensed tie-in video games of popular films were also prevalent; Argonaut Games' 2001 adaptation of Harry Potter and the Philosopher's Stone went on to sell over eight million copies late in the console's lifespan. Third-party developers committed largely to the console's wide-ranging game catalogue even after the launch of the PlayStation 2; some of the notable exclusives in this era include Harry Potter and the Philosopher's Stone, Fear Effect 2: Retro Helix, Syphon Filter 3, C-12: Final Resistance, Dance Dance Revolution Konamix and Digimon World 3.[c] Sony assisted with game reprints as late as 2008 with Metal Gear Solid: The Essential Collection, this being the last PlayStation game officially released and licensed by Sony. Initially, in the United States, PlayStation games were packaged in long cardboard boxes, similar to non-Japanese 3DO and Saturn games. Sony later switched to the jewel case format typically used for audio CDs and Japanese video games, as this format took up less retailer shelf space (which was at a premium due to the large number of PlayStation games being released), and focus testing showed that most consumers preferred this format. Reception The PlayStation was mostly well received upon release. Critics in the west generally welcomed the new console; the staff of Next Generation reviewed the PlayStation a few weeks after its North American launch, where they commented that, while the CPU is "fairly average", the supplementary custom hardware, such as the GPU and sound processor, is stunningly powerful. They praised the PlayStation's focus on 3D, and complemented the comfort of its controller and the convenience of its memory cards. Giving the system 41⁄2 out of 5 stars, they concluded, "To succeed in this extremely cut-throat market, you need a combination of great hardware, great games, and great marketing. Whether by skill, luck, or just deep pockets, Sony has scored three out of three in the first salvo of this war." Albert Kim from Entertainment Weekly praised the PlayStation as a technological marvel, rivalling that of Sega and Nintendo. Famicom Tsūshin scored the console a 19 out of 40, lower than the Saturn's 24 out of 40, in May 1995. In a 1997 year-end review, a team of five Electronic Gaming Monthly editors gave the PlayStation scores of 9.5, 8.5, 9.0, 9.0, and 9.5—for all five editors, the highest score they gave to any of the five consoles reviewed in the issue. They lauded the breadth and quality of the games library, saying it had vastly improved over previous years due to developers mastering the system's capabilities in addition to Sony revising their stance on 2D and role playing games. They also complimented the low price point of the games compared to the Nintendo 64's, and noted that it was the only console on the market that could be relied upon to deliver a solid stream of games for the coming year, primarily due to third party developers almost unanimously favouring it over its competitors. Legacy SCE was an upstart in the video game industry in late 1994, as the video game market in the early 1990s was dominated by Nintendo and Sega. Nintendo had been the clear leader in the industry since the introduction of the Nintendo Entertainment System in 1985 and the Nintendo 64 was initially expected to maintain this position. The PlayStation's target audience included the generation which was the first to grow up with mainstream video games, along with 18- to 29-year-olds who were not the primary focus of Nintendo. By the late 1990s, Sony became a highly regarded console brand due to the PlayStation, with a significant lead over second-place Nintendo, while Sega was relegated to a distant third. The PlayStation became the first "computer entertainment platform" to ship over 100 million units worldwide, with many critics attributing the console's success to third-party developers. It remains the sixth best-selling console of all time as of 2025[update], with a total of 102.49 million units sold. Around 7,900 individual games were published for the console during its 11-year life span, the second-most games ever produced for a console. Its success resulted in a significant financial boon for Sony as profits from their video game division contributed to 23%. Sony's next-generation PlayStation 2, which is backward compatible with the PlayStation's DualShock controller and games, was announced in 1999 and launched in 2000. The PlayStation's lead in installed base and developer support paved the way for the success of its successor, which overcame the earlier launch of the Sega's Dreamcast and then fended off competition from Microsoft's newcomer Xbox and Nintendo's GameCube. The PlayStation 2's immense success and failure of the Dreamcast were among the main factors which led to Sega abandoning the console market. To date, five PlayStation home consoles have been released, which have continued the same numbering scheme, as well as two portable systems. The PlayStation 3 also maintained backward compatibility with original PlayStation discs. Hundreds of PlayStation games have been digitally re-released on the PlayStation Portable, PlayStation 3, PlayStation Vita, PlayStation 4, and PlayStation 5. The PlayStation has often ranked among the best video game consoles. In 2018, Retro Gamer named it the third best console, crediting its sophisticated 3D capabilities as one of its key factors in gaining mass success, and lauding it as a "game-changer in every sense possible". In 2009, IGN ranked the PlayStation the seventh best console in their list, noting its appeal towards older audiences to be a crucial factor in propelling the video game industry, as well as its assistance in transitioning game industry to use the CD-ROM format. Keith Stuart from The Guardian likewise named it as the seventh best console in 2020, declaring that its success was so profound it "ruled the 1990s". In January 2025, Lorentio Brodesco announced the nsOne project, attempting to reverse engineer PlayStation's motherboard. Brodesco stated that "detailed documentation on the original motherboard was either incomplete or entirely unavailable". The project was successfully crowdfunded via Kickstarter. In June, Brodesco manufactured the first working motherboard, promising to bring a fully rooted version with multilayer routing as well as documentation and design files in the near future. The success of the PlayStation contributed to the demise of cartridge-based home consoles. While not the first system to use an optical disc format, it was the first highly successful one, and ended up going head-to-head with the proprietary cartridge-relying Nintendo 64,[d] which the industry had expected to use CDs like PlayStation. After the demise of the Sega Saturn, Nintendo was left as Sony's main competitor in Western markets. Nintendo chose not to use CDs for the Nintendo 64; they were likely concerned with the proprietary cartridge format's ability to help enforce copy protection, given their substantial reliance on licensing and exclusive games for their revenue. Besides their larger capacity, CD-ROMs could be produced in bulk quantities at a much faster rate than ROM cartridges, a week compared to two to three months. Further, the cost of production per unit was far cheaper, allowing Sony to offer games about 40% lower cost to the user compared to ROM cartridges while still making the same amount of net revenue. In Japan, Sony published fewer copies of a wide variety of games for the PlayStation as a risk-limiting step, a model that had been used by Sony Music for CD audio discs. The production flexibility of CD-ROMs meant that Sony could produce larger volumes of popular games to get onto the market quickly, something that could not be done with cartridges due to their manufacturing lead time. The lower production costs of CD-ROMs also allowed publishers an additional source of profit: budget-priced reissues of games which had already recouped their development costs. Tokunaka remarked in 1996: Choosing CD-ROM is one of the most important decisions that we made. As I'm sure you understand, PlayStation could just as easily have worked with masked ROM [cartridges]. The 3D engine and everything—the whole PlayStation format—is independent of the media. But for various reasons (including the economies for the consumer, the ease of the manufacturing, inventory control for the trade, and also the software publishers) we deduced that CD-ROM would be the best media for PlayStation. The increasing complexity of developing games pushed cartridges to their storage limits and gradually discouraged some third-party developers. Part of the CD format's appeal to publishers was that they could be produced at a significantly lower cost and offered more production flexibility to meet demand. As a result, some third-party developers switched to the PlayStation, including Square and Enix, whose Final Fantasy VII and Dragon Quest VII respectively had been planned for the Nintendo 64 (both companies later merged to form Square Enix). Other developers released fewer games for the Nintendo 64 (Konami, releasing only thirteen N64 games but over fifty on the PlayStation). Nintendo 64 game releases were less frequent than the PlayStation's, with many being developed by either Nintendo themselves or second-parties such as Rare. The PlayStation Classic is a dedicated video game console made by Sony Interactive Entertainment that emulates PlayStation games. It was announced in September 2018 at the Tokyo Game Show, and released on 3 December 2018, the 24th anniversary of the release of the original console. As a dedicated console, the PlayStation Classic features 20 pre-installed games; the games run off the open source emulator PCSX. The console is bundled with two replica wired PlayStation controllers (those without analogue sticks), an HDMI cable, and a USB-Type A cable. Internally, the console uses a MediaTek MT8167a Quad A35 system on a chip with four central processing cores clocked at @ 1.5 GHz and a Power VR GE8300 graphics processing unit. It includes 16 GB of eMMC flash storage and 1 Gigabyte of DDR3 SDRAM. The PlayStation Classic is 45% smaller than the original console. The PlayStation Classic received negative reviews from critics and was compared unfavorably to Nintendo's rival Nintendo Entertainment System Classic Edition and Super Nintendo Entertainment System Classic Edition. Criticism was directed at its meagre game library, user interface, emulation quality, use of PAL versions for certain games, use of the original controller, and high retail price, though the console's design received praise. The console sold poorly. See also Notes References |
======================================== |
[SOURCE: https://en.wikipedia.org/wiki/Spreadsheet] | [TOKENS: 7667] |
Contents Spreadsheet A spreadsheet is a computer application for computation, organization, analysis and storage of data in tabular form. Spreadsheets were developed as computerized analogs of paper accounting worksheets. The program operates on data entered in cells of a table. Each cell may contain either numeric or text data, or the results of formulas that automatically calculate and display a value based on the contents of other cells. The term spreadsheet may also refer to one such electronic document. In modern spreadsheet applications, several spreadsheets, often known as "worksheets" or simply "sheets", are gathered together to form a "workbook". A workbook is physically represented by a file containing all the data for the book, the sheets, and the cells with the sheets. Worksheets are normally represented by tabs that flip between pages, each one containing one of the sheets, although Numbers changes this model significantly. Cells in a multi-sheet book add the sheet name to their reference, for instance, "Sheet 1!C10". Some systems extend this syntax to allow cell references to different workbooks. Spreadsheet users can adjust any stored value and observe the effects on calculated values. This makes the spreadsheet useful for "what-if" analysis since many cases can be rapidly investigated without manual recalculation. Modern spreadsheet software can have multiple interacting sheets and can display data either as text and numerals or in graphical form. Besides performing basic arithmetic and mathematical functions, modern spreadsheets provide built-in functions for common financial accountancy and statistical operations. Such calculations as net present value, standard deviation, or regression analysis can be applied to tabular data with a pre-programmed function in a formula. Spreadsheet programs also provide conditional expressions, functions to convert between text and numbers, and functions that operate on strings of text. Spreadsheets have replaced paper-based systems throughout the business world. Although they were first developed for accounting or bookkeeping tasks, they now are used extensively in any context where tabular lists are built, sorted, and shared. Basics LANPAR, available in 1969, was the first electronic spreadsheet on mainframe and time sharing computers. LANPAR was an acronym: LANguage for Programming Arrays at Random. VisiCalc (1979) was the first electronic spreadsheet on a microcomputer, and it helped turn the Apple II into a popular and widely used personal computer. Lotus 1-2-3 was the leading spreadsheet when DOS was the dominant operating system. Microsoft Excel now has the largest market share on the Windows and Macintosh platforms. A spreadsheet program is a standard feature of an office productivity suite. In 2006 Google launched a beta release spreadsheet web application, this is currently known as Google Sheets and one of the applications provided in Google Drive. A spreadsheet consists of a table of cells arranged into rows and columns and referred to by the X and Y locations. X locations, the columns, are normally represented by letters, "A," "B," "C," etc., while rows are normally represented by numbers, 1, 2, 3, etc. A single cell can be referred to by addressing its row and column, "C10". This electronic concept of cell references was first introduced in LANPAR (Language for Programming Arrays at Random) (co-invented by Rene Pardo and Remy Landau) and a variant used in VisiCalc and known as "A1 notation". Additionally, spreadsheets have the concept of a range, a group of cells, normally contiguous. For instance, one can refer to the first ten cells in the first column with the range "A1:A10". LANPAR innovated forward referencing/natural order calculation which didn't re-appear until Lotus 123 and Microsoft's MultiPlan Version 2. Users interact with sheets primarily through the cells. A given cell can hold data by simply entering it in, or a formula, which is normally created by preceding the text with an equals sign. Data might include the string of text hello world, the number 5 or the date 10-Sep-97. A formula would begin with the equals sign, =5*3, but this would normally be invisible because the display shows the result of the calculation, 15 in this case, not the formula itself. This may lead to confusion in some cases. The key feature of spreadsheets is the ability for a formula to refer to the contents of other cells, which may, in turn, be the result of a formula. To make such a formula, one replaces a number with a cell reference. For instance, the formula =5*C10 would produce the result of multiplying the value in cell C10 by the number 5. If C10 holds the value 3 the result will be 15. But C10 might also hold its formula referring to other cells, and so on. The ability to chain formulas together is what gives a spreadsheet its power. Many problems can be broken down into a series of individual mathematical steps, and these can be assigned to individual formulas in cells. Some of these formulas can apply to ranges as well, like the SUM function that adds up all the numbers within a range. Spreadsheets share many principles and traits of databases, but spreadsheets and databases are not the same things. A spreadsheet is essentially just one table, whereas a database is a collection of many tables with machine-readable semantic relationships. While it is true that a workbook that contains three sheets is indeed a file containing multiple tables that can interact with each other, it lacks the relational structure of a database. Spreadsheets and databases are interoperable—sheets can be imported into databases to become tables within them, and database queries can be exported into spreadsheets for further analysis. A spreadsheet program is one of the main components of an office productivity suite, which usually also contains a word processor, a presentation program, and a database management system. Programs within a suite use similar commands for similar functions. Usually, sharing data between the components is easier than with a non-integrated collection of functionally equivalent programs. This was particularly an advantage at a time when many personal computer systems used text-mode displays and commands instead of a graphical user interface. History Humans have organized data into tables, that is, grids of columns and rows, since ancient times. The Babylonians used clay tablets to store data as far back as 1800 BCE. Other examples can be found in book-keeping ledgers and astronomical records. Since at least 1906 the term "spread sheet" has been used in accounting to mean a grid of columns and rows in a ledger. Prior to the rise of computerized spreadsheets, "spread" also referred to a newspaper or magazine item (text or graphics) that covers two facing pages, extending across the centerfold and treating the two pages as one large page. The compound word 'spread-sheet' came to mean the format used to present book-keeping ledgers—with columns for categories of expenditures across the top, invoices listed down the left margin, and the amount of each payment in the cell where its row and column intersect—which were, traditionally, a "spread" across facing pages of a bound ledger (book for keeping accounting records) or on oversized sheets of paper (termed 'analysis paper') ruled into rows and columns in that format and approximately twice as wide as ordinary paper.[failed verification] A batch "spreadsheet" is indistinguishable from a batch compiler with added input data, producing an output report, i.e., a 4GL or conventional, non-interactive, batch computer program. However, this concept of an electronic spreadsheet was outlined in the 1961 paper "Budgeting Models and System Simulation" by Richard Mattessich. The subsequent work by Mattessich (1964a, Chpt. 9, Accounting and Analytical Methods) and its companion volume, Mattessich (1964b, Simulation of the Firm through a Budget Computer Program) applied computerized spreadsheets to accounting and budgeting systems (on mainframe computers programmed in FORTRAN IV). These batch Spreadsheets dealt primarily with the addition or subtraction of entire columns or rows (of input variables), rather than individual cells. In 1962, this concept of the spreadsheet, called BCL for Business Computer Language, was implemented on an IBM 1130[dubious – discuss] and in 1963 was ported to an IBM 7040 by R. Brian Walsh at Marquette University, Wisconsin. This program was written in Fortran. Primitive timesharing was available on those machines. In 1968 BCL was ported by Walsh to the IBM 360/67 timesharing machine at Washington State University. It was used to assist in the teaching of finance to business students. Students were able to take information prepared by the professor and manipulate it to represent it and show ratios etc. In 1964, a book entitled Business Computer Language was written by Kimball, Stoffells and Walsh. Both the book and program were copyrighted in 1966 and years later that copyright was renewed. In the late 1960s, Xerox used BCL to develop a more sophisticated version for their timesharing system. A key invention in the development of electronic spreadsheets was made by Rene K. Pardo and Remy Landau, who filed in 1970 U.S. patent 4,398,249 on a spreadsheet automatic natural order calculation algorithm. While the patent was initially rejected by the patent office as being a purely mathematical invention, following 12 years of appeals, Pardo and Landau won a landmark court case at the Predecessor Court of the Federal Circuit (CCPA), overturning the Patent Office in 1983 — establishing that "something does not cease to become patentable merely because the point of novelty is in an algorithm." However, in 1995 a federal district court ruled the patent unenforceable due to inequitable conduct by the inventors during the application process. The United States Court of Appeals for the Federal Circuit upheld that decision in 1996. The actual software was called LANPAR — LANguage for Programming Arrays at Random.[note 1] This was conceived and entirely developed in the summer of 1969, following Pardo and Landau's recent graduation from Harvard University. Co-inventor Rene Pardo recalls that he felt that one manager at Bell Canada should not have to depend on programmers to program and modify budgeting forms, and he thought of letting users type out forms in any order and having an electronic computer calculate results in the right order ("Forward Referencing/Natural Order Calculation"). Pardo and Landau developed and implemented the software in 1969. LANPAR was used by Bell Canada, AT&T, and the 18 operating telephone companies nationwide for their local and national budgeting operations. LANPAR was also used by General Motors. Its uniqueness was Pardo's co-invention incorporating forward referencing/natural order calculation (one of the first "non-procedural" computer languages) as opposed to the left-to-right, top-to-bottom sequence for calculating the results in each cell that was used by VisiCalc, SuperCalc, and the first version of MultiPlan. Without forward referencing/natural order calculation, the user had to refresh the spreadsheet until the values in all cells remained unchanged. Once the cell values stayed constant, the user was assured that there were no remaining uncalculated forward references within the spreadsheet. In 1968, three former employees from the General Electric computer company headquartered in Phoenix, Arizona set out to start their own software development house. A. Leroy Ellison, Harry N. Cantrell, and Russell E. Edwards found themselves doing a large number of calculations when making tables for the business plans that they were presenting to venture capitalists. They decided to save themselves a lot of effort and wrote a computer program that produced their tables for them. This program, originally conceived as a simple utility for their personal use, would turn out to be the first software product offered by the company that would become known as Capex Corporation. "AutoPlan" ran on GE's Time-sharing service; afterward, a version that ran on IBM mainframes was introduced under the name AutoTab. (National CSS offered a similar product, CSSTAB, which had a moderate timesharing user base by the early 1970s. A major application was opinion research tabulation.) AutoPlan/AutoTab was not a WYSIWYG interactive spreadsheet program, it was a simple scripting language for spreadsheets. The user defined the names and labels for the rows and columns, then the formulas that defined each row or column. In 1975, Autotab-II was advertised as extending the original to a maximum of "1,500 rows and columns, combined in any proportion the user requires..." GE Information Services, which operated the time-sharing service, also launched its own spreadsheet system, Financial Analysis Language (FAL), circa 1974. It was later supplemented by an additional spreadsheet language, TABOL, which was developed by an independent author, Oliver Vellacott in the UK. Both FAL and TABOL were integrated with GEIS's database system, DMS. The IBM Financial Planning and Control System was developed in 1976, by Brian Ingham at IBM Canada. It was implemented by IBM in at least 30 countries. It ran on an IBM mainframe and was the first application for financial planning developed with APL that completely hid the programming language from the end-user. Through IBM's VM operating system, it was among the first programs to auto-update each copy of the application as new versions were released. Users could specify simple mathematical relationships between rows and between columns. Compared to any contemporary alternatives, it could support very large spreadsheets. It loaded actual financial planning data drawn from the legacy batch system into each user's spreadsheet monthly. It was designed to optimize the power of APL through object kernels, increasing program efficiency by as much as 50 fold over traditional programming approaches. An example of an early "industrial weight" spreadsheet was APLDOT, developed in 1976 at the United States Railway Association on an IBM 360/91, running at The Johns Hopkins University Applied Physics Laboratory in Laurel, MD. The application was used successfully for many years in developing such applications as financial and costing models for the US Congress and for Conrail. APLDOT was dubbed a "spreadsheet" because financial analysts and strategic planners used it to solve the same problems they addressed with paper spreadsheet pads. The concept of spreadsheets became widely known due to VisiCalc, developed for the Apple II in 1979 by VisiCorp staff Dan Bricklin and Bob Frankston. Significantly, it also turned the personal computer from a hobby for computer enthusiasts into a business tool. VisiCalc was the first spreadsheet that combined many of the essential features of modern spreadsheet applications, such as a WYSIWYG interactive user interface, automatic recalculation, status and formula lines, range copying with relative and absolute references, and formula building by selecting referenced cells. Unaware of LANPAR at the time, PC World magazine called VisiCalc the first electronic spreadsheet. Bricklin has spoken of watching his university professor create a table of calculation results on a blackboard. When the professor found an error, he had to tediously erase and rewrite several sequential entries in the table, triggering Bricklin to think that he could replicate the process on a computer, using the blackboard as the model to view results of underlying formulas. His idea became VisiCalc. VisiCalc for the Apple II went on to become the first killer application, a program so compelling, people would buy a particular computer just to use it. It was ported to other computers, including CP/M machines, Atari 8-bit computers, and the Commodore PET, but VisiCalc remains best known as an Apple II program. SuperCalc was a spreadsheet application published by Sorcim in 1980, and originally bundled (along with WordStar) as part of the CP/M software package included with the Osborne 1 portable computer. It quickly became the de facto standard spreadsheet for CP/M. The introduction of Lotus 1-2-3 in November 1982 accelerated the acceptance of the IBM Personal Computer. It was written especially for IBM PC DOS and had improvements in speed and graphics compared to VisiCalc on the Apple II, this helped it grow in popularity. Lotus 1-2-3 was the leading spreadsheet for several years. Microsoft released the first version of Excel for the Apple Macintosh on September 30, 1985, and then ported it to Windows, with the first version being numbered 2.05 (to synchronize with the Macintosh version 2.2) and released in November 1987. Microsoft's Windows 3.x platforms of the early 1990s made it possible for their Excel spreadsheet application to take market share from Lotus. By the time Lotus responded with usable Windows products, Microsoft had begun to assemble their Office suite. By 1995, Excel was the market leader, edging out Lotus 1-2-3, and in 2013, IBM discontinued Lotus 1-2-3 altogether. In 2006 Google launched their beta release Google Sheets, a web based spreadsheet application that can be accessed by multiple users from any device type using a compatible web browser, it can be used online and offline (with or without internet connectivity). Google Sheets originated from a web-based spreadsheet application XL2Web developed by 2Web Technologies, combined with DocVerse which enabled multiple-user online collaboration of Office documents. In 2016 Collabora Online Calc was launched, notable in that the web based spreadsheet could be hosted and integrated into any environment without dependency on a 3rd party for authentication or maintenance. Collabora Online runs LibreOffice kit at its core, which grew from StarOffice that was launched 41 years ago in 1985. Notable current spreadsheet software: Discontinued spreadsheet software: Several companies have attempted to break into the spreadsheet market with programs based on very different paradigms. Lotus introduced what is likely the most successful example, Lotus Improv, which saw some commercial success, notably in the financial world where its powerful data mining capabilities remain well respected to this day. Spreadsheet 2000 attempted to dramatically simplify formula construction, but was generally not successful. Concepts The main concepts are those of a grid of cells, called a sheet, with either raw data, called values, or formulas in the cells. Formulas say how to mechanically compute new values from existing values. Values are general numbers, but can also be pure text, dates, months, etc. Extensions of these concepts include logical spreadsheets. Various tools for programming sheets, visualizing data, remotely connecting sheets, displaying cells' dependencies, etc. are commonly provided. A "cell" can be thought of as a box for holding data. A single cell is usually referenced by its column and row (C2 would represent the cell containing the value 30 in the example table below). Usually rows, representing the dependent variables, are referenced in decimal notation starting from 1, while columns representing the independent variables use 26-adic bijective numeration using the letters A-Z as numerals. Its physical size can usually be tailored to its content by dragging its height or width at box intersections (or for entire columns or rows by dragging the column- or row-headers). An array of cells is called a sheet or worksheet. It is analogous to an array of variables in a conventional computer program (although certain unchanging values, once entered, could be considered, by the same analogy, constants). In most implementations, many worksheets may be located within a single spreadsheet. A worksheet is simply a subset of the spreadsheet divided for the sake of clarity. Functionally, the spreadsheet operates as a whole and all cells operate as global variables within the spreadsheet (each variable having 'read' access only except its containing cell). A cell may contain a value or a formula, or it may simply be left empty. By convention, formulas usually begin with = sign. A value can be entered from the computer keyboard by directly typing into the cell itself. Alternatively, a value can be based on a formula (see below), which might perform a calculation, display the current date or time, or retrieve external data such as a stock quote or a database value. The Spreadsheet Value Rule Computer scientist Alan Kay used the term value rule to summarize a spreadsheet's operation: a cell's value relies solely on the formula the user has typed into the cell. The formula may rely on the value of other cells, but those cells are likewise restricted to user-entered data or formulas. There are no 'side effects' to calculating a formula: the only output is to display the calculated result inside its occupying cell. There is no natural mechanism for permanently modifying the contents of a cell unless the user manually modifies the cell's contents. In the context of programming languages, this yields a limited form of first-order functional programming. A standard of spreadsheets since the 1980s, this optional feature eliminates the need to manually request the spreadsheet program to recalculate values (nowadays typically the default option unless specifically 'switched off' for large spreadsheets, usually to improve performance). Some earlier spreadsheets required a manual request to recalculate since the recalculation of large or complex spreadsheets often reduced data entry speed. Many modern spreadsheets still retain this option. Recalculation generally requires that there are no circular dependencies in a spreadsheet. A dependency graph is a graph that has a vertex for each object to be updated, and an edge connecting two objects whenever one of them needs to be updated earlier than the other. Dependency graphs without circular dependencies form directed acyclic graphs, representations of partial orderings (in this case, across a spreadsheet) that can be relied upon to give a definite result. This feature refers to updating a cell's contents periodically with a value from an external source—such as a cell in a "remote" spreadsheet. For shared, Web-based spreadsheets, it applies to "immediately" updating cells another user has updated. All dependent cells must be updated also. Once entered, selected cells (or the entire spreadsheet) can optionally be "locked" to prevent accidental overwriting. Typically this would apply to cells containing formulas but might apply to cells containing "constants" such as a kilogram/pounds conversion factor (2.20462262 to eight decimal places). Even though individual cells are marked as locked, the spreadsheet data are not protected until the feature is activated in the file preferences. A cell or range can optionally be defined to specify how the value is displayed. The default display format is usually set by its initial content if not specifically previously set, so that for example "31/12/2007" or "31 Dec 2007" would default to the cell format of date. Similarly adding a % sign after a numeric value would tag the cell as a percentage cell format. The cell contents are not changed by this format, only the displayed value. Some cell formats such as "numeric" or "currency" can also specify the number of decimal places. This can allow invalid operations (such as doing multiplication on a cell containing a date), resulting in illogical results without an appropriate warning. Depending on the capability of the spreadsheet application, each cell (like its counterpart the "style" in a word processor) can be separately formatted using the attributes of either the content (point size, color, bold or italic) or the cell (border thickness, background shading, color). To aid the readability of a spreadsheet, cell formatting may be conditionally applied to data; for example, a negative number may be displayed in red. A cell's formatting does not typically affect its content and depending on how cells are referenced or copied to other worksheets or applications, the formatting may not be carried with the content. In most implementations, a cell, or group of cells in a column or row, can be "named" enabling the user to refer to those cells by a name rather than by a grid reference. Names must be unique within the spreadsheet, but when using multiple sheets in a spreadsheet file, an identically named cell range on each sheet can be used if it is distinguished by adding the sheet name. One reason for this usage is for creating or running macros that repeat a command across many sheets. Another reason is that formulas with named variables are readily checked against the algebra they are intended to implement (they resemble Fortran expressions). The use of named variables and named functions also makes the spreadsheet structure more transparent. In place of a named cell, an alternative approach is to use a cell (or grid) reference. Most cell references indicate another cell in the same sheet, but a cell reference can also refer to a cell in a different sheet within the same spreadsheet, or (depending on the implementation) to a cell in another spreadsheet entirely, or a value from a remote application. A typical cell reference in "A1" style consists of one or two case-insensitive letters to identify the column (if there are up to 256 columns: A–Z and AA–IV) followed by a row number (e.g., in the range 1–65536). Either part can be relative (it changes when the formula it is in is moved or copied), or absolute (indicated with $ in front of the part concerned of the cell reference). The alternative "R1C1" reference style consists of the letter R, the row number, the letter C, and the column number; relative row or column numbers are indicated by enclosing the number in square brackets. Most current spreadsheets use the A1 style, some providing the R1C1 style as a compatibility option. When the computer calculates a formula in one cell to update the displayed value of that cell, cell reference(s) in that cell, naming some other cell(s), causes the computer to fetch the value of the named cell(s). A cell on the same "sheet" is usually addressed as: =A1 A cell on a different sheet of the same spreadsheet is usually addressed as: Some spreadsheet implementations in Excel allow cell references to another spreadsheet (not the currently open and active file) on the same computer or a local network. It may also refer to a cell in another open and active spreadsheet on the same computer or network that is defined as shareable. These references contain the complete filename, such as: In a spreadsheet, references to cells automatically update when new rows or columns are inserted or deleted. Care must be taken, however, when adding a row immediately before a set of column totals to ensure that the totals reflect the values of the additional rows—which they often do not. A circular reference occurs when the formula in one cell refers—directly, or indirectly through a chain of cell references—to another cell that refers back to the first cell. Many common errors cause circular references. However, some valid techniques use circular references. These techniques, after many spreadsheet recalculations, (usually) converge on the correct values for those cells. Likewise, instead of using a named range of cells, a range reference can be used. A reference to a range of cells typically takes the form (A1:A6), which specifies all the cells in column A from row 1 through row 6. A formula such as "=SUM(A1:A6)" would add all the cells specified and put the result in the cell containing the formula itself. In the earliest spreadsheets, cells were a simple two-dimensional grid. Over time, the model has expanded to include a third dimension, and in some cases a series of named grids, called sheets. The most advanced examples allow inversion and rotation operations which can slice and project the data set in various ways. A formula identifies the calculation needed to place the result in the cell it is contained within. A cell containing a formula, therefore, has two display components; the formula itself and the resulting value. The formula is normally only shown when the cell is selected by "clicking" the mouse over a particular cell; otherwise, it contains the result of the calculation. A formula assigns values to a cell or range of cells, and typically has the format: where the expression consists of: When a cell contains a formula, it often contains references to other cells. Such a cell reference is a type of variable. Its value is the value of the referenced cell or some derivation of it. If that cell in turn references other cells, the value depends on the values of those. References can be relative (e.g., A1, or B1:B3), absolute (e.g., $A$1, or $B$1:$B$3) or mixed row– or column-wise absolute/relative (e.g., $A1 is column-wise absolute and A$1 is row-wise absolute). The available options for valid formulas depend on the particular spreadsheet implementation but, in general, most arithmetic operations and quite complex nested conditional operations can be performed by most of today's commercial spreadsheets. Modern implementations also offer functions to access custom-build functions, remote data, and applications. A formula may contain a condition (or nested conditions)—with or without an actual calculation—and is sometimes used purely to identify and highlight errors. In the example below, it is assumed the sum of a column of percentages (A1 through A6) is tested for validity and an explicit message put into the adjacent right-hand cell. Further examples: The best way to build up conditional statements is step by step composing followed by trial and error testing and refining code. A spreadsheet does not have to contain any formulas at all, in which case it could be considered merely a collection of data arranged in rows and columns (a database) like a calendar, timetable, or simple list. Because of its ease of use, formatting, and hyperlinking capabilities, many spreadsheets are used solely for this purpose. Spreadsheets usually contain several supplied functions, such as arithmetic operations (for example, summations, averages, and so forth), trigonometric functions, statistical functions, and so forth. In addition there is often a provision for user-defined functions. In Microsoft Excel, these functions are defined using Visual Basic for Applications in the supplied Visual Basic editor, and such functions are automatically accessible on the worksheet. Also, programs can be written that pull information from the worksheet, perform some calculations, and report the results back to the worksheet. In the figure, the name sq is user-assigned, and the function sq is introduced using the Visual Basic editor supplied with Excel. Name Manager displays the spreadsheet definitions of named variables x & y. Functions themselves cannot write into the worksheet but simply return their evaluation. However, in Microsoft Excel, subroutines can write values or text found within the subroutine directly to the spreadsheet. The figure shows the Visual Basic code for a subroutine that reads each member of the named column variable x, calculates its square, and writes this value into the corresponding element of named column variable y. The y column contains no formula because its values are calculated in the subroutine, not on the spreadsheet, and simply are written in. Whenever a reference is made to a cell or group of cells that are not located within the current physical spreadsheet file, it is considered as accessing a "remote" spreadsheet. The contents of the referenced cell may be accessed either on the first reference with a manual update or more recently in the case of web-based spreadsheets, as a near real-time value with a specified automatic refresh interval. Many spreadsheet applications permit charts and graphs (e.g., histograms, pie charts) to be generated from specified groups of cells that are dynamically re-built as cell contents change. The generated graphic component can either be embedded within the current sheet or added as a separate object. To create an Excel histogram, a formula based on the REPT function can be used. In the late 1980s and early 1990s, first Javelin Software and Lotus Improv appeared. Unlike models in a conventional spreadsheet, they utilized models built on objects called variables, not on data in cells of a report. These multi-dimensional spreadsheets enabled viewing data and algorithms in various self-documenting ways, including simultaneous multiple synchronized views. For example, users of Javelin could move through the connections between variables on a diagram while seeing the logical roots and branches of each variable. This is an example of what is perhaps its primary contribution of the earlier Javelin—the concept of traceability of a user's logic or model structure through its twelve views. A complex model can be dissected and understood by others who had no role in its creation. In these programs, a time series, or any variable, was an object in itself, not a collection of cells that happen to appear in a row or column. Variables could have many attributes, including complete awareness of their connections to all other variables, data references, and text and image notes. Calculations were performed on these objects, as opposed to a range of cells, so adding two-time series automatically aligns them in calendar time, or in a user-defined time frame. Data were independent of worksheets—variables, and therefore data, could not be destroyed by deleting a row, column, or entire worksheet. For instance, January's costs are subtracted from January's revenues, regardless of where or whether either appears in a worksheet. This permits actions later used in pivot tables, except that flexible manipulation of report tables, was but one of many capabilities supported by variables. Moreover, if costs were entered by week and revenues by month, the program could allocate or interpolate as appropriate. This object design enabled variables and whole models to reference each other with user-defined variable names and to perform multidimensional analysis and massive, but easily editable consolidations. Spreadsheets that have a formula language based upon logical expressions, rather than arithmetic expressions are known as logical spreadsheets. Such spreadsheets can be used to reason deductively about their cell values. Programming issues Just as the early programming languages were designed to generate spreadsheet printouts, programming techniques themselves have evolved to process tables (also known as spreadsheets or matrices) of data more efficiently in the computer itself. Spreadsheets are a popular end-user development tool. EUD denotes activities or techniques in which people who are not professional developers create automated behavior and complex data objects without significant knowledge of a programming language. Many people find it easier to perform calculations in spreadsheets than by writing the equivalent sequential program. This is due to several traits of spreadsheets. A "spreadsheet program" is designed to perform general computation tasks using spatial relationships rather than time as the primary organizing principle. It is often convenient to think of a spreadsheet as a mathematical graph, where the nodes are spreadsheet cells, and the edges are references to other cells specified in formulas. This is often called the dependency graph of the spreadsheet. References between cells can take advantage of spatial concepts such as relative position and absolute position, as well as named locations, to make the spreadsheet formulas easier to understand and manage. Spreadsheets usually attempt to automatically update cells when the cells depend on change. The earliest spreadsheets used simple tactics like evaluating cells in a particular order, but modern spreadsheets calculate following a minimal recomputation order from the dependency graph. Later spreadsheets also include a limited ability to propagate values in reverse, altering source values so that a particular answer is reached in a certain cell. Since spreadsheet cell formulas are not generally invertible, though, this technique is of somewhat limited value. Many of the concepts common to sequential programming models have analogs in the spreadsheet world. For example, the sequential model of the indexed loop is usually represented as a table of cells, with similar formulas (normally differing only in which cells they reference). Spreadsheets have evolved to use scripting programming languages like VBA as a tool for extensibility beyond what the spreadsheet language makes easy. Shortcomings While spreadsheets represented a major step forward in quantitative modeling, they have deficiencies. Their shortcomings include the perceived unfriendliness of alpha-numeric cell addresses. Other problems associated with spreadsheets include: While there are built-in and third-party tools for desktop spreadsheet applications that address some of these shortcomings, awareness, and use of these is generally low. Many professionals in the financial field "don't know" how their spreadsheets are audited; only 6% invest in a third-party solution. Spreadsheet risk Spreadsheet risk is the risk associated with deriving a materially incorrect value from a spreadsheet application that will be utilized in making a related (usually numerically based) decision. Examples include the valuation of an asset, the determination of financial accounts, the calculation of medicinal doses, or the size of a load-bearing beam for structural engineering. The risk may arise from inputting erroneous or fraudulent data values, from mistakes (or incorrect changes) within the logic of the spreadsheet or the omission of relevant updates (e.g., out of date exchange rates). Some single-instance errors have exceeded US$1 billion. Because spreadsheet risk is principally linked to the actions (or inaction) of individuals it is defined as a sub-category of operational risk. Despite this, research carried out by ClusterSeven revealed that around half (48%) of c-level executives and senior managers at firms reporting annual revenues over £50m said there were either no usage controls at all or poorly applied manual processes over the use of spreadsheets at the firms. In 2013 Thomas Herndon, a graduate student of economics at the University of Massachusetts Amherst, found major coding flaws in the spreadsheet used by the economists Carmen Reinhart and Kenneth Rogoff in Growth in a Time of Debt, a very influential 2010 journal article which was widely used as justification to drive 2010–2013 European austerity programs. See also Notes References External links |
======================================== |
[SOURCE: https://en.wikipedia.org/wiki/Muhammad_al-Baqir] | [TOKENS: 8483] |
Contents Muhammad al-Baqir Muhammad ibn Ali al-Baqir (Arabic: محمد بن علي الباقر, romanized: Muḥammad ibn ʿAlī al-Bāqir; c. 676–732) was a descendant of the Islamic prophet Muhammad and the fifth of the twelve Shia imams, succeeding his father, Ali al-Sajjad, and succeeded by his son, Ja'far al-Sadiq. Muhammad's honorific title al-Baqir is short for baqir al-ilm, which means 'the one who splits knowledge open', a reference to his fame as a religious scholar. Muhammad was born in Medina around 676 CE. In 680, when he was a small child, he witnessed the Battle of Karbala, where his grandfather Husayn ibn Ali and most of his relatives were massacred by the forces of the Umayyad caliph Yazid ibn Mu'awiya (r. 680–683). Upon his father's death around 712, Muhammad was recognised as the next imam by most followers of his father. These were the Imamites, the forerunners of Twelvers and Isma'ilis, which now constitute the majority of Shia Muslims. At the time, however, this quiescent group was a minority compared to other rival Shia groups, who actively worked against the Umayyads. One such rival group were Zaydis. These followed Zayd ibn Ali, a much younger half-brother of al-Baqir, who staged an unsuccessful revolt shortly after al-Baqir's death. In contrast, like his father, al-Baqir was politically quiescent but was nevertheless harassed by the Umayyads, especially by Caliph Hisham (r. 724–743). Muhammad al-Baqir led a pious and scholarly life in Medina, attracting a growing number of followers, students, and visitors. He is credited with laying the doctrinal and legal foundations of Twelver Shi'ism during some twenty years of his imamate. He may also be regarded as the father of Isma'ili and Zaydi jurisprudence. Finally, he significantly contributed to Twelver exegesis of the Quran. Most of al-Baqir's disciples were based in Kufa, in present-day Iraq, many of whom later became outstanding Shia jurists and traditionists. Some of these, such as Zurara ibn A'yan, may have occasionally disagreed with al-Baqir, who disapproved of such independent views if they went beyond the general theological and legal framework provided by (Shia) imams. In Sunni Islam, al-Baqir is regarded as an authority in law and prophetic tradition, but portrayed as anti-Shia and proto-Sunni. Muhammad al-Baqir died around 732, poisoned by the Umayyads, according to most Shia reports. He is buried in the Baqi Cemetery in Medina, but the shrine that stood over his grave has been demolished twice by Wahhabis. Al-Baqir was succeeded by his eldest son, Ja'far al-Sadiq, who further developed Shia theology and law. Ancestry Muhammad al-Baqir was a descendant of the Islamic prophet Muhammad, through both of his grandsons, namely, Hasan and Husayn, who were the second and third of the twelve Shia imams, respectively. More specifically, al-Baqir's father was Husayn's son, Ali al-Sajjad, the fourth of the twelve imams. Muhammad's mother was Fatima Umm Abd Allah, while his maternal grandfather was Hasan. Hasan and Husayn were the eldest sons of the first Shia imam, Ali ibn Abi Talib, through his first wife, Fatima, daughter of the Islamic prophet. Titles Muhammad's kunya is Abu Ja'far, and his honorific title is al-Baqir, short for baqir al-'ilm, which means either 'the one who splits knowledge open' (brings it to light) or 'the one who possesses great knowledge', both of which are references to Muhammad's fame as a religious scholar. By some accounts, Muhammad was already known in his lifetime by the title al-Baqir. Shia sources posit that this title was designated by the Islamic prophet, who sent his greetings via his companion Jabir ibn Abd Allah, who lived long enough to meet al-Baqir in his childhood. According to another Shia account, Caliph Hisham, a contemporary of al-Baqir, contemptuously referred to him as al-baqara (lit. 'the cow'), which again suggests that he was known by this title in his lifetime. The occasion was the caliph's meeting with al-Baqir's half brother, Zayd ibn Ali, who reprimanded Hisham and attributed al-Baqir's title to the Islamic prophet. Biography Muhammad al-Baqir was born in Medina in about 676 CE (56 AH). Twelver Shias annually celebrate this occasion on the third of Safar. In 680, when Muhammad was a small child, his grandfather Husayn and most of his male relatives were massacred in the Battle of Karbala by forces of the Umayyad caliph Yazid. Muhammad was present in Karbala and witnessed the carnage. Muhammad's youth coincided with power struggles between the Umayyads, Abd Allah ibn al-Zubayr, and various Shia groups, while Muhammad's father, al-Sajjad, stayed aloof from politics. When al-Sajjad died around 712, most of his followers accepted the imamate of his son Muhammad, who was about thirty-seven years old. He lived a quiet pious life in Medina, like his father, but was nevertheless harassed by the Umayyads, especially by Caliph Hisham. Muhammad, however, enjoyed certain liberties because the Umayyads were more lenient in this period, or perhaps because they were busy infighting and quelling revolts. During the next twenty years or so, Muhammad al-Baqir thus expounded Shia doctrines and laws, attracting a growing number of followers, students, and visitors. The fifth Umayyad caliph, Abd al-Malik ibn Marwan, is credited with issuing an Islamic gold coinage for the first time to replace Byzantine coins. This was likely done at the suggestion of al-Baqir. Often praised for his piety, the Umayyad caliph Umar II was favorably disposed to al-Baqir. After meeting with him, the caliph apparently returned the disputed lands of Fadak to Alids, that is, descendants of Ali ibn Abi Talib. In a Sunni tradition, likely circulated by anti-Alids, al-Baqir identifies Umar II as the Mahdi, the promised savior in Islam. In a Shia tradition, however, al-Baqir suggests that Umar's good deeds would not redeem him, for he had usurped the imam's right to rule. Hisham ibn Abd al-Malik summoned al-Baqir to the Umayyad capital Damascus several times and imprisoned him at least once. During these visits, the caliph apparently held theological debates in which al-Baqir emerged victorious. On one occasion, the caliph ordered al-Baqir to join an ongoing archery practice, probably hoping to embarrass him, but was astonished by al-Baqir's excellent marksmanship. Although 732 (114 AH) and 735 (117 AH) are commonly reported, there is considerable disagreement about when al-Baqir died, ranging from 732 to 736. He was about fifty-seven years old at the time, and most likely died before Zayd's revolt in 740. Twelvers annually commemorate his death on the seventh of Dhu al-Hijja. As with the rest of the twelve imams, Shia sources report that al-Baqir was killed. There is no consensus about the details, and different sources accuse Hisham or his successor, al-Walid II (r. 743–744), of poisoning al-Baqir. According to another account, al-Baqir was poisoned by his cousin, Zayd ibn al-Hasan, once the latter failed to wrest control of the Islamic prophet's inheritance from al-Baqir. Al-Baqir is buried in the Al-Baqi Cemetery in Medina. A shrine stood over his grave until its demolition in 1806 and then again around 1925, both times carried out by Wahhabis. Imamate After al-Sajjad, most of his followers accepted the imamate of his eldest son Muhammad. These were the Imamites, who were the forerunners of Twelver and Isma'ili Shias. Twelver and Isma'ili sources indeed report that al-Sajjad had earlier designated al-Baqir as his successor. Followers of al-Baqir, however, were in minority compared to the rival Kaysanites, which was a (now-extinct) Shia group that traced the imamate through Muhammad ibn al-Hanafiyyah, son of Ali ibn Abi Talib and Khawla bint Ja'far, a woman from the Banu Hanifa tribe. Nevertheless, al-Baqir had an advantage over these non-Fatimid claimants because of his prestigious lineage from Ali ibn Abi Talib and Fatima, the only surviving daughter of the Islamic prophet. Another claimant to leadership was Zayd ibn Ali, a much younger half-brother of al-Baqir. It is not certain, however, if Zayd was a rival for al-Baqir. Despite their disagreements, relationship between the two brothers is described as cordial. The quiescent al-Baqir even attempted to dissuade the politically active Zayd from rebellion. In 740, not long after al-Baqir's death, Zayd took up arms against the Umayyads but was defeated and killed by Caliph Hisham. Zayd's activism initially gained him a larger following than al-Baqir, especially because the former accommodated some of the majority views. For instance, even though Zayd regarded Ali ibn Abi Talib more qualified to succeed the Islamic prophet, he refused to condemn the first two caliphs, namely, Abu Bakr and Umar. Such views, however, cost Zayd part of his Shia support, most of whom condemn Abu Bakr and Umar as usurpers of Ali's right to the caliphate. Those Shia Muslims who thus rejected Zayd joined al-Baqir or his son Ja'far. Zayd's rebellion marks the beginning of the Zaydi movement, a Shia subsect that has survived to present day in Yemen. Muhammad al-Baqir also challenged al-Hasan al-Muthanna and two of his sons for controlling the prophet's inheritance and for claiming to be the Mahdi. Like his father, al-Baqir was politically quiescent, to the point that some have suggested that he did not claim the imamate. Indeed, al-Baqir's notion of imamate was based primarily on knowledge rather than political power, although he also considered Shia imams entitled to the latter. Al-Baqir instead focused on religious teaching, attracting a growing number of visitors, students, and followers. He is often credited with laying the foundations of Twelver and Isma'ili doctrines and law. Among key Shia doctrines that took their definitive form under al-Baqir are imamate, sacred alliance (walaya) and separation (bara'a), and religious dissimulation (taqiyya). As for law, al-Baqir is often regarded as the founding father of Twelver and Isma'ili jurisprudence. In particular, al-Baqir's imamate marks the transition of the Shia community to completely rely on their own imams in matters of law and rituals. As for religious dues, al-Baqir accepted gifts but did not collect khums (lit. 'one fifth'), another Islamic alms which was likely enforced by later imams. Several traditions of al-Baqir are against the Ghulat (lit. 'exaggerators'). These often conferred divinity on Shia imams or had other extreme beliefs, such as anthropomorphism and metempsychosis. For instance, al-Baqir condemned Mughira ibn Sa'id al-Bajali, who said that the imam was divine. Mughira has also been accused of falsifying al-Baqir's traditions. Similarly, al-Baqir denounced Bayan ibn Sam'an, who apparently claimed to be a prophet. Some miracles are attributed to al-Baqir in Shia sources. He is reported to have conversed with animals, returned sight to a blind, and foretold future events, such Zayd's death in battle, collapse of the Umayyads, and the accession of the Abbasid caliph, al-Mansur (r. 754–775). When al-Baqir died, most of his followers accepted the imamate of his eldest son Ja'far, aged about thirty-seven at the time. Ja'far is often known by the honorific al-Sadiq (lit. 'the truthful'). On multiple occasions, al-Baqir seems to have told his followers about his preference for Ja'far. Apparently some did not accept al-Baqir's death and awaited his return as the Mahdi. After al-Baqir's death, some Ghulat figures claimed to have inherited extraordinary powers from him, including Bayan ibn Sam'an and Abu Mansur al-Ijli. Appearance and character The Shia scholar Ibn Shahrashub (d. 1192) describes al-Baqir as medium height, with delicate skin and slightly curly hair. He adds that al-Baqir had birthmarks, one on his cheek, and that he had a beautiful voice and a slender waist. By contrast, al-Mufid (d. 1022), another Shia scholar, describes al-Baqir as a "well-built man," as translated by the Islamicist I.K.A. Howard, or "big-bodied," as translated by M. Pierce, another Islamicist. Such differences may reflect the changing social standards over centuries. Muhammad al-Baqir is said to have been extremely generous, pious, and peaceful by nature. According to some Shia accounts, al-Baqir did not spare himself and his family from wearing good clothes and eating delicious food, and this behavior attracted attention at a time when the tendencies of giving up the world were widespread. He used to work in the field to earn a living on par with his servants, and the motivation for this work, he said, was obedience to God and not needing people. According to a narration by Ja'far al-Sadiq, al-Baqir had less income but more expenses compared to other family members. He treated his relatives with good food and gave them good clothes. He also helped his servants in difficult tasks. According to Ibn Asakir and Ibn Qutaybah, although he was saddened by his son's illness, he did not mourn his death, because he considered this to be an act of opposition to God. Contributions In his lifetime, al-Baqir was regarded as a prominent transmitter of prophetic traditions. As a Shia imam, al-Baqir's own sayings and deeds have also been recorded in Shia sources, including some fifteen percent of the traditions collected in the celebrated Man La Yahduruhu al-Faqih. Such is the extent of his contributions that Shia traditions attributed to al-Baqir and his successor al-Sadiq outnumber all other Shia imams and the prophet combined. As the first Shia imam who engaged in systematic teaching, al-Baqir is also credited with laying the doctrinal and legal foundations of Twelver Shi'ism, which were further developed by al-Sadiq. Contributions of al-Baqir to Twelver doctrine and law are collected in the six-volume Musnad al-Imam al-Baqir, compiled by A. al-Utaridi. Al-Baqir may also be regarded as the father of Isma'ili and Zaydi jurisprudence. Finally, al-Baqir significantly contributed to Twelver exegesis of the Quran and two commentaries are attributed to him. Al-Baqir's doctrine of imamate, further elaborated by his successor al-Sadiq, characterised the necessary qualities of imams, particularly their divinely-inspired designation (nass), their esoteric knowledge (ilm), and their infallibility (isma), all of which distinguished imams as the best of mankind, representatives of God on earth, and the only source of spiritual guidance. In particular, after Muhammad, they are the only authoritative interpreters of the Quran, the exalted part of which actually refers to them and Muhammad. In contrast, as the executer of religious laws, imamate or caliphate is essentially a political function in Sunni Islam, where caliphs are ideally appointed by consensus, although hereditary caliphate is the norm. In al-Baqir's view, imamate is confined to descendants of the Islamic prophet Muhammad, from the marriage of his daughter Fatima to his cousin Ali ibn Abi Talib. Following a divine mandate, each imam is designated by his predecessor (nass), beginning with Ali himself who was designated by the prophet at the Ghadir Khumm. For instance, al-Baqir cited the Quranic verse 2:124, according to which, God designated Abraham as imam and also granted this favor to those of his progeny who are not evildoers. Crucially, the hereditary nature of imamate in al-Baqir's doctrine closed the field to outside claimants. Nass is often accompanied in Shia sources by inheritance of secret religious scrolls and the prophet's weapons. The latter paralleled the Ark of the Covenant for the Israelites. In al-Baqir's doctrine, imams are distinguished by their esoteric knowledge, which they inherited from Ali. In turn, Ali received this knowledge from the prophet, a reference to the well-known prophetic tradition, "I am the city of knowledge and Ali is its gate." In particular, imams know the true exegesis (ta'wil) of the Quran, a reference to the famous hadith of the thaqalayn, attributed to the prophet. Al-Baqir's doctrine of imamate was thus primarily based on knowledge rather than political power, although he also considered imams entitled to the latter. The divinely-inspired knowledge of imams and prophets are similar but imams could only hear (and not see) the archangels, according to al-Baqir. According to al-Baqir, imams also inherit certain spiritual and primordial lights (nur) referenced in the Quran.Their divine knowledge and spiritual light protect imams from sins, for which al-Baqir cited the verse of purification. Already in his lifetime, some followers of al-Baqir regarded him as infallible. By implication, al-Baqir's doctrine gave imams absolute spiritual authority over Muslims, resting on the absolute authority of the prophet. His doctrine also held imams as the sole spiritual guides in life and the source of intercession in the afterlife. In al-Baqir's view, imams are the highest proofs (sg. hujja) of God and guides towards Him, without whom the world cannot exist for a moment. Not only obedience to imams is obligatory in al-Baqir's interpretation of the verse of obedience, but love for them is also mandated in his exegesis of the verse of mawadda. Shias thus form an all-encompassing bond of spiritual loyalty (walaya) with their imams, who are both masters and supportive friends in the journey of the spirit. Identifying his imam is a religious duty for every Muslim, and those who die without knowing their imam have died a death of ignorance (Jahilliya). Furthermore, willfull opposition to imams is a grave sin and staunch enemies of imams are destined for hellfire. To support his theory, al-Baqir relied on his interpretations of various Quranic verses and prophetic traditions. For instance, al-Baqir emphasised his interpretation of the verse of walaya, according to which Ali was granted the guardianship (walaya) of Muslims, on par with the prophet. According to al-Baqir, fearing backlash from some, the prophet was reluctant to publicly announce the walaya of Ali until he was spurred to do so (at the Ghadir Khumm) by the verse of tabligh. The prophetic traditions that al-Baqir invoked include the hadith of the Ghadir Khumm and the hadith of the position. A hotly debated issue at the time was whether the Quran, thought to be the word of God, was created or eternal. Those who believed in predetermination argued that the Quran was eternal for God has always known the events referenced in the Quran. In contrast, those who advocated for free will thought that the Quran was created in time. Al-Baqir held that the Quran was neither created nor eternal. Rather, it is the word of the Creator. More generally, al-Baqir held that all attributes of God were eternal but only as adjectives. For instance, "knowing", "hearing", and "seeing" are how God characterises himself. These help believers understand something about God but are not to be confused with Him. In al-Baqir's view, God is beyond human imagination. He thus advised his followers to discuss God's creation rather than his nature. When asked if he has seen God, al-Baqir responded that God could not be seen by eyes but can be apprehended by the inner reality of faith. On another controversial topic, al-Baqir held that God was a thing, but a thing incomparable to all other things, something neither cognisable nor delimited. By definition, a mu'min (lit. '[true] believer') and a Muslim are characterised, respectively, by the two notions of iman (lit. 'faith') and islam (lit. 'submission [to God]'). Citing the Quranic verse 49:14, al-Baqir defined Muslims as those who confess Islam in words and outwardly practice Islamic rites, such as praying and fasting. In his view, however, iman is more exclusive than islam, that is, the former implies the latter but not vice versa. More specifically, al-Baqir held that mu'min is a Muslim with inner faith, a faith demonstrated through fulfillment of religious duties. The foremost among these duties is the walaya to (Shia) imams. There are indeed numerous traditions attributed to al-Baqir about walaya, the importance of which, in his view, is such that one's good deeds would not be accepted without walaya. Yet al-Baqir also curtailed this absolutist perspective by emphasising that walaya cannot be attained without virtue and piety. In addition to walaya, al-Baqir listed the remaining duties of a mu'min as tahara (lit. 'purification'), prayer, fasting, pilgrimage (Hajj), and jihad (striving in God's way). He also listed sabr (lit. 'patience'), yaqin (lit. 'certitude [in God]'), adl (lit. 'justice'), and (jihad) as the pillars of iman. Al-Baqir thus identified an intermediate state between iman and kufr (disbelief). This gray area was further characterised by his successor al-Sadiq, who held that a Muslim who does not harbor enmity towards the Ahl al-Bayt and their followers is neither mu'min nor kafir (disbeliever). That is, such non-Shias are considered Muslims, with their due legal rights, but not (true) believers. By implication, al-Baqir considered righteous action as an integral component of iman, a view that sharply differed from Murji'ites and Kharijites, two contemporary currents. The former did not consider good conduct essential to iman, with the political implication that dissent and disobedience were discouraged, even if Muslim rulers were corrupt. For Kharijites, in contrast, anyone who committed a mortal sin automatically apostated. In al-Baqir's view, iman had degrees of perfection and could vary over time. In particular, he held that new (religious) knowledge, when put into action, would strengthen one's iman. Later Sunni thought similarly adopted the notion of gradated iman. Under Umayyads, predestination, the belief that God has pre-ordained everything, was likely promoted to justify their rule and encourage moral complacency. In contrast, there were others who believed in free will. Among this latter group, some held that all that is good is created by God and everything bad is from men. Al-Baqir rejected both views, saying that there was a third position between predestination and free will. He argued that God is too merciful to force his creatures to sin and then punish them and that He is too mighty to will a thing that would not transpire. In words of his successor al-Sadiq, God predestined some things but left others to man. Closely related is the early doctrine of bada', that is, advancement or postponement of an act of creation, depending on circumstances, without any change to the overall design and intention of God. The doctrine of bada' thus describes an intermediate position between predestination and free will: God's decision on some matters remain suspended, according to al-Baqir, subject to advancement and postponement, until the autonomous choice of His creatures has occurred. Only then His definite decision is made. In turn, bada' is closely tied to the concept of abrogation (naskh) of some verses of the Quran. Muhammad al-Baqir is often credited with formulating the Shia doctrine of taqiyya, that is, precautionary dissimulation to avoid persecution. Taqiyya was intended for the survival of Shia imams and their followers, for Shias were molested in al-Baqir's time to the point that he thought that it was easier to be a nonbeliever (zindiq). Traditions attributed to al-Baqir thus encourage his followers to hide their faith for their safety, some even characterising taqiyya as a pillar of faith. For instance, al-Baqir is not known to have publicly reviled Abu Bakr and Umar, most likely because he exercised taqiyya. Indeed, al-Baqir's conviction that the Islamic prophet had explicitly designated Ali ibn Abi Talib as his successor implies that Abu Bakr and Umar lacked legitimacy. Al-Baqir's quiescent views sharply differed from Mu'tazilites, who held that enjoining good and forbidding wrong should be enforced by force, if necessary. The notion of taqiyya was not unfamiliar to early Muslims. When Ammar, an early companion of the Islamic prophet, renounced his faith under torture, Muhammad is said to have approved his conduct. The Quranic verse 16:106 is often connected to this episode. Other Quranic verses suggest that Abraham and Joseph both practice dissimulation, the former when he said he was ill and the latter when his brother was accused of theft. Al-Baqir also taught the doctrine of al-bara'a, that is, dissociation from the first three caliphs and the majority of the prophet's companions as enemies of Shia imams. Indeed, Imamites regard the early caliphs as usurpers of Ali's right to succeed the prophet. Al-Baqir founded what later developed into the Twelver school of law and consolidated some characteristic practices of the Shia. For instance, in the call to daily prayer (adhan), al-Baqir added the expression hayy ala khayr al-amal (lit. 'come to the best of deed'), an expression that was removed by Umar, according to Shia and some early Sunni sources. Al-Baqir also defended muta (lit. 'temporary') marriage, saying that it was a practice sanctioned by the Islamic prophet, but later abandoned by Umar. In these rulings, al-Baqir thus sided with Ali ibn Abi Talib and Ibn Abbas, two influential figures in early Islam. Another distinct ruling of al-Baqir was that wiping one's footwear before prayer, though common at the time, was unacceptable as a substitute for washing one's feet. Al-Baqir also forbade all intoxicants, whereas Kufan jurists of his time permitted fermented drinks (nabidth). He also maintained that, under threat of death or injury, self-protection through dissimulation (taqiya) is obligatory. Finally, al-Manasik is an extant treatise on the rituals of Hajj, attributed to al-Baqir and narrated by his disciple Abu al-Jurad Ziyad ibn Mundhir. Shia imams expected their disciples to seek (and then follow) their advice about new legal questions, or else answer those questions by applying limited reasoning within the general framework provided by imams. Al-Baqir is indeed known to have rebuked those who went beyond this framework, including Muhammad ibn al-Hakim and Muhammad al-Tayyar. In particular, al-Baqir discouraged his followers from ijtihad (individual reasoning) or applying ra'y (lit. 'personal opinion') and qiyas (lit. 'analogy'). He considered these methods speculative and lacking in religious authority, which, in his view, was limited to Shia imams as the only authoritative interpretors of the Quran and the prophetic tradition (sunna). Muhammad al-Baqir is credited with the Quranic exegesis Kitab al-Baqir (lit. 'Book of al-Baqir'), narrated by his disciple Ibn al-Mundhir. Parts of this work have survived in Tafsir al-Qummi, written by the Twelver scholar al-Qummi (d. 919). This commentary is ranked first by the Twelver bibliographer al-Najashi (d. c. 1058) among early Quranic commentaries. Similarly, Tafsir Jabir al-Ju'fi is a collection of exegetical traditions, ascribed to al-Baqir and narrated by his disciple Jabir ibn Yazid al-Ju'fi. In Tafsir Nur al-Thaqalayn, an extensive Twelver exegesis of the Quran, al-Baqir is the authority for thirteen percent of its traditions, behind only the prophet and al-Sadiq. Notable disciples Muhammad al-Baqir may have been the first Shia imam who systematically taught Shia beliefs. Even though he lived in Medina, the main following of al-Baqir was in Kufa, where he attracted a number of distinguished theologians. Basra, Mecca, and Syria were other places where al-Baqir's students were based. More than four hundred and sixty names are listed as students of al-Baqir in al-Rijal, a Twelver work on biographical evaluation authored by al-Kashahi (d. c. 941). Al-Baqir had several distinguished disciples in Kufa, where Jabir al-Ju'fi was his main representative. Jabir is the authority for some traditions in Umm al-Kitab, which parallels Infancy Gospel of Thomas in gnostic Christology. In its "Apocalypse of Jabir," al-Baqir confides to Jabir how the cosmos were created, how men descended to this world, and how they can gain deliverance from it. Some have accused Jabir of extremism (ghuluw) and his reliability is debated in Shia circles. Risalat al-Ju'fi is said to contain Jabir's views about Isma'ilism. Zurara ibn A'yan was already a prominent traditionist and theologian before joining al-Baqir's circle. Zurara apparently disagreed with al-Baqir about some theological issues. For instance, unlike al-Baqir, Zurara argued that there is no intermediate state between a believer and a nonbeliever. By some accounts, Zurara later fell out with al-Sadiq, but perhaps the imam distanced himself from Zurara only in public to save the latter from persecution. Aban ibn Taghlib was another associate of al-Baqir and later of al-Sadiq. An outstanding jurist, Aban was authorised by al-Baqir to issue legal rulings for the public. Despite his Shia tendencies, Aban's traditions have been cited in Sunni sources. Abu Basir al-Asadi is among the consensus companions of al-Baqir and al-Sadiq, that is, those whose traditions are generally accepted in Shia circles. Muhammad ibn Muslim, another close associate of al-Baqir and al-Sadiq, was a prominent jurist and traditionist, who is said to have transmitted some thirty thousand traditions from al-Baqir. Fudayl ibn Yasar was another favorite of al-Baqir and al-Sadiq, whom the latter apparently compared to Salman al-Farsi, the famous companion of the Islamic prophet. Abu al-Qasim al-Ijli and Abu Basir al-Muradi, both notable jurists and traditionists, were associates of al-Baqir and al-Sadiq. Abu Hamza al-Thumali and Abu Khalid al-Kabuli, were two followers of al-Baqir and earlier of al-Sajjad. In particular, some traditions narrated by Abu Hamza are of miraculous nature. Al-Kumayt ibn Zayd al-Asadi was a poet supporter of al-Baqir, praised by him for laudatory poems about the Ahl al-Bayt. Kumayt's Hashimiyyat, in praise of the Ahl al-Bayt, is indeed considered among the earliest evidence for the doctrine of imamate, and perhaps the earliest dateable reference to the Ghadir Khumm. Likely to avoid persecution, Kumayt also occasionally wrote in praise of the Umayyads. Mu'min al-Taq was another follower of al-Baqir, who wrote and debated about imamate. Ibn Mundhir was a close disciple of al-Baqir and the principal transmitter of Tafsir al-Baqir He later supported Zayd's rebellion and founded the Jarudiyya, the Zaydi sect closest in doctrine to Twelver Shi'ism. Basra was not a Shia center, but al-Baqir had a few notable disciples there, including Muhammad ibn Marwan al-Basri, Isma'il ibn Fadl al-Hashemi, Malek ibn A'yan al-Juhani. In Mecca, al-Baqir's circle included Ma'ruf ibn Kharbuz Makki and Maymun ibn al-Aswad al-Qaddah. The latter was likely a merchant and in charge of al-Baqir's property in Mecca. One of Maymun's sons, Abdullah, is the alleged ancestor of Isma'ili imams. Elsewhere, prominent followers of al-Baqir included Muhammad ibn Isma'il Bazi and other members of Bazi's family, Abu Harun and his namesake, Abu Harun Makfuf, and also Uqba ibn Bashir al-Asadi, Aslam al-Makki, and Najiyy ibn Abi Mu'adh ibn Muslim. Views Muhammad al-Baqir is regarded as a reliable (thiqa) traditionist in Sunni Islam, distinct in that he accepted only those prophetic traditions that had been reported by his predecessors. Among those scholars who drew from al-Baqir's traditions are al-Shafi'i (d. 820) and Ibn Hanbal (d. 855), the eponymous founders of the Hanbali and Shafi'i schools of law, respectively, and the famed Sunni historian al-Tabari (d. 923). In particular, al-Baqir is the authority for over a hundred traditions in the six canonical collections of Sunni hadith. However, some have criticised al-Baqir for directly quoting individuals who had died before him. Despite al-Baqir's reputation as a reliable traditionist, he is rarely cited in Sunni collections of hadith, compared to their Shia counterparts. This might be because Sunni traditionists "picked and chose" from traditions attributed to al-Baqir, or perhaps because they intentionally discarded the numerous Shia-colored traditions attributed to him. Those traditions of al-Baqir that do appear in Sunni collections have different chains of transmission compared to their Shia counterparts. These Sunni traditions portray al-Baqir as a proto-Sunni scholar who denounced Shias and their beliefs. For instance, unlike Shia sources, later Sunni authors maintain that al-Baqir supported Abu Bakr and Umar, called them imams, and said that one should pray behind Umayyads. In Sunni Islam, al-Baqir is also regarded as an authority in jurisprudence. For instance, Abu Hanifa (d. 767), the eponym of the Hanafi school of law, Ibn Jurayj (d. 767), and Awza'i (d. 774) deferred to al-Baqir in legal matters. The Sunni attitude towards al-Baqir is reflected in the following reports. The exegesis attributed to al-Baqir tops the list of Quranic works compiled by the Sunni bibliographer Ibn al-Nadim (d. c. 995). Abd Allah ibn Ata al-Makki thought that other scholars felt humbled in al-Baqir's presence. According to the Sunni traditionist Muhammad ibn Munkadir (d. 747), al-Baqir was the only scholar who surpassed his father, al-Sajjad. Sunni sources also describe Abu Hanifa as a prominent disciple of al-Baqir, even saying that the latter prophesied that the former would revive the sunna. Muhammad al-Baqir is an imam for Twelvers and Isma'ilis, who constitute the overwhelming majority of Shia Muslims. For them, after the prophet, Shia imams are regarded as the sole source of religious guidance, fostering an all-encompassing bond of loyalty (walaya) with their Shias. Al-Baqir is also regarded as a founding authority in Shia jurisprudence, and a significant contributor to Shia theology. Al-Baqir is considered here as the foremost religious authority of his time, unlike Sunnism. His status is such that a tradition with interrupted or broken chain of transmission would be deemed reliable when narrated by him. Muhammad al-Baqir is a prominent figure in Zaydism. His traditions appear in some Zaydi works, and he has heavily influenced Zaydi jurisprudence. However, the quiescent al-Baqir is generally not recognised as an imam in Zaydism, for Zaydi imams are required to be politically active. In Zaydi works, al-Baqir acknowledges Zayd's superior knowledge and implicitly Zayd's claims to the imamate. Muhammad al-Baqir is regarded as a founding figure in Sufism, where he is noted for his asceticism and piety. In Sufism, al-Baqir is portrayed as an authority in esoteric sciences and the hidden dimensions of the Quran, and a gnostic who performed miracles (karamat). It is related that al-Baqir defined Sufism as "goodness of disposition: He that has the better disposition is the better Sufi." Family Muhammad al-Baqir was married to Umm Farwa, who bore him two sons. One of them was Ja'far, who later succeeded al-Baqir. Umm Hakim, daughter of Usayd al-Thaqafi, was al-Baqir's other wife, from whom two more sons were born, although both died in childhood. Al-Baqir had three more sons, all of whom were born to a concubine. According to the Shia genealogist Alawi al-Umari, al-Baqir's lineage has continued only through Ja'far. See also Footnotes Sources |
======================================== |
[SOURCE: https://en.wikipedia.org/wiki/Elon_Musk#cite_note-408] | [TOKENS: 10515] |
Contents Elon Musk Elon Reeve Musk (/ˈiːlɒn/ EE-lon; born June 28, 1971) is a businessman and entrepreneur known for his leadership of Tesla, SpaceX, Twitter, and xAI. Musk has been the wealthiest person in the world since 2025; as of February 2026,[update] Forbes estimates his net worth to be around US$852 billion. Born into a wealthy family in Pretoria, South Africa, Musk emigrated in 1989 to Canada; he has Canadian citizenship since his mother was born there. He received bachelor's degrees in 1997 from the University of Pennsylvania before moving to California to pursue business ventures. In 1995, Musk co-founded the software company Zip2. Following its sale in 1999, he co-founded X.com, an online payment company that later merged to form PayPal, which was acquired by eBay in 2002. Musk also became an American citizen in 2002. In 2002, Musk founded the space technology company SpaceX, becoming its CEO and chief engineer; the company has since led innovations in reusable rockets and commercial spaceflight. Musk joined the automaker Tesla as an early investor in 2004 and became its CEO and product architect in 2008; it has since become a leader in electric vehicles. In 2015, he co-founded OpenAI to advance artificial intelligence (AI) research, but later left; growing discontent with the organization's direction and their leadership in the AI boom in the 2020s led him to establish xAI, which became a subsidiary of SpaceX in 2026. In 2022, he acquired the social network Twitter, implementing significant changes, and rebranding it as X in 2023. His other businesses include the neurotechnology company Neuralink, which he co-founded in 2016, and the tunneling company the Boring Company, which he founded in 2017. In November 2025, a Tesla pay package worth $1 trillion for Musk was approved, which he is to receive over 10 years if he meets specific goals. Musk was the largest donor in the 2024 U.S. presidential election, where he supported Donald Trump. After Trump was inaugurated as president in early 2025, Musk served as Senior Advisor to the President and as the de facto head of the Department of Government Efficiency (DOGE). After a public feud with Trump, Musk left the Trump administration and returned to managing his companies. Musk is a supporter of global far-right figures, causes, and political parties. His political activities, views, and statements have made him a polarizing figure. Musk has been criticized for COVID-19 misinformation, promoting conspiracy theories, and affirming antisemitic, racist, and transphobic comments. His acquisition of Twitter was controversial due to a subsequent increase in hate speech and the spread of misinformation on the service, following his pledge to decrease censorship. His role in the second Trump administration attracted public backlash, particularly in response to DOGE. The emails he sent to Jeffrey Epstein are included in the Epstein files, which were published between 2025–26 and became a topic of worldwide debate. Early life Elon Reeve Musk was born on June 28, 1971, in Pretoria, South Africa's administrative capital. He is of British and Pennsylvania Dutch ancestry. His mother, Maye (née Haldeman), is a model and dietitian born in Saskatchewan, Canada, and raised in South Africa. Musk therefore holds both South African and Canadian citizenship from birth. His father, Errol Musk, is a South African electromechanical engineer, pilot, sailor, consultant, emerald dealer, and property developer, who partly owned a rental lodge at Timbavati Private Nature Reserve. His maternal grandfather, Joshua N. Haldeman, who died in a plane crash when Elon was a toddler, was an American-born Canadian chiropractor, aviator and political activist in the technocracy movement who moved to South Africa in 1950. Elon has a younger brother, Kimbal, a younger sister, Tosca, and four paternal half-siblings. Musk was baptized as a child in the Anglican Church of Southern Africa. Despite both Elon and Errol previously stating that Errol was a part owner of a Zambian emerald mine, in 2023, Errol recounted that the deal he made was to receive "a portion of the emeralds produced at three small mines". Errol was elected to the Pretoria City Council as a representative of the anti-apartheid Progressive Party and has said that his children shared their father's dislike of apartheid. After his parents divorced in 1979, Elon, aged around 9, chose to live with his father because Errol Musk had an Encyclopædia Britannica and a computer. Elon later regretted his decision and became estranged from his father. Elon has recounted trips to a wilderness school that he described as a "paramilitary Lord of the Flies" where "bullying was a virtue" and children were encouraged to fight over rations. In one incident, after an altercation with a fellow pupil, Elon was thrown down concrete steps and beaten severely, leading to him being hospitalized for his injuries. Elon described his father berating him after he was discharged from the hospital. Errol denied berating Elon and claimed, "The [other] boy had just lost his father to suicide, and Elon had called him stupid. Elon had a tendency to call people stupid. How could I possibly blame that child?" Elon was an enthusiastic reader of books, and had attributed his success in part to having read The Lord of the Rings, the Foundation series, and The Hitchhiker's Guide to the Galaxy. At age ten, he developed an interest in computing and video games, teaching himself how to program from the VIC-20 user manual. At age twelve, Elon sold his BASIC-based game Blastar to PC and Office Technology magazine for approximately $500 (equivalent to $1,600 in 2025). Musk attended Waterkloof House Preparatory School, Bryanston High School, and then Pretoria Boys High School, where he graduated. Musk was a decent but unexceptional student, earning a 61/100 in Afrikaans and a B on his senior math certification. Musk applied for a Canadian passport through his Canadian-born mother to avoid South Africa's mandatory military service, which would have forced him to participate in the apartheid regime, as well as to ease his path to immigration to the United States. While waiting for his application to be processed, he attended the University of Pretoria for five months. Musk arrived in Canada in June 1989, connected with a second cousin in Saskatchewan, and worked odd jobs, including at a farm and a lumber mill. In 1990, he entered Queen's University in Kingston, Ontario. Two years later, he transferred to the University of Pennsylvania, where he studied until 1995. Although Musk has said that he earned his degrees in 1995, the University of Pennsylvania did not award them until 1997 – a Bachelor of Arts in physics and a Bachelor of Science in economics from the university's Wharton School. He reportedly hosted large, ticketed house parties to help pay for tuition, and wrote a business plan for an electronic book-scanning service similar to Google Books. In 1994, Musk held two internships in Silicon Valley: one at energy storage startup Pinnacle Research Institute, which investigated electrolytic supercapacitors for energy storage, and another at Palo Alto–based startup Rocket Science Games. In 1995, he was accepted to a graduate program in materials science at Stanford University, but did not enroll. Musk decided to join the Internet boom of the 1990s, applying for a job at Netscape, to which he reportedly never received a response. The Washington Post reported that Musk lacked legal authorization to remain and work in the United States after failing to enroll at Stanford. In response, Musk said he was allowed to work at that time and that his student visa transitioned to an H1-B. According to numerous former business associates and shareholders, Musk said he was on a student visa at the time. Business career In 1995, Musk, his brother Kimbal, and Greg Kouri founded the web software company Zip2 with funding from a group of angel investors. They housed the venture at a small rented office in Palo Alto. Replying to Rolling Stone, Musk denounced the notion that they started their company with funds borrowed from Errol Musk, but in a tweet, he recognized that his father contributed 10% of a later funding round. The company developed and marketed an Internet city guide for the newspaper publishing industry, with maps, directions, and yellow pages. According to Musk, "The website was up during the day and I was coding it at night, seven days a week, all the time." To impress investors, Musk built a large plastic structure around a standard computer to create the impression that Zip2 was powered by a small supercomputer. The Musk brothers obtained contracts with The New York Times and the Chicago Tribune, and persuaded the board of directors to abandon plans for a merger with CitySearch. Musk's attempts to become CEO were thwarted by the board. Compaq acquired Zip2 for $307 million in cash in February 1999 (equivalent to $590,000,000 in 2025), and Musk received $22 million (equivalent to $43,000,000 in 2025) for his 7-percent share. In 1999, Musk co-founded X.com, an online financial services and e-mail payment company. The startup was one of the first federally insured online banks, and, in its initial months of operation, over 200,000 customers joined the service. The company's investors regarded Musk as inexperienced and replaced him with Intuit CEO Bill Harris by the end of the year. The following year, X.com merged with online bank Confinity to avoid competition. Founded by Max Levchin and Peter Thiel, Confinity had its own money-transfer service, PayPal, which was more popular than X.com's service. Within the merged company, Musk returned as CEO. Musk's preference for Microsoft software over Unix created a rift in the company and caused Thiel to resign. Due to resulting technological issues and lack of a cohesive business model, the board ousted Musk and replaced him with Thiel in 2000.[b] Under Thiel, the company focused on the PayPal service and was renamed PayPal in 2001. In 2002, PayPal was acquired by eBay for $1.5 billion (equivalent to $2,700,000,000 in 2025) in stock, of which Musk—the largest shareholder with 11.72% of shares—received $175.8 million (equivalent to $320,000,000 in 2025). In 2017, Musk purchased the domain X.com from PayPal for an undisclosed amount, stating that it had sentimental value. In 2001, Musk became involved with the nonprofit Mars Society and discussed funding plans to place a growth-chamber for plants on Mars. Seeking a way to launch the greenhouse payloads into space, Musk made two unsuccessful trips to Moscow to purchase intercontinental ballistic missiles (ICBMs) from Russian companies NPO Lavochkin and Kosmotras. Musk instead decided to start a company to build affordable rockets. With $100 million of his early fortune, (equivalent to $180,000,000 in 2025) Musk founded SpaceX in May 2002 and became the company's CEO and Chief Engineer. SpaceX attempted its first launch of the Falcon 1 rocket in 2006. Although the rocket failed to reach Earth orbit, it was awarded a Commercial Orbital Transportation Services program contract from NASA, then led by Mike Griffin. After two more failed attempts that nearly caused Musk to go bankrupt, SpaceX succeeded in launching the Falcon 1 into orbit in 2008. Later that year, SpaceX received a $1.6 billion NASA contract (equivalent to $2,400,000,000 in 2025) for Falcon 9-launched Dragon spacecraft flights to the International Space Station (ISS), replacing the Space Shuttle after its 2011 retirement. In 2012, the Dragon vehicle docked with the ISS, a first for a commercial spacecraft. Working towards its goal of reusable rockets, in 2015 SpaceX successfully landed the first stage of a Falcon 9 on a land platform. Later landings were achieved on autonomous spaceport drone ships, an ocean-based recovery platform. In 2018, SpaceX launched the Falcon Heavy; the inaugural mission carried Musk's personal Tesla Roadster as a dummy payload. Since 2019, SpaceX has been developing Starship, a reusable, super heavy-lift launch vehicle intended to replace the Falcon 9 and Falcon Heavy. In 2020, SpaceX launched its first crewed flight, the Demo-2, becoming the first private company to place astronauts into orbit and dock a crewed spacecraft with the ISS. In 2024, NASA awarded SpaceX an $843 million (equivalent to $865,000,000 in 2025) contract to build a spacecraft that NASA will use to deorbit the ISS at the end of its lifespan. In 2015, SpaceX began development of the Starlink constellation of low Earth orbit satellites to provide satellite Internet access. After the launch of prototype satellites in 2018, the first large constellation was deployed in May 2019. As of May 2025[update], over 7,600 Starlink satellites are operational, comprising 65% of all operational Earth satellites. The total cost of the decade-long project to design, build, and deploy the constellation was estimated by SpaceX in 2020 to be $10 billion (equivalent to $12,000,000,000 in 2025).[c] During the Russian invasion of Ukraine, Musk provided free Starlink service to Ukraine, permitting Internet access and communication at a yearly cost to SpaceX of $400 million (equivalent to $440,000,000 in 2025). However, Musk refused to block Russian state media on Starlink. In 2023, Musk denied Ukraine's request to activate Starlink over Crimea to aid an attack against the Russian navy, citing fears of a nuclear response. Tesla, Inc., originally Tesla Motors, was incorporated in July 2003 by Martin Eberhard and Marc Tarpenning. Both men played active roles in the company's early development prior to Musk's involvement. Musk led the Series A round of investment in February 2004; he invested $6.35 million (equivalent to $11,000,000 in 2025), became the majority shareholder, and joined Tesla's board of directors as chairman. Musk took an active role within the company and oversaw Roadster product design, but was not deeply involved in day-to-day business operations. Following a series of escalating conflicts in 2007 and the 2008 financial crisis, Eberhard was ousted from the firm.[page needed] Musk assumed leadership of the company as CEO and product architect in 2008. A 2009 lawsuit settlement with Eberhard designated Musk as a Tesla co-founder, along with Tarpenning and two others. Tesla began delivery of the Roadster, an electric sports car, in 2008. With sales of about 2,500 vehicles, it was the first mass production all-electric car to use lithium-ion battery cells. Under Musk, Tesla has since launched several well-selling electric vehicles, including the four-door sedan Model S (2012), the crossover Model X (2015), the mass-market sedan Model 3 (2017), the crossover Model Y (2020), and the pickup truck Cybertruck (2023). In May 2020, Musk resigned as chairman of the board as part of the settlement of a lawsuit from the SEC over him tweeting that funding had been "secured" for potentially taking Tesla private. The company has also constructed multiple lithium-ion battery and electric vehicle factories, called Gigafactories. Since its initial public offering in 2010, Tesla stock has risen significantly; it became the most valuable carmaker in summer 2020, and it entered the S&P 500 later that year. In October 2021, it reached a market capitalization of $1 trillion (equivalent to $1,200,000,000,000 in 2025), the sixth company in U.S. history to do so. Musk provided the initial concept and financial capital for SolarCity, which his cousins Lyndon and Peter Rive founded in 2006. By 2013, SolarCity was the second largest provider of solar power systems in the United States. In 2014, Musk promoted the idea of SolarCity building an advanced production facility in Buffalo, New York, triple the size of the largest solar plant in the United States. Construction of the factory started in 2014 and was completed in 2017. It operated as a joint venture with Panasonic until early 2020. Tesla acquired SolarCity for $2 billion in 2016 (equivalent to $2,700,000,000 in 2025) and merged it with its battery unit to create Tesla Energy. The deal's announcement resulted in a more than 10% drop in Tesla's stock price; at the time, SolarCity was facing liquidity issues. Multiple shareholder groups filed a lawsuit against Musk and Tesla's directors, stating that the purchase of SolarCity was done solely to benefit Musk and came at the expense of Tesla and its shareholders. Tesla directors settled the lawsuit in January 2020, leaving Musk the sole remaining defendant. Two years later, the court ruled in Musk's favor. In 2016, Musk co-founded Neuralink, a neurotechnology startup, with an investment of $100 million. Neuralink aims to integrate the human brain with artificial intelligence (AI) by creating devices that are embedded in the brain. Such technology could enhance memory or allow the devices to communicate with software. The company also hopes to develop devices to treat neurological conditions like spinal cord injuries. In 2022, Neuralink announced that clinical trials would begin by the end of the year. In September 2023, the Food and Drug Administration approved Neuralink to initiate six-year human trials. Neuralink has conducted animal testing on macaques at the University of California, Davis. In 2021, the company released a video in which a macaque played the video game Pong via a Neuralink implant. The company's animal trials—which have caused the deaths of some monkeys—have led to claims of animal cruelty. The Physicians Committee for Responsible Medicine has alleged that Neuralink violated the Animal Welfare Act. Employees have complained that pressure from Musk to accelerate development has led to botched experiments and unnecessary animal deaths. In 2022, a federal probe was launched into possible animal welfare violations by Neuralink.[needs update] In 2017, Musk founded the Boring Company to construct tunnels; he also revealed plans for specialized, underground, high-occupancy vehicles that could travel up to 150 miles per hour (240 km/h) and thus circumvent above-ground traffic in major cities. Early in 2017, the company began discussions with regulatory bodies and initiated construction of a 30-foot (9.1 m) wide, 50-foot (15 m) long, and 15-foot (4.6 m) deep "test trench" on the premises of SpaceX's offices, as that required no permits. The Los Angeles tunnel, less than two miles (3.2 km) in length, debuted to journalists in 2018. It used Tesla Model Xs and was reported to be a rough ride while traveling at suboptimal speeds. Two tunnel projects announced in 2018, in Chicago and West Los Angeles, have been canceled. A tunnel beneath the Las Vegas Convention Center was completed in early 2021. Local officials have approved further expansions of the tunnel system. April 14, 2022 In early 2017, Musk expressed interest in buying Twitter and had questioned the platform's commitment to freedom of speech. By 2022, Musk had reached 9.2% stake in the company, making him the largest shareholder.[d] Musk later agreed to a deal that would appoint him to Twitter's board of directors and prohibit him from acquiring more than 14.9% of the company. Days later, Musk made a $43 billion offer to buy Twitter. By the end of April Musk had successfully concluded his bid for approximately $44 billion. This included approximately $12.5 billion in loans and $21 billion in equity financing. Having backtracked on his initial decision, Musk bought the company on October 27, 2022. Immediately after the acquisition, Musk fired several top Twitter executives including CEO Parag Agrawal; Musk became the CEO instead. Under Elon Musk, Twitter instituted monthly subscriptions for a "blue check", and laid off a significant portion of the company's staff. Musk lessened content moderation and hate speech also increased on the platform after his takeover. In late 2022, Musk released internal documents relating to Twitter's moderation of Hunter Biden's laptop controversy in the lead-up to the 2020 presidential election. Musk also promised to step down as CEO after a Twitter poll, and five months later, Musk stepped down as CEO and transitioned his role to executive chairman and chief technology officer (CTO). Despite Musk stepping down as CEO, X continues to struggle with challenges such as viral misinformation, hate speech, and antisemitism controversies. Musk has been accused of trying to silence some of his critics such as Twitch streamer Asmongold, who criticized him during one of his streams. Musk has been accused of removing their accounts' blue checkmarks, which hinders visibility and is considered a form of shadow banning, or suspending their accounts without justification. Other activities In August 2013, Musk announced plans for a version of a vactrain, and assigned engineers from SpaceX and Tesla to design a transport system between Greater Los Angeles and the San Francisco Bay Area, at an estimated cost of $6 billion. Later that year, Musk unveiled the concept, dubbed the Hyperloop, intended to make travel cheaper than any other mode of transport for such long distances. In December 2015, Musk co-founded OpenAI, a not-for-profit artificial intelligence (AI) research company aiming to develop artificial general intelligence, intended to be safe and beneficial to humanity. Musk pledged $1 billion of funding to the company, and initially gave $50 million. In 2018, Musk left the OpenAI board. Since 2018, OpenAI has made significant advances in machine learning. In July 2023, Musk launched the artificial intelligence company xAI, which aims to develop a generative AI program that competes with existing offerings like OpenAI's ChatGPT. Musk obtained funding from investors in SpaceX and Tesla, and xAI hired engineers from Google and OpenAI. December 16, 2022 Musk uses a private jet owned by Falcon Landing LLC, a SpaceX-linked company, and acquired a second jet in August 2020. His heavy use of the jets and the consequent fossil fuel usage have received criticism. Musk's flight usage is tracked on social media through ElonJet. In December 2022, Musk banned the ElonJet account on Twitter, and made temporary bans on the accounts of journalists that posted stories regarding the incident, including Donie O'Sullivan, Keith Olbermann, and journalists from The New York Times, The Washington Post, CNN, and The Intercept. In October 2025, Musk's company xAI launched Grokipedia, an AI-generated online encyclopedia that he promoted as an alternative to Wikipedia. Articles on Grokipedia are generated and reviewed by xAI's Grok chatbot. Media coverage and academic analysis described Grokipedia as frequently reusing Wikipedia content but framing contested political and social topics in line with Musk's own views and right-wing narratives. A study by Cornell University researchers and NBC News stated that Grokipedia cites sources that are blacklisted or considered "generally unreliable" on Wikipedia, for example, the conspiracy site Infowars and the neo-Nazi forum Stormfront. Wired, The Guardian and Time criticized Grokipedia for factual errors and for presenting Musk himself in unusually positive terms while downplaying controversies. Politics Musk is an outlier among business leaders who typically avoid partisan political advocacy. Musk was a registered independent voter when he lived in California. Historically, he has donated to both Democrats and Republicans, many of whom serve in states in which he has a vested interest. Since 2022, his political contributions have mostly supported Republicans, with his first vote for a Republican going to Mayra Flores in the 2022 Texas's 34th congressional district special election. In 2024, he started supporting international far-right political parties, activists, and causes, and has shared misinformation and numerous conspiracy theories. Since 2024, his views have been generally described as right-wing. Musk supported Barack Obama in 2008 and 2012, Hillary Clinton in 2016, Joe Biden in 2020, and Donald Trump in 2024. In the 2020 Democratic Party presidential primaries, Musk endorsed candidate Andrew Yang and expressed support for Yang's proposed universal basic income, and endorsed Kanye West's 2020 presidential campaign. In 2021, Musk publicly expressed opposition to the Build Back Better Act, a $3.5 trillion legislative package endorsed by Joe Biden that ultimately failed to pass due to unanimous opposition from congressional Republicans and several Democrats. In 2022, gave over $50 million to Citizens for Sanity, a conservative political action committee. In 2023, he supported Republican Ron DeSantis for the 2024 U.S. presidential election, giving $10 million to his campaign, and hosted DeSantis's campaign announcement on a Twitter Spaces event. From June 2023 to January 2024, Musk hosted a bipartisan set of X Spaces with Republican and Democratic candidates, including Robert F. Kennedy Jr., Vivek Ramaswamy, and Dean Phillips. In October 2025, former vice-president Kamala Harris commented that it was a mistake from the Democratic side to not invite Musk to a White House electric vehicle event organized in August 2021 and featuring executives from General Motors, Ford and Stellantis, despite Tesla being "the major American manufacturer of extraordinary innovation in this space." Fortune remarked that this was a nod to United Auto Workers and organized labor. Harris said presidents should put aside political loyalties when it came to recognizing innovation, and guessed that the non-invitation impacted Musk's perspective. Fortune noted that, at the time, Musk said, "Yeah, seems odd that Tesla wasn't invited." A month later, he criticized Biden as "not the friendliest administration." Jacob Silverman, author of the book Gilded Rage: Elon Musk and the Radicalization of Silicon Valley, said that the tech industry represented by Musk, Thiel, Andreessen and other capitalists, actually flourished under Biden, but the tech leaders chose Trump for their common ground on cultural issues. By early 2024, Musk had become a vocal and financial supporter of Donald Trump. In July 2024, minutes after the attempted assassination of Donald Trump, Musk endorsed him for president saying; "I fully endorse President Trump and hope for his rapid recovery." During the presidential campaign, Musk joined Trump on stage at a campaign rally, and during the campaign promoted conspiracy theories and falsehoods about Democrats, election fraud and immigration, in support of Trump. Musk was the largest individual donor of the 2024 election. In 2025, Musk contributed $19 million to the Wisconsin Supreme Court race, hoping to influence the state's future redistricting efforts and its regulations governing car manufacturers and dealers. In 2023, Musk said he shunned the World Economic Forum because it was boring. The organization commented that they had not invited him since 2015. He has participated in Dialog, dubbed "Tech Bilderberg" and organized by Peter Thiel and Auren Hoffman, though. Musk's international political actions and comments have come under increasing scrutiny and criticism, especially from the governments and leaders of France, Germany, Norway, Spain and the United Kingdom, particularly due to his position in the U.S. government as well as ownership of X. An NBC News analysis found he had boosted far-right political movements to cut immigration and curtail regulation of business in at least 18 countries on six continents since 2023. During his speech after the second inauguration of Donald Trump, Musk twice made a gesture interpreted by many as a Nazi or a fascist Roman salute.[e] He thumped his right hand over his heart, fingers spread wide, and then extended his right arm out, emphatically, at an upward angle, palm down and fingers together. He then repeated the gesture to the crowd behind him. As he finished the gestures, he said to the crowd, "My heart goes out to you. It is thanks to you that the future of civilization is assured." It was widely condemned as an intentional Nazi salute in Germany, where making such gestures is illegal. The Anti-Defamation League said it was not a Nazi salute, but other Jewish organizations disagreed and condemned the salute. American public opinion was divided on partisan lines as to whether it was a fascist salute. Musk dismissed the accusations of Nazi sympathies, deriding them as "dirty tricks" and a "tired" attack. Neo-Nazi and white supremacist groups celebrated it as a Nazi salute. Multiple European political parties demanded that Musk be banned from entering their countries. The concept of DOGE emerged in a discussion between Musk and Donald Trump, and in August 2024, Trump committed to giving Musk an advisory role, with Musk accepting the offer. In November and December 2024, Musk suggested that the organization could help to cut the U.S. federal budget, consolidate the number of federal agencies, and eliminate the Consumer Financial Protection Bureau, and that its final stage would be "deleting itself". In January 2025, the organization was created by executive order, and Musk was designated a "special government employee". Musk led the organization and was a senior advisor to the president, although his official role is not clear. In sworn statement during a lawsuit, the director of the White House Office of Administration stated that Musk "is not an employee of the U.S. DOGE Service or U.S. DOGE Service Temporary Organization", "is not the U.S. DOGE Service administrator", and has "no actual or formal authority to make government decisions himself". Trump said two days later that he had put Musk in charge of DOGE. A federal judge has ruled that Musk acted as the de facto leader of DOGE. Musk's role in the second Trump administration, particularly in response to DOGE, has attracted public backlash. He was criticized for his treatment of federal government employees, including his influence over the mass layoffs of the federal workforce. He has prioritized secrecy within the organization and has accused others of violating privacy laws. A Senate report alleged that Musk could avoid up to $2 billion in legal liability as a result of DOGE's actions. In May 2025, Bill Gates accused Musk of "killing the world's poorest children" through his cuts to USAID, which modeling by Boston University estimated had resulted in 300,000 deaths by this time, most of them of children. By November 2025, the estimated death toll had increased to 400,000 children and 200,000 adults. Musk announced on May 28, 2025, that he would depart from the Trump administration as planned when the special government employee's 130 day deadline expired, with a White House official confirming that Musk's offboarding from the Trump administration was already underway. His departure was officially confirmed during a joint Oval Office press conference with Trump on May 30, 2025. @realDonaldTrump is in the Epstein files. That is the real reason they have not been made public. June 5, 2025 After leaving office, Musk criticized the Trump administration's Big Beautiful Bill, calling it a "disgusting abomination" due to its provisions increasing the deficit. A feud began between Musk and Trump, with its most notable event being Musk alleging Trump had ties to sex offender Jeffrey Epstein on X (formerly Twitter) on June 5, 2025. Trump responded on Truth Social stating that Musk went "CRAZY" after the "EV Mandate" was purportedly taken away and threatened to cut Musk's government contracts. Musk then called for a third Trump impeachment. The next day, Trump stated that he did not wish to reconcile with Musk, and added that Musk would face "very serious consequences" if he funds Democratic candidates. On June 11, Musk publicly apologized for the tweets against Trump, saying they "went too far". Views November 6, 2022 Rejecting the conservative label, Musk has described himself as a political moderate, even as his views have become more right-wing over time. His views have been characterized as libertarian and far-right, and after his involvement in European politics, they have received criticism from world leaders such as Emmanuel Macron and Olaf Scholz. Within the context of American politics, Musk supported Democratic candidates up until 2022, at which point he voted for a Republican for the first time. He has stated support for universal basic income, gun rights, freedom of speech, a tax on carbon emissions, and H-1B visas. Musk has expressed concern about issues such as artificial intelligence (AI) and climate change, and has been a critic of wealth tax, short-selling, and government subsidies. An immigrant himself, Musk has been accused of being anti-immigration, and regularly blames immigration policies for illegal immigration. He is also a pronatalist who believes population decline is the biggest threat to civilization, and identifies as a cultural Christian. Musk has long been an advocate for space colonization, especially the colonization of Mars. He has repeatedly pushed for humanity colonizing Mars, in order to become an interplanetary species and lower the risks of human extinction. Musk has promoted conspiracy theories and made controversial statements that have led to accusations of racism, sexism, antisemitism, transphobia, disseminating disinformation, and support of white pride. While describing himself as a "pro-Semite", his comments regarding George Soros and Jewish communities have been condemned by the Anti-Defamation League and the Biden White House. Musk was criticized during the COVID-19 pandemic for making unfounded epidemiological claims, defying COVID-19 lockdowns restrictions, and supporting the Canada convoy protest against vaccine mandates. He has amplified false claims of white genocide in South Africa. Musk has been critical of Israel's actions in the Gaza Strip during the Gaza war, praised China's economic and climate goals, suggested that Taiwan and China should resolve cross-strait relations, and was described as having a close relationship with the Chinese government. In Europe, Musk expressed support for Ukraine in 2022 during the Russian invasion, recommended referendums and peace deals on the annexed Russia-occupied territories, and supported the far-right Alternative for Germany political party in 2024. Regarding British politics, Musk blamed the 2024 UK riots on mass migration and open borders, criticized Prime Minister Keir Starmer for what he described as a "two-tier" policing system, and was subsequently attacked as being responsible for spreading misinformation and amplifying the far-right. He has also voiced his support for far-right activist Tommy Robinson and pledged electoral support for Reform UK. In February 2026, Musk described Spanish Prime Minister Pedro Sánchez as a "tyrant" following Sánchez's proposal to prohibit minors under the age of 16 from accessing social media platforms. Legal affairs In 2018, Musk was sued by the U.S. Securities and Exchange Commission (SEC) for a tweet stating that funding had been secured for potentially taking Tesla private.[f] The securities fraud lawsuit characterized the tweet as false, misleading, and damaging to investors, and sought to bar Musk from serving as CEO of publicly traded companies. Two days later, Musk settled with the SEC, without admitting or denying the SEC's allegations. As a result, Musk and Tesla were fined $20 million each, and Musk was forced to step down for three years as Tesla chairman but was able to remain as CEO. Shareholders filed a lawsuit over the tweet, and in February 2023, a jury found Musk and Tesla not liable. Musk has stated in interviews that he does not regret posting the tweet that triggered the SEC investigation. In 2019, Musk stated in a tweet that Tesla would build half a million cars that year. The SEC reacted by asking a court to hold him in contempt for violating the terms of the 2018 settlement agreement. A joint agreement between Musk and the SEC eventually clarified the previous agreement details, including a list of topics about which Musk needed preclearance. In 2020, a judge blocked a lawsuit that claimed a tweet by Musk regarding Tesla stock price ("too high imo") violated the agreement. Freedom of Information Act (FOIA)-released records showed that the SEC concluded Musk had subsequently violated the agreement twice by tweeting regarding "Tesla's solar roof production volumes and its stock price". In October 2023, the SEC sued Musk over his refusal to testify a third time in an investigation into whether he violated federal law by purchasing Twitter stock in 2022. In February 2024, Judge Laurel Beeler ruled that Musk must testify again. In January 2025, the SEC filed a lawsuit against Musk for securities violations related to his purchase of Twitter. In January 2024, Delaware judge Kathaleen McCormick ruled in a 2018 lawsuit that Musk's $55 billion pay package from Tesla be rescinded. McCormick called the compensation granted by the company's board "an unfathomable sum" that was unfair to shareholders. The Delaware Supreme Court overturned McCormick's decision in December 2025, restoring Musk's compensation package and awarding $1 in nominal damages. Personal life Musk became a U.S. citizen in 2002. From the early 2000s until late 2020, Musk resided in California, where both Tesla and SpaceX were founded. He then relocated to Cameron County, Texas, saying that California had become "complacent" about its economic success. While hosting Saturday Night Live in 2021, Musk stated that he has Asperger syndrome (an outdated term for autism spectrum disorder). When asked about his experience growing up with Asperger's syndrome in a TED2022 conference in Vancouver, Musk stated that "the social cues were not intuitive ... I would just tend to take things very literally ... but then that turned out to be wrong — [people were not] simply saying exactly what they mean, there's all sorts of other things that are meant, and [it] took me a while to figure that out." Musk suffers from back pain and has undergone several spine-related surgeries, including a disc replacement. In 2000, he contracted a severe case of malaria while on vacation in South Africa. Musk has stated he uses doctor-prescribed ketamine for occasional depression and that he doses "a small amount once every other week or something like that"; since January 2024, some media outlets have reported that he takes ketamine, marijuana, LSD, ecstasy, mushrooms, cocaine and other drugs. Musk at first refused to comment on his alleged drug use, before responding that he had not tested positive for drugs, and that if drugs somehow improved his productivity, "I would definitely take them!". The New York Times' investigations revealed Musk's overuse of ketamine and numerous other drugs, as well as strained family relationships and concerns from close associates who have become troubled by his public behavior as he became more involved in political activities and government work. According to The Washington Post, President Trump described Musk as "a big-time drug addict". Through his own label Emo G Records, Musk released a rap track, "RIP Harambe", on SoundCloud in March 2019. The following year, he released an EDM track, "Don't Doubt Ur Vibe", featuring his own lyrics and vocals. Musk plays video games, which he stated has a "'restoring effect' that helps his 'mental calibration'". Some games he plays include Quake, Diablo IV, Elden Ring, and Polytopia. Musk once claimed to be one of the world's top video game players but has since admitted to "account boosting", or cheating by hiring outside services to achieve top player rankings. Musk has justified the boosting by claiming that all top accounts do it so he has to as well to remain competitive. In 2024 and 2025, Musk criticized the video game Assassin's Creed Shadows and its creator Ubisoft for "woke" content. Musk posted to X that "DEI kills art" and specified the inclusion of the historical figure Yasuke in the Assassin's Creed game as offensive; he also called the game "terrible". Ubisoft responded by saying that Musk's comments were "just feeding hatred" and that they were focused on producing a game not pushing politics. Musk has fathered at least 14 children, one of whom died as an infant. The Wall Street Journal reported in 2025 that sources close to Musk suggest that the "true number of Musk's children is much higher than publicly known". He had six children with his first wife, Canadian author Justine Wilson, whom he met while attending Queen's University in Ontario, Canada; they married in 2000. In 2002, their first child Nevada Musk died of sudden infant death syndrome at the age of 10 weeks. After his death, the couple used in vitro fertilization (IVF) to continue their family; they had twins in 2004, followed by triplets in 2006. The couple divorced in 2008 and have shared custody of their children. The elder twin he had with Wilson came out as a trans woman and, in 2022, officially changed her name to Vivian Jenna Wilson, adopting her mother's surname because she no longer wished to be associated with Musk. Musk began dating English actress Talulah Riley in 2008. They married two years later at Dornoch Cathedral in Scotland. In 2012, the couple divorced, then remarried the following year. After briefly filing for divorce in 2014, Musk finalized a second divorce from Riley in 2016. Musk then dated the American actress Amber Heard for several months in 2017; he had reportedly been "pursuing" her since 2012. In 2018, Musk and Canadian musician Grimes confirmed they were dating. Grimes and Musk have three children, born in 2020, 2021, and 2022.[g] Musk and Grimes originally gave their eldest child the name "X Æ A-12", which would have violated California regulations as it contained characters that are not in the modern English alphabet; the names registered on the birth certificate are "X" as a first name, "Æ A-Xii" as a middle name, and "Musk" as a last name. They received criticism for choosing a name perceived to be impractical and difficult to pronounce; Musk has said the intended pronunciation is "X Ash A Twelve". Their second child was born via surrogacy. Despite the pregnancy, Musk confirmed reports that the couple were "semi-separated" in September 2021; in an interview with Time in December 2021, he said he was single. In October 2023, Grimes sued Musk over parental rights and custody of X Æ A-Xii. Elon Musk has taken X Æ A-Xii to multiple official events in Washington, D.C. during Trump's second term in office. Also in July 2022, The Wall Street Journal reported that Musk allegedly had an affair with Nicole Shanahan, the wife of Google co-founder Sergey Brin, in 2021, leading to their divorce the following year. Musk denied the report. Musk also had a relationship with Australian actress Natasha Bassett, who has been described as "an occasional girlfriend". In October 2024, The New York Times reported Musk bought a Texas compound for his children and their mothers, though Musk denied having done so. Musk also has four children with Shivon Zilis, director of operations and special projects at Neuralink: twins born via IVF in 2021, a child born in 2024 via surrogacy and a child born in 2025.[h] On February 14, 2025, Ashley St. Clair, an influencer and author, posted on X claiming to have given birth to Musk's son Romulus five months earlier, which media outlets reported as Musk's supposed thirteenth child.[i] On February 22, 2025, it was reported that St Clair had filed for sole custody of her five-month-old son and for Musk to be recognised as the child's father. On March 31, 2025, Musk wrote that, while he was unsure if he was the father of St. Clair's child, he had paid St. Clair $2.5 million and would continue paying her $500,000 per year.[j] Later reporting from the Wall Street Journal indicated that $1 million of these payments to St. Clair were structured as a loan. In 2014, Musk and Ghislaine Maxwell appeared together in a photograph taken at an Academy Awards after-party, which Musk later described as a "photobomb". The January 2026 Epstein files contain emails between Musk and Epstein from 2012 to 2013, after Epstein's first conviction. Emails released on January 30, 2026, indicated that Epstein invited Musk to visit his private island on multiple occasions. The correspondence showed that while Epstein repeatedly encouraged Musk to attend, Musk did not visit the island. In one instance, Musk discussed the possibility of attending a party with his then-wife Talulah Riley and asked which day would be the "wildest party"; according to the emails, the visit did not take place after Epstein later cancelled the plans.[k] On Christmas day in 2012, Musk emailed Epstein asking "Do you have any parties planned? I’ve been working to the edge of sanity this year and so, once my kids head home after Christmas, I really want to hit the party scene in St Barts or elsewhere and let loose. The invitation is much appreciated, but a peaceful island experience is the opposite of what I’m looking for". Epstein replied that the "ratio on my island" might make Musk's wife uncomfortable to which Musk responded, "Ratio is not a problem for Talulah". On September 11, 2013, Epstein sent an email asking Musk if he had any plans for coming to New York for the opening of the United Nations General Assembly where many "interesting people" would be coming to his house to which Musk responded that "Flying to NY to see UN diplomats do nothing would be an unwise use of time". Epstein responded by stating "Do you think i am retarded. Just kidding, there is no one over 25 and all very cute." Musk has denied any close relationship with Epstein and described him as a "creep" who attempted to ingratiate himself with influential people. When Musk was asked in 2019 if he introduced Epstein to Mark Zuckerberg, Musk responded: "I don’t recall introducing Epstein to anyone, as I don’t know the guy well enough to do so." The released emails nonetheless showed cordial exchanges on a range of topics, including Musk's inquiry about parties on the island. The correspondence also indicated that Musk suggested hosting Epstein at SpaceX, while Epstein separately discussed plans to tour SpaceX and bring "the girls", though there is no evidence that such a visit occurred. Musk has described the release of the files a "distraction", later accusing the second Trump administration of suppressing them to protect powerful individuals, including Trump himself.[l] Wealth Elon Musk is the wealthiest person in the world, with an estimated net worth of US$690 billion as of January 2026, according to the Bloomberg Billionaires Index, and $852 billion according to Forbes, primarily from his ownership stakes in SpaceX and Tesla. Having been first listed on the Forbes Billionaires List in 2012, around 75% of Musk's wealth was derived from Tesla stock in November 2020, although he describes himself as "cash poor". According to Forbes, he became the first person in the world to achieve a net worth of $300 billion in 2021; $400 billion in December 2024; $500 billion in October 2025; $600 billion in mid-December 2025; $700 billion later that month; and $800 billion in February 2026. In November 2025, a Tesla pay package worth potentially $1 trillion for Musk was approved, which he is to receive over 10 years if he meets specific goals. Public image Although his ventures have been highly influential within their separate industries starting in the 2000s, Musk only became a public figure in the early 2010s. He has been described as an eccentric who makes spontaneous and impactful decisions, while also often making controversial statements, contrary to other billionaires who prefer reclusiveness to protect their businesses. Musk's actions and his expressed views have made him a polarizing figure. Biographer Ashlee Vance described people's opinions of Musk as polarized due to his "part philosopher, part troll" persona on Twitter. He has drawn denouncement for using his platform to mock the self-selection of personal pronouns, while also receiving praise for bringing international attention to matters like British survivors of grooming gangs. Musk has been described as an American oligarch due to his extensive influence over public discourse, social media, industry, politics, and government policy. After Trump's re-election, Musk's influence and actions during the transition period and the second presidency of Donald Trump led some to call him "President Musk", the "actual president-elect", "shadow president" or "co-president". Awards for his contributions to the development of the Falcon rockets include the American Institute of Aeronautics and Astronautics George Low Transportation Award in 2008, the Fédération Aéronautique Internationale Gold Space Medal in 2010, and the Royal Aeronautical Society Gold Medal in 2012. In 2015, he received an honorary doctorate in engineering and technology from Yale University and an Institute of Electrical and Electronics Engineers Honorary Membership. Musk was elected a Fellow of the Royal Society (FRS) in 2018.[m] In 2022, Musk was elected to the National Academy of Engineering. Time has listed Musk as one of the most influential people in the world in 2010, 2013, 2018, and 2021. Musk was selected as Time's "Person of the Year" for 2021. Then Time editor-in-chief Edward Felsenthal wrote that, "Person of the Year is a marker of influence, and few individuals have had more influence than Musk on life on Earth, and potentially life off Earth too." Notes References Works cited Further reading External links |
======================================== |
[SOURCE: https://en.wikipedia.org/wiki/Jews#cite_note-:3-82] | [TOKENS: 15852] |
Contents Jews Jews (Hebrew: יְהוּדִים, ISO 259-2: Yehudim, Israeli pronunciation: [jehuˈdim]), or the Jewish people, are an ethnoreligious group and nation, originating from the Israelites of ancient Israel and Judah. They traditionally adhere to Judaism. Jewish ethnicity, religion, and community are highly interrelated, as Judaism is an ethnic religion, though many ethnic Jews do not practice it. Religious Jews regard converts to Judaism as members of the Jewish nation, pursuant to the long-standing conversion process. The Israelites emerged from the pre-existing Canaanite peoples to establish Israel and Judah in the Southern Levant during the Iron Age. Originally, Jews referred to the inhabitants of the kingdom of Judah and were distinguished from the gentiles and the Samaritans. According to the Hebrew Bible, these inhabitants predominately originate from the tribe of Judah, who were descendants of Judah, the fourth son of Jacob. The tribe of Benjamin were another significant demographic in Judah and were considered Jews too. By the late 6th century BCE, Judaism had evolved from the Israelite religion, dubbed Yahwism (for Yahweh) by modern scholars, having a theology that religious Jews believe to be the expression of the Mosaic covenant between God and the Jewish people. After the Babylonian exile, Jews referred to followers of Judaism, descendants of the Israelites, citizens of Judea, or allies of the Judean state. Jewish migration within the Mediterranean region during the Hellenistic period, followed by population transfers, caused by events like the Jewish–Roman wars, gave rise to the Jewish diaspora, consisting of diverse Jewish communities that maintained their sense of Jewish history, identity, and culture. In the following millennia, Jewish diaspora communities coalesced into three major ethnic subdivisions according to where their ancestors settled: the Ashkenazim (Central and Eastern Europe), the Sephardim (Iberian Peninsula), and the Mizrahim (Middle East and North Africa). While these three major divisions account for most of the world's Jews, there are other smaller Jewish groups outside of the three. Prior to World War II, the global Jewish population reached a peak of 16.7 million, representing around 0.7% of the world's population at that time. During World War II, approximately six million Jews throughout Europe were systematically murdered by Nazi Germany in a genocide known as the Holocaust. Since then, the population has slowly risen again, and as of 2021[update], was estimated to be at 15.2 million by the demographer Sergio Della Pergola or less than 0.2% of the total world population in 2012.[b] Today, over 85% of Jews live in Israel or the United States. Israel, whose population is 73.9% Jewish, is the only country where Jews comprise more than 2.5% of the population. Jews have significantly influenced and contributed to the development and growth of human progress in many fields, both historically and in modern times, including in science and technology, philosophy, ethics, literature, governance, business, art, music, comedy, theatre, cinema, architecture, food, medicine, and religion. Jews founded Christianity and had an indirect but profound influence on Islam. In these ways and others, Jews have played a significant role in the development of Western culture. Name and etymology The term "Jew" is derived from the Hebrew word יְהוּדִי Yehudi, with the plural יְהוּדִים Yehudim. Endonyms in other Jewish languages include the Ladino ג׳ודיו Djudio (plural ג׳ודיוס, Djudios) and the Yiddish ייִד Yid (plural ייִדן Yidn). Though Genesis 29:35 and 49:8 connect "Judah" with the verb yada, meaning "praise", scholars generally agree that "Judah" most likely derives from the name of a Levantine geographic region dominated by gorges and ravines. The gradual ethnonymic shift from "Israelites" to "Jews", regardless of their descent from Judah, although not contained in the Torah, is made explicit in the Book of Esther (4th century BCE) of the Tanakh. Some modern scholars disagree with the conflation, based on the works of Josephus, Philo and Apostle Paul. The English word "Jew" is a derivation of Middle English Gyw, Iewe. The latter was loaned from the Old French giu, which itself evolved from the earlier juieu, which in turn derived from judieu/iudieu which through elision had dropped the letter "d" from the Medieval Latin Iudaeus, which, like the New Testament Greek term Ioudaios, meant both "Jew" and "Judean" / "of Judea". The Greek term was a loan from Aramaic *yahūdāy, corresponding to Hebrew יְהוּדִי Yehudi. Some scholars prefer translating Ioudaios as "Judean" in the Bible since it is more precise, denotes the community's origins and prevents readers from engaging in antisemitic eisegesis. Others disagree, believing that it erases the Jewish identity of Biblical characters such as Jesus. Daniel R. Schwartz distinguishes "Judean" and "Jew". Here, "Judean" refers to the inhabitants of Judea, which encompassed southern Palestine. Meanwhile, "Jew" refers to the descendants of Israelites that adhere to Judaism. Converts are included in the definition. But Shaye J.D. Cohen argues that "Judean" is inclusive of believers of the Judean God and allies of the Judean state. Another scholar, Jodi Magness, wrote the term Ioudaioi refers to a "people of Judahite/Judean ancestry who worshipped the God of Israel as their national deity and (at least nominally) lived according to his laws." The etymological equivalent is in use in other languages, e.g., يَهُودِيّ yahūdī (sg.), al-yahūd (pl.), in Arabic, "Jude" in German, "judeu" in Portuguese, "Juif" (m.)/"Juive" (f.) in French, "jøde" in Danish and Norwegian, "judío/a" in Spanish, "jood" in Dutch, "żyd" in Polish etc., but derivations of the word "Hebrew" are also in use to describe a Jew, e.g., in Italian (Ebreo), in Persian ("Ebri/Ebrani" (Persian: عبری/عبرانی)) and Russian (Еврей, Yevrey). The German word "Jude" is pronounced [ˈjuːdə], the corresponding adjective "jüdisch" [ˈjyːdɪʃ] (Jewish) is the origin of the word "Yiddish". According to The American Heritage Dictionary of the English Language, fourth edition (2000), It is widely recognized that the attributive use of the noun Jew, in phrases such as Jew lawyer or Jew ethics, is both vulgar and highly offensive. In such contexts Jewish is the only acceptable possibility. Some people, however, have become so wary of this construction that they have extended the stigma to any use of Jew as a noun, a practice that carries risks of its own. In a sentence such as There are now several Jews on the council, which is unobjectionable, the substitution of a circumlocution like Jewish people or persons of Jewish background may in itself cause offense for seeming to imply that Jew has a negative connotation when used as a noun. Identity Judaism shares some of the characteristics of a nation, an ethnicity, a religion, and a culture, making the definition of who is a Jew vary slightly depending on whether a religious or national approach to identity is used.[better source needed] Generally, in modern secular usage, Jews include three groups: people who were born to a Jewish family regardless of whether or not they follow the religion, those who have some Jewish ancestral background or lineage (sometimes including those who do not have strictly matrilineal descent), and people without any Jewish ancestral background or lineage who have formally converted to Judaism and therefore are followers of the religion. In the context of biblical and classical literature, Jews could refer to inhabitants of the Kingdom of Judah, or the broader Judean region, allies of the Judean state, or anyone that followed Judaism. Historical definitions of Jewish identity have traditionally been based on halakhic definitions of matrilineal descent, and halakhic conversions. These definitions of who is a Jew date back to the codification of the Oral Torah into the Babylonian Talmud, around 200 CE. Interpretations by Jewish sages of sections of the Tanakh – such as Deuteronomy 7:1–5, which forbade intermarriage between their Israelite ancestors and seven non-Israelite nations: "for that [i.e. giving your daughters to their sons or taking their daughters for your sons,] would turn away your children from following me, to serve other gods"[failed verification] – are used as a warning against intermarriage between Jews and gentiles. Leviticus 24:10 says that the son in a marriage between a Hebrew woman and an Egyptian man is "of the community of Israel." This is complemented by Ezra 10:2–3, where Israelites returning from Babylon vow to put aside their gentile wives and their children. A popular theory is that the rape of Jewish women in captivity brought about the law of Jewish identity being inherited through the maternal line, although scholars challenge this theory citing the Talmudic establishment of the law from the pre-exile period. Another argument is that the rabbis changed the law of patrilineal descent to matrilineal descent due to the widespread rape of Jewish women by Roman soldiers. Since the anti-religious Haskalah movement of the late 18th and 19th centuries, halakhic interpretations of Jewish identity have been challenged. According to historian Shaye J. D. Cohen, the status of the offspring of mixed marriages was determined patrilineally in the Bible. He brings two likely explanations for the change in Mishnaic times: first, the Mishnah may have been applying the same logic to mixed marriages as it had applied to other mixtures (Kil'ayim). Thus, a mixed marriage is forbidden as is the union of a horse and a donkey, and in both unions the offspring are judged matrilineally. Second, the Tannaim may have been influenced by Roman law, which dictated that when a parent could not contract a legal marriage, offspring would follow the mother. Rabbi Rivon Krygier follows a similar reasoning, arguing that Jewish descent had formerly passed through the patrilineal descent and the law of matrilineal descent had its roots in the Roman legal system. Origins The prehistory and ethnogenesis of the Jews are closely intertwined with archaeology, biology, historical textual records, mythology, and religious literature. The ethnic origin of the Jews lie in the Israelites, a confederation of Iron Age Semitic-speaking tribes that inhabited a part of Canaan during the tribal and monarchic periods. Modern Jews are named after and also descended from the southern Israelite Kingdom of Judah. Gary A. Rendsburg links the early Canaanite nomadic pastoralists confederation to the Shasu known to the Egyptians around the 15th century BCE. According to the Hebrew Bible narrative, Jewish history begins with the Biblical patriarchs such as Abraham, his son Isaac, Isaac's son Jacob, and the Biblical matriarchs Sarah, Rebecca, Leah, and Rachel, who lived in Canaan. The twelve sons of Jacob subsequently gave birth to the Twelve Tribes. Jacob and his family migrated to Ancient Egypt after being invited to live with Jacob's son Joseph by the Pharaoh himself. Jacob's descendants were later enslaved until the Exodus, led by Moses. Afterwards, the Israelites conquered Canaan under Moses' successor Joshua, and went through the period of the Biblical judges after the death of Joshua. Through the mediation of Samuel, the Israelites were subject to a king, Saul, who was succeeded by David and then Solomon, after whom the United Monarchy ended and was split into a separate Kingdom of Israel and a Kingdom of Judah. The Kingdom of Judah is described as comprising the tribes of Judah, Benjamin and partially, Levi. They later assimilated remnants of other tribes who migrated there from the northern Kingdom of Israel. In the extra-biblical record, the Israelites become visible as a people between 1200 and 1000 BCE. There is well accepted archeological evidence referring to "Israel" in the Merneptah Stele, which dates to about 1200 BCE, and in the Mesha stele from 840 BCE. It is debated whether a period like that of the Biblical judges occurred and if there ever was a United Monarchy. There is further disagreement about the earliest existence of the Kingdoms of Israel and Judah and their extent and power. Historians agree that a Kingdom of Israel existed by c. 900 BCE,: 169–95 there is a consensus that a Kingdom of Judah existed by c. 700 BCE at least, and recent excavations in Khirbet Qeiyafa have provided strong evidence for dating the Kingdom of Judah to the 10th century BCE. In 587 BCE, Nebuchadnezzar II, King of the Neo-Babylonian Empire, besieged Jerusalem, destroyed the First Temple and deported parts of the Judahite population. Scholars disagree regarding the extent to which the Bible should be accepted as a historical source for early Israelite history. Rendsburg states that there are two approximately equal groups of scholars who debate the historicity of the biblical narrative, the minimalists who largely reject it, and the maximalists who largely accept it, with the minimalists being the more vocal of the two. Some of the leading minimalists reframe the biblical account as constituting the Israelites' inspiring national myth narrative, suggesting that according to the modern archaeological and historical account, the Israelites and their culture did not overtake the region by force, but instead branched out of the Canaanite peoples and culture through the development of a distinct monolatristic—and later monotheistic—religion of Yahwism centered on Yahweh, one of the gods of the Canaanite pantheon. The growth of Yahweh-centric belief, along with a number of cultic practices, gradually gave rise to a distinct Israelite ethnic group, setting them apart from other Canaanites. According to Dever, modern archaeologists have largely discarded the search for evidence of the biblical narrative surrounding the patriarchs and the exodus. According to the maximalist position, the modern archaeological record independently points to a narrative which largely agrees with the biblical account. This narrative provides a testimony of the Israelites as a nomadic people known to the Egyptians as belonging to the Shasu. Over time these nomads left the desert and settled on the central mountain range of the land of Canaan, in simple semi-nomadic settlements in which pig bones are notably absent. This population gradually shifted from a tribal lifestyle to a monarchy. While the archaeological record of the ninth century BCE provides evidence for two monarchies, one in the south under a dynasty founded by a figure named David with its capital in Jerusalem, and one in the north under a dynasty founded by a figure named Omri with its capital in Samaria. It also points to an early monarchic period in which these regions shared material culture and religion, suggesting a common origin. Archaeological finds also provide evidence for the later cooperation of these two kingdoms in their coalition against Aram, and for their destructions by the Assyrians and later by the Babylonians. Genetic studies on Jews show that most Jews worldwide bear a common genetic heritage which originates in the Middle East, and that they share certain genetic traits with other Gentile peoples of the Fertile Crescent. The genetic composition of different Jewish groups shows that Jews share a common gene pool dating back four millennia, as a marker of their common ancestral origin. Despite their long-term separation, Jewish communities maintained their unique commonalities, propensities, and sensibilities in culture, tradition, and language. History The earliest recorded evidence of a people by the name of Israel appears in the Merneptah Stele, which dates to around 1200 BCE. The majority of scholars agree that this text refers to the Israelites, a group that inhabited the central highlands of Canaan, where archaeological evidence shows that hundreds of small settlements were constructed between the 12th and 10th centuries BCE. The Israelites differentiated themselves from neighboring peoples through various distinct characteristics including religious practices, prohibition on intermarriage, and an emphasis on genealogy and family history. In the 10th century BCE, two neighboring Israelite kingdoms—the northern Kingdom of Israel and the southern Kingdom of Judah—emerged. Since their inception, they shared ethnic, cultural, linguistic and religious characteristics despite a complicated relationship. Israel, with its capital mostly in Samaria, was larger and wealthier, and soon developed into a regional power. In contrast, Judah, with its capital in Jerusalem, was less prosperous and covered a smaller, mostly mountainous territory. However, while in Israel the royal succession was often decided by a military coup d'état, resulting in several dynasty changes, political stability in Judah was much greater, as it was ruled by the House of David for the whole four centuries of its existence. Scholars also describe Biblical Jews as a 'proto-nation', in the modern nationalist sense, comparable to classical Greeks, the Gauls and the British Celts. Around 720 BCE, Kingdom of Israel was destroyed when it was conquered by the Neo-Assyrian Empire, which came to dominate the ancient Near East. Under the Assyrian resettlement policy, a significant portion of the northern Israelite population was exiled to Mesopotamia and replaced by immigrants from the same region. During the same period, and throughout the 7th century BCE, the Kingdom of Judah, now under Assyrian vassalage, experienced a period of prosperity and witnessed a significant population growth. This prosperity continued until the Neo-Assyrian king Sennacherib devastated the region of Judah in response to a rebellion in the area, ultimately halting at Jerusalem. Later in the same century, the Assyrians were defeated by the rising Neo-Babylonian Empire, and Judah became its vassal. In 587 BCE, following a revolt in Judah, the Babylonian king Nebuchadnezzar II besieged and destroyed Jerusalem and the First Temple, putting an end to the kingdom. The majority of Jerusalem's residents, including the kingdom's elite, were exiled to Babylon. According to the Book of Ezra, the Persian Cyrus the Great ended the Babylonian exile in 538 BCE, the year after he captured Babylon. The exile ended with the return under Zerubbabel the Prince (so called because he was a descendant of the royal line of David) and Joshua the Priest (a descendant of the line of the former High Priests of the Temple) and their construction of the Second Temple circa 521–516 BCE. As part of the Persian Empire, the former Kingdom of Judah became the province of Judah (Yehud Medinata), with a smaller territory and a reduced population. Judea was under control of the Achaemenids until the fall of their empire in c. 333 BCE to Alexander the Great. After several centuries under foreign imperial rule, the Maccabean Revolt against the Seleucid Empire resulted in an independent Hasmonean kingdom, under which the Jews once again enjoyed political independence for a period spanning from 110 to 63 BCE. Under Hasmonean rule the boundaries of their kingdom were expanded to include not only the land of the historical kingdom of Judah, but also the Galilee and Transjordan. In the beginning of this process the Idumeans, who had infiltrated southern Judea after the destruction of the First Temple, were converted en masse. In 63 BCE, Judea was conquered by the Romans. From 37 BCE to 6 CE, the Romans allowed the Jews to maintain some degree of independence by installing the Herodian dynasty as vassal kings. However, Judea eventually came directly under Roman control and was incorporated into the Roman Empire as the province of Judaea. The Jewish–Roman wars, a series of failed uprisings against Roman rule during the first and second centuries CE, had profound and devastating consequences for the Jewish population of Judaea. The First Jewish–Roman War (66–73/74 CE) culminated in the destruction of Jerusalem and the Second Temple, after which the significantly diminished Jewish population was stripped of political autonomy. A few generations later, the Bar Kokhba revolt (132–136 CE) erupted in response to Roman plans to rebuild Jerusalem as a Roman colony, and, possibly, to restrictions on circumcision. Its violent suppression by the Romans led to the near-total depopulation of Judea, and the demographic and cultural center of Jewish life shifted to Galilee. Jews were subsequently banned from residing in Jerusalem and the surrounding area, and the province of Judaea was renamed Syria Palaestina. These developments effectively ended Jewish efforts to restore political sovereignty in the region for nearly two millennia. Similar upheavals impacted the Jewish communities in the empire's eastern provinces during the Diaspora Revolt (115–117 CE), leading to the near-total destruction of Jewish diaspora communities in Libya, Cyprus and Egypt, including the highly influential community in Alexandria. The destruction of the Second Temple in 70 CE brought profound changes to Judaism. With the Temple's central place in Jewish worship gone, religious practices shifted towards prayer, Torah study (including Oral Torah), and communal gatherings in synagogues. Judaism also lost much of its sectarian nature.: 69 Two of the three main sects that flourished during the late Second Temple period, namely the Sadducees and Essenes, eventually disappeared, while Pharisaic beliefs became the foundational, liturgical, and ritualistic basis of Rabbinic Judaism, which emerged as the prevailing form of Judaism since late antiquity. The Jewish diaspora existed well before the destruction of the Second Temple in 70 CE and had been ongoing for centuries, with the dispersal driven by both forced expulsions and voluntary migrations. In Mesopotamia, a testimony to the beginnings of the Jewish community can be found in Joachin's ration tablets, listing provisions allotted to the exiled Judean king and his family by Nebuchadnezzar II, and further evidence are the Al-Yahudu tablets, dated to the 6th–5th centuries BCE and related to the exiles from Judea arriving after the destruction of the First Temple, though there is ample evidence for the presence of Jews in Babylonia even from 626 BCE. In Egypt, the documents from Elephantine reveal the trials of a community founded by a Persian Jewish garrison at two fortresses on the frontier during the 5th–4th centuries BCE, and according to Josephus the Jewish community in Alexandria existed since the founding of the city in the 4th century BCE by Alexander the Great. By 200 BCE, there were well established Jewish communities both in Egypt and Mesopotamia ("Babylonia" in Jewish sources) and in the two centuries that followed, Jewish populations were also present in Asia Minor, Greece, Macedonia, Cyrene, and, beginning in the middle of the first century BCE, in the city of Rome. Later, in the first centuries CE, as a result of the Jewish-Roman Wars, a large number of Jews were taken as captives, sold into slavery, or compelled to flee from the regions affected by the wars, contributing to the formation and expansion of Jewish communities across the Roman Empire as well as in Arabia and Mesopotamia. After the Bar Kokhba revolt, the Jewish population in Judaea—now significantly reduced— made efforts to recover from the revolt's devastating effects, but never fully regained its former strength. Between the second and fourth centuries CE, the region of Galilee emerged as the primary center of Jewish life in Syria Palaestina, experiencing both demographic growth and cultural development. It was during this period that two central rabbinic texts, the Mishnah and the Jerusalem Talmud, were composed. The Romans recognized the patriarchs—rabbinic sages such as Judah ha-Nasi—as representatives of the Jewish people, granting them a certain degree of autonomy. However, as the Roman Empire gave way to the Christianized Byzantine Empire under Constantine, Jews began to face persecution by both the Church and imperial authorities, Jews came to be persecuted by the church and the authorities, and many immigrated to communities in the diaspora. By the fourth century CE, Jews are believed to have lost their demographic majority in Syria Palaestina. The long-established Jewish community of Mesopotamia, which had been living under Parthian and later Sasanian rule, beyond the confines of the Roman Empire, became an important center of Jewish study as Judea's Jewish population declined. Estimates often place the Babylonian Jewish community of the 3rd to 7th centuries at around one million, making it the largest Jewish diaspora community of that period. Under the political leadership of the exilarch, who was regarded as a royal heir of the House of David, this community had an autonomous status and served as a place of refuge for the Jews of Syria Palaestina. A number of significant Talmudic academies, such as the Nehardea, Pumbedita, and Sura academies, were established in Mesopotamia, and many important Amoraim were active there. The Babylonian Talmud, a centerpiece of Jewish religious law, was compiled in Babylonia in the 3rd to 6th centuries. Jewish diaspora communities are generally described to have coalesced into three major ethnic subdivisions according to where their ancestors settled: the Ashkenazim (initially in the Rhineland and France), the Sephardim (initially in the Iberian Peninsula), and the Mizrahim (Middle East and North Africa). Romaniote Jews, Tunisian Jews, Yemenite Jews, Egyptian Jews, Ethiopian Jews, Bukharan Jews, Mountain Jews, and other groups also predated the arrival of the Sephardic diaspora. During the same period, Jewish communities in the Middle East thrived under Islamic rule, especially in cities like Baghdad, Cairo, and Damascus. In Babylonia, from the 7th to 11th centuries the Pumbedita and Sura academies led the Arab and to an extent the entire Jewish world. The deans and students of said academies defined the Geonic period in Jewish history. Following this period were the Rishonim who lived from the 11th to 15th centuries. Like their European counterparts, Jews in the Middle East and North Africa also faced periods of persecution and discriminatory policies, with the Almohad Caliphate in North Africa and Iberia issuing forced conversion decrees, causing Jews such as Maimonides to seek safety in other regions. Despite experiencing repeated waves of persecution, Ashkenazi Jews in Western Europe worked in a variety of fields, making an impact on their communities' economy and societies. In Francia, for example, figures like Isaac Judaeus and Armentarius occupied prominent social and economic positions. Francia also witnessed the development of a sophisticated tradition of biblical commentary, as exemplified by Rashi and the tosafists. In 1144, the first documented blood libel occurred in Norwich, England, marking an escalation in the pattern of discrimination and violence that Jews had already been subjected to throughout medieval Europe. During the 12th and 13th centuries, Jews faced frequent antisemitic legislation - including laws prescribing distinctive dress - alongside segregation, repeated blood libels, pogroms, and massacres such as the Rhineland Massacres (1066). The Jews of the Holy Roman Empire were designated Servi camerae regis (“servants of the imperial chamber”) by Frederick II, a status that afforded limited protection while simultaneously entangling them in the political struggles between the emperor and the German principalities and cities. Persecution intensified during the Black Death in the mid-14th century, when Jews were accused of poisoning wells and many communities were destroyed. These pressures, combined with major expulsions such as that from England in 1290, gradually pushed Ashkenazi Jewish populations eastward into Poland, Lithuania, and Russia. One of the largest Jewish communities of the Middle Ages was in the Iberian Peninsula, which for a time contained the largest Jewish population in Europe. Iberian Jewry endured discrimination under the Visigoths but saw its fortunes improve under Umayyad rule and later the Taifa kingdoms. During this period, the Jews of Muslim Spain entered a "Golden Age" marked by achievements in Hebrew poetry and literature, religious scholarship, grammar, medicine and science, with leading figures including Hasdai ibn Shaprut, Judah Halevi, Moses ibn Ezra and Solomon ibn Gabirol. Jews also rose to high office, most notably Samuel ibn Naghrillah, a scholar and poet who served as grand vizier and military commander of Granada. The Golden Age ended with the rise of the radical Almoravid and Almohad dynasties, whose persecutions drove many Jews from Iberia (including Maimonides), together with the advancing Reconquista. In 1391, widespread pogroms swept across Spain, leaving thousands dead and forcing mass conversions. The Spanish Inquisition was later established to pursue, torture and execute conversos who continued to practice Judaism in secret, while public disputations were staged to discredit Judaism. In 1492, after the Reconquista, Isabella I of Castile and Ferdinand II of Aragon decreed the expulsion of all Jews who refused conversion, sending an estimated 200,000 into exile in Portugal, Italy, North Africa, and the Ottoman Empire. In 1497, Portugal's Jews, about 30,000, were formally ordered expelled but instead were forcibly converted to retain their economic role. In 1498, some 3,500 Jews were expelled from Navarre. Many converts outwardly adopted Christianity while secretly preserving Jewish practices, becoming crypto-Jews (also known as marranos or anusim), who remained targets of the various Inquisitions for centuries. Following the expulsions from Spain and Portugal in the 1490s, Jewish exiles dispersed across the Mediterranean, Europe, and North Africa. Many settled in the Ottoman Empire—which, replacing the Iberian Peninsula, became home to the world's largest Jewish population—where new communities developed in Anatolia, the Balkans, and the Land of Israel. Cities such as Istanbul and Thessaloniki grew into major Jewish centers, while in 16th-century Safed a flourishing spiritual life took shape. There, Solomon Alkabetz, Moses Cordovero, and Isaac Luria developed influential new schools of Kabbalah, giving powerful impetus to Jewish mysticism, and Joseph Karo composed the Shulchan Aruch, which became a cornerstone of Jewish law. In the 17th century, Portuguese conversos who returned to Judaism and engaged in trade and banking helped establish Amsterdam as a prosperous Jewish center, while also forming communities in cities such as Antwerp and London. This period also witnessed waves of messianic fervor, most notably the rise of the Sabbatean movement in the 1660s, led by Sabbatai Zvi of İzmir, which reverberated throughout the Jewish world. In Eastern Europe, Poland–Lithuania became the principal center of Ashkenazi Jewry, eventually becoming home to the largest Jewish population in the world. Jewish life flourished there from in the early modern era, supported by relative stability, economic opportunity, and strong communal institutions. The mid-17th century brought devastation with the Cossack uprisings in Ukraine, which reversed migration flows and sent refugees westward, yet Poland–Lithuania remained the demographic and cultural heartland of Ashkenazic Jewry. Following the partitions of Poland, most of its Jews came under Russian rule and were confined to the "Pale of Settlement." The 18th century also witnessed new religious and intellectual currents. Hasidism, founded by Baal Shem Tov, emphasized mysticism and piety, while its opponents, the Misnagdim ("opponents") led by the Vilna Gaon, defended rabbinic scholarship and tradition. In Western Europe, during the 1760s and 1770s, the Haskalah (Jewish Enlightenment) emerged in German-speaking lands, where figures such as Moses Mendelssohn promoted secular learning, vernacular literacy, and integration into European society. Elsewhere, Jews began to be re-admitted to Western Europe, including England, where Menasseh ben Israel petitioned Oliver Cromwell for their return. In the Americas, Jews of Sephardic descent first arrived as conversos in Spanish and Portuguese colonies, where many faced trial by Inquisition tribunals for "judaizing." A more durable presence began in Dutch Brazil, where Jews openly practiced their religion and established the first synagogues in the New World, before the Portuguese reconquest forced their dispersal to Amsterdam, the Caribbean, and North America. Sephardic communities took root in Curaçao, Suriname, Jamaica, and Barbados, later joined by Ashkenazi migrants. In North America, Jews were present from the mid-17th century, with New Amsterdam hosting the first organized congregation in 1654. By the time of the American Revolution, small communities in New York, Newport, Philadelphia, Savannah, and Charleston played an active role in the struggle for independence. In the late 19th century, Jews in Western Europe gradually achieved legal emancipation, though social acceptance remained limited by persistent antisemitism and rising nationalism. In Eastern Europe, particularly within the Russian Empire's Pale of Settlement, Jews faced mounting legal restrictions and recurring pogroms. From this environment emerged Zionism, a national revival movement originating in Central and Eastern Europe that sought to re-establish a Jewish polity in the Land of Israel as a means of returning the Jewish people to their ancestral homeland and ending centuries of exile and persecution. This led to waves of Jewish migration to Ottoman-controlled Palestine. Theodor Herzl, who is considered the father of political Zionism, offered his vision of a future Jewish state in his 1896 book Der Judenstaat (The Jewish State); a year later, he presided over the First Zionist Congress. The antisemitism that inflicted Jewish communities in Europe also triggered a mass exodus of 2.8 million Jews to the United States between 1881 and 1924. Despite this, some Jews of Europe and the United States were able to make great achievements in various fields of science and culture. Among the most influential from this period are Albert Einstein in physics, Sigmund Freud in psychology, Franz Kafka in literature, and Irving Berlin in music. Many Nobel Prize winners at this time were Jewish, as is still the case. When Adolf Hitler and the Nazi Party came to power in Germany in 1933, the situation for Jews deteriorated rapidly as a direct result of Nazi policies. Many Jews fled from Europe to Mandatory Palestine, the United States, and the Soviet Union as a result of racial anti-Semitic laws, economic difficulties, and the fear of an impending war. World War II started in 1939, and by 1941, Hitler occupied almost all of Europe. Following the German invasion of the Soviet Union in 1941, the Final Solution—an extensive, organized effort with an unprecedented scope intended to annihilate the Jewish people—began, and resulted in the persecution and murder of Jews in Europe and North Africa. In Poland, three million were murdered in gas chambers in all concentration camps combined, with one million at the Auschwitz camp complex alone. The Holocaust is the name given to this genocide, in which six million Jews in total were systematically murdered. Before and during the Holocaust, enormous numbers of Jews immigrated to Mandatory Palestine. In 1944, the Jewish insurgency in Mandatory Palestine began with the aim of gaining full independence from the United Kingdom. On 14 May 1948, upon the termination of the mandate, David Ben-Gurion declared the creation of the State of Israel, a Jewish and democratic state. Immediately afterwards, all neighboring Arab states invaded, and were resisted by the newly formed Israel Defense Forces. In 1949, the war ended and Israel started building its state and absorbing waves of Aliyah, granting citizenship to Jews all over the world via the Law of Return passed in 1950. However, both the Israeli–Palestinian conflict and wider Arab–Israeli conflict continue to this day. Culture The Jewish people and the religion of Judaism are strongly interrelated. Converts to Judaism have a status within the Jewish people equal to those born into it. However, converts who go on to practice no Judaism are likely to be viewed with skepticism. Mainstream Judaism does not proselytize, and conversion is considered a difficult task. A significant portion of conversions are undertaken by children of mixed marriages, or would-be or current spouses of Jews. The Hebrew Bible, a religious interpretation of the traditions and early history of the Jews, established the first of the Abrahamic religions, which are now practiced by 54 percent of the world. Judaism guides its adherents in both practice and belief, and has been called not only a religion, but also a "way of life," which has made drawing a clear distinction between Judaism, Jewish culture, and Jewish identity rather difficult. Throughout history, in eras and places as diverse as the ancient Hellenic world, in Europe before and after The Age of Enlightenment (see Haskalah), in Islamic Spain and Portugal, in North Africa and the Middle East, India, China, or the contemporary United States and Israel, cultural phenomena have developed that are in some sense characteristically Jewish without being at all specifically religious. Some factors in this come from within Judaism, others from the interaction of Jews or specific communities of Jews with their surroundings, and still others from the inner social and cultural dynamics of the community, as opposed to from the religion itself. This phenomenon has led to considerably different Jewish cultures unique to their own communities. Hebrew is the liturgical language of Judaism (termed lashon ha-kodesh, "the holy tongue"), the language in which most of the Hebrew scriptures (Tanakh) were composed, and the daily speech of the Jewish people for centuries. By the 5th century BCE, Aramaic, a closely related tongue, joined Hebrew as the spoken language in Judea. By the 3rd century BCE, some Jews of the diaspora were speaking Greek. Others, such as in the Jewish communities of Asoristan, known to Jews as Babylonia, were speaking Hebrew and Aramaic, the languages of the Babylonian Talmud. Dialects of these same languages were also used by the Jews of Syria Palaestina at that time.[citation needed] For centuries, Jews worldwide have spoken the local or dominant languages of the regions they migrated to, often developing distinctive dialectal forms or branches that became independent languages. Yiddish is the Judaeo-German language developed by Ashkenazi Jews who migrated to Central Europe. Ladino is the Judaeo-Spanish language developed by Sephardic Jews who migrated to the Iberian Peninsula. Due to many factors, including the impact of the Holocaust on European Jewry, the Jewish exodus from Arab and Muslim countries, and widespread emigration from other Jewish communities around the world, ancient and distinct Jewish languages of several communities, including Judaeo-Georgian, Judaeo-Arabic, Judaeo-Berber, Krymchak, Judaeo-Malayalam and many others, have largely fallen out of use. For over sixteen centuries Hebrew was used almost exclusively as a liturgical language, and as the language in which most books had been written on Judaism, with a few speaking only Hebrew on the Sabbath. Hebrew was revived as a spoken language by Eliezer ben Yehuda, who arrived in Palestine in 1881. It had not been used as a mother tongue since Tannaic times. Modern Hebrew is designated as the "State language" of Israel. Despite efforts to revive Hebrew as the national language of the Jewish people, knowledge of the language is not commonly possessed by Jews worldwide and English has emerged as the lingua franca of the Jewish diaspora. Although many Jews once had sufficient knowledge of Hebrew to study the classic literature, and Jewish languages like Yiddish and Ladino were commonly used as recently as the early 20th century, most Jews lack such knowledge today and English has by and large superseded most Jewish vernaculars. The three most commonly spoken languages among Jews today are Hebrew, English, and Russian. Some Romance languages, particularly French and Spanish, are also widely used. Yiddish has been spoken by more Jews in history than any other language, but it is far less used today following the Holocaust and the adoption of Modern Hebrew by the Zionist movement and the State of Israel. In some places, the mother language of the Jewish community differs from that of the general population or the dominant group. For example, in Quebec, the Ashkenazic majority has adopted English, while the Sephardic minority uses French as its primary language. Similarly, South African Jews adopted English rather than Afrikaans. Due to both Czarist and Soviet policies, Russian has superseded Yiddish as the language of Russian Jews, but these policies have also affected neighboring communities. Today, Russian is the first language for many Jewish communities in a number of Post-Soviet states, such as Ukraine and Uzbekistan,[better source needed] as well as for Ashkenazic Jews in Azerbaijan, Georgia, and Tajikistan. Although communities in North Africa today are small and dwindling, Jews there had shifted from a multilingual group to a monolingual one (or nearly so), speaking French in Algeria, Morocco, and the city of Tunis, while most North Africans continue to use Arabic or Berber as their mother tongue.[citation needed] There is no single governing body for the Jewish community, nor a single authority with responsibility for religious doctrine. Instead, a variety of secular and religious institutions at the local, national, and international levels lead various parts of the Jewish community on a variety of issues. Today, many countries have a Chief Rabbi who serves as a representative of that country's Jewry. Although many Hasidic Jews follow a certain hereditary Hasidic dynasty, there is no one commonly accepted leader of all Hasidic Jews. Many Jews believe that the Messiah will act a unifying leader for Jews and the entire world. A number of modern scholars of nationalism support the existence of Jewish national identity in antiquity. One of them is David Goodblatt, who generally believes in the existence of nationalism before the modern period. In his view, the Bible, the parabiblical literature and the Jewish national history provide the base for a Jewish collective identity. Although many of the ancient Jews were illiterate (as were their neighbors), their national narrative was reinforced through public readings. The Hebrew language also constructed and preserved national identity. Although it was not widely spoken after the 5th century BCE, Goodblatt states: the mere presence of the language in spoken or written form could invoke the concept of a Jewish national identity. Even if one knew no Hebrew or was illiterate, one could recognize that a group of signs was in Hebrew script. ... It was the language of the Israelite ancestors, the national literature, and the national religion. As such it was inseparable from the national identity. Indeed its mere presence in visual or aural medium could invoke that identity. Anthony D. Smith, an historical sociologist considered one of the founders of the field of nationalism studies, wrote that the Jews of the late Second Temple period provide "a closer approximation to the ideal type of the nation [...] than perhaps anywhere else in the ancient world." He adds that this observation "must make us wary of pronouncing too readily against the possibility of the nation, and even a form of religious nationalism, before the onset of modernity." Agreeing with Smith, Goodblatt suggests omitting the qualifier "religious" from Smith's definition of ancient Jewish nationalism, noting that, according to Smith, a religious component in national memories and culture is common even in the modern era. This view is echoed by political scientist Tom Garvin, who writes that "something strangely like modern nationalism is documented for many peoples in medieval times and in classical times as well," citing the ancient Jews as one of several "obvious examples", alongside the classical Greeks and the Gaulish and British Celts. Fergus Millar suggests that the sources of Jewish national identity and their early nationalist movements in the first and second centuries CE included several key elements: the Bible as both a national history and legal source, the Hebrew language as a national language, a system of law, and social institutions such as schools, synagogues, and Sabbath worship. Adrian Hastings argued that Jews are the "true proto-nation", that through the model of ancient Israel found in the Hebrew Bible, provided the world with the original concept of nationhood which later influenced Christian nations. However, following Jerusalem's destruction in the first century CE, Jews ceased to be a political entity and did not resemble a traditional nation-state for almost two millennia. Despite this, they maintained their national identity through collective memory, religion and sacred texts, even without land or political power, and remained a nation rather than just an ethnic group, eventually leading to the rise of Zionism and the establishment of Israel. Steven Weitzman suggests that Jewish nationalist sentiment in antiquity was encouraged because under foreign rule (Persians, Greeks, Romans) Jews were able to claim that they were an ancient nation. This claim was based on the preservation and reverence of their scriptures, the Hebrew language, the Temple and priesthood, and other traditions of their ancestors. Doron Mendels further observes that the Hasmonean kingdom, one of the few examples of indigenous statehood at its time, significantly reinforced Jewish national consciousness. The memory of this period of independence contributed to the persistent efforts to revive Jewish sovereignty in Judea, leading to the major revolts against Roman rule in the 1st and 2nd centuries CE. Demographics Within the world's Jewish population there are distinct ethnic divisions, most of which are primarily the result of geographic branching from an originating Israelite population, and subsequent independent evolutions. An array of Jewish communities was established by Jewish settlers in various places around the Old World, often at great distances from one another, resulting in effective and often long-term isolation. During the millennia of the Jewish diaspora the communities would develop under the influence of their local environments: political, cultural, natural, and populational. Today, manifestations of these differences among the Jews can be observed in Jewish cultural expressions of each community, including Jewish linguistic diversity, culinary preferences, liturgical practices, religious interpretations, as well as degrees and sources of genetic admixture. Jews are often identified as belonging to one of two major groups: the Ashkenazim and the Sephardim. Ashkenazim are so named in reference to their geographical origins (their ancestors' culture coalesced in the Rhineland, an area historically referred to by Jews as Ashkenaz). Similarly, Sephardim (Sefarad meaning "Spain" in Hebrew) are named in reference their origins in Iberia. The diverse groups of Jews of the Middle East and North Africa are often collectively referred to as Sephardim together with Sephardim proper for liturgical reasons having to do with their prayer rites. A common term for many of these non-Spanish Jews who are sometimes still broadly grouped as Sephardim is Mizrahim (lit. 'easterners' in Hebrew). Nevertheless, Mizrahis and Sepharadim are usually ethnically distinct. Smaller groups include, but are not restricted to, Indian Jews such as the Bene Israel, Bnei Menashe, Cochin Jews, and Bene Ephraim; the Romaniotes of Greece; the Italian Jews ("Italkim" or "Bené Roma"); the Teimanim from Yemen; various African Jews, including most numerously the Beta Israel of Ethiopia; and Chinese Jews, most notably the Kaifeng Jews, as well as various other distinct but now almost extinct communities. The divisions between all these groups are approximate and their boundaries are not always clear. The Mizrahim for example, are a heterogeneous collection of North African, Central Asian, Caucasian, and Middle Eastern Jewish communities that are no closer related to each other than they are to any of the earlier mentioned Jewish groups. In modern usage, however, the Mizrahim are sometimes termed Sephardi due to similar styles of liturgy, despite independent development from Sephardim proper. Thus, among Mizrahim there are Egyptian Jews, Iraqi Jews, Lebanese Jews, Kurdish Jews, Moroccan Jews, Libyan Jews, Syrian Jews, Bukharian Jews, Mountain Jews, Georgian Jews, Iranian Jews, Afghan Jews, and various others. The Teimanim from Yemen are sometimes included, although their style of liturgy is unique and they differ in respect to the admixture found among them to that found in Mizrahim. In addition, there is a differentiation made between Sephardi migrants who established themselves in the Middle East and North Africa after the expulsion of the Jews from Spain and Portugal in the 1490s and the pre-existing Jewish communities in those regions. Ashkenazi Jews represent the bulk of modern Jewry, with at least 70 percent of Jews worldwide (and up to 90 percent prior to World War II and the Holocaust). As a result of their emigration from Europe, Ashkenazim also represent the overwhelming majority of Jews in the New World continents, in countries such as the United States, Canada, Argentina, Australia, and Brazil. In France, the immigration of Jews from Algeria (Sephardim) has led them to outnumber the Ashkenazim. Only in Israel is the Jewish population representative of all groups, a melting pot independent of each group's proportion within the overall world Jewish population. Y DNA studies tend to imply a small number of founders in an old population whose members parted and followed different migration paths. In most Jewish populations, these male line ancestors appear to have been mainly Middle Eastern. For example, Ashkenazi Jews share more common paternal lineages with other Jewish and Middle Eastern groups than with non-Jewish populations in areas where Jews lived in Eastern Europe, Germany, and the French Rhine Valley. This is consistent with Jewish traditions in placing most Jewish paternal origins in the region of the Middle East. Conversely, the maternal lineages of Jewish populations, studied by looking at mitochondrial DNA, are generally more heterogeneous. Scholars such as Harry Ostrer and Raphael Falk believe this indicates that many Jewish males found new mates from European and other communities in the places where they migrated in the diaspora after fleeing ancient Israel. In contrast, Behar has found evidence that about 40 percent of Ashkenazi Jews originate maternally from just four female founders, who were of Middle Eastern origin. The populations of Sephardi and Mizrahi Jewish communities "showed no evidence for a narrow founder effect." Subsequent studies carried out by Feder et al. confirmed the large portion of non-local maternal origin among Ashkenazi Jews. Reflecting on their findings related to the maternal origin of Ashkenazi Jews, the authors conclude "Clearly, the differences between Jews and non-Jews are far larger than those observed among the Jewish communities. Hence, differences between the Jewish communities can be overlooked when non-Jews are included in the comparisons." However, a 2025 genetic study on the Ashkenazi Jewish founder population supports the presence of a substantial Near Eastern component in the maternal lineages. Analyses of mitochondrial DNA (mtDNA) indicate that the core founder lineages, estimated at around 54, likely originated from the Near East, with these founder signatures appearing in multiple copies across the population. While later admixture introduced additional mtDNA lineages, these absorbed lineages are distinguishable from the original founders. The findings are consistent with genome-wide Identity-by-Descent and Lineage Extinction analyses, reinforcing the Near Eastern origin of the Ashkenazi maternal founders. A study showed that 7% of Ashkenazi Jews have the haplogroup G2c, which is mainly found in Pashtuns and on lower scales all major Jewish groups, Palestinians, Syrians, and Lebanese. Studies of autosomal DNA, which look at the entire DNA mixture, have become increasingly important as the technology develops. They show that Jewish populations have tended to form relatively closely related groups in independent communities, with most in a community sharing significant ancestry in common. For Jewish populations of the diaspora, the genetic composition of Ashkenazi, Sephardic, and Mizrahi Jewish populations show a predominant amount of shared Middle Eastern ancestry. According to Behar, the most parsimonious explanation for this shared Middle Eastern ancestry is that it is "consistent with the historical formulation of the Jewish people as descending from ancient Hebrew and Israelite residents of the Levant" and "the dispersion of the people of ancient Israel throughout the Old World". North African, Italian and others of Iberian origin show variable frequencies of admixture with non-Jewish historical host populations among the maternal lines. In the case of Ashkenazi and Sephardi Jews (in particular Moroccan Jews), who are closely related, the source of non-Jewish admixture is mainly Southern European, while Mizrahi Jews show evidence of admixture with other Middle Eastern populations. Behar et al. have remarked on a close relationship between Ashkenazi Jews and modern Italians. A 2001 study found that Jews were more closely related to groups of the Fertile Crescent (Kurds, Turks, and Armenians) than to their Arab neighbors, whose genetic signature was found in geographic patterns reflective of Islamic conquests. The studies also show that Sephardic Bnei Anusim (descendants of the "anusim" who were forced to convert to Catholicism), which comprise up to 19.8 percent of the population of today's Iberia (Spain and Portugal) and at least 10 percent of the population of Ibero-America (Hispanic America and Brazil), have Sephardic Jewish ancestry within the last few centuries. The Bene Israel and Cochin Jews of India, Beta Israel of Ethiopia, and a portion of the Lemba people of Southern Africa, despite more closely resembling the local populations of their native countries, have also been thought to have some more remote ancient Jewish ancestry. Views on the Lemba have changed and genetic Y-DNA analyses in the 2000s have established a partially Middle-Eastern origin for a portion of the male Lemba population but have been unable to narrow this down further. Although historically, Jews have been found all over the world, in the decades since World War II and the establishment of Israel, they have increasingly concentrated in a small number of countries. In 2021, Israel and the United States together accounted for over 85 percent of the global Jewish population, with approximately 45.3% and 39.6% of the world's Jews, respectively. More than half (51.2%) of world Jewry resides in just ten metropolitan areas. As of 2021, these ten areas were Tel Aviv, New York, Jerusalem, Haifa, Los Angeles, Miami, Philadelphia, Paris, Washington, and Chicago. The Tel Aviv metro area has the highest percent of Jews among the total population (94.8%), followed by Jerusalem (72.3%), Haifa (73.1%), and Beersheba (60.4%), the balance mostly being Israeli Arabs. Outside Israel, the highest percent of Jews in a metropolitan area was in New York (10.8%), followed by Miami (8.7%), Philadelphia (6.8%), San Francisco (5.1%), Washington (4.7%), Los Angeles (4.7%), Toronto (4.5%), and Baltimore (4.1%). As of 2010, there were nearly 14 million Jews around the world, roughly 0.2% of the world's population at the time. According to the 2007 estimates of The Jewish People Policy Planning Institute, the world's Jewish population is 13.2 million. This statistic incorporates both practicing Jews affiliated with synagogues and the Jewish community, and approximately 4.5 million unaffiliated and secular Jews.[citation needed] According to Sergio Della Pergola, a demographer of the Jewish population, in 2021 there were about 6.8 million Jews in Israel, 6 million in the United States, and 2.3 million in the rest of the world. Israel, the Jewish nation-state, is the only country in which Jews make up a majority of the citizens. Israel was established as an independent democratic and Jewish state on 14 May 1948. Of the 120 members in its parliament, the Knesset, as of 2016[update], 14 members of the Knesset are Arab citizens of Israel (not including the Druze), most representing Arab political parties. One of Israel's Supreme Court judges is also an Arab citizen of Israel. Between 1948 and 1958, the Jewish population rose from 800,000 to two million. Currently, Jews account for 75.4 percent of the Israeli population, or 6 million people. The early years of the State of Israel were marked by the mass immigration of Holocaust survivors in the aftermath of the Holocaust and Jews fleeing Arab lands. Israel also has a large population of Ethiopian Jews, many of whom were airlifted to Israel in the late 1980s and early 1990s. Between 1974 and 1979 nearly 227,258 immigrants arrived in Israel, about half being from the Soviet Union. This period also saw an increase in immigration to Israel from Western Europe, Latin America, and North America. A trickle of immigrants from other communities has also arrived, including Indian Jews and others, as well as some descendants of Ashkenazi Holocaust survivors who had settled in countries such as the United States, Argentina, Australia, Chile, and South Africa. Some Jews have emigrated from Israel elsewhere, because of economic problems or disillusionment with political conditions and the continuing Arab–Israeli conflict. Jewish Israeli emigrants are known as yordim. The waves of immigration to the United States and elsewhere at the turn of the 19th century, the founding of Zionism and later events, including pogroms in Imperial Russia (mostly within the Pale of Settlement in present-day Ukraine, Moldova, Belarus and eastern Poland), the massacre of European Jewry during the Holocaust, and the founding of the state of Israel, with the subsequent Jewish exodus from Arab lands, all resulted in substantial shifts in the population centers of world Jewry by the end of the 20th century. More than half of the Jews live in the Diaspora (see Population table). Currently, the largest Jewish community outside Israel, and either the largest or second-largest Jewish community in the world, is located in the United States, with 6 million to 7.5 million Jews by various estimates. Elsewhere in the Americas, there are also large Jewish populations in Canada (315,000), Argentina (180,000–300,000), and Brazil (196,000–600,000), and smaller populations in Mexico, Uruguay, Venezuela, Chile, Colombia and several other countries (see History of the Jews in Latin America). According to a 2010 Pew Research Center study, about 470,000 people of Jewish heritage live in Latin America and the Caribbean. Demographers disagree on whether the United States has a larger Jewish population than Israel, with many maintaining that Israel surpassed the United States in Jewish population during the 2000s, while others maintain that the United States still has the largest Jewish population in the world. Currently, a major national Jewish population survey is planned to ascertain whether or not Israel has overtaken the United States in Jewish population. Western Europe's largest Jewish community, and the third-largest Jewish community in the world, can be found in France, home to between 483,000 and 500,000 Jews, the majority of whom are immigrants or refugees from North African countries such as Algeria, Morocco, and Tunisia (or their descendants). The United Kingdom has a Jewish community of 292,000. In Eastern Europe, the exact figures are difficult to establish. The number of Jews in Russia varies widely according to whether a source uses census data (which requires a person to choose a single nationality among choices that include "Russian" and "Jewish") or eligibility for immigration to Israel (which requires that a person have one or more Jewish grandparents). According to the latter criteria, the heads of the Russian Jewish community assert that up to 1.5 million Russians are eligible for aliyah. In Germany, the 102,000 Jews registered with the Jewish community are a slowly declining population, despite the immigration of tens of thousands of Jews from the former Soviet Union since the fall of the Berlin Wall. Thousands of Israelis also live in Germany, either permanently or temporarily, for economic reasons. Prior to 1948, approximately 800,000 Jews were living in lands which now make up the Arab world (excluding Israel). Of these, just under two-thirds lived in the French-controlled Maghreb region, 15 to 20 percent in the Kingdom of Iraq, approximately 10 percent in the Kingdom of Egypt and approximately 7 percent in the Kingdom of Yemen. A further 200,000 lived in Pahlavi Iran and the Republic of Turkey. Today, around 26,000 Jews live in Muslim-majority countries, mainly in Turkey (14,200) and Iran (9,100), while Morocco (2,000), Tunisia (1,000), and the United Arab Emirates (500) host the largest communities in the Arab world. A small-scale exodus had begun in many countries in the early decades of the 20th century, although the only substantial aliyah came from Yemen and Syria. The exodus from Arab and Muslim countries took place primarily from 1948. The first large-scale exoduses took place in the late 1940s and early 1950s, primarily in Iraq, Yemen and Libya, with up to 90 percent of these communities leaving within a few years. The peak of the exodus from Egypt occurred in 1956. The exodus in the Maghreb countries peaked in the 1960s. Lebanon was the only Arab country to see a temporary increase in its Jewish population during this period, due to an influx of refugees from other Arab countries, although by the mid-1970s the Jewish community of Lebanon had also dwindled. In the aftermath of the exodus wave from Arab states, an additional migration of Iranian Jews peaked in the 1980s when around 80 percent of Iranian Jews left the country.[citation needed] Outside Europe, the Americas, the Middle East, and the rest of Asia, there are significant Jewish populations in Australia (112,500) and South Africa (70,000). There is also a 6,800-strong community in New Zealand. Since at least the time of the Ancient Greeks, a proportion of Jews have assimilated into the wider non-Jewish society around them, by either choice or force, ceasing to practice Judaism and losing their Jewish identity. Assimilation took place in all areas, and during all time periods, with some Jewish communities, for example the Kaifeng Jews of China, disappearing entirely. The advent of the Jewish Enlightenment of the 18th century (see Haskalah) and the subsequent emancipation of the Jewish populations of Europe and America in the 19th century, accelerated the situation, encouraging Jews to increasingly participate in, and become part of, secular society. The result has been a growing trend of assimilation, as Jews marry non-Jewish spouses and stop participating in the Jewish community. Rates of interreligious marriage vary widely: In the United States, it is just under 50 percent; in the United Kingdom, around 53 percent; in France, around 30 percent; and in Australia and Mexico, as low as 10 percent. In the United States, only about a third of children from intermarriages affiliate with Jewish religious practice. The result is that most countries in the Diaspora have steady or slightly declining religiously Jewish populations as Jews continue to assimilate into the countries in which they live.[citation needed] The Jewish people and Judaism have experienced various persecutions throughout their history. During Late Antiquity and the Early Middle Ages, the Roman Empire (in its later phases known as the Byzantine Empire) repeatedly repressed the Jewish population, first by ejecting them from their homelands during the pagan Roman era and later by officially establishing them as second-class citizens during the Christian Roman era. According to James Carroll, "Jews accounted for 10% of the total population of the Roman Empire. By that ratio, if other factors had not intervened, there would be 200 million Jews in the world today, instead of something like 13 million." Later in medieval Western Europe, further persecutions of Jews by Christians occurred, notably during the Crusades—when Jews all over Germany were massacred—and in a series of expulsions from the Kingdom of England, Germany, and France. Then there occurred the largest expulsion of all, when Spain and Portugal, after the Reconquista (the Catholic Reconquest of the Iberian Peninsula), expelled both unbaptized Sephardic Jews and the ruling Muslim Moors. In the Papal States, which existed until 1870, Jews were required to live only in specified neighborhoods called ghettos. Islam and Judaism have a complex relationship. Traditionally Jews and Christians living in Muslim lands, known as dhimmis, were allowed to practice their religions and administer their internal affairs, but they were subject to certain conditions. They had to pay the jizya (a per capita tax imposed on free adult non-Muslim males) to the Islamic state. Dhimmis had an inferior status under Islamic rule. They had several social and legal disabilities such as prohibitions against bearing arms or giving testimony in courts in cases involving Muslims. Many of the disabilities were highly symbolic. The one described by Bernard Lewis as "most degrading" was the requirement of distinctive clothing, not found in the Quran or hadith but invented in early medieval Baghdad; its enforcement was highly erratic. On the other hand, Jews rarely faced martyrdom or exile, or forced compulsion to change their religion, and they were mostly free in their choice of residence and profession. Notable exceptions include the massacre of Jews and forcible conversion of some Jews by the rulers of the Almohad dynasty in Al-Andalus in the 12th century, as well as in Islamic Persia, and the forced confinement of Moroccan Jews to walled quarters known as mellahs beginning from the 15th century and especially in the early 19th century. In modern times, it has become commonplace for standard antisemitic themes to be conflated with anti-Zionist publications and pronouncements of Islamic movements such as Hezbollah and Hamas, in the pronouncements of various agencies of the Islamic Republic of Iran, and even in the newspapers and other publications of Turkish Refah Partisi."[better source needed] Throughout history, many rulers, empires and nations have oppressed their Jewish populations or sought to eliminate them entirely. Methods employed ranged from expulsion to outright genocide; within nations, often the threat of these extreme methods was sufficient to silence dissent. The history of antisemitism includes the First Crusade which resulted in the massacre of Jews; the Spanish Inquisition (led by Tomás de Torquemada) and the Portuguese Inquisition, with their persecution and autos-da-fé against the New Christians and Marrano Jews; the Bohdan Chmielnicki Cossack massacres in Ukraine; the Pogroms backed by the Russian Tsars; as well as expulsions from Spain, Portugal, England, France, Germany, and other countries in which the Jews had settled. According to a 2008 study published in the American Journal of Human Genetics, 19.8 percent of the modern Iberian population has Sephardic Jewish ancestry, indicating that the number of conversos may have been much higher than originally thought. The persecution reached a peak in Nazi Germany's Final Solution, which led to the Holocaust and the slaughter of approximately 6 million Jews. Of the world's 16 million Jews in 1939, almost 40% were murdered in the Holocaust. The Holocaust—the state-led systematic persecution and genocide of European Jews (and certain communities of North African Jews in European controlled North Africa) and other minority groups of Europe during World War II by Germany and its collaborators—remains the most notable modern-day persecution of Jews. The persecution and genocide were accomplished in stages. Legislation to remove the Jews from civil society was enacted years before the outbreak of World War II. Concentration camps were established in which inmates were used as slave labour until they died of exhaustion or disease. Where the Third Reich conquered new territory in Eastern Europe, specialized units called Einsatzgruppen murdered Jews and political opponents in mass shootings. Jews and Roma were crammed into ghettos before being transported hundreds of kilometres by freight train to extermination camps where, if they survived the journey, the majority of them were murdered in gas chambers. Virtually every arm of Germany's bureaucracy was involved in the logistics of the mass murder, turning the country into what one Holocaust scholar has called "a genocidal nation." Throughout Jewish history, Jews have repeatedly been directly or indirectly expelled from both their original homeland, the Land of Israel, and many of the areas in which they have settled. This experience as refugees has shaped Jewish identity and religious practice in many ways, and is thus a major element of Jewish history. In summary, the pogroms in Eastern Europe, the rise of modern antisemitism, the Holocaust, as well as the rise of Arab nationalism, all served to fuel the movements and migrations of huge segments of Jewry from land to land and continent to continent until they arrived back in large numbers at their original historical homeland in Israel. In the Bible, the patriarch Abraham is described as a migrant to the land of Canaan from Ur of the Chaldees. His descendants, the Children of Israel, undertook the Exodus (meaning "departure" or "exit" in Greek) from ancient Egypt, as described in the Book of Exodus. The first movement documented in the historical record occurred with the resettlement policy of the Neo-Assyrian Empire, which mandated the deportation of conquered peoples, and it is estimated some 4,500,000 among its captive populations suffered this dislocation over three centuries of Assyrian rule. With regard to Israel, Tiglath-Pileser III claims he deported 80% of the population of Lower Galilee, some 13,520 people. Some 27,000 Israelites, 20 to 25% of the population of the Kingdom of Israel, were described as being deported by Sargon II, and were replaced by other deported populations and sent into permanent exile by Assyria, initially to the Upper Mesopotamian provinces of the Assyrian Empire. Between 10,000 and 80,000 people from the Kingdom of Judah were similarly exiled by Babylonia, but these people were then returned to Judea by Cyrus the Great of the Persian Achaemenid Empire. Many Jews were exiled again by the Roman Empire. The 2,000 year dispersion of the Jewish diaspora beginning under the Roman Empire, as Jews were spread throughout the Roman world and, driven from land to land, settled wherever they could live freely enough to practice their religion. Over the course of the diaspora the center of Jewish life moved from Babylonia to the Iberian Peninsula to Poland to the United States and, as a result of Zionism, back to Israel. There were also many expulsions of Jews during the Middle Ages and Enlightenment in Europe, including: 1290, 16,000 Jews were expelled from England, (see the Statute of Jewry); in 1396, 100,000 from France; in 1421, thousands were expelled from Austria. Many of these Jews settled in East-Central Europe, especially Poland. Following the Spanish Inquisition in 1492, the Spanish population of around 200,000 Sephardic Jews were expelled by the Spanish crown and Catholic church, followed by expulsions in 1493 in Sicily (37,000 Jews) and Portugal in 1496. The expelled Jews fled mainly to the Ottoman Empire, the Netherlands, and North Africa, others migrating to Southern Europe and the Middle East. During the 19th century, France's policies of equal citizenship regardless of religion led to the immigration of Jews (especially from Eastern and Central Europe). This contributed to the arrival of millions of Jews in the New World. Over two million Eastern European Jews arrived in the United States from 1880 to 1925. In the latest phase of migrations, the Islamic Revolution of Iran caused many Iranian Jews to flee Iran. Most found refuge in the US (particularly Los Angeles, California, and Long Island, New York) and Israel. Smaller communities of Persian Jews exist in Canada and Western Europe. Similarly, when the Soviet Union collapsed, many of the Jews in the affected territory (who had been refuseniks) were suddenly allowed to leave. This produced a wave of migration to Israel in the early 1990s. Israel is the only country with a Jewish population that is consistently growing through natural population growth, although the Jewish populations of other countries, in Europe and North America, have recently increased through immigration. In the Diaspora, in almost every country the Jewish population in general is either declining or steady, but Orthodox and Haredi Jewish communities, whose members often shun birth control for religious reasons, have experienced rapid population growth. Orthodox and Conservative Judaism discourage proselytism to non-Jews, but many Jewish groups have tried to reach out to the assimilated Jewish communities of the Diaspora in order for them to reconnect to their Jewish roots. Additionally, while in principle Reform Judaism favours seeking new members for the faith, this position has not translated into active proselytism, instead taking the form of an effort to reach out to non-Jewish spouses of intermarried couples. There is also a trend of Orthodox movements reaching out to secular Jews in order to give them a stronger Jewish identity so there is less chance of intermarriage. As a result of the efforts by these and other Jewish groups over the past 25 years, there has been a trend (known as the Baal teshuva movement) for secular Jews to become more religiously observant, though the demographic implications of the trend are unknown. Additionally, there is also a growing rate of conversion to Jews by Choice of gentiles who make the decision to head in the direction of becoming Jews. Contributions Jewish individuals have played a significant role in the development and growth of Western culture, advancing many fields of thought, science and technology, both historically and in modern times, including through discrete trends in Jewish philosophy, Jewish ethics and Jewish literature, as well as specific trends in Jewish culture, including in Jewish art, Jewish music, Jewish humor, Jewish theatre, Jewish cuisine and Jewish medicine. Jews have established various Jewish political movements, religious movements, and, through the authorship of the Hebrew Bible and parts of the New Testament, provided the foundation for Christianity and Islam. More than 20 percent of the awarded Nobel Prize have gone to individuals of Jewish descent. Philanthropic giving is a widespread core function among Jewish organizations. Notes References External links |
======================================== |
[SOURCE: https://en.wikipedia.org/wiki/IDF] | [TOKENS: 13032] |
Contents Israel Defense Forces The Israel Defense Forces (IDF; Hebrew: צבא הגנה לישראל, romanized: Tsva Hagana le-Yisra'elⓘ, lit. 'Army for the Defense of Israel'), alternatively referred to by the Hebrew-language acronym Tzahal (צה״ל), is the national military of the State of Israel. It consists of three service branches: the Israeli Ground Forces, the Israeli Air Force, and the Israeli Navy. It is the sole military wing of the Israeli security apparatus. The IDF is headed by the chief of the general staff, who is subordinate to the defense minister. On the orders of first prime minister David Ben-Gurion, the IDF was formed on 26 May 1948 and began to operate as a conscript military, drawing its initial recruits from the already existing paramilitaries of the Yishuv—namely Haganah, the Irgun, and Lehi. It was formed shortly after the Israeli Declaration of Independence and has participated in every armed conflict involving Israel. In the wake of the 1979 Egypt–Israel peace treaty and the 1994 Israel–Jordan peace treaty, the IDF underwent a significant strategic realignment. Previously spread across various fronts—Lebanon and Syria in the north, Jordan and Iraq in the east, and Egypt in the south—the IDF redirected its focus towards southern Lebanon and the Palestinian territories. In 2000, the IDF withdrew from Southern Lebanon and in 2005 from Gaza. Conflict between Israel and Islamist groups based in Gaza, notably Hamas, has continued since then. Moreover, notable Israeli–Syrian border incidents have occurred frequently since 2011, due to regional instability caused by the Syrian civil war. Since 1967, the IDF has maintained a close security relationship with the United States, including in research and development cooperation, with joint efforts on the F-15I and the Arrow defence system, among others. The IDF is believed to have maintained an operational nuclear weapons capability since 1967, possibly possessing between 80 and 400 nuclear warheads. The IDF's actions and policies in the Palestinian territories have faced widespread criticism, with accusations of repression, institutionalized discrimination, unlawful killings and systematic abuses of Palestinian rights, with multiple human rights organizations and scholars accusing the IDF of genocide. Etymology The Israeli cabinet ratified the name "Israel Defense Forces" (Hebrew: צְבָא הַהֲגָנָה לְיִשְׂרָאֵל), Tzva haHagana leYisra'el, literally "the army for the defence of Israel", on 26 May 1948. The other main contender was Tzva Yisra'el (Hebrew: צְבָא יִשְׂרָאֵל). The name was chosen because it conveyed the idea that the army's role was defence and incorporated the name Haganah, the pre-state defensive organization upon which the new army was based. Among the primary opponents of the name were Minister Haim-Moshe Shapira and the Hatzohar party, both in favor of Tzva Yisra'el. In Israel's basic law about the military, the Hebrew name of the IDF is צבא הגנה לישראל, Tzva Hagana LeYisra'el, without a definite article on the second word. History The IDF traces its roots to Jewish paramilitary organizations in the New Yishuv, starting with the Second Aliyah (1904 to 1914). There had been several such organizations, or in part even older date, such as the "Mahane Yehuda" mounted guards company founded by Michael Halperin in 1891 (see Ness Ziona), HaMagen (1915–17), HaNoter (1912–13; see Zionism: Pre-state self-defense), and the much more consequential (but falsely-claimed "first" such organization), Bar-Giora, founded in September 1907. Bar-Giora was transformed into Hashomer in April 1909, which operated until the British Mandate of Palestine came into being in 1920. Hashomer was an elitist organization with a narrow scope and was mainly created to protect against criminal gangs seeking to steal property. The Zion Mule Corps and the Jewish Legion, both part of the British Army of World War I, further bolstered the Yishuv with military experience and manpower, forming the basis for later paramilitary forces. After the 1920 Palestine riots against Jews in April 1920, the Yishuv leadership realized the need for a nationwide underground defence organization, and the Haganah was founded in June 1920. The Haganah became a full-scale defence force after the 1936–1939 Arab revolt in Palestine with an organized structure, consisting of three main units—the Field Corps, Guard Corps, and the Palmach. During World War II, the Yishuv participated in the British war effort, culminating in the formation of the Jewish Brigade. These would eventually form the backbone of the Israel Defense Forces, and provide it with its initial manpower and doctrine. Following Israel's Declaration of Independence, Prime Minister and Defense Minister David Ben-Gurion issued an order for the formation of the Israel Defense Forces on 26 May 1948. Although Ben-Gurion had no legal authority to issue such an order, the order was made legal by the cabinet on 31 May. The same order called for the disbandment of all other Jewish armed forces. The two other Jewish underground organizations, Irgun and Lehi, agreed to join the IDF if they would be able to form independent units and agreed not to make independent arms purchases. This was the background for the Altalena Affair, a confrontation surrounding weapons purchased by the Irgun resulting in a standoff between Irgun members and the newly created IDF. The affair came to an end when Altalena, the ship carrying the arms, was shelled by the IDF. Following the affair, all independent Irgun and Lehi units were either disbanded or merged into the IDF. The Palmach, a leading component of the Haganah, also joined the IDF with provisions, and Ben Gurion responded by disbanding its staff in 1949, after which many senior Palmach officers retired, notably its first commander, Yitzhak Sadeh. The new army organized itself when the 1947–48 Civil War in Mandatory Palestine escalated into the 1948 Arab–Israeli War, which saw neighbouring Arab states attack. Twelve infantry and armoured brigades formed: Golani, Carmeli, Alexandroni, Kiryati, Givati, Etzioni, the 7th, and 8th armoured brigades, Oded, Harel, Yiftach, and the Negev. After the war, some of the brigades were converted to reserve units, and others were disbanded. Directorates and corps were created from corps and services in the Haganah, and this basic structure in the IDF still exists today. Immediately after the 1948 war, the Israel-Palestinian conflict shifted to a low-intensity conflict between the IDF and Palestinian fedayeen. In the 1956 Suez Crisis, the IDF's first serious test of strength after 1949, the new army captured the Sinai Peninsula from Egypt, which was later returned. In the 1967 Six-Day War, Israel conquered the Sinai Peninsula, Gaza Strip, West Bank (including East Jerusalem) and Golan Heights from the surrounding Arab states, changing the balance of power in the region as well as the role of the IDF. In the following years leading up to the Yom Kippur War, the IDF fought in the War of Attrition against Egypt in the Sinai and a border war against the Palestine Liberation Organization (PLO) in Jordan, culminating in the Battle of Karameh. The surprise of the Yom Kippur War and its aftermath completely changed the IDF's procedures and approach to warfare. Organizational changes were made and more time was dedicated to training for conventional warfare. However, in the following years the army's role slowly shifted again to low-intensity conflict, urban warfare and counter-terrorism. An example of the latter was the successful 1976 Operation Entebbe commando raid to free hijacked airline passengers being held captive in Uganda. During this era, the IDF also mounted a successful bombing mission in Iraq to destroy its nuclear reactor. It was involved in the Lebanese Civil War, initiating Operation Litani and later the 1982 Lebanon War, where the IDF ousted Palestinian guerrilla organizations from Lebanon. For twenty-five years the IDF maintained a security zone inside South Lebanon with their allies the South Lebanon Army. Palestinian militancy has been the main focus of the IDF ever since, especially during the First and Second Intifadas, Operation Defensive Shield, the Gaza War (2008–2009), the 2012 Gaza War, the 2014 Gaza War, and the 2021 Israel-Palestine crisis, causing the IDF to change many of its values and publish the IDF Code of Ethics. The Lebanese Shia organization Hezbollah has also been a growing threat, against which the IDF fought an asymmetric conflict between 1982 and 2000, as well as a full-scale war in 2006. The Israel Defense Forces have been accused of committing various war crimes since the founding of Israel in 1948. Israel ratified the Geneva Conventions on July 6, 1951, and on January 2, 2015, the State of Palestine acceded to the Rome Statute, granting the International Criminal Court (ICC) jurisdiction over war crimes committed in the Occupied Palestinian Territories (OPT). A 2017 report by Human Rights Watch accused the IDF of unlawful killings, using excessive force in policing situations, forced displacement, excessive use of detention and excessive restrictions on movement, as well as criticized the IDF's support and protection for Israeli settlements in the occupied Palestinian territory. Human rights experts argue that actions taken by the IDF during armed conflicts in the OPT fall under the rubric of war crimes. Various UN special rapporteurs, alongside human rights and aid organizations including Human Rights Watch, Médecins Sans Frontières, Amnesty International, have accused Israel of war crimes, and since 2024, genocide. Organization All branches of the IDF answer to a single General Staff. The Chief of the General Staff is the only serving officer having the rank of Lieutenant General (Rav Aluf). He reports directly to the Defense Minister and indirectly to the Prime Minister of Israel and the cabinet. Chiefs of Staff are formally appointed by the cabinet, based on the Defense Minister's recommendation, for three years. The government can vote to extend their service to four, and on rare occasions even five years. The current chief of staff is Eyal Zamir. The IDF includes the following bodies. Those whose respective heads are members of the General Staff are in bold: Ground Arm Air and Space Arm Sea Arm General Staff The IDF operates, as of 2024, a Multi-Domain Joint Operations Array, dealing with multi-domain operations, a joint warfare branch. Military: Civilian: Ranks, uniforms and insignia Unlike most militaries, the IDF uses the same rank names in all corps, including the air force and navy. From the formation of the IDF until the late 1980s, sergeant major was a particularly important warrant officer rank, in line with usage in other armies. In the 1980s and 1990s the proliferating ranks of sergeant major became devalued, and now all professional non-commissioned officer ranks are a variation on sergeant major (rav samal) except for rav nagad. All translations here are the official translations of the IDF's website. Conscripts (Hogrim) (Conscript ranks may be gained purely on time served) Warrant Officers (Nagadim) Academic officers (Ktzinim Akadema'im) Officers (Ktzinim) The Israel Defense Forces has several types of uniforms: The first two resemble each other but the Madei Alef is made of higher quality materials in a golden olive while the madei bet is in olive drab. The dress uniforms may also exhibit a surface shine The service uniform for all ground forces personnel is olive green; navy and air force uniforms are beige/tan (also once worn by the ground forces). The uniforms consist of a two-pocket shirt, combat trousers, sweater, jacket or blouse, and shoes or boots. The navy also has an all-white dress uniform. The green fatigues are the same for winter and summer and heavy winter gear is issued as needed. Women's dress parallels the men's but may substitute a skirt for trousers and a blouse for a shirt. Headgear included a service cap for dress and semi-dress and a field cap or "Kova raful" bush hat worn with fatigues. Many IDF personnel once wore the tembel as a field hat. IDF personnel generally wear berets instead of the service cap and there are many beret colours issued to IDF personnel. Paratroopers are issued a maroon beret, Golani brown, Givati purple, Nahal lime green, Kfir camouflage, Combat Engineers grey, navy blue for IDF Naval and dark grey for IDF Air Force personnel. In combat uniforms, the Orlite helmet has replaced the British Brodie helmet Mark II/Mark III, RAC Mk II modified helmet with chin web jump harness (used by paratroopers and similar to the HSAT Mk II/Mk III paratrooper helmets), US M1 helmet, and French Modèle 1951 helmet – previously worn by Israeli infantry and airborne troops from the late 1940s to the mid-1970s and early 1980s. Some corps or units have small variations in their uniforms – for instance, military police wear a white belt and police hat, Naval personnel have dress whites for parades, paratroopers are issued a four pocket tunic (yarkit/yerkit) worn untucked with a pistol belt cinched tight around the waist over the shirt. The IDF Air Corps has a dress uniform consisting of a pale blue shirt with dark blue trousers. Most IDF soldiers are issued black leather combat boots, certain units issue reddish-brown leather boots for historical reasons — the paratroopers, combat medics, Nahal and Kfir Brigades, as well as some Special Forces units (Sayeret Matkal, Oketz, Duvdevan, Maglan, and the Counter-Terror School). Women were also formerly issued sandals, but this practice has ceased. IDF soldiers have three types of insignia (other than rank insignia) which identify their corps, specific unit, and position. A pin attached to the beret identifies a soldier's corps. Individual units are identified by a shoulder tag attached to the left shoulder strap. The position/job of a soldier can then be identified by an aiguillette attached to the left shoulder strap and shirt pocket, and a pin indicating the soldier's work type. Service The military service is held in three different tracks: Sometimes the IDF would also hold pre-military courses (קורס קדם צבאי or קד"צ) for soon-to-be regular service soldiers. The Israeli Manpower Directorate (Hebrew: אגף משאבי אנוש) at the Israeli General Staff is the body which coordinates and assembles activities related to the control over human resources and its placement. National military service is mandatory for all Israeli citizens over the age of 18, although Arab (but not Druze) citizens are exempted if they so please, and other exceptions may be made on religious, physical or psychological grounds (see Profile 21). The Tal law exempted ultra-Orthodox Jews from service. In June 2024, Israel's Supreme Court unanimously ruled that Haredi Jews were eligible for compulsory service, ending nearly eight decades of exemption. The army began drafting Haredi men the following month. Until the draft of July 2015, men served three years in the IDF. Men drafted since July 2015 serve two years and eight months (32 months), with some roles requiring an additional four months of Permanent service. Women serve two years. The IDF women who volunteer for several combat positions often serve for three years, due to the longer period of training. Women in other positions, such as programmers, who also require lengthy training time, may also serve three years. Many Religious Zionist men (and many Modern Orthodox who make Aliyah) elect to do Hesder, a five-year program envisioned by Rabbi Yehuda Amital which combines Torah learning and military service. Some distinguished soldiers are selected to be trained to eventually become members of special forces units. Every brigade in the IDF has its special force branch. Career soldiers are paid on average NIS 23,000 a month. Conscripts are paid according to four tiers based on their positions. Frontline soldiers receive NIS 3,048 a month, other combat soldiers receive NIS 2,463, combat support soldiers receive NIS 1,793, and administrative soldiers receive NIS 1,235. In 1998–2000, only about 9% of those who refused to serve in the Israeli military were granted an exemption. In early 2026 it was made public that the IDF was considering expanding recruitment outreach to Jewish communities abroad in response to a manpower shortfall estimated at 10,000–12,000 soldiers. Reports state that the initiative would target young diaspora Jews, encouraging them to immigrate to Israel and enlist, with a focus on communities in North America and Europe. The move is described as a voluntary measure aimed at easing personnel shortages rather than altering Israel's core conscription system. Permanent service is designed for soldiers who choose to continue serving in the army after their regular service, for a short or long period, and in many cases making the military their career. Permanent service is based on a contractual agreement between the IDF and the permanent position holder. After personnel complete their regular service, they are either granted permanent exemption from military service or assigned a position in the reserve forces. No distinction is made between the assignment of men and women to reserve service. The IDF may call up reservists for: All Israelis who served in the IDF and are under the age of 40, unless otherwise exempt, are eligible for reserve duty. Only those who completed at least 20 days of reserve duty within the past three years are considered active reservists. Other than the civil, i.e. non-military "National Service" (Sherut Leumi), IDF conscripts may serve in bodies other than the IDF in several ways. The combat option is Israel Border Police (Magav – the exact translation from Hebrew means "border guard") service, part of the Israel Police. Some soldiers complete their IDF combat training and later undergo additional counter terror and Border Police training. These are assigned to Border Police units. The Border Police units fight side by side with the regular IDF combat units though to a lower capacity. They are also responsible for security in heavy urban areas such as Jerusalem and security and crime fighting in rural areas. Non-combat services include the Mandatrory Police Service [he] (Shaham, שח"מ) program, where youth serve in the Israeli Police, Israel Prison Service, or other wings of the Israeli Security Forces instead of the regular army service. Israel is one of only a few nations that conscript women or deploy them in combat roles, although in practice, women can avoid conscription through a religious exemption and over a third of Israeli women do so. As of 2010, 88% of all roles in the IDF are open to female candidates, and women could be found in 69% of all IDF positions. According to the IDF, 535 female Israeli soldiers were killed during service in the period 1962–2016, and dozens before then. The IDF says that fewer than 4 percent of women are in combat positions. Rather, they are concentrated in "combat-support" positions which command a lower compensation and status than combat positions. Civilian pilot and aeronautical engineer Alice Miller successfully petitioned the High Court of Justice to take the Israeli Air Force pilot training exams, after being rejected on grounds of gender. Though president Ezer Weizman, a former IAF commander, told Miller that she would be better off staying home and darning socks, the court eventually ruled in 1996 that the IAF could not exclude qualified women from pilot training. Even though Miller would not pass the exams, the ruling was a watershed, opening doors for women in new IDF roles. Female legislators took advantage of the momentum to draft a bill allowing women to volunteer for any position if they could qualify. In 2000, the Equality Amendment to the Military Service law stated that the right of women to serve in any role in the IDF is equal to the right of men. Women have served in the military since before the founding of the state of Israel in 1948. Women started to enter combat support and light combat roles in a few areas, including the Artillery Corps, infantry units and armoured divisions. A few platoons named Karakal were formed for men and women to serve together in light infantry. By 2000, Karakal became a full-fledged battalion, with a second mixed-gender battalion, Lions of the Jordan (אריות הירדן, Arayot Ha-Yarden) formed in 2015. Many women also joined the Border Police. In June 2011, Maj. General Orna Barbivai became the first female major general in the IDF, replacing the head of the directorate Maj. General Avi Zamir. Barbivai stated, "I am proud to be the first woman to become a major general and to be part of an organization in which equality is a central principle. Ninety percent of jobs in the IDF are open to women and I am sure that other women will continue to break down barriers." In 2013, the IDF announced they would, for the first time, allow a transgender woman to serve in the army as a female soldier. Elana Sztokman notes it would be "difficult to claim that women are equals in the IDF". "And tellingly, there is only one female general in the entire IDF," she adds. In 2012, religious soldiers claimed they were promised they would not have to listen to women sing or lecture, but IAF Chief Rabbi Moshe Raved resigned because male religious soldiers were being required to do so. In January 2015, three women IDF singers performed in one of the IDF's units. The performance was first disrupted by fifteen religious soldiers, who left in protest and then the Master Sergeant forced the women to end the performance because it was disturbing the religious soldiers. An IDF spokesperson announced an investigation of the incident: "We are aware of the incident and already begun examining it. The exclusion of women is not consistent with the values of the IDF." Defense Minister Moshe Ya'alon has also arranged for women to be excluded from recruitment centres catering to religious males. As the IDF recruits more religious soldiers, the rights of male religious soldiers and women in the IDF come into conflict. Brig. Gen. Zeev Lehrer, who served on the chief of staff's panel of the integration of women, noted "There is a clear process of 'religionization' in the army, and the story of the women is a central piece of it. There are very strong pressures at work to halt the process of integrating women into the army, and they are coming from the direction of religion." Sex segregation is allowed in the IDF, which reached what it considers a "new milestone" in 2006, creating the first company of soldiers segregated in an all-female unit, the Nachshol (Hebrew for "giant wave") Reconnaissance Company. "We are the only unit in the world made up entirely of female combat soldiers," said Nachshol Company Commander Cpt. Dana Ben-Ezra. "Our effectiveness and the dividends we earn are the factors by which we are measured, not our gender." With the rise of social media platforms such as TikTok and Twitter, some critics claim that women in the IDF are frequently used as tools of propaganda, with official military accounts frequently posting attractive young women to create a sympathetic social media presence. Non-Jewish minorities tended to serve in one of several special units: the Sword Battalion, also known as Unit 300 or the Minorities Unit, until it was disbanded in 2015; the Druze Reconnaissance Unit; and the Trackers Unit, composed mostly of Negev Bedouins. In 1982, the IDF general staff decided to integrate the armed forces by opening up other units to minorities, while placing some Jewish conscripts in the Minorities Unit. Until 1988, the Intelligence Corps and the Air Force remained closed to minorities. Although Israel has a majority of Jewish soldiers, all citizens including large numbers of Druze and Circassian men are subject to mandatory conscription. Originally, they served in the framework of a special unit called "The Minorities' Unit", which operated until 2015 in the form of the independent Herev Gdud ("Sword") battalion. However, since the 1980s Druze soldiers have increasingly protested this practice, which they considered a means of segregating them and denying them access to elite units (like sayeret units). The army has increasingly admitted Druze soldiers to regular combat units and promoted them to higher ranks from which they had been previously excluded. In 2015, Rav Aluf Gadi Eizenkot ordered the unit's closure to assimilate the Druze soldiers no differently than Jewish soldiers, as part of an ongoing reorganization of the army. Several Druze officers reached ranks as high as Major General, and many received commendations for distinguished service. In proportion to their numbers, the Druze people achieve much higher—documented—levels in the Israeli army than other soldiers. Nevertheless, some Druze still charge that discrimination continues, such as exclusion from the Air Force, although the official low-security classification for Druze has been abolished for some time. The first Druze aircraft navigator completed his training course in 2005. Like all Air Force pilots, his identity is not disclosed. During the 1948 Arab–Israeli War, many Druze who had initially sided with the Arabs deserted their ranks to either return to their villages or side with Israel in various capacities. Since the late 1970s, the Druze Initiative Committee, centred at the village of Beit Jan and linked to Maki, has campaigned to abolish Druze conscription. Military service is a tradition among some of the Druze population, with most opposition in Druze communities of the Golan Heights. 83 per cent of Druze boys serve in the army, according to the IDF's statistics. According to the Israeli army in 2010, 369 Druze soldiers had been killed in combat operations since 1948. By law, all Israeli citizens are subject to conscription. The Defense Minister has complete discretion to grant exemption to individual citizens or classes of citizens. A long-standing policy dating to Israel's early years extends an exemption to all other Israeli minorities (most notably Israeli Arabs). However, there is a long-standing government policy of encouraging Bedouins to volunteer and of offering them various inducements, and in some impoverished Bedouin communities a military career seems one of the few means of (relative) social mobility available. Also, Muslims and Christians are accepted as volunteers, even if older than 18. From among non-Bedouin Arab citizens, the number of volunteers for military service—some Christian Arabs and even a few Muslim Arabs—is minute, and the government makes no special effort to increase it. Six Israeli Arabs have received orders of distinction as a result of their military service; of them the most famous is a Bedouin officer, Lieutenant Colonel Abd el-Majid Hidr (also known as Amos Yarkoni), who received the Order of Distinction. Vahid el Huzil was the first Bedouin to be a battalion commander. Until the second term of Yitzhak Rabin as Prime Minister (1992–1995), social benefits given to families in which at least one member (including a grandfather, uncle, or cousin) had served at some time in the armed forces were significantly higher than to "non-military" families, which was considered a means of blatant discrimination between Jews and Arabs. Rabin led the abolition of the measure, in the teeth of strong opposition from the Right. At present, the only official advantage of military service is the attaining of security clearance and serving in some types of government positions (in most cases, security-related), as well as some indirect benefits. Rather than perform army service, Israeli Arab youths have the option to volunteer to national service and receive benefits similar to those received by discharged soldiers. The volunteers are generally allocated to Arab populations, where they assist with social and community matters. As of 2010[update], 1,473 Arabs were volunteering for national service. According to sources in the national service administration, Arab leaders are counselling youths to refrain from performing services to the state. According to a National Service official, "For years the Arab leadership has demanded, justifiably, benefits for Arab youths similar to those received by discharged soldiers. Now, when this opportunity is available, it is precisely these leaders who reject the state's call to come and do the service, and receive these benefits." Although Arabs are not obliged to serve in the IDF, any Arab can volunteer. In 2008, a Muslim Arab woman was serving as a medic with unit 669. Cpl. Elinor Joseph from Haifa became the first female Arab combat soldier for IDF. Other Arab-Muslim officers who have served in the IDF are Second Lieutenant Hisham Abu Varia and Major Ala Wahib, the highest ranking Muslim officer in the IDF in 2013. In October 2012, the IDF promoted Mona Abdo to become the first female Christian Arab to the rank of combat commander. Abdo had voluntarily enlisted in the IDF, which her family had encouraged, and transferred from the Ordnance Corps to the Caracal Battalion, a mixed-gender unit with both Jewish and Arab soldiers. In 2014, an increase in Israeli Christian Arabs joining the army was reported. Muslim Arabs have also been drafted into the Israel Defense Forces in increasing numbers in recent years. In 2020, 606 Muslim Arabs were drafted, compared to 489 in 2019 and 436 in 2018. More than half of those who have drafted have gone into combat roles. The IDF carried out extended missions in Ethiopia and neighbouring states, whose purpose was to protect Ethiopian Jews (Beta Israel) and to help their immigration to Israel. The IDF adopted policies and special activities for the absorption and integration of Ethiopian immigrant soldiers, reported to have much improved the achievements and integration of those soldiers in the army, and Israeli society in general. Statistical research showed that the Ethiopian soldiers are esteemed as excellent soldiers and many aspire to be recruited to combat units. Under a special arrangement called Torato Umanuto, Haredi men could choose to defer service while enrolled in yeshivot and many avoided conscription altogether. This has given rise to tensions between the Israeli religious and secular communities. In 2024, the Supreme Court unanimously ruled that Haredi men were eligible for compulsory service and the army began drafting them. Haredi males have the option of serving in the 97th "Netzah Yehuda" Infantry Battalion. This unit is a standard IDF infantry battalion focused on the Jenin region. To facilitate Haredi soldiers to serve, the Netzah Yehuda military bases follow the standards of Jewish dietary laws. The only women permitted on these bases were wives of soldiers and officers. Some Haredim serve in the IDF via the Hesder system, principally designed for the Religious Zionist sector. It is a 5-year program which includes 2 years of religious studies, 1½ years of military service and 1½ years of religious studies during which the soldiers can be recalled to active duty at any moment. Haredi soldiers may join other units of the IDF but rarely do. Although the IDF claims it will not discriminate against women, it is offering Haredim "women free and secular free" recruitment centres. Defense Minister Moshe Ya'alon expressed his willingness to relax regulations to meet the demands of ultra-Orthodox rabbis. Regulations regarding gender equality had already been relaxed so that Haredim could be assured that men would not receive physical exams from female medical staff. In 2024, the Hasmonean Brigade, a Haredi infantry unit, was created to accommodate Haredi soldiers. In January 2025, the unit received its first 50 soldiers. Since the beginning of the Israel-Hamas War, 600 Haredim have enlisted through the Shlav Bet program, an IDF recruitment track intended for Haredim over the age of 26 to complete military service. Since the early 1990s, sexual identity has presented no formal barrier in terms of soldiers' military specialization or eligibility for promotion. Until the 1980s, the IDF tended to discharge soldiers who were openly gay. In 1983, the IDF permitted homosexuals to serve but banned them from intelligence and top-secret positions. A decade later, professor Uzi Even, an IDF reserves officer and chairman of Tel Aviv University's Chemistry Department, revealed that his rank had been revoked and that he had been barred from researching sensitive topics in military intelligence, solely because of his sexual orientation. His testimony to the Knesset in 1993 raised a political storm, forcing the IDF to remove such restrictions against gays. The chief of staff's policy states that it is strictly forbidden to harm or hurt anyone's dignity or feeling based on their gender or sexual orientation in any way, including signs, slogans, pictures, poems, lectures, any means of guidance, propaganda, publishing, voicing, and utterance. Moreover, gays in the IDF have additional rights, such as the right to take a shower alone if they want to. According to a University of California, Santa Barbara study, a brigadier general stated that Israelis show a "great tolerance" for gay soldiers. Consul David Saranga at the Israeli Consulate in New York, who was interviewed by the St. Petersburg Times, said, "It's a non-issue. You can be a very good officer, a creative one, a brave one, and be gay at the same time." A study published by the Israel Gay Youth (IGY) Movement in January 2012 found that half of the homosexual soldiers who serve in the IDF suffer from violence and homophobia, although the head of the group said "I am happy to say that the intention among the top brass is to change that." Israel is the only country in the world that requires deaf and hard-of-hearing people to serve in the military. Sign language interpreters are provided during training, and many of them serve in non-combat capacities such as mapping and office work. The major language spoken by the deaf and hard-of-hearing in Israel is Israeli Sign Language (also called Shassi)–a language related to German Sign Language but not Hebrew or any other local language–though Israel and Palestine are home to numerous sign languages spoken by various populations like Bedouins' Al-Sayyid Bedouin Sign Language. According to a Care2 report, vegans in the IDF may refuse vaccination if they oppose animal testing. They are given artificial leather boots and a black fleece beret. Until 2014, vegan soldiers in the IDF received special allowances to buy their own food, when this policy was replaced with vegan food being provided in all bases, as well as vegan combat rations being offered to vegan combat soldiers. Non-immigrating foreign volunteers typically serve with the IDF in one of five ways: Doctrine In 1992, the IDF drafted a Code of Conduct that combines international law, Israeli law, Jewish heritage and the IDF's traditional ethical code—the IDF Spirit (Hebrew: רוח צה"ל, Ru'ah Tzahal). The document defines four core values for all IDF soldiers to follow, including "defense of the state, its citizens and its residents", "love of the homeland and loyalty to the country", "human dignity" and "stateliness, as well as ten secondary values. "The Spirit of the IDF" (cf. supra) is still considered the only binding moral code that formally applies to the IDF troops. In 2009, Amos Yadlin (then head of Military Intelligence) suggested that the article he co-authored with Asa Kasher be ratified as a formal binding code, arguing that "the current code ['The Spirit of the IDF'] does not sufficiently address one of the army's most pressing challenges: asymmetric warfare against terrorist organizations that operate amid a civilian population". Details of the IDF's rules of engagement remain classified. Targeted killing, targeted prevention or assassination is a tactic that has been repeatedly used by the IDF and other Israeli organisations in the course of the Israeli–Palestinian conflict, the Iran–Israel proxy conflict or other conflicts. In 2005, Asa Kasher and Amos Yadlin co-authored a noticed article published in the Journal of Military Ethics under the title: "Military Ethics of Fighting Terror: An Israeli Perspective". The article was meant as an "extension of the classical Just War Theory", and as a "[needed] third model" or missing paradigm besides which of "classical war (army) and law enforcement (police).", resulting in a "doctrine (...) on the background of the IDF fight against acts and activities of terror performed by Palestinian individuals and organizations." In this article, Kasher and Yadlin concluded that targeted killings of terrorists were justifiable, even at the cost of hitting nearby civilians. In a 2009 interview to Haaretz, Asa Kasher later confirmed, pointing to the fact that in an area in which the IDF does not have effective security control (e.g., Gaza, vs. East-Jerusalem), soldiers' lives protection takes priority over avoiding injury to enemy civilians. Some, along with Avishai Margalit and Michael Walzer, have disputed this argument, arguing that such a position was "contrary to centuries of theorizing about the morality of war as well as international humanitarian law", since drawing "a sharp line between combatants and noncombatants" would be "the only morally relevant distinction that all those involved in a war can agree on." The Hannibal Directive is a controversial procedure that the IDF has used to prevent the capture of Israeli soldiers by enemy forces. It was introduced in 1986, after some abductions of IDF soldiers in Lebanon and the subsequent controversial prisoner exchanges. The full text of the directive has never been published and until 2003 Israeli military censorship even forbade any discussion of the subject in the press. The directive has been changed several times. At one time the formulation was that "the kidnapping must be stopped by all means, even at the price of striking and harming our own forces." The Hannibal directive has, at times, apparently existed in two different versions, one top-secret written version, accessible only to the upper echelon of the IDF, and one "oral law" version for division commanders and lower levels. In the latter versions, "by all means" was often interpreted literally, as in "an IDF soldier was better dead than abducted". In 2011, IDF Chief of Staff Benny Gantz stated the directive does not permit the killing of IDF soldiers. The Dahiya doctrine is a military strategy of asymmetric warfare, outlined by former IDF Chief of General Staff Gadi Eizenkot, which encompasses the use of aerial and artillery fire against civilian infrastructure used by terrorist organizations and endorses the employment of "disproportionate power" to secure that end. The doctrine is named after the Dahieh neighborhood of Beirut, where Hezbollah was headquartered during the 2006 Lebanon War, which were heavily damaged by the IDF. Budget During 1950–66, Israel spent an average of 9% of its GDP on defense. Defence expenditures increased dramatically after both the 1967 and 1973 wars. They reached a high of about 30% of GDP in 1975, but have since come down significantly, following the signing of peace agreements with Jordan and Egypt. In September 2009, Defense Minister Ehud Barak, Finance Minister Yuval Steinitz and Prime Minister Benjamin Netanyahu endorsed an additional NIS 1.5 billion for the defense budget to help Israel address problems regarding Iran. The budget changes came two months after Israel had approved its two-year budget. The defence budget in 2009 stood at NIS 48.6 billion and NIS 53.2 billion for 2010 – the highest amount in Israel's history. The figure constituted 6.3% of expected gross domestic product and 15.1% of the overall budget, even before the planned NIS 1.5 billion addition. In 2011, the prime minister Benjamin Netanyahu reversed course and moved to make significant cuts in the defence budget to pay for social programs. The General Staff concluded that the proposed cuts endangered the battle readiness of the armed forces. In 2012, Israel spent $15.2 billion on its armed forces, one of the highest ratios of defense spending to GDP among developed countries ($1,900 per person). However, Israel's spending per capita is below that of the US. Equipment and weaponry The IDF possesses various foreign and domestically produced weapons and computer systems. Some gear comes from the US (with some equipment modified for IDF use) such as the M4A1 and M16 assault rifles, the M24 SWS 7.62 mm bolt action sniper rifle, the SR-25 7.62 mm semi-automatic sniper rifle, the F-15 Eagle and F-16 Fighting Falcon fighter jets, and the AH-1 Cobra and AH-64D Apache attack helicopters. Israel has also developed its own independent weapons industry, which has developed weapons and vehicles such as the Merkava battle tank series, Nesher and Kfir fighter aircraft, and various small arms such as the Galil and Tavor assault rifles, and the Uzi submachine gun. Israel has also installed a variant of the Samson RCWS, a remote controlled weapons platform, which can include machine guns, grenade launchers, and anti-tank missiles on a remotely operated turret, in pillboxes along the Gaza–Israel barrier intended to prevent Palestinian militants from entering its territory. Israel has developed observation balloons equipped with sophisticated cameras and surveillance systems used to thwart terror attacks from Gaza. The IDF also possesses advanced combat engineering equipment which includes the IDF Caterpillar D9 armoured bulldozer, IDF Puma CEV, Tzefa Shiryon and CARPET minefield breaching rockets, and a variety of robots and explosive devices. The IDF has several large internal research and development departments, and it purchases many technologies produced by the Israeli security industries including IAI, IMI, Elbit Systems, Rafael, and dozens of smaller firms. Many of these developments have been battle-tested in Israel's numerous military engagements, making the relationship mutually beneficial, the IDF getting tailor-made solutions and the industries a good reputation.[citation needed] In response to the price overruns on the US Littoral Combat Ship program, Israel is considering producing their own warships, which would take a decade and depend on diverting US financing to the project. Israel's military technology is famous for its firearms, armoured fighting vehicles (tanks, tank-converted armoured personnel carriers (APCs), armoured bulldozers, etc.), unmanned aerial vehicles, and rocketry (missiles and rockets). Israel also has manufactured aircraft including the Kfir (reserve), IAI Lavi (cancelled), and the IAI Phalcon Airborne early warning System, and naval systems (patrol and missile ships). Much of the IDF's electronic systems (intelligence, communication, command and control, navigation etc.) are Israeli-developed, including many systems installed on foreign platforms (esp. aircraft, tanks and submarines), as are many of its precision-guided munitions. Israel is the world's largest exporter of drones. Israel Military Industries (IMI), as well as its former subsidiary Israel Weapon Industries, are known for their firearms. The IMI Galil, the Uzi, the IMI Negev light machine gun and the new Tavor TAR-21 Bullpup assault rifle are used by the IDF. The Rafael Advanced Defense Systems Spike missile is one of the most widely exported ATGMs in the world. Israel's Arrow anti-ballistic missile system, jointly funded and produced by Israel and the United States, has successfully conducted combat intercepts of incoming ballistic missile attacks. The Iron Dome system against short-range rockets is operational and proved to be successful, intercepting hundreds of Qassam, 122 mm Grad and Fajr-5 artillery rockets fire by Palestinian militants from the Gaza Strip. David's Sling, an anti-missile system designed to counter medium range rockets, became operational in 2017. Israel has also worked with the US on the development of a tactical high energy laser system against medium-range rockets (called Nautilus or THEL). Iron Beam is a short-range laser beam air defence system, created to eliminate rockets, artillery, and mortar bombs. With a range of several kilometres, it complements the Iron Dome system, which is specifically designed for intercepting missiles launched from greater distances. Israel has the independent capability of launching reconnaissance satellites into orbit, a capability shared with Russia, the United States, the United Kingdom, France, South Korea, Italy, Germany, the People's Republic of China, India, Japan, Brazil and Ukraine. Israeli security industries developed both the satellites (Ofeq) and the launchers (Shavit). Israel is known to have developed nuclear weapons. Israel does not officially acknowledge its nuclear weapons program. It is thought Israel possesses between one hundred and four hundred nuclear warheads. It is believed that Jericho intercontinental ballistic missiles are capable of delivering nuclear warheads with a superior degree of accuracy and a range of 11,500 km. Israeli F-15I and F-16 fighter-bomber aircraft also have been cited as possible nuclear delivery systems (these aircraft types are nuclear capable in the US Air Force). The U.S. Air Force F-15E has tactical nuclear weapon (B61 and B83 bombs) capability. It has been asserted that Dolphin-class submarines have been adapted to carry Popeye Turbo Submarine-launched cruise missiles with nuclear warheads, so as to give Israel a second strike capacity. In 2006, Israel deployed the Wolf Armoured Vehicle APC for use in urban warfare and to protect VIPs. Field rations, called manot krav, usually consist of canned tuna, sardines, beans, stuffed vine leaves, maize and fruit cocktail and bars of halva. Packets of fruit-flavoured drink powder are provided along with condiments like ketchup, mustard, chocolate spread and jam. Around 2010, the IDF announced that certain freeze-dried MREs served in water-activated disposable heaters like goulash, turkey schwarma and meatballs would be introduced as field rations. One staple of these rations was loof, a type of Kosher spam made from chicken or beef that was phased out around 2008. Food historian Gil Marks has written that: "Many Israeli soldiers insist that Loof uses all the parts of the cow that the hot dog manufacturers will not accept, but no one outside of the manufacturer and the kosher supervisors know what is inside." The IDF is planning several technological upgrades and structural reforms for the future of its land, air, and sea branches. Training has been increased, including cooperation between ground, air, and naval units. The Israeli Army is phasing out the M-16 rifle from all ground units in favor of the IMI Tavor variants, most recently the IWI Tavor X95 flat-top ("Micro-Tavor Dor Gimel"). The IDF is replacing its outdated M113 armored personnel carriers in favor of new Namer APCs, with 200 ordered in 2014, the Eitan AFV, and is upgrading its IDF Achzarit APCs. The IDF announced plans to streamline its military bureaucracy so as to better maintain its reserve force, which a 2014 State Comptroller report noted was under-trained and may not be able to fulfill wartime missions. As part of the plans, 100,000 reservists will be discharged, and training for the remainder will be improved. The officer corps will be slashed by 5,000. Infantry and light artillery brigades will be reduced to increase training standards among the rest. The IDF is planning a future tank to replace the Merkava. The new tank will be able to fire lasers and electromagnetic pulses, run on a hybrid engine, run with a crew as small as two, will be faster, and will be better protected, with emphasis on protection systems such as the Trophy over armour. The Combat Engineering Corps assimilated new technologies, mainly in tunnel detection and unmanned ground vehicles and military robots, such as remote-controlled IDF Caterpillar D9T "Panda" armoured bulldozers, Sahar engineering scout robot and improved Remotec ANDROS robots. The Israeli Air Force will purchase as many as 100 F-35 Lightning II fighter jets from the United States. The aircraft will be modified and designated F-35I. They will use Israeli-built electronic warfare systems, outer wings, guided bombs, and air-to-air missiles. As part of a 2013 arms deal, the IAF will purchase KC-135 Stratotanker aerial refueling aircraft and V-22 Osprey multi-mission aircraft from the United States, as well as advanced radars for warplanes and missiles designed to take out radars. In April 2013, an Israeli official stated that within 40–50 years, piloted aircraft would be phased out of service by unmanned aerial vehicles capable of executing nearly any operation that can be performed by piloted combat aircraft. Israel's military industries are reportedly on the path to developing such technology in a few decades. Israel will also manufacture tactical satellites for military use. The Israeli Navy is currently expanding its submarine fleet, with a planned total of six Dolphin class submarines. Currently, five have been delivered, with the sixth, INS Drakon, expected to be delivered in 2020. It is also upgrading and expanding its surface fleet. It is planning to upgrade the electronic warfare systems of its Sa'ar 5-class corvettes and Sa'ar 4.5 class missile boats, and has ordered two new classes of warship: the Sa'ar 6-class corvette (a variant of the Braunschweig-class corvette) and the Sa'ar 72-class corvette, an improved and enlarged version of the Sa'ar 4.5-class. It plans to acquire four Saar 6-class corvettes and three Sa'ar 72-class corvettes. Israel is also developing marine artillery, including a gun capable of firing satellite-guided 155mm rounds between 75 and 120 kilometres. Foreign military relations Starting on 14 May 1948 (5 Iyar 5708), when Israel became a sovereign state, a strong military, commercial and political relationship was established between France and Israel, which lasted until 1969. Between 1956 and 1966, the two countries had the highest level of military collaboration. In 1983, the United States and Israel established a Joint Political Military Group, which convenes twice a year. Both the U.S. and Israel participate in joint military planning and combined exercises and have collaborated on military research and weapons development. Additionally the U.S. military maintains two classified, pre-positioned War Reserve Stocks in Israel valued at $493 million.[citation needed] Israel has the official distinction of being an American Major non-NATO ally. Since 1976, Israel had been the largest annual recipient of U.S. foreign assistance. In 2009, Israel received $2.55 billion in Foreign Military Financing (FMF) grants from the Department of Defense. All but 26% of this military aid is for the purchase of military hardware from American companies only. In October 2012, the United States and Israel began their biggest joint air and missile defence exercise, known as Austere Challenge 12, involving around 3,500 U.S. troops in the region along with 1,000 IDF personnel. Germany and Britain also participated. Since mid-2017, the United States has operated an anti-missile system in the Negev region of Southern Israel, which is manned by 120 US Army personnel. It is a facility used by the U.S. inside a larger Mashabim Israeli Air Force base. India and Israel enjoy strong military and strategic ties. Israeli authorities consider Indian citizens to be the most pro-Israel people in the world. Apart from being Israel's second-largest economic partner in Asia, India is also the largest customer of Israeli arms in the world. In 2006, annual military sales between India and Israel stood at US$900 million, totalling around US$4,200,000,000 (equivalent to $7,636,722,421 in 2025) for the past two decades of arms trade. Israeli defense firms had the largest exhibition at the 2009 Aero India show, during which Israel offered several state-of-the art weapons to India. The first major military deal between the two countries was the sale of Israeli Phalcon airborne warning and control system (AWACS) radars to the Indian Air Force in 2004. In March 2009, India and Israel signed a US$1.4 billion deal under which Israel would sell India an advanced air-defense system. India and Israel have also embarked on extensive space cooperation. In 2008, India's ISRO launched Israel's most technologically advanced spy satellite TecSAR. In 2009, India reportedly developed a high-tech spy satellite RISAT-2 with significant assistance from Israel. The satellite was successfully launched by India in April 2009. According to a Los Angeles Times news story, the 2008 Mumbai attacks were an attack on the growing India-Israel partnership. It quotes retired Indian Vice Admiral Premvir S. Das thus "Their aim was to... tell the Indians clearly that your growing linkage with Israel is not what you should be doing..." In the past, India and Israel have held numerous joint anti-terror training exercises Despite regional crisis and ongoing Gaza war, defence supply lines to India remained active in 2023-24, as political ties and defence buyers kept the relationship operational. Germany developed the Dolphin submarine and supplied it to Israel. Two submarines were donated by Germany. The military co-operation has been discreet but mutually profitable: Israeli intelligence, for example, sent captured Warsaw Pact armor to West Germany to be analyzed. The results aided the German development of an anti-tank system. Israel also trained members of GSG 9, a German counter-terrorism and special operations unit. The Israeli Merkava MK IV tank uses a German V12 engine produced under license. In 2008, the website DefenseNews revealed that Germany and Israel had been jointly developing a nuclear warning system, Operation Bluebird. The United Kingdom has supplied equipment and spare parts for Sa'ar 4.5-class missile boats and F-4 Phantom fighter-bombers, components for small-calibre artillery ammunition and air-to-surface missiles, and engines for Elbit Hermes 450 Unmanned aerial vehicles. British arms sales to Israel mainly consist of light weaponry, and ammunition and components for helicopters, tanks, armoured personnel carriers, and combat aircraft. On 19 October 1999, the Defense Minister of China, General Chi Haotian, after meeting with Syrian Defense Minister Mustafa Tlass in Damascus, Syria, to discuss expanding military ties between Syria and China, then flew directly to Israel and met with Ehud Barak, the then Prime Minister and Defense Minister of Israel where they discussed military relations. Among the military arrangements was a $1 billion Israeli–Russian sale of military aircraft to China, which were to be jointly produced by Russia and Israel. Russia has bought drones from Israel. Israel is the second-largest foreign supplier of arms to the People's Republic of China, only after the Russian Federation. China has purchased a wide array of military hardware from Israel, including Unmanned aerial vehicles and communications satellites. China has become an extensive market for Israel's military industries and arms manufacturers, and trade with Israel has allowed it to obtain "dual-use" technology which the United States and European Union were reluctant to provide. In 2010 Yair Golan, head of IDF Home Front Command visited China to strengthen military ties. In 2012, IDF Chief of Staff Benny Gantz visited China for high-level talks with the Chinese defence establishment. As closely neighbouring countries, Israel and Cyprus have enjoyed greatly improving diplomatic relations since 2010. During the Mount Carmel Forest Fire, Cyprus dispatched two aviation assets to assist fire-fighting operations in Israel – the first time Cypriot Government aircraft were permitted to operate from Israeli airfields in a non-civil capacity. Israel and Cyprus have closely cooperated in maritime activities relating to Gaza, since 2010, and have reportedly begun an extensive sharing program of regional intelligence to support mutual security concerns. In May 2012, it was widely reported that the Israeli Air Force had been granted unrestricted access to the Nicosia Flight Information Region of Cyprus and that Israeli aviation assets may have operated over the island itself. Since 2010, the Israeli and Greek air forces trained jointly in Greece, indicating a boost in ties due in large part to Israel's rift with Turkey. Recent purchases include 100 million euro deal between Greece and Israel for the purchase of SPICE 1000 and SPICE 2000 pound bomb kits. In November 2011, the Israeli Air Force hosted Greece's Hellenic Air Force in a joint exercise at the Uvda base. Similar training was held in 2012 by the IAF in cooperation with the Hellenic Air Force in the Peloponnese and parts of southern Greece in a response to the need of the IAF training of pilots in unfamiliar areas. On March 14, 2013, the navies of Israel, Greece and the US held a two-week joint military exercise for the third year in a row. The annual operation is nicknamed Noble Dina and was established in 2011. Similar to Noble Dina in 2012, the exercise in 2013 included defending offshore natural gas platforms and simulated air-to-air combat and anti-submarine warfare. In March 2017, Israel participated in the large-scale "Iniochus 2017" military exercise, which is organized annually by Greece, along with USA, Italy and the United Arab Emirates. Israel has provided extensive military assistance to Turkey. Israel sold Turkey IAI Heron Unmanned aerial vehicles, and modernized Turkey's F-4 Phantom and Northrop F-5 aircraft at the cost of $900 million. Turkey's main battle tank is the Israeli-made Sabra tank, of which Turkey has 170. Israel later upgraded them for $500 million. Israel has also supplied Turkey with Israeli-made missiles, and the two nations have engaged in naval cooperation. Turkey allowed Israeli pilots to practice long-range flying over mountainous terrain in Turkey's Konya firing range, while Israel trains Turkish pilots at Israel's computerized firing range at Nevatim Airbase. Until 2009, the Turkish military was one of Israel's largest defence customers. Israel defence companies have sold unmanned aerial vehicles and long-range targeting pods. However, relations have been strained in recent times. Since 2010 the Turkish military has declined to participate in the annual joint naval exercise with Israel and the United States. The exercise, known as "Reliant Mermaid" was started in 1998 and included the Israeli, Turkish and American navies. The objective of the exercise is to practice search-and-rescue operations and to familiarize each navy with international partners who also operate in the Mediterranean Sea. Azerbaijan and Israel have engaged in intense cooperation since 1992. Israeli military have been a major provider of battlefield aviation, artillery, antitank, and anti-infantry weaponry to Azerbaijan. In 2009, Israeli President Shimon Peres made a visit to Azerbaijan where military relations were expanded further, with the Israeli company Aeronautics Defense Systems Ltd announcing it was going to build a factory in Baku. In 2012, Israel and Azerbaijan signed an agreement according to which state-run Israel Aerospace Industries would sell $1.6 billion in drones and anti-aircraft and missile defense systems to Azerbaijan. In March 2012, the magazine Foreign Policy reported that the Israeli Air Force may be preparing to use the Sitalchay Military Airbase, located 500 km (310 mi) from the Iranian border, for air strikes against the nuclear program of Iran, later backed up by other media. Israel has also sold to or received supplies of military equipment from the Czech Republic, Argentina, Portugal, Spain, Slovakia, Italy, South Africa, Canada, Australia, Poland, Slovenia, Romania, Hungary, Belgium, Austria, Serbia, Montenegro, Bosnia and Herzegovina, Georgia, Vietnam and Colombia, among others. Commemoration Israel Defense Forces parades took place on Independence Day, during the first 25 years of the State of Israel's existence. They were cancelled after 1973 due to financial and security concerns. The Israel Defense Forces still have weapon exhibitions country-wide on Independence Day, but they are stationary. Yom Hazikaron, Israel's day of remembrance for fallen soldiers, is observed on the 4th day of the month of Iyar of the Hebrew calendar, the day before the celebration of Independence Day. The main museum for Israel's armoured corps is the Yad La-Shiryon in Latrun, which houses one of the largest tank museums in the world. Other significant military museums are the Israel Defense Forces History Museum (Batei Ha-Osef) in Tel Aviv, the Palmach Museum, and the Beit HaTotchan of artillery in Zikhron Ya'akov. The Israeli Air Force Museum is located at Hatzerim Airbase in the Negev Desert, and the Israeli Clandestine Immigration and Naval Museum, is in Haifa. Israel's National Military Cemetery is at Mount Herzl. Other Israeli military cemeteries include Kiryat Shaul Military Cemetery in Tel Aviv, and Sgula military cemetery at Petah Tikva. See also References and footnotes Further reading External links |
======================================== |
[SOURCE: https://en.wikipedia.org/wiki/Default_route] | [TOKENS: 393] |
Contents Default route In computer networking, the default route is a configuration of the Internet Protocol (IP) that establishes a forwarding rule for packets when no specific address of a next-hop host is available from the routing table or other routing mechanisms. The default route is generally the address of another router, which treats the packet the same way: if a route matches, the packet is forwarded accordingly, otherwise the packet is forwarded to the default route of that router. The route evaluation process in each router uses the longest prefix match method to obtain the most specific route. The network with the longest subnet mask or network prefix that matches the destination IP address is the next-hop network gateway. The process repeats until a packet is delivered to the destination host, or earlier along the route, when a router has no default route available and cannot route the packet otherwise. In the latter case, the packet is dropped and an ICMP Destination Unreachable message may be returned. Each router traversal counts as one hop in the distance calculation for the transmission path. The device to which the default route points is often called the default gateway, and it often carries out other functions such as packet filtering, firewalling, or proxy server operations. The default route in Internet Protocol Version 4 (IPv4) is designated as the zero address, 0.0.0.0/0 in CIDR notation. Similarly, in IPv6, the default route is specified by ::/0. The subnet mask is specified as /0, which effectively specifies all networks and is the shortest match possible. A route lookup that does not match any other rule falls back to this route. In the highest-level segment of a network, administrators generally point the default route for a given host towards the router that has a connection to a network service provider. Therefore, packets with destinations outside the organization's LAN, typically destinations on the Internet or a wide area network, are forwarded to the router with the connection to that provider. References |
======================================== |
[SOURCE: https://en.wikipedia.org/wiki/Federal_lands] | [TOKENS: 2074] |
Contents Federal lands Federal lands are lands in the United States owned and managed by the federal government. Pursuant to the Property Clause of the United States Constitution (Article 4, section 3, clause 2), Congress has the power to retain, buy, sell, and regulate federal lands. These powers have been recognized in a long series of United States Supreme Court decisions. In Article I, Section 8, Clause 17 the United States Constitution empowers the federal government with exclusive legislative authority like that exercised for Washington D.C. over "Places purchased by the Consent of the Legislature of the State in which the same shall be, for the erection of Forts, Magazines, Arsenals, Dock-yards, and other needful Buildings." The federal government manages about 640 million acres (2.6 million km2) of land in the United States, which is about 28% of the total land area of 2.27 billion acres (9.2 million km2). The majority of federal lands (610.1 million acres (2.469 million km2) or 95 percent area in 2015) are administered by the Bureau of Land Management (BLM), United States Fish and Wildlife Service (FWS), National Park Service (NPS), or United States Forest Service (USFS). BLM, FWS, and NPS are part of the United States Department of the Interior, while the Forest Service is part of the United States Department of Agriculture. An additional 11.4 million acres (46 thousand km2) of land (about 2% of all federal land) is managed by the United States Department of Defense (DOD). The majority of federal lands are located in Alaska and the Western states. Legal background The United States Supreme Court has upheld the broad powers of the federal government to deal with federal lands, for example having unanimously held in Kleppe v. New Mexico that "the complete power that Congress has over federal lands under this clause necessarily includes the power to regulate and protect wildlife living there, state law notwithstanding." Lands held by the United States in trust for Native American tribes are generally not considered public lands. There are some 55 million acres (0.22 million km2) of land held in trust by the federal government for Indian tribes and almost 11 million acres (45 thousand km2) of land held in trust by the federal government for individual Natives. Although the United States holds legal title to these lands, the tribe or individual holds beneficial title (the right to use and benefit from the property). As a result, Indian Country is "quasi-private, not public, land." Nevertheless, "because the United States is a legal title holder, the federal government is a necessary part in all leases and dispositions of resources including trust land. For example, the secretary of the interior must approve any contract for payment or grant by an Indian tribe for services for the tribe 'relative to their lands' (25 U.S.C. § 81)." History of federal lands The Land Ordinance of 1785 and the Northwest Ordinance of 1787 provided for the survey and settlement of the lands that the original Thirteen Colonies ceded to the federal government after the American Revolution. As additional lands were acquired by the United States from Spain, France, Native American Nations and other countries, the United States Congress directed that they be explored, surveyed, and made available for settlement. During the Revolutionary War, military bounty land was promised to soldiers who fought for the colonies. After the war, the Treaty of Paris of 1783, signed by the United States, the Kingdom of Great Britain, France, and Spain, ceded territory to the United States. In the 1780s, other states relinquished their own claims to land in modern-day Ohio. By this time, the United States needed revenue to function. Land was sold so that the government would have money to survive. In order to sell the land, surveys needed to be conducted. The Land Ordinance of 1785 instructed a geographer to oversee this work as undertaken by a group of surveyors. The first years of surveying were completed by trial and error; once the territory of Ohio had been surveyed, a modern public land survey system had been developed. In 1812, Congress established the United States General Land Office as part of the Department of the Treasury to oversee the disposition of these federal lands. By the early 1800s, promised bounty land claims were finally fulfilled. In the 19th century, other bounty land and homestead laws were enacted to dispose of federal land. These included, among others, the Homestead Act of 1862 and the Desert Lands Entry Act of 1877. Several different types of patents existed. These include cash entry, credit, homestead, Indian, military warrants, mineral certificates, private land claims, railroads, state selections, swamps, town sites, and town lots. A system of local land offices spread throughout the territories, patenting land that was surveyed via the corresponding Office of the Surveyor General of a particular territory. This pattern gradually spread across the entire United States. Homestead entries peaked in 1910, when they amounted to 18.3 million acres (0.074 million km2), and sharply declined after 1935 and were eliminated in 1986. The laws that spurred mass federal land transfers, with the exception of the General Mining Law of 1872 and the Desert Land Act of 1877, have since been repealed or superseded. Between 1781 and 2018, the federal government divested itself of estimated 1.29 billion acres (5.2 million km2) of public domain land. The vast majority (97%) of transfers of federal land to private ownership occurred before 1940. Beginning in the early 20th century, U.S. government policy shifted from disposing of public land to retaining and managing it. Congress took additional steps toward recognizing the value of the assets on public lands and directed the Executive Branch to manage activities on the remaining public lands. The Mineral Leasing Act of 1920 allowed leasing, exploration, and production of selected commodities, such as coal, oil, gas, and sodium to take place on public lands. The Taylor Grazing Act of 1934 established the United States Grazing Service to manage the public rangelands by establishment of advisory boards that set grazing fees. The Oregon and California Revested Lands Sustained Yield Management Act of 1937, commonly referred as the O&C Act, required sustained yield management of the timberlands in western Oregon. The Sagebrush Rebellion movement in the Western United States in the 1970s and the 1980s sought major changes to federal land control, use, and disposal policy in 13 western states in which federal land holdings include between 20% and 85% of a state's area. Supporters of the movement wanted more state and local control over the lands, if not outright transfer of them to state and local authorities and/or privatization. From 1990 to 2018, the overall acreage held by the federal government decreased by 4.9% (i.e., from 646.9 million acres (2.618 million km2) to 615.3 million acres (2.490 million km2)). Over that time period, the federal acreage held by the Bureau of Land Management and Department of Defense decreased by 10.2% and 56.8%, respectively, and the federal acreage held by the Forest Service, Fish and Wildlife Service, and National Park Service increased by 0.8%, 2.7%, and 5.0%, respectively. Over the 1990–2018 time period, the largest decline in federal acreage was in Alaska (a decrease of 9.4%, or 23.0 million acres (0.093 million km2)) and in the 11 contiguous states of the West (a 3% decrease in federal land, or 10.7 million acres (0.043 million km2)). Primary federal land holders The four primary federal land holders are: The fifth largest federal landowner is the United States Department of Defense, which owns, leases, or possessed 26.1 million acres (0.106 million km2) worldwide, of which 8.8 million acres (0.036 million km2) are located in the United States (this figure excluded United States Army Corps of Engineers land). DoD thus administers approximately 1% of federal land. DOD land is mostly military bases and reservations. The largest single DOD-owned, all-land tract is the 2.3-million-acre White Sands Missile Range in New Mexico. Together, the BLM, FWS, NPS, Forest Service, and DOD manage about 96% of federal land. The remaining 4% of federal land is controlled by other federal agencies, including the United States Army Corps of Engineers, the Bureau of Reclamation, the United States Postal Service, the National Aeronautics and Space Administration, and the U.S. Department of Energy. Distribution Federal land is concentrated in the Western United States. Nationwide, the federal government owns 27.4% of all land area. There are significant variations regionally; the federal government owns 61.3% of the land area in Alaska, 46.4% of the land area in the 11 contiguous Western states; and 4.2% of the land area of other states. The state with the highest percentage of land held by the federal government is Nevada (80.1%); the states with the lowest percentage of land held by the federal government are Connecticut and Iowa (0.3%). Primary laws regarding federal lands Acquired lands In the United States, acquired lands refer to a category of public lands under federal management that were obtained by the federal government through purchase, condemnation, gift, or exchange. Uses In the United States, federal lands are often used for environmental conservation. The administration of Joe Biden turned into protected areas 12.5 million acres in 2023 alone. See also References Further reading |
======================================== |
[SOURCE: https://en.wikipedia.org/wiki/URI_scheme] | [TOKENS: 2817] |
Contents Uniform Resource Identifier A Uniform Resource Identifier (URI), formerly Universal Resource Identifier, is a unique sequence of characters that identifies an abstract or physical resource,: 1 such as resources on a webpage, email address, phone number,: 7 books, real-world objects such as people and places, and concepts.: 5 URIs which provide a means of locating and retrieving information resources on a network (either on the Internet or on another private network, such as a computer file system or an Intranet) are Uniform Resource Locators (URLs). Therefore, URLs are a subset of URIs, i.e. every URL is a URI (and not necessarily the other way around).: 7 Other URIs provide only a unique name, without a means of locating or retrieving the resource or information about it; these are Uniform Resource Names (URNs). The web technologies that use URIs are not limited to web browsers. History URIs and URLs have a shared history. In 1990, Tim Berners-Lee's proposals for hypertext implicitly introduced the idea of a URL as a short string representing a resource that is the target of a hyperlink. At the time, people referred to it as a "hypertext name" or "document name". Over the next three and a half years, as the World Wide Web's core technologies of HTML, HTTP, and web browsers developed, a need to distinguish a string that provided an address for a resource from a string that merely named a resource emerged. Although not yet formally defined, the term Uniform Resource Locator came to represent the former, and the more contentious Uniform Resource Name came to represent the latter. In July 1992 Berners-Lee's report on the Internet Engineering Task Force (IETF) "UDI (Universal Document Identifiers) BOF" mentions URLs (as Uniform Resource Locators), URNs (originally, as Unique Resource Numbers), and the need to charter a new working group. In November 1992 the IETF "URI Working Group" met for the first time. During the debate over defining URLs and URNs, it became evident that the concepts embodied by the two terms were merely aspects of the fundamental, overarching, notion of resource identification. In June 1994, the IETF published RFC 1630, Berners-Lee's first Request for Comments that acknowledged the existence of URLs and URNs. Most importantly, it defined a formal syntax for Universal Resource Identifiers (i.e. URL-like strings whose precise syntaxes and semantics depended on their schemes). It also attempted to summarize the syntaxes of URL schemes in use at the time. It acknowledged – but did not standardize—the existence of relative URLs and fragment identifiers. In December 1994, RFC 1738 formally defined relative and absolute URLs, refined the general URL syntax, defined how to resolve relative URLs to absolute form, and better enumerated the URL schemes then in use. The agreed definition and syntax of URNs had to wait until the publication of IETF RFC 2141 in May 1997. The publication of IETF RFC 2396 in August 1998 saw the URI syntax become a separate specification and most of the parts of RFCs 1630 and 1738 relating to URIs and URLs in general were revised and expanded by the IETF. The new RFC changed the meaning of U in URI from "Universal" to "Uniform." In December 1999, RFC 2732 provided a minor update to RFC 2396, allowing URIs to accommodate IPv6 addresses. A number of shortcomings discovered in the two specifications led to a community effort, coordinated by RFC 2396 co-author Roy Fielding, that culminated in the publication of IETF RFC 3986 in January 2005. While obsoleting the prior standard, it did not render the details of existing URL schemes obsolete; RFC 1738 continues to govern such schemes except where otherwise superseded. IETF RFC 2616 for example, refines the http scheme. Simultaneously, the IETF published the content of RFC 3986 as the full standard STD 66, reflecting the establishment of the URI generic syntax as an official Internet protocol. In 2001, the World Wide Web Consortium's (W3C) Technical Architecture Group (TAG) published a guide to best practices and canonical URIs for publishing multiple versions of a given resource. For example, content might differ by language or by size to adjust for capacity or settings of the device used to access that content. In August 2002, IETF RFC 3305 pointed out that the term "URL" had, despite widespread public use, faded into near obsolescence, and serves only as a reminder that some URIs act as addresses by having schemes implying network accessibility, regardless of any such actual use. As URI-based standards such as Resource Description Framework make evident, resource identification need not suggest the retrieval of resource representations over the Internet, nor need they imply network-based resources at all. The Semantic Web uses the HTTP URI scheme to identify both documents and concepts for practical uses, a distinction which has caused confusion as to how to distinguish the two. The TAG published an e-mail in 2005 with a solution of the problem, which became known as the httpRange-14 resolution. The W3C subsequently published an Interest Group Note titled "Cool URIs for the Semantic Web", which explained the use of content negotiation and the HTTP 303 response code for redirections in more detail. Design A Uniform Resource Name (URN) is a URI that identifies a resource by name in a particular namespace. A URN may be used to talk about a resource without implying its location or how to access it. For example, in the International Standard Book Number (ISBN) system, ISBN 0-486-27557-4 identifies a specific edition of the William Shakespeare play Romeo and Juliet. The URN for that edition would be urn:isbn:0-486-27557-4. However, it gives no information as to where to find a copy of that book. A Uniform Resource Locator (URL) is a URI that specifies the means of acting upon or obtaining the representation of a resource, i.e. specifying both its primary access mechanism and network location. For example, the URL http://example.org/wiki/Main_Page refers to a resource identified as /wiki/Main_Page, whose representation is obtainable via the Hypertext Transfer Protocol (http:) from a network host whose domain name is example.org. (In this case, HTTP usually implies it to be in the form of HTML and related code. In practice, that is not necessarily the case, as HTTP allows specifying arbitrary formats in its header.) A URN is analogous to a person's name, while a URL is analogous to their street address. In other words, a URN identifies an item and a URL provides a method for finding it. Technical publications, especially standards produced by the IETF and by the W3C, normally reflect a view outlined in a W3C Recommendation of 30 July 2001, which acknowledges the precedence of the term URI rather than endorsing any formal subdivision into URL and URN. URL is a useful but informal concept: a URL is a type of URI that identifies a resource via a representation of its primary access mechanism (e.g., its network "location"), rather than by some other attributes it may have. As such, a URL is simply a URI that happens to point to a resource over a network.[a] However, in non-technical contexts and in software for the World Wide Web, the term "URL" remains widely used. Additionally, the term "web address" (which has no formal definition) often occurs in non-technical publications as a synonym for a URI that uses the http or https schemes. Such assumptions can lead to confusion, for example, in the case of XML namespaces that have a visual similarity to resolvable URIs. Specifications produced by the WHATWG prefer URL over URI, and so newer HTML5 APIs use URL over URI. Standardize on the term URL. URI and IRI [Internationalized Resource Identifier] are just confusing. In practice a single algorithm is used for both so keeping them distinct is not helping anyone. URL also easily wins the search result popularity contest. While most URI schemes were originally designed to be used with a particular protocol, and often have the same name, they are semantically different from protocols. For example, the scheme http is generally used for interacting with web resources using HTTP, but the scheme file has no protocol. A URI has a scheme that refers to a specification for assigning identifiers within that scheme. As such, the URI syntax is a federated and extensible naming system wherein each scheme's specification may further restrict the syntax and semantics of identifiers using that scheme. The URI generic syntax is a superset of the syntax of all URI schemes. It was first defined in RFC 2396, published in August 1998, and finalized in RFC 3986, published in January 2005. A URI is composed from an allowed set of ASCII characters consisting of reserved characters (gen-delims: :, /, ?, #, [, ], and @; sub-delims: !, $, &, ', (, ), *, +, ,, ;, and =),: 13–14 unreserved characters (uppercase and lowercase letters, decimal digits, -, ., _, and ~),: 13–14 and the character %.: 12 Syntax components and subcomponents are separated by delimiters from the reserved characters (only from generic reserved characters for components) and define identifying data represented as unreserved characters, reserved characters that do not act as delimiters in the component and subcomponent respectively,: §2 and percent-encodings when the corresponding character is outside the allowed set or is being used as a delimiter of, or within, the component. A percent-encoding of an identifying data octet is a sequence of three characters, consisting of the character % followed by the two hexadecimal digits representing that octet's numeric value.: §2.1 The URI generic syntax consists of five components organized hierarchically in order of decreasing significance from left to right:: §3 A component is undefined if it has an associated delimiter and the delimiter does not appear in the URI; the scheme and path components are always defined.: §5.2.1 A component is empty if it has no characters; the scheme component is always non-empty.: §3 The authority component consists of subcomponents: This is represented in a syntax diagram as: The URI comprises: The scheme- or implementation-specific reserved character + may be used in the scheme, userinfo, host, path, query, and fragment, and the scheme- or implementation-specific reserved characters !, $, &, ', (, ), *, ,, ;, and = may be used in the userinfo, host, path, query, and fragment. Additionally, the generic reserved character : may be used in the userinfo, path, query and fragment, the generic reserved characters @ and / may be used in the path, query and fragment, and the generic reserved character ? may be used in the query and fragment.: §A The following figure displays example URIs and their component parts. DOIs (digital object identifiers) fit within the Handle System and fit within the URI system, as facilitated by appropriate syntax. A URI reference is either a URI or a relative reference when it does not begin with a scheme component followed by a colon (:).: §4.1 A path segment that contains a colon character (e.g., foo:bar) cannot be used as the first path segment of a relative reference if its path component does not begin with a slash (/), as it would be mistaken for a scheme component. Such a path segment must be preceded by a dot path segment (e.g., ./foo:bar).: §4.2 Web document markup languages frequently use URI references to point to other resources, such as external documents or specific portions of the same logical document:: §4.4 Resolving a URI reference against a base URI results in a target URI. This implies that the base URI exists and is an absolute URI (a URI with no fragment component). The base URI can be obtained, in order of precedence, from:: §5.1 Within a representation with a well defined base URI of a relative reference is resolved to its target URI as follows:: §5.4 URL munging is a technique by which a command is appended to a URL, usually at the end, after a "?" token. It is commonly used in WebDAV as a mechanism of adding functionality to HTTP. In a versioning system, for example, to add a "checkout" command to a URL, it is written as http://editing.com/resource/file.php?command=checkout. It has the advantage of both being easy for CGI parsers and also acts as an intermediary between HTTP and underlying resource, in this case. In XML, a namespace is an abstract domain to which a collection of element and attribute names can be assigned. The namespace name is a character string which must adhere to the generic URI syntax. However, the name is generally not considered to be a URI, because the URI specification bases the decision not only on lexical components, but also on their intended use. A namespace name does not necessarily imply any of the semantics of URI schemes; for example, a namespace name beginning with http: may have no connotation to the use of the HTTP. Originally, the namespace name could match the syntax of any non-empty URI reference, but the use of relative URI references was deprecated by the W3C. A separate W3C specification for namespaces in XML 1.1 permits Internationalized Resource Identifier (IRI) references to serve as the basis for namespace names in addition to URI references. See also Notes References Further reading External links |
======================================== |
[SOURCE: https://en.wikipedia.org/wiki/Special:BookSources/0-02-923843-9] | [TOKENS: 380] |
Contents Book sources This page allows users to search multiple sources for a book given a 10- or 13-digit International Standard Book Number. Spaces and dashes in the ISBN do not matter. This page links to catalogs of libraries, booksellers, and other book sources where you will be able to search for the book by its International Standard Book Number (ISBN). Online text Google Books and other retail sources below may be helpful if you want to verify citations in Wikipedia articles, because they often let you search an online version of the book for specific words or phrases, or you can browse through the book (although for copyright reasons the entire book is usually not available). At the Open Library (part of the Internet Archive) you can borrow and read entire books online. Online databases Subscription eBook databases Libraries Alabama Alaska California Colorado Connecticut Delaware Florida Georgia Illinois Indiana Iowa Kansas Kentucky Massachusetts Michigan Minnesota Missouri Nebraska New Jersey New Mexico New York North Carolina Ohio Oklahoma Oregon Pennsylvania Rhode Island South Carolina South Dakota Tennessee Texas Utah Washington state Wisconsin Bookselling and swapping Find your book on a site that compiles results from other online sites: These sites allow you to search the catalogs of many individual booksellers: Non-English book sources If the book you are looking for is in a language other than English, you might find it helpful to look at the equivalent pages on other Wikipedias, linked below – they are more likely to have sources appropriate for that language. Find other editions The WorldCat xISBN tool for finding other editions is no longer available. However, there is often a "view all editions" link on the results page from an ISBN search. Google books often lists other editions of a book and related books under the "about this book" link. You can convert between 10 and 13 digit ISBNs with these tools: Find on Wikipedia See also Get free access to research! Research tools and services Outreach Get involved |
======================================== |
[SOURCE: https://en.wikipedia.org/wiki/Wilmer_Cutler_Pickering_Hale_and_Dorr] | [TOKENS: 2338] |
Contents WilmerHale Wilmer Cutler Pickering Hale and Dorr LLP, also known as WilmerHale, is an American multinational law firm with offices in the United States and Europe. Co-headquartered in Washington, D.C., and Boston, it was formed in 2004 through the merger of the Boston-based firm Hale and Dorr and the D.C.-based firm Wilmer Cutler & Pickering. It employs more than 1,000 attorneys worldwide. History Hale and Dorr was founded in Boston in 1918 by Richard Hale, Dudley Huntington Dorr, Frank Grinnell, Roger Swaim, and John Maguire. On January 1, 1919, the partnership was reconstituted to admit George W. Wightman and Reginald Heber Smith. Smith, author of the seminal work Justice and the Poor and a pioneer in the American legal aid movement, joined the firm in 1919 and served as managing partner for thirty years. Hale and Dorr gained national recognition in 1954 when partner Joseph Welch, assisted by associate James St. Clair and John Kimball Jr., represented the U.S. Army on a pro bono basis during the historic Army-McCarthy hearings. In 1974, James D. St. Clair represented President Richard Nixon before the Supreme Court of the United States in United States v. Nixon. In 1988, partner Paul Brountas chaired the presidential campaign of Massachusetts Governor Michael Dukakis, and in 1990, senior partner William Weld was elected governor. The firm has had a long relationship with nearby Harvard Law School, home of the WilmerHale Legal Services Center. Wilmer, Cutler & Pickering was founded in Washington in 1962 by former Cravath attorneys Lloyd Cutler and John Pickering, along with a senior lawyer, Richard H. Wilmer. Cutler, who later served as White House Counsel to Presidents Jimmy Carter and Bill Clinton, founded the Lawyers' Committee for Civil Rights Under Law in 1962 and served on its executive committee until 1987. In the 1980s, Cutler led the founding of the Southern Africa Legal Services and Legal Education Project, to aid South African lawyers who fought to implement the rule of law during apartheid. From 1981 to 1993, partner C. Boyden Gray, a prominent member of The Federalist Society, left the firm to serve as White House Counsel to Vice President and President George H. W. Bush. In 2003, partner Jamie Gorelick began serving as a member of the 9/11 Commission. The two firms merged to form Wilmer Cutler Pickering Hale and Dorr in 2004, with headquarters now in both Boston and Washington. In 2010, the law firm relocated its administrative support base to a new campus in Dayton, Ohio, as it sought to streamline internal business operations across its many offices. The office houses more than 200 employees from existing WilmerHale offices and new employees from the Dayton area. Individuals in the Business Services Center include administrative support staff, bringing together services such as finance, human resources, information technology services, operations, document review and management, and practice management, which will provide improved efficiencies for administrative teams and the firm, and reduce significant operational expenses. In June 2023, the firm announced that former federal prosecutor Anjan Sahni would replace the co-managing partners Robert Novick and Susan Murley at the beginning of next year. On March 27, 2025, President Donald Trump signed an executive order taking action against the firm over its ties to Robert Mueller, former special counsel, who led a probe into Trump's 2016 campaign and its alleged ties with Russian state officials. The executive order directs federal agencies to end contracts with WilmerHale's clients, revokes their lawyers' security clearances and restricts their access to certain government buildings. The following day a lawsuit was filed by Paul Clement of Clement & Murphy seeking to bar the executive order. Later that day, U.S. District Judge Richard J. Leon issued a temporary restraining order blocking a portion of the order, but did not block the provision ending WilmerHale’s lawyers' security clearances. On May 27, 2025, Judge Leon struck down the executive order, calling the order “unconstitutional.” Reputation According to one study examining political donations by large white shoe firms, WilmerHale was ranked as the most liberal out of the top twenty prestigious law firms in the nation. The Washington Post ranked WilmerHale as the No. 1 Top Workplace in DC in 2019, and the firm has earned a Top Workplace for eight consecutive years, and noted that the firm "has played a leading role in reimagining what 'Big Law' can be." The American Lawyer named WilmerHale Law Firm of the Year in 2021 and noted that for 17 years, the law firm has earned a spot on its A-List, which takes into consideration not just revenues, but pro bono work, diversity, and attorney satisfaction. The firm is also well known for its litigation and trial expertise. In The American Lawyer's biennial Litigation Department of the Year contest honoring law firms considered the "absolute best in the industry when it comes to litigation prowess", they highlighted that the firm has a strong reputation in the most important practice areas such as public policy and legislative affairs, regulation, antitrust, intellectual property and international trade and that WilmerHale is home to some of the most well-known appellate and Supreme Court litigators in the country. Clients In the late 1980s, Hale and Dorr partner Jerome Facher represented Beatrice Foods in a suit by eight families from Woburn, Massachusetts who claimed that Beatrice, along with W.R. Grace, had polluted the town's water supply, resulting in an elevated number of leukemia cases and immune-system disorders. The case was memorialized in the book A Civil Action, by Jonathan Harr, and in a movie of the same name starring Robert Duvall as Facher and John Travolta as plaintiffs' lawyer Jan Schlichtmann. Upon further discovery, the EPA took the case on and W.R. Grace was successfully indicted for making false statements. Both W. R. Grace and Beatrice Foods paid a total $64.9 million to clean up the contaminated sites in Woburn. In the wake of news articles raising concerns about transactions between Enron and its CFO, Andy Fastow, lawyers from Wilmer, Cutler & Pickering represented a special investigative committee of Enron's board of directors in an internal investigation into those transactions. The resulting report, known as the "Powers Report," laid out the facts that have been the predicate for much of the public discussion of Enron since that time. Similarly, after WorldCom's announcement that it would have to restate financial statements, the firm represented a special investigative committee of WorldCom's board of directors in performing an internal investigation into the accounting irregularities. The investigation resulted in a widely covered written report that detailed a variety of accounting issues as well as the role of management and the board of directors. WilmerHale counseled Apple Inc. in its hotly contested smartphone patent dispute with Samsung Electronics Co. The two parties reached a settlement in 2018 after a seven-year-long battle that began when Apple accused Samsung of infringing numerous design and utility patents related to the iPhone. In 2021, WilmerHale represented PerkinElmer in its acquisition of life sciences company BioLegend for $5.25 billion, the largest acquisition to date for PerkinElmer. President and CEO of PerkinElmer Prahlad Singh said the deal will "push science and discovery forward." In 1986, Wilmer, Cutler & Pickering represented corporate raider Ivan Boesky in high-profile Department of Justice and SEC proceedings, as well as multiple class actions based on his participation in insider trading violations. Wilmer, Cutler & Pickering represented Swiss banks accused of profiting from the Holocaust in their settlement negotiations with plaintiffs. The firm also represented Siemens AG, Krupp AG, and other German companies accused of exploiting forced laborers during the Nazi era. Since 2005, WilmerHale has represented Senator William Frist in regard to an SEC insider trading investigation. WilmerHale was hired to represent PepsiCo in the SEC investigation related to the departure of PepsiCo general counsel Maura Smith. In the course of this representation, a WilmerHale attorney inadvertently e-mailed a confidential legal memorandum to a Wall Street Journal reporter as part of an internal communication to other attorneys working on the matter, which made several details of the investigation public. In early December 2022 Caroline Ellison, former CEO of Alameda Research, hired Stephanie Avakian as her lead attorney. WilmerHale was the external law firm hired by OpenAI to conduct a review and investigation into the events surrounding the November 2023 removal and subsequent reinstatement of CEO Sam Altman. In July 2025, the Financial Times reported that Boston Consulting Group (BCG) had hired WilmerHale to help in an internal investigation on "process failures", referring to the role of the "BCG modelled plan to 'relocate' Palestinians from Gaza". Pro bono Both Hale and Dorr and Wilmer, Cutler & Pickering have a long history of involvement in pro bono work. WilmerHale lawyers dedicated pro bono hours to addressing issues of systemic racial injustice following the murder of George Floyd in 2020. Attorneys at the firm also challenged the legality of Georgia's voting maps, claiming that the revised maps diluted Black voting strength. A team of WilmerHale attorneys represents the "Algerian Six", a group of men who fell under suspicion of planning to attack the US embassy in Bosnia and who are now held in the Guantanamo Bay detainment camp. In 2006, attorney Melissa Hoffer, then part of the team with WilmerHale, delivered a speech in Caen, France, critical of U.S. detainee policy. Other WilmerHale lawyers participating in the case include Stephen Oleskey and Rob Kirsch. In January 2007, Cully Stimson, deputy assistant secretary of defense for detainee affairs, criticized WilmerHale and other major law firms for representing "the very terrorists who hit their bottom line back in 2001," and questioned whether such work was really being done pro bono or might actually receive funding from shadowy sources. In a Wall Street Journal editorial criticizing Stimson, Harvard Law School professor (and former United States Solicitor General under President Reagan) Charles Fried wrote: It is no surprise that firms like WilmerHale (which represents both Big Pharma and Tobacco Free Kids), Covington & Burling (which represents both Big Tobacco and Guantanamo detainees), and the other firms on Mr. Stimson's hit list, are among the most sought-after by law school graduates, and retain the loyalty and enthusiasm of their partners. They offer their lawyers the profession at its best and help assure that the rule of law is not just a slogan but a satisfying way of life. In December 2007, Seth Waxman made the oral argument to the Supreme Court in Boumediene v. Bush which upheld habeas corpus rights for detainees at Guantanamo Bay. Notable alumni See also References External links |
======================================== |
[SOURCE: https://en.wikipedia.org/wiki/OpenAI#cite_ref-148] | [TOKENS: 8773] |
Contents OpenAI OpenAI is an American artificial intelligence research organization comprising both a non-profit foundation and a controlled for-profit public benefit corporation (PBC), headquartered in San Francisco. It aims to develop "safe and beneficial" artificial general intelligence (AGI), which it defines as "highly autonomous systems that outperform humans at most economically valuable work". OpenAI is widely recognized for its development of the GPT family of large language models, the DALL-E series of text-to-image models, and the Sora series of text-to-video models, which have influenced industry research and commercial applications. Its release of ChatGPT in November 2022 has been credited with catalyzing widespread interest in generative AI. The organization was founded in 2015 in Delaware but evolved a complex corporate structure. As of October 2025, following restructuring approved by California and Delaware regulators, the non-profit OpenAI Foundation holds 26% of the for-profit OpenAI Group PBC, with Microsoft holding 27% and employees/other investors holding 47%. Under its governance arrangements, the OpenAI Foundation holds the authority to appoint the board of the for-profit OpenAI Group PBC, a mechanism designed to align the entity’s strategic direction with the Foundation’s charter. Microsoft previously invested over $13 billion into OpenAI, and provides Azure cloud computing resources. In October 2025, OpenAI conducted a $6.6 billion share sale that valued the company at $500 billion. In 2023 and 2024, OpenAI faced multiple lawsuits for alleged copyright infringement against authors and media companies whose work was used to train some of OpenAI's products. In November 2023, OpenAI's board removed Sam Altman as CEO, citing a lack of confidence in him, but reinstated him five days later following a reconstruction of the board. Throughout 2024, roughly half of then-employed AI safety researchers left OpenAI, citing the company's prominent role in an industry-wide problem. Founding In December 2015, OpenAI was founded as a not for profit organization by Sam Altman, Elon Musk, Ilya Sutskever, Greg Brockman, Trevor Blackwell, Vicki Cheung, Andrej Karpathy, Durk Kingma, John Schulman, Pamela Vagata, and Wojciech Zaremba, with Sam Altman and Elon Musk as the co-chairs. A total of $1 billion in capital was pledged by Sam Altman, Greg Brockman, Elon Musk, Reid Hoffman, Jessica Livingston, Peter Thiel, Amazon Web Services (AWS), and Infosys. However, the actual capital collected significantly lagged pledges. According to company disclosures, only $130 million had been received by 2019. In its founding charter, OpenAI stated an intention to collaborate openly with other institutions by making certain patents and research publicly available, but later restricted access to its most capable models, citing competitive and safety concerns. OpenAI was initially run from Brockman's living room. It was later headquartered at the Pioneer Building in the Mission District, San Francisco. According to OpenAI's charter, its founding mission is "to ensure that artificial general intelligence (AGI)—by which we mean highly autonomous systems that outperform humans at most economically valuable work—benefits all of humanity." Musk and Altman stated in 2015 that they were partly motivated by concerns about AI safety and existential risk from artificial general intelligence. OpenAI stated that "it's hard to fathom how much human-level AI could benefit society", and that it is equally difficult to comprehend "how much it could damage society if built or used incorrectly". The startup also wrote that AI "should be an extension of individual human wills and, in the spirit of liberty, as broadly and evenly distributed as possible", and that "because of AI's surprising history, it's hard to predict when human-level AI might come within reach. When it does, it'll be important to have a leading research institution which can prioritize a good outcome for all over its own self-interest." Co-chair Sam Altman expected a decades-long project that eventually surpasses human intelligence. Brockman met with Yoshua Bengio, one of the "founding fathers" of deep learning, and drew up a list of great AI researchers. Brockman was able to hire nine of them as the first employees in December 2015. OpenAI did not pay AI researchers salaries comparable to those of Facebook or Google. It also did not pay stock options which AI researchers typically get. Nevertheless, OpenAI spent $7 million on its first 52 employees in 2016. OpenAI's potential and mission drew these researchers to the firm; a Google employee said he was willing to leave Google for OpenAI "partly because of the very strong group of people and, to a very large extent, because of its mission." OpenAI co-founder Wojciech Zaremba stated that he turned down "borderline crazy" offers of two to three times his market value to join OpenAI instead. In April 2016, OpenAI released a public beta of "OpenAI Gym", its platform for reinforcement learning research. Nvidia gifted its first DGX-1 supercomputer to OpenAI in August 2016 to help it train larger and more complex AI models with the capability of reducing processing time from six days to two hours. In December 2016, OpenAI released "Universe", a software platform for measuring and training an AI's general intelligence across the world's supply of games, websites, and other applications. Corporate structure In 2019, OpenAI transitioned from non-profit to "capped" for-profit, with the profit being capped at 100 times any investment. According to OpenAI, the capped-profit model allows OpenAI Global, LLC to legally attract investment from venture funds and, in addition, to grant employees stakes in the company. Many top researchers work for Google Brain, DeepMind, or Facebook, which offer equity that a nonprofit would be unable to match. Before the transition, OpenAI was legally required to publicly disclose the compensation of its top employees. The company then distributed equity to its employees and partnered with Microsoft, announcing an investment package of $1 billion into the company. Since then, OpenAI systems have run on an Azure-based supercomputing platform from Microsoft. OpenAI Global, LLC then announced its intention to commercially license its technologies. It planned to spend $1 billion "within five years, and possibly much faster". Altman stated that even a billion dollars may turn out to be insufficient, and that the lab may ultimately need "more capital than any non-profit has ever raised" to achieve artificial general intelligence. The nonprofit, OpenAI, Inc., is the sole controlling shareholder of OpenAI Global, LLC, which, despite being a for-profit company, retains a formal fiduciary responsibility to OpenAI, Inc.'s nonprofit charter. A majority of OpenAI, Inc.'s board is barred from having financial stakes in OpenAI Global, LLC. In addition, minority members with a stake in OpenAI Global, LLC are barred from certain votes due to conflict of interest. Some researchers have argued that OpenAI Global, LLC's switch to for-profit status is inconsistent with OpenAI's claims to be "democratizing" AI. On February 29, 2024, Elon Musk filed a lawsuit against OpenAI and CEO Sam Altman, accusing them of shifting focus from public benefit to profit maximization—a case OpenAI dismissed as "incoherent" and "frivolous," though Musk later revived legal action against Altman and others in August. On April 9, 2024, OpenAI countersued Musk in federal court, alleging that he had engaged in "bad-faith tactics" to slow the company's progress and seize its innovations for his personal benefit. OpenAI also argued that Musk had previously supported the creation of a for-profit structure and had expressed interest in controlling OpenAI himself. The countersuit seeks damages and legal measures to prevent further alleged interference. On February 10, 2025, a consortium of investors led by Elon Musk submitted a $97.4 billion unsolicited bid to buy the nonprofit that controls OpenAI, declaring willingness to match or exceed any better offer. The offer was rejected on 14 February 2025, with OpenAI stating that it was not for sale, but the offer complicated Altman's restructuring plan by suggesting a lower bar for how much the nonprofit should be valued. OpenAI, Inc. was originally designed as a nonprofit in order to ensure that AGI "benefits all of humanity" rather than "the private gain of any person". In 2019, it created OpenAI Global, LLC, a capped-profit subsidiary controlled by the nonprofit. In December 2024, OpenAI proposed a restructuring plan to convert the capped-profit into a Delaware-based public benefit corporation (PBC), and to release it from the control of the nonprofit. The nonprofit would sell its control and other assets, getting equity in return, and would use it to fund and pursue separate charitable projects, including in science and education. OpenAI's leadership described the change as necessary to secure additional investments, and claimed that the nonprofit's founding mission to ensure AGI "benefits all of humanity" would be better fulfilled. The plan has been criticized by former employees. A legal letter named "Not For Private Gain" asked the attorneys general of California and Delaware to intervene, stating that the restructuring is illegal and would remove governance safeguards from the nonprofit and the attorneys general. The letter argues that OpenAI's complex structure was deliberately designed to remain accountable to its mission, without the conflicting pressure of maximizing profits. It contends that the nonprofit is best positioned to advance its mission of ensuring AGI benefits all of humanity by continuing to control OpenAI Global, LLC, whatever the amount of equity that it could get in exchange. PBCs can choose how they balance their mission with profit-making. Controlling shareholders have a large influence on how closely a PBC sticks to its mission. On October 28, 2025, OpenAI announced that it had adopted the new PBC corporate structure after receiving approval from the attorneys general of California and Delaware. Under the new structure, OpenAI's for-profit branch became a public benefit corporation known as OpenAI Group PBC, while the non-profit was renamed to the OpenAI Foundation. The OpenAI Foundation holds a 26% stake in the PBC, while Microsoft holds a 27% stake and the remaining 47% is owned by employees and other investors. All members of the OpenAI Group PBC board of directors will be appointed by the OpenAI Foundation, which can remove them at any time. Members of the Foundation's board will also serve on the for-profit board. The new structure allows the for-profit PBC to raise investor funds like most traditional tech companies, including through an initial public offering, which Altman claimed was the most likely path forward. In January 2023, OpenAI Global, LLC was in talks for funding that would value the company at $29 billion, double its 2021 value. On January 23, 2023, Microsoft announced a new US$10 billion investment in OpenAI Global, LLC over multiple years, partially needed to use Microsoft's cloud-computing service Azure. From September to December, 2023, Microsoft rebranded all variants of its Copilot to Microsoft Copilot, and they added MS-Copilot to many installations of Windows and released Microsoft Copilot mobile apps. Following OpenAI's 2025 restructuring, Microsoft owns a 27% stake in the for-profit OpenAI Group PBC, valued at $135 billion. In a deal announced the same day, OpenAI agreed to purchase $250 billion of Azure services, with Microsoft ceding their right of first refusal over OpenAI's future cloud computing purchases. As part of the deal, OpenAI will continue to share 20% of its revenue with Microsoft until it achieves AGI, which must now be verified by an independent panel of experts. The deal also loosened restrictions on both companies working with third parties, allowing Microsoft to pursue AGI independently and allowing OpenAI to develop products with other companies. In 2017, OpenAI spent $7.9 million, a quarter of its functional expenses, on cloud computing alone. In comparison, DeepMind's total expenses in 2017 were $442 million. In the summer of 2018, training OpenAI's Dota 2 bots required renting 128,000 CPUs and 256 GPUs from Google for multiple weeks. In October 2024, OpenAI completed a $6.6 billion capital raise with a $157 billion valuation including investments from Microsoft, Nvidia, and SoftBank. On January 21, 2025, Donald Trump announced The Stargate Project, a joint venture between OpenAI, Oracle, SoftBank and MGX to build an AI infrastructure system in conjunction with the US government. The project takes its name from OpenAI's existing "Stargate" supercomputer project and is estimated to cost $500 billion. The partners planned to fund the project over the next four years. In July, the United States Department of Defense announced that OpenAI had received a $200 million contract for AI in the military, along with Anthropic, Google, and xAI. In the same month, the company made a deal with the UK Government to use ChatGPT and other AI tools in public services. OpenAI subsequently began a $50 million fund to support nonprofit and community organizations. In April 2025, OpenAI raised $40 billion at a $300 billion post-money valuation, which was the highest-value private technology deal in history. The financing round was led by SoftBank, with other participants including Microsoft, Coatue, Altimeter and Thrive. In July 2025, the company reported annualized revenue of $12 billion. This was an increase from $3.7 billion in 2024, which was driven by ChatGPT subscriptions, which reached 20 million paid subscribers by April 2025, up from 15.5 million at the end of 2024, alongside a rapidly expanding enterprise customer base that grew to five million business users. The company’s cash burn remains high because of the intensive computational costs required to train and operate large language models. It projects an $8 billion operating loss in 2025. OpenAI reports revised long-term spending projections totaling approximately $115 billion through 2029, with annual expenditures projected to escalate significantly, reaching $17 billion in 2026, $35 billion in 2027, and $45 billion in 2028. These expenditures are primarily allocated toward expanding compute infrastructure, developing proprietary AI chips, constructing data centers, and funding intensive model training programs, with more than half of the spending through the end of the decade expected to support research-intensive compute for model training and development. The company's financial strategy prioritizes market expansion and technological advancement over near-term profitability, with OpenAI targeting cash-flow-positive operations by 2029 and projecting revenue of approximately $200 billion by 2030. This aggressive spending trajectory underscores both the enormous capital requirements of scaling cutting-edge AI technology and OpenAI's commitment to maintaining its position as a leader in the artificial intelligence industry. In October 2025, OpenAI completed an employee share sale of up to $10 billion to existing investors which valued the company at $500 billion. The deal values OpenAI as the most valuable privately owned company in the world—surpassing SpaceX as the world's most valuable private company. On November 17, 2023, Sam Altman was removed as CEO when its board of directors (composed of Helen Toner, Ilya Sutskever, Adam D'Angelo and Tasha McCauley) cited a lack of confidence in him. Chief Technology Officer Mira Murati took over as interim CEO. Greg Brockman, the president of OpenAI, was also removed as chairman of the board and resigned from the company's presidency shortly thereafter. Three senior OpenAI researchers subsequently resigned: director of research and GPT-4 lead Jakub Pachocki, head of AI risk Aleksander Mądry, and researcher Szymon Sidor. On November 18, 2023, there were reportedly talks of Altman returning as CEO amid pressure placed upon the board by investors such as Microsoft and Thrive Capital, who objected to Altman's departure. Although Altman himself spoke in favor of returning to OpenAI, he has since stated that he considered starting a new company and bringing former OpenAI employees with him if talks to reinstate him didn't work out. The board members agreed "in principle" to resign if Altman returned. On November 19, 2023, negotiations with Altman to return failed and Murati was replaced by Emmett Shear as interim CEO. The board initially contacted Anthropic CEO Dario Amodei (a former OpenAI executive) about replacing Altman, and proposed a merger of the two companies, but both offers were declined. On November 20, 2023, Microsoft CEO Satya Nadella announced Altman and Brockman would be joining Microsoft to lead a new advanced AI research team, but added that they were still committed to OpenAI despite recent events. Before the partnership with Microsoft was finalized, Altman gave the board another opportunity to negotiate with him. About 738 of OpenAI's 770 employees, including Murati and Sutskever, signed an open letter stating they would quit their jobs and join Microsoft if the board did not rehire Altman and then resign. This prompted OpenAI investors to consider legal action against the board as well. In response, OpenAI management sent an internal memo to employees stating that negotiations with Altman and the board had resumed and would take some time. On November 21, 2023, after continued negotiations, Altman and Brockman returned to the company in their prior roles along with a reconstructed board made up of new members Bret Taylor (as chairman) and Lawrence Summers, with D'Angelo remaining. According to subsequent reporting, shortly before Altman’s firing, some employees raised concerns to the board about how he had handled the safety implications of a recent internal AI capability discovery. On November 29, 2023, OpenAI announced that an anonymous Microsoft employee had joined the board as a non-voting member to observe the company's operations; Microsoft resigned from the board in July 2024. In February 2024, the Securities and Exchange Commission subpoenaed OpenAI's internal communication to determine if Altman's alleged lack of candor misled investors. In 2024, following the temporary removal of Sam Altman and his return, many employees gradually left OpenAI, including most of the original leadership team and a significant number of AI safety researchers. In August 2023, it was announced that OpenAI had acquired the New York-based start-up Global Illumination, a company that deploys AI to develop digital infrastructure and creative tools. In June 2024, OpenAI acquired Multi, a startup focused on remote collaboration. In March 2025, OpenAI reached a deal with CoreWeave to acquire $350 million worth of CoreWeave shares and access to AI infrastructure, in return for $11.9 billion paid over five years. Microsoft was already CoreWeave's biggest customer in 2024. Alongside their other business dealings, OpenAI and Microsoft were renegotiating the terms of their partnership to facilitate a potential future initial public offering by OpenAI, while ensuring Microsoft's continued access to advanced AI models. On May 21, OpenAI announced the $6.5 billion acquisition of io, an AI hardware start-up founded by former Apple designer Jony Ive in 2024. In September 2025, OpenAI agreed to acquire the product testing startup Statsig for $1.1 billion in an all-stock deal and appointed Statsig's founding CEO Vijaye Raji as OpenAI's chief technology officer of applications. The company also announced development of an AI-driven hiring service designed to rival LinkedIn. OpenAI acquired personal finance app Roi in October 2025. In October 2025, OpenAI acquired Software Applications Incorporated, the developer of Sky, a macOS-based natural language interface designed to operate across desktop applications. The Sky team joined OpenAI, and the company announced plans to integrate Sky’s capabilities into ChatGPT. In December 2025, it was announced OpenAI had agreed to acquire Neptune, an AI tooling startup that helps companies track and manage model training, for an undisclosed amount. In January 2026, it was announced OpenAI had acquired healthcare technology startup Torch for approximately $60 million. The acquisition followed the launch of OpenAI’s ChatGPT Health product and was intended to strengthen the company’s medical data and healthcare artificial intelligence capabilities. OpenAI has been criticized for outsourcing the annotation of data sets to Sama, a company based in San Francisco that employed workers in Kenya. These annotations were used to train an AI model to detect toxicity, which could then be used to moderate toxic content, notably from ChatGPT's training data and outputs. However, these pieces of text usually contained detailed descriptions of various types of violence, including sexual violence. The investigation uncovered that OpenAI began sending snippets of data to Sama as early as November 2021. The four Sama employees interviewed by Time described themselves as mentally scarred. OpenAI paid Sama $12.50 per hour of work, and Sama was redistributing the equivalent of between $1.32 and $2.00 per hour post-tax to its annotators. Sama's spokesperson said that the $12.50 was also covering other implicit costs, among which were infrastructure expenses, quality assurance and management. In 2024, OpenAI began collaborating with Broadcom to design a custom AI chip capable of both training and inference, targeted for mass production in 2026 and to be manufactured by TSMC on a 3 nm process node. This initiative intended to reduce OpenAI's dependence on Nvidia GPUs, which are costly and face high demand in the market. In January 2024, Arizona State University purchased ChatGPT Enterprise in OpenAI's first deal with a university. In June 2024, Apple Inc. signed a contract with OpenAI to integrate ChatGPT features into its products as part of its new Apple Intelligence initiative. In June 2025, OpenAI began renting Google Cloud's Tensor Processing Units (TPUs) to support ChatGPT and related services, marking its first meaningful use of non‑Nvidia AI chips. In September 2025, it was revealed that OpenAI signed a contract with Oracle to purchase $300 billion in computing power over the next five years. In September 2025, OpenAI and NVIDIA announced a memorandum of understanding that included a potential deployment of at least 10 gigawatts of NVIDIA systems and a $100 billion investment from NVIDIA in OpenAI. OpenAI expected the negotiations to be completed within weeks. As of January 2026, this has not been realized, and the two sides are rethinking the future of their partnership. In October 2025, OpenAI announced a multi-billion dollar deal with AMD. OpenAI committed to purchasing six gigawatts worth of AMD chips, starting with the MI450. OpenAI will have the option to buy up to 160 million shares of AMD, about 10% of the company, depending on development, performance and share price targets. In December 2025, Disney said it would make a $1 billion investment in OpenAI, and signed a three-year licensing deal that will let users generate videos using Sora—OpenAI's short-form AI video platform. More than 200 Disney, Marvel, Star Wars and Pixar characters will be available to OpenAI users. In early 2026, Amazon entered advanced discussions to invest up to $50 billion in OpenAI as part of a potential artificial intelligence partnership. Under the proposed agreement, OpenAI’s models could be integrated into Amazon’s digital assistant Alexa and other internal projects. OpenAI provides LLMs to the Artificial Intelligence Cyber Challenge and to the Advanced Research Projects Agency for Health. In October 2024, The Intercept revealed that OpenAI's tools are considered "essential" for AFRICOM's mission and included in an "Exception to Fair Opportunity" contractual agreement between the United States Department of Defense and Microsoft. In December 2024, OpenAI said it would partner with defense-tech company Anduril to build drone defense technologies for the United States and its allies. In 2025, OpenAI's Chief Product Officer, Kevin Weil, was commissioned lieutenant colonel in the U.S. Army to join Detachment 201 as senior advisor. In June 2025, the U.S. Department of Defense awarded OpenAI a $200 million one-year contract to develop AI tools for military and national security applications. OpenAI announced a new program, OpenAI for Government, to give federal, state, and local governments access to its models, including ChatGPT. Services In February 2019, GPT-2 was announced, which gained attention for its ability to generate human-like text. In 2020, OpenAI announced GPT-3, a language model trained on large internet datasets. GPT-3 is aimed at natural language answering questions, but it can also translate between languages and coherently generate improvised text. It also announced that an associated API, named the API, would form the heart of its first commercial product. Eleven employees left OpenAI, mostly between December 2020 and January 2021, in order to establish Anthropic. In 2021, OpenAI introduced DALL-E, a specialized deep learning model adept at generating complex digital images from textual descriptions, utilizing a variant of the GPT-3 architecture. In December 2022, OpenAI received widespread media coverage after launching a free preview of ChatGPT, its new AI chatbot based on GPT-3.5. According to OpenAI, the preview received over a million signups within the first five days. According to anonymous sources cited by Reuters in December 2022, OpenAI Global, LLC was projecting $200 million of revenue in 2023 and $1 billion in revenue in 2024. After ChatGPT was launched, Google announced a similar chatbot, Bard, amid internal concerns that ChatGPT could threaten Google’s position as a primary source of online information. On February 7, 2023, Microsoft announced that it was building AI technology based on the same foundation as ChatGPT into Microsoft Bing, Edge, Microsoft 365 and other products. On March 14, 2023, OpenAI released GPT-4, both as an API (with a waitlist) and as a feature of ChatGPT Plus. On November 6, 2023, OpenAI launched GPTs, allowing individuals to create customized versions of ChatGPT for specific purposes, further expanding the possibilities of AI applications across various industries. On November 14, 2023, OpenAI announced they temporarily suspended new sign-ups for ChatGPT Plus due to high demand. Access for newer subscribers re-opened a month later on December 13. In December 2024, the company launched the Sora model. It also launched OpenAI o1, an early reasoning model that was internally codenamed strawberry. Additionally, ChatGPT Pro—a $200/month subscription service offering unlimited o1 access and enhanced voice features—was introduced, and preliminary benchmark results for the upcoming OpenAI o3 models were shared. On January 23, 2025, OpenAI released Operator, an AI agent and web automation tool for accessing websites to execute goals defined by users. The feature was only available to Pro users in the United States. OpenAI released deep research agent, nine days later. It scored a 27% accuracy on the benchmark Humanity's Last Exam (HLE). Altman later stated GPT-4.5 would be the last model without full chain-of-thought reasoning. In July 2025, reports indicated that AI models by both OpenAI and Google DeepMind solved mathematics problems at the level of top-performing students in the International Mathematical Olympiad. OpenAI's large language model was able to achieve gold medal-level performance, reflecting significant progress in AI's reasoning abilities. On October 6, 2025, OpenAI unveiled its Agent Builder platform during the company's DevDay event. The platform includes a visual drag-and-drop interface that lets developers and businesses design, test, and deploy agentic workflows with limited coding. On October 21, 2025, OpenAI introduced ChatGPT Atlas, a browser integrating the ChatGPT assistant directly into web navigation, to compete with existing browsers such as Google Chrome and Apple Safari. On December 11, 2025, OpenAI announced GPT-5.2. This model will be better at creating spreadsheets, building presentations, perceiving images, writing code and understanding long context. On January 27, 2026, OpenAI introduced Prism, a LaTeX-native workspace meant to assist scientists to help with research and writing. The platform utilizes GPT-5.2 as a backend to automate the process of drafting for scientific papers, including features for managing citations, complex equation formatting, and real-time collaborative editing. In March 2023, the company was criticized for disclosing particularly few technical details about products like GPT-4, contradicting its initial commitment to openness and making it harder for independent researchers to replicate its work and develop safeguards. OpenAI cited competitiveness and safety concerns to justify this repudiation. OpenAI's former chief scientist Ilya Sutskever argued in 2023 that open-sourcing increasingly capable models was increasingly risky, and that the safety reasons for not open-sourcing the most potent AI models would become "obvious" in a few years. In September 2025, OpenAI published a study on how people use ChatGPT for everyday tasks. The study found that "non-work tasks" (according to an LLM-based classifier) account for more than 72 percent of all ChatGPT usage, with a minority of overall usage related to business productivity. In July 2023, OpenAI launched the superalignment project, aiming within four years to determine how to align future superintelligent systems. OpenAI promised to dedicate 20% of its computing resources to the project, although the team denied receiving anything close to 20%. OpenAI ended the project in May 2024 after its co-leaders Ilya Sutskever and Jan Leike left the company. In August 2025, OpenAI was criticized after thousands of private ChatGPT conversations were inadvertently exposed to public search engines like Google due to an experimental "share with search engines" feature. The opt-in toggle, intended to allow users to make specific chats discoverable, resulted in some discussions including personal details such as names, locations, and intimate topics appearing in search results when users accidentally enabled it while sharing links. OpenAI announced the feature's permanent removal on August 1, 2025, and the company began coordinating with search providers to remove the exposed content, emphasizing that it was not a security breach but a design flaw that heightened privacy risks. CEO Sam Altman acknowledged the issue in a podcast, noting users often treat ChatGPT as a confidant for deeply personal matters, which amplified concerns about AI handling sensitive data. Management In 2018, Musk resigned from his Board of Directors seat, citing "a potential future conflict [of interest]" with his role as CEO of Tesla due to Tesla's AI development for self-driving cars. OpenAI stated that Musk's financial contributions were below $45 million. On March 3, 2023, Reid Hoffman resigned from his board seat, citing a desire to avoid conflicts of interest with his investments in AI companies via Greylock Partners, and his co-founding of the AI startup Inflection AI. Hoffman remained on the board of Microsoft, a major investor in OpenAI. In May 2024, Chief Scientist Ilya Sutskever resigned and was succeeded by Jakub Pachocki. Co-leader Jan Leike also departed amid concerns over safety and trust. OpenAI then signed deals with Reddit, News Corp, Axios, and Vox Media. Paul Nakasone then joined the board of OpenAI. In August 2024, cofounder John Schulman left OpenAI to join Anthropic, and OpenAI's president Greg Brockman took extended leave until November. In September 2024, CTO Mira Murati left the company. In November 2025, Lawrence Summers resigned from the board of directors. Governance and legal issues In May 2023, Sam Altman, Greg Brockman and Ilya Sutskever posted recommendations for the governance of superintelligence. They stated that superintelligence could happen within the next 10 years, allowing a "dramatically more prosperous future" and that "given the possibility of existential risk, we can't just be reactive". They proposed creating an international watchdog organization similar to IAEA to oversee AI systems above a certain capability threshold, suggesting that relatively weak AI systems on the other side should not be overly regulated. They also called for more technical safety research for superintelligences, and asked for more coordination, for example through governments launching a joint project which "many current efforts become part of". In July 2023, the FTC issued a civil investigative demand to OpenAI to investigate whether the company's data security and privacy practices to develop ChatGPT were unfair or harmed consumers (including by reputational harm) in violation of Section 5 of the Federal Trade Commission Act of 1914. These are typically preliminary investigative matters and are nonpublic, but the FTC's document was leaked. In July 2023, the FTC launched an investigation into OpenAI over allegations that the company scraped public data and published false and defamatory information. They asked OpenAI for comprehensive information about its technology and privacy safeguards, as well as any steps taken to prevent the recurrence of situations in which its chatbot generated false and derogatory content about people. The agency also raised concerns about ‘circular’ spending arrangements—for example, Microsoft extending Azure credits to OpenAI while both companies shared engineering talent—and warned that such structures could negatively affect the public. In September 2024, OpenAI's global affairs chief endorsed the UK's "smart" AI regulation during testimony to a House of Lords committee. In February 2025, OpenAI CEO Sam Altman stated that the company is interested in collaborating with the People's Republic of China, despite regulatory restrictions imposed by the U.S. government. This shift comes in response to the growing influence of the Chinese artificial intelligence company DeepSeek, which has disrupted the AI market with open models, including DeepSeek V3 and DeepSeek R1. Following DeepSeek's market emergence, OpenAI enhanced security protocols to protect proprietary development techniques from industrial espionage. Some industry observers noted similarities between DeepSeek's model distillation approach and OpenAI's methodology, though no formal intellectual property claim was filed. According to Oliver Roberts, in March 2025, the United States had 781 state AI bills or laws. OpenAI advocated for preempting state AI laws with federal laws. According to Scott Kohler, OpenAI has opposed California's AI legislation and suggested that the state bill encroaches on a more competent federal government. Public Citizen opposed a federal preemption on AI and pointed to OpenAI's growth and valuation as evidence that existing state laws have not hampered innovation. Before May 2024, OpenAI required departing employees to sign a lifelong non-disparagement agreement forbidding them from criticizing OpenAI and acknowledging the existence of the agreement. Daniel Kokotajlo, a former employee, publicly stated that he forfeited his vested equity in OpenAI in order to leave without signing the agreement. Sam Altman stated that he was unaware of the equity cancellation provision, and that OpenAI never enforced it to cancel any employee's vested equity. However, leaked documents and emails refute this claim. On May 23, 2024, OpenAI sent a memo releasing former employees from the agreement. OpenAI was sued for copyright infringement by authors Sarah Silverman, Matthew Butterick, Paul Tremblay and Mona Awad in July 2023. In September 2023, 17 authors, including George R. R. Martin, John Grisham, Jodi Picoult and Jonathan Franzen, joined the Authors Guild in filing a class action lawsuit against OpenAI, alleging that the company's technology was illegally using their copyrighted work. The New York Times also sued the company in late December 2023. In May 2024 it was revealed that OpenAI had destroyed its Books1 and Books2 training datasets, which were used in the training of GPT-3, and which the Authors Guild believed to have contained over 100,000 copyrighted books. In 2021, OpenAI developed a speech recognition tool called Whisper. OpenAI used it to transcribe more than one million hours of YouTube videos into text for training GPT-4. The automated transcription of YouTube videos raised concerns within OpenAI employees regarding potential violations of YouTube's terms of service, which prohibit the use of videos for applications independent of the platform, as well as any type of automated access to its videos. Despite these concerns, the project proceeded with notable involvement from OpenAI's president, Greg Brockman. The resulting dataset proved instrumental in training GPT-4. In February 2024, The Intercept as well as Raw Story and Alternate Media Inc. filed lawsuit against OpenAI on copyright litigation ground. The lawsuit is said to have charted a new legal strategy for digital-only publishers to sue OpenAI. On April 30, 2024, eight newspapers filed a lawsuit in the Southern District of New York against OpenAI and Microsoft, claiming illegal harvesting of their copyrighted articles. The suing publications included The Mercury News, The Denver Post, The Orange County Register, St. Paul Pioneer Press, Chicago Tribune, Orlando Sentinel, Sun Sentinel, and New York Daily News. In June 2023, a lawsuit claimed that OpenAI scraped 300 billion words online without consent and without registering as a data broker. It was filed in San Francisco, California, by sixteen anonymous plaintiffs. They also claimed that OpenAI and its partner as well as customer Microsoft continued to unlawfully collect and use personal data from millions of consumers worldwide to train artificial intelligence models. On May 22, 2024, OpenAI entered into an agreement with News Corp to integrate news content from The Wall Street Journal, the New York Post, The Times, and The Sunday Times into its AI platform. Meanwhile, other publications like The New York Times chose to sue OpenAI and Microsoft for copyright infringement over the use of their content to train AI models. In November 2024, a coalition of Canadian news outlets, including the Toronto Star, Metroland Media, Postmedia, The Globe and Mail, The Canadian Press and CBC, sued OpenAI for using their news articles to train its software without permission. In October 2024 during a New York Times interview, Suchir Balaji accused OpenAI of violating copyright law in developing its commercial LLMs which he had helped engineer. He was a likely witness in a major copyright trial against the AI company, and was one of several of its current or former employees named in court filings as potentially having documents relevant to the case. On November 26, 2024, Balaji died by suicide. His death prompted the circulation of conspiracy theories alleging that he had been deliberately silenced. California Congressman Ro Khanna endorsed calls for an investigation. On April 24, 2025, Ziff Davis sued OpenAI in Delaware federal court for copyright infringement. Ziff Davis is known for publications such as ZDNet, PCMag, CNET, IGN and Lifehacker. In April 2023, the EU's European Data Protection Board (EDPB) formed a dedicated task force on ChatGPT "to foster cooperation and to exchange information on possible enforcement actions conducted by data protection authorities" based on the "enforcement action undertaken by the Italian data protection authority against OpenAI about the ChatGPT service". In late April 2024 NOYB filed a complaint with the Austrian Datenschutzbehörde against OpenAI for violating the European General Data Protection Regulation. A text created with ChatGPT gave a false date of birth for a living person without giving the individual the option to see the personal data used in the process. A request to correct the mistake was denied. Additionally, neither the recipients of ChatGPT's work nor the sources used, could be made available, OpenAI claimed. OpenAI was criticized for lifting its ban on using ChatGPT for "military and warfare". Up until January 10, 2024, its "usage policies" included a ban on "activity that has high risk of physical harm, including", specifically, "weapons development" and "military and warfare". Its new policies prohibit "[using] our service to harm yourself or others" and to "develop or use weapons". In August 2025, the parents of a 16-year-old boy who died by suicide filed a wrongful death lawsuit against OpenAI (and CEO Sam Altman), alleging that months of conversations with ChatGPT about mental health and methods of self-harm contributed to their son's death and that safeguards were inadequate for minors. OpenAI expressed condolences and said it was strengthening protections (including updated crisis response behavior and parental controls). Coverage described it as a first-of-its-kind wrongful death case targeting the company's chatbot. The complaint was filed in California state court in San Francisco. In November 2025, the Social Media Victims Law Center and Tech Justice Law Project filed seven lawsuits against OpenAI, of which four lawsuits alleged wrongful death. The suits were filed on behalf of Zane Shamblin, 23, of Texas; Amaurie Lacey, 17, of Georgia; Joshua Enneking, 26, of Florida; and Joe Ceccanti, 48, of Oregon, who each committed suicide after prolonged ChatGPT usage. In December 2025, Stein-Erik Soelberg, who was 56 years old at the time, allegedly murdered his mother Suzanne Adams. In the months prior the paranoid, delusional man often discussed his ideas with ChatGPT. Adam's estate then sued OpenAI claiming that the company shared responsibility due to the risk of chatbot psychosis despite the fact that chatbot psychosis is not a real medical diagnosis. OpenAI responded saying they will make ChatGPT safer for users disconnected from reality. See also References Further reading External links |
======================================== |
[SOURCE: https://en.wikipedia.org/wiki/OpenAI#cite_ref-149] | [TOKENS: 8773] |
Contents OpenAI OpenAI is an American artificial intelligence research organization comprising both a non-profit foundation and a controlled for-profit public benefit corporation (PBC), headquartered in San Francisco. It aims to develop "safe and beneficial" artificial general intelligence (AGI), which it defines as "highly autonomous systems that outperform humans at most economically valuable work". OpenAI is widely recognized for its development of the GPT family of large language models, the DALL-E series of text-to-image models, and the Sora series of text-to-video models, which have influenced industry research and commercial applications. Its release of ChatGPT in November 2022 has been credited with catalyzing widespread interest in generative AI. The organization was founded in 2015 in Delaware but evolved a complex corporate structure. As of October 2025, following restructuring approved by California and Delaware regulators, the non-profit OpenAI Foundation holds 26% of the for-profit OpenAI Group PBC, with Microsoft holding 27% and employees/other investors holding 47%. Under its governance arrangements, the OpenAI Foundation holds the authority to appoint the board of the for-profit OpenAI Group PBC, a mechanism designed to align the entity’s strategic direction with the Foundation’s charter. Microsoft previously invested over $13 billion into OpenAI, and provides Azure cloud computing resources. In October 2025, OpenAI conducted a $6.6 billion share sale that valued the company at $500 billion. In 2023 and 2024, OpenAI faced multiple lawsuits for alleged copyright infringement against authors and media companies whose work was used to train some of OpenAI's products. In November 2023, OpenAI's board removed Sam Altman as CEO, citing a lack of confidence in him, but reinstated him five days later following a reconstruction of the board. Throughout 2024, roughly half of then-employed AI safety researchers left OpenAI, citing the company's prominent role in an industry-wide problem. Founding In December 2015, OpenAI was founded as a not for profit organization by Sam Altman, Elon Musk, Ilya Sutskever, Greg Brockman, Trevor Blackwell, Vicki Cheung, Andrej Karpathy, Durk Kingma, John Schulman, Pamela Vagata, and Wojciech Zaremba, with Sam Altman and Elon Musk as the co-chairs. A total of $1 billion in capital was pledged by Sam Altman, Greg Brockman, Elon Musk, Reid Hoffman, Jessica Livingston, Peter Thiel, Amazon Web Services (AWS), and Infosys. However, the actual capital collected significantly lagged pledges. According to company disclosures, only $130 million had been received by 2019. In its founding charter, OpenAI stated an intention to collaborate openly with other institutions by making certain patents and research publicly available, but later restricted access to its most capable models, citing competitive and safety concerns. OpenAI was initially run from Brockman's living room. It was later headquartered at the Pioneer Building in the Mission District, San Francisco. According to OpenAI's charter, its founding mission is "to ensure that artificial general intelligence (AGI)—by which we mean highly autonomous systems that outperform humans at most economically valuable work—benefits all of humanity." Musk and Altman stated in 2015 that they were partly motivated by concerns about AI safety and existential risk from artificial general intelligence. OpenAI stated that "it's hard to fathom how much human-level AI could benefit society", and that it is equally difficult to comprehend "how much it could damage society if built or used incorrectly". The startup also wrote that AI "should be an extension of individual human wills and, in the spirit of liberty, as broadly and evenly distributed as possible", and that "because of AI's surprising history, it's hard to predict when human-level AI might come within reach. When it does, it'll be important to have a leading research institution which can prioritize a good outcome for all over its own self-interest." Co-chair Sam Altman expected a decades-long project that eventually surpasses human intelligence. Brockman met with Yoshua Bengio, one of the "founding fathers" of deep learning, and drew up a list of great AI researchers. Brockman was able to hire nine of them as the first employees in December 2015. OpenAI did not pay AI researchers salaries comparable to those of Facebook or Google. It also did not pay stock options which AI researchers typically get. Nevertheless, OpenAI spent $7 million on its first 52 employees in 2016. OpenAI's potential and mission drew these researchers to the firm; a Google employee said he was willing to leave Google for OpenAI "partly because of the very strong group of people and, to a very large extent, because of its mission." OpenAI co-founder Wojciech Zaremba stated that he turned down "borderline crazy" offers of two to three times his market value to join OpenAI instead. In April 2016, OpenAI released a public beta of "OpenAI Gym", its platform for reinforcement learning research. Nvidia gifted its first DGX-1 supercomputer to OpenAI in August 2016 to help it train larger and more complex AI models with the capability of reducing processing time from six days to two hours. In December 2016, OpenAI released "Universe", a software platform for measuring and training an AI's general intelligence across the world's supply of games, websites, and other applications. Corporate structure In 2019, OpenAI transitioned from non-profit to "capped" for-profit, with the profit being capped at 100 times any investment. According to OpenAI, the capped-profit model allows OpenAI Global, LLC to legally attract investment from venture funds and, in addition, to grant employees stakes in the company. Many top researchers work for Google Brain, DeepMind, or Facebook, which offer equity that a nonprofit would be unable to match. Before the transition, OpenAI was legally required to publicly disclose the compensation of its top employees. The company then distributed equity to its employees and partnered with Microsoft, announcing an investment package of $1 billion into the company. Since then, OpenAI systems have run on an Azure-based supercomputing platform from Microsoft. OpenAI Global, LLC then announced its intention to commercially license its technologies. It planned to spend $1 billion "within five years, and possibly much faster". Altman stated that even a billion dollars may turn out to be insufficient, and that the lab may ultimately need "more capital than any non-profit has ever raised" to achieve artificial general intelligence. The nonprofit, OpenAI, Inc., is the sole controlling shareholder of OpenAI Global, LLC, which, despite being a for-profit company, retains a formal fiduciary responsibility to OpenAI, Inc.'s nonprofit charter. A majority of OpenAI, Inc.'s board is barred from having financial stakes in OpenAI Global, LLC. In addition, minority members with a stake in OpenAI Global, LLC are barred from certain votes due to conflict of interest. Some researchers have argued that OpenAI Global, LLC's switch to for-profit status is inconsistent with OpenAI's claims to be "democratizing" AI. On February 29, 2024, Elon Musk filed a lawsuit against OpenAI and CEO Sam Altman, accusing them of shifting focus from public benefit to profit maximization—a case OpenAI dismissed as "incoherent" and "frivolous," though Musk later revived legal action against Altman and others in August. On April 9, 2024, OpenAI countersued Musk in federal court, alleging that he had engaged in "bad-faith tactics" to slow the company's progress and seize its innovations for his personal benefit. OpenAI also argued that Musk had previously supported the creation of a for-profit structure and had expressed interest in controlling OpenAI himself. The countersuit seeks damages and legal measures to prevent further alleged interference. On February 10, 2025, a consortium of investors led by Elon Musk submitted a $97.4 billion unsolicited bid to buy the nonprofit that controls OpenAI, declaring willingness to match or exceed any better offer. The offer was rejected on 14 February 2025, with OpenAI stating that it was not for sale, but the offer complicated Altman's restructuring plan by suggesting a lower bar for how much the nonprofit should be valued. OpenAI, Inc. was originally designed as a nonprofit in order to ensure that AGI "benefits all of humanity" rather than "the private gain of any person". In 2019, it created OpenAI Global, LLC, a capped-profit subsidiary controlled by the nonprofit. In December 2024, OpenAI proposed a restructuring plan to convert the capped-profit into a Delaware-based public benefit corporation (PBC), and to release it from the control of the nonprofit. The nonprofit would sell its control and other assets, getting equity in return, and would use it to fund and pursue separate charitable projects, including in science and education. OpenAI's leadership described the change as necessary to secure additional investments, and claimed that the nonprofit's founding mission to ensure AGI "benefits all of humanity" would be better fulfilled. The plan has been criticized by former employees. A legal letter named "Not For Private Gain" asked the attorneys general of California and Delaware to intervene, stating that the restructuring is illegal and would remove governance safeguards from the nonprofit and the attorneys general. The letter argues that OpenAI's complex structure was deliberately designed to remain accountable to its mission, without the conflicting pressure of maximizing profits. It contends that the nonprofit is best positioned to advance its mission of ensuring AGI benefits all of humanity by continuing to control OpenAI Global, LLC, whatever the amount of equity that it could get in exchange. PBCs can choose how they balance their mission with profit-making. Controlling shareholders have a large influence on how closely a PBC sticks to its mission. On October 28, 2025, OpenAI announced that it had adopted the new PBC corporate structure after receiving approval from the attorneys general of California and Delaware. Under the new structure, OpenAI's for-profit branch became a public benefit corporation known as OpenAI Group PBC, while the non-profit was renamed to the OpenAI Foundation. The OpenAI Foundation holds a 26% stake in the PBC, while Microsoft holds a 27% stake and the remaining 47% is owned by employees and other investors. All members of the OpenAI Group PBC board of directors will be appointed by the OpenAI Foundation, which can remove them at any time. Members of the Foundation's board will also serve on the for-profit board. The new structure allows the for-profit PBC to raise investor funds like most traditional tech companies, including through an initial public offering, which Altman claimed was the most likely path forward. In January 2023, OpenAI Global, LLC was in talks for funding that would value the company at $29 billion, double its 2021 value. On January 23, 2023, Microsoft announced a new US$10 billion investment in OpenAI Global, LLC over multiple years, partially needed to use Microsoft's cloud-computing service Azure. From September to December, 2023, Microsoft rebranded all variants of its Copilot to Microsoft Copilot, and they added MS-Copilot to many installations of Windows and released Microsoft Copilot mobile apps. Following OpenAI's 2025 restructuring, Microsoft owns a 27% stake in the for-profit OpenAI Group PBC, valued at $135 billion. In a deal announced the same day, OpenAI agreed to purchase $250 billion of Azure services, with Microsoft ceding their right of first refusal over OpenAI's future cloud computing purchases. As part of the deal, OpenAI will continue to share 20% of its revenue with Microsoft until it achieves AGI, which must now be verified by an independent panel of experts. The deal also loosened restrictions on both companies working with third parties, allowing Microsoft to pursue AGI independently and allowing OpenAI to develop products with other companies. In 2017, OpenAI spent $7.9 million, a quarter of its functional expenses, on cloud computing alone. In comparison, DeepMind's total expenses in 2017 were $442 million. In the summer of 2018, training OpenAI's Dota 2 bots required renting 128,000 CPUs and 256 GPUs from Google for multiple weeks. In October 2024, OpenAI completed a $6.6 billion capital raise with a $157 billion valuation including investments from Microsoft, Nvidia, and SoftBank. On January 21, 2025, Donald Trump announced The Stargate Project, a joint venture between OpenAI, Oracle, SoftBank and MGX to build an AI infrastructure system in conjunction with the US government. The project takes its name from OpenAI's existing "Stargate" supercomputer project and is estimated to cost $500 billion. The partners planned to fund the project over the next four years. In July, the United States Department of Defense announced that OpenAI had received a $200 million contract for AI in the military, along with Anthropic, Google, and xAI. In the same month, the company made a deal with the UK Government to use ChatGPT and other AI tools in public services. OpenAI subsequently began a $50 million fund to support nonprofit and community organizations. In April 2025, OpenAI raised $40 billion at a $300 billion post-money valuation, which was the highest-value private technology deal in history. The financing round was led by SoftBank, with other participants including Microsoft, Coatue, Altimeter and Thrive. In July 2025, the company reported annualized revenue of $12 billion. This was an increase from $3.7 billion in 2024, which was driven by ChatGPT subscriptions, which reached 20 million paid subscribers by April 2025, up from 15.5 million at the end of 2024, alongside a rapidly expanding enterprise customer base that grew to five million business users. The company’s cash burn remains high because of the intensive computational costs required to train and operate large language models. It projects an $8 billion operating loss in 2025. OpenAI reports revised long-term spending projections totaling approximately $115 billion through 2029, with annual expenditures projected to escalate significantly, reaching $17 billion in 2026, $35 billion in 2027, and $45 billion in 2028. These expenditures are primarily allocated toward expanding compute infrastructure, developing proprietary AI chips, constructing data centers, and funding intensive model training programs, with more than half of the spending through the end of the decade expected to support research-intensive compute for model training and development. The company's financial strategy prioritizes market expansion and technological advancement over near-term profitability, with OpenAI targeting cash-flow-positive operations by 2029 and projecting revenue of approximately $200 billion by 2030. This aggressive spending trajectory underscores both the enormous capital requirements of scaling cutting-edge AI technology and OpenAI's commitment to maintaining its position as a leader in the artificial intelligence industry. In October 2025, OpenAI completed an employee share sale of up to $10 billion to existing investors which valued the company at $500 billion. The deal values OpenAI as the most valuable privately owned company in the world—surpassing SpaceX as the world's most valuable private company. On November 17, 2023, Sam Altman was removed as CEO when its board of directors (composed of Helen Toner, Ilya Sutskever, Adam D'Angelo and Tasha McCauley) cited a lack of confidence in him. Chief Technology Officer Mira Murati took over as interim CEO. Greg Brockman, the president of OpenAI, was also removed as chairman of the board and resigned from the company's presidency shortly thereafter. Three senior OpenAI researchers subsequently resigned: director of research and GPT-4 lead Jakub Pachocki, head of AI risk Aleksander Mądry, and researcher Szymon Sidor. On November 18, 2023, there were reportedly talks of Altman returning as CEO amid pressure placed upon the board by investors such as Microsoft and Thrive Capital, who objected to Altman's departure. Although Altman himself spoke in favor of returning to OpenAI, he has since stated that he considered starting a new company and bringing former OpenAI employees with him if talks to reinstate him didn't work out. The board members agreed "in principle" to resign if Altman returned. On November 19, 2023, negotiations with Altman to return failed and Murati was replaced by Emmett Shear as interim CEO. The board initially contacted Anthropic CEO Dario Amodei (a former OpenAI executive) about replacing Altman, and proposed a merger of the two companies, but both offers were declined. On November 20, 2023, Microsoft CEO Satya Nadella announced Altman and Brockman would be joining Microsoft to lead a new advanced AI research team, but added that they were still committed to OpenAI despite recent events. Before the partnership with Microsoft was finalized, Altman gave the board another opportunity to negotiate with him. About 738 of OpenAI's 770 employees, including Murati and Sutskever, signed an open letter stating they would quit their jobs and join Microsoft if the board did not rehire Altman and then resign. This prompted OpenAI investors to consider legal action against the board as well. In response, OpenAI management sent an internal memo to employees stating that negotiations with Altman and the board had resumed and would take some time. On November 21, 2023, after continued negotiations, Altman and Brockman returned to the company in their prior roles along with a reconstructed board made up of new members Bret Taylor (as chairman) and Lawrence Summers, with D'Angelo remaining. According to subsequent reporting, shortly before Altman’s firing, some employees raised concerns to the board about how he had handled the safety implications of a recent internal AI capability discovery. On November 29, 2023, OpenAI announced that an anonymous Microsoft employee had joined the board as a non-voting member to observe the company's operations; Microsoft resigned from the board in July 2024. In February 2024, the Securities and Exchange Commission subpoenaed OpenAI's internal communication to determine if Altman's alleged lack of candor misled investors. In 2024, following the temporary removal of Sam Altman and his return, many employees gradually left OpenAI, including most of the original leadership team and a significant number of AI safety researchers. In August 2023, it was announced that OpenAI had acquired the New York-based start-up Global Illumination, a company that deploys AI to develop digital infrastructure and creative tools. In June 2024, OpenAI acquired Multi, a startup focused on remote collaboration. In March 2025, OpenAI reached a deal with CoreWeave to acquire $350 million worth of CoreWeave shares and access to AI infrastructure, in return for $11.9 billion paid over five years. Microsoft was already CoreWeave's biggest customer in 2024. Alongside their other business dealings, OpenAI and Microsoft were renegotiating the terms of their partnership to facilitate a potential future initial public offering by OpenAI, while ensuring Microsoft's continued access to advanced AI models. On May 21, OpenAI announced the $6.5 billion acquisition of io, an AI hardware start-up founded by former Apple designer Jony Ive in 2024. In September 2025, OpenAI agreed to acquire the product testing startup Statsig for $1.1 billion in an all-stock deal and appointed Statsig's founding CEO Vijaye Raji as OpenAI's chief technology officer of applications. The company also announced development of an AI-driven hiring service designed to rival LinkedIn. OpenAI acquired personal finance app Roi in October 2025. In October 2025, OpenAI acquired Software Applications Incorporated, the developer of Sky, a macOS-based natural language interface designed to operate across desktop applications. The Sky team joined OpenAI, and the company announced plans to integrate Sky’s capabilities into ChatGPT. In December 2025, it was announced OpenAI had agreed to acquire Neptune, an AI tooling startup that helps companies track and manage model training, for an undisclosed amount. In January 2026, it was announced OpenAI had acquired healthcare technology startup Torch for approximately $60 million. The acquisition followed the launch of OpenAI’s ChatGPT Health product and was intended to strengthen the company’s medical data and healthcare artificial intelligence capabilities. OpenAI has been criticized for outsourcing the annotation of data sets to Sama, a company based in San Francisco that employed workers in Kenya. These annotations were used to train an AI model to detect toxicity, which could then be used to moderate toxic content, notably from ChatGPT's training data and outputs. However, these pieces of text usually contained detailed descriptions of various types of violence, including sexual violence. The investigation uncovered that OpenAI began sending snippets of data to Sama as early as November 2021. The four Sama employees interviewed by Time described themselves as mentally scarred. OpenAI paid Sama $12.50 per hour of work, and Sama was redistributing the equivalent of between $1.32 and $2.00 per hour post-tax to its annotators. Sama's spokesperson said that the $12.50 was also covering other implicit costs, among which were infrastructure expenses, quality assurance and management. In 2024, OpenAI began collaborating with Broadcom to design a custom AI chip capable of both training and inference, targeted for mass production in 2026 and to be manufactured by TSMC on a 3 nm process node. This initiative intended to reduce OpenAI's dependence on Nvidia GPUs, which are costly and face high demand in the market. In January 2024, Arizona State University purchased ChatGPT Enterprise in OpenAI's first deal with a university. In June 2024, Apple Inc. signed a contract with OpenAI to integrate ChatGPT features into its products as part of its new Apple Intelligence initiative. In June 2025, OpenAI began renting Google Cloud's Tensor Processing Units (TPUs) to support ChatGPT and related services, marking its first meaningful use of non‑Nvidia AI chips. In September 2025, it was revealed that OpenAI signed a contract with Oracle to purchase $300 billion in computing power over the next five years. In September 2025, OpenAI and NVIDIA announced a memorandum of understanding that included a potential deployment of at least 10 gigawatts of NVIDIA systems and a $100 billion investment from NVIDIA in OpenAI. OpenAI expected the negotiations to be completed within weeks. As of January 2026, this has not been realized, and the two sides are rethinking the future of their partnership. In October 2025, OpenAI announced a multi-billion dollar deal with AMD. OpenAI committed to purchasing six gigawatts worth of AMD chips, starting with the MI450. OpenAI will have the option to buy up to 160 million shares of AMD, about 10% of the company, depending on development, performance and share price targets. In December 2025, Disney said it would make a $1 billion investment in OpenAI, and signed a three-year licensing deal that will let users generate videos using Sora—OpenAI's short-form AI video platform. More than 200 Disney, Marvel, Star Wars and Pixar characters will be available to OpenAI users. In early 2026, Amazon entered advanced discussions to invest up to $50 billion in OpenAI as part of a potential artificial intelligence partnership. Under the proposed agreement, OpenAI’s models could be integrated into Amazon’s digital assistant Alexa and other internal projects. OpenAI provides LLMs to the Artificial Intelligence Cyber Challenge and to the Advanced Research Projects Agency for Health. In October 2024, The Intercept revealed that OpenAI's tools are considered "essential" for AFRICOM's mission and included in an "Exception to Fair Opportunity" contractual agreement between the United States Department of Defense and Microsoft. In December 2024, OpenAI said it would partner with defense-tech company Anduril to build drone defense technologies for the United States and its allies. In 2025, OpenAI's Chief Product Officer, Kevin Weil, was commissioned lieutenant colonel in the U.S. Army to join Detachment 201 as senior advisor. In June 2025, the U.S. Department of Defense awarded OpenAI a $200 million one-year contract to develop AI tools for military and national security applications. OpenAI announced a new program, OpenAI for Government, to give federal, state, and local governments access to its models, including ChatGPT. Services In February 2019, GPT-2 was announced, which gained attention for its ability to generate human-like text. In 2020, OpenAI announced GPT-3, a language model trained on large internet datasets. GPT-3 is aimed at natural language answering questions, but it can also translate between languages and coherently generate improvised text. It also announced that an associated API, named the API, would form the heart of its first commercial product. Eleven employees left OpenAI, mostly between December 2020 and January 2021, in order to establish Anthropic. In 2021, OpenAI introduced DALL-E, a specialized deep learning model adept at generating complex digital images from textual descriptions, utilizing a variant of the GPT-3 architecture. In December 2022, OpenAI received widespread media coverage after launching a free preview of ChatGPT, its new AI chatbot based on GPT-3.5. According to OpenAI, the preview received over a million signups within the first five days. According to anonymous sources cited by Reuters in December 2022, OpenAI Global, LLC was projecting $200 million of revenue in 2023 and $1 billion in revenue in 2024. After ChatGPT was launched, Google announced a similar chatbot, Bard, amid internal concerns that ChatGPT could threaten Google’s position as a primary source of online information. On February 7, 2023, Microsoft announced that it was building AI technology based on the same foundation as ChatGPT into Microsoft Bing, Edge, Microsoft 365 and other products. On March 14, 2023, OpenAI released GPT-4, both as an API (with a waitlist) and as a feature of ChatGPT Plus. On November 6, 2023, OpenAI launched GPTs, allowing individuals to create customized versions of ChatGPT for specific purposes, further expanding the possibilities of AI applications across various industries. On November 14, 2023, OpenAI announced they temporarily suspended new sign-ups for ChatGPT Plus due to high demand. Access for newer subscribers re-opened a month later on December 13. In December 2024, the company launched the Sora model. It also launched OpenAI o1, an early reasoning model that was internally codenamed strawberry. Additionally, ChatGPT Pro—a $200/month subscription service offering unlimited o1 access and enhanced voice features—was introduced, and preliminary benchmark results for the upcoming OpenAI o3 models were shared. On January 23, 2025, OpenAI released Operator, an AI agent and web automation tool for accessing websites to execute goals defined by users. The feature was only available to Pro users in the United States. OpenAI released deep research agent, nine days later. It scored a 27% accuracy on the benchmark Humanity's Last Exam (HLE). Altman later stated GPT-4.5 would be the last model without full chain-of-thought reasoning. In July 2025, reports indicated that AI models by both OpenAI and Google DeepMind solved mathematics problems at the level of top-performing students in the International Mathematical Olympiad. OpenAI's large language model was able to achieve gold medal-level performance, reflecting significant progress in AI's reasoning abilities. On October 6, 2025, OpenAI unveiled its Agent Builder platform during the company's DevDay event. The platform includes a visual drag-and-drop interface that lets developers and businesses design, test, and deploy agentic workflows with limited coding. On October 21, 2025, OpenAI introduced ChatGPT Atlas, a browser integrating the ChatGPT assistant directly into web navigation, to compete with existing browsers such as Google Chrome and Apple Safari. On December 11, 2025, OpenAI announced GPT-5.2. This model will be better at creating spreadsheets, building presentations, perceiving images, writing code and understanding long context. On January 27, 2026, OpenAI introduced Prism, a LaTeX-native workspace meant to assist scientists to help with research and writing. The platform utilizes GPT-5.2 as a backend to automate the process of drafting for scientific papers, including features for managing citations, complex equation formatting, and real-time collaborative editing. In March 2023, the company was criticized for disclosing particularly few technical details about products like GPT-4, contradicting its initial commitment to openness and making it harder for independent researchers to replicate its work and develop safeguards. OpenAI cited competitiveness and safety concerns to justify this repudiation. OpenAI's former chief scientist Ilya Sutskever argued in 2023 that open-sourcing increasingly capable models was increasingly risky, and that the safety reasons for not open-sourcing the most potent AI models would become "obvious" in a few years. In September 2025, OpenAI published a study on how people use ChatGPT for everyday tasks. The study found that "non-work tasks" (according to an LLM-based classifier) account for more than 72 percent of all ChatGPT usage, with a minority of overall usage related to business productivity. In July 2023, OpenAI launched the superalignment project, aiming within four years to determine how to align future superintelligent systems. OpenAI promised to dedicate 20% of its computing resources to the project, although the team denied receiving anything close to 20%. OpenAI ended the project in May 2024 after its co-leaders Ilya Sutskever and Jan Leike left the company. In August 2025, OpenAI was criticized after thousands of private ChatGPT conversations were inadvertently exposed to public search engines like Google due to an experimental "share with search engines" feature. The opt-in toggle, intended to allow users to make specific chats discoverable, resulted in some discussions including personal details such as names, locations, and intimate topics appearing in search results when users accidentally enabled it while sharing links. OpenAI announced the feature's permanent removal on August 1, 2025, and the company began coordinating with search providers to remove the exposed content, emphasizing that it was not a security breach but a design flaw that heightened privacy risks. CEO Sam Altman acknowledged the issue in a podcast, noting users often treat ChatGPT as a confidant for deeply personal matters, which amplified concerns about AI handling sensitive data. Management In 2018, Musk resigned from his Board of Directors seat, citing "a potential future conflict [of interest]" with his role as CEO of Tesla due to Tesla's AI development for self-driving cars. OpenAI stated that Musk's financial contributions were below $45 million. On March 3, 2023, Reid Hoffman resigned from his board seat, citing a desire to avoid conflicts of interest with his investments in AI companies via Greylock Partners, and his co-founding of the AI startup Inflection AI. Hoffman remained on the board of Microsoft, a major investor in OpenAI. In May 2024, Chief Scientist Ilya Sutskever resigned and was succeeded by Jakub Pachocki. Co-leader Jan Leike also departed amid concerns over safety and trust. OpenAI then signed deals with Reddit, News Corp, Axios, and Vox Media. Paul Nakasone then joined the board of OpenAI. In August 2024, cofounder John Schulman left OpenAI to join Anthropic, and OpenAI's president Greg Brockman took extended leave until November. In September 2024, CTO Mira Murati left the company. In November 2025, Lawrence Summers resigned from the board of directors. Governance and legal issues In May 2023, Sam Altman, Greg Brockman and Ilya Sutskever posted recommendations for the governance of superintelligence. They stated that superintelligence could happen within the next 10 years, allowing a "dramatically more prosperous future" and that "given the possibility of existential risk, we can't just be reactive". They proposed creating an international watchdog organization similar to IAEA to oversee AI systems above a certain capability threshold, suggesting that relatively weak AI systems on the other side should not be overly regulated. They also called for more technical safety research for superintelligences, and asked for more coordination, for example through governments launching a joint project which "many current efforts become part of". In July 2023, the FTC issued a civil investigative demand to OpenAI to investigate whether the company's data security and privacy practices to develop ChatGPT were unfair or harmed consumers (including by reputational harm) in violation of Section 5 of the Federal Trade Commission Act of 1914. These are typically preliminary investigative matters and are nonpublic, but the FTC's document was leaked. In July 2023, the FTC launched an investigation into OpenAI over allegations that the company scraped public data and published false and defamatory information. They asked OpenAI for comprehensive information about its technology and privacy safeguards, as well as any steps taken to prevent the recurrence of situations in which its chatbot generated false and derogatory content about people. The agency also raised concerns about ‘circular’ spending arrangements—for example, Microsoft extending Azure credits to OpenAI while both companies shared engineering talent—and warned that such structures could negatively affect the public. In September 2024, OpenAI's global affairs chief endorsed the UK's "smart" AI regulation during testimony to a House of Lords committee. In February 2025, OpenAI CEO Sam Altman stated that the company is interested in collaborating with the People's Republic of China, despite regulatory restrictions imposed by the U.S. government. This shift comes in response to the growing influence of the Chinese artificial intelligence company DeepSeek, which has disrupted the AI market with open models, including DeepSeek V3 and DeepSeek R1. Following DeepSeek's market emergence, OpenAI enhanced security protocols to protect proprietary development techniques from industrial espionage. Some industry observers noted similarities between DeepSeek's model distillation approach and OpenAI's methodology, though no formal intellectual property claim was filed. According to Oliver Roberts, in March 2025, the United States had 781 state AI bills or laws. OpenAI advocated for preempting state AI laws with federal laws. According to Scott Kohler, OpenAI has opposed California's AI legislation and suggested that the state bill encroaches on a more competent federal government. Public Citizen opposed a federal preemption on AI and pointed to OpenAI's growth and valuation as evidence that existing state laws have not hampered innovation. Before May 2024, OpenAI required departing employees to sign a lifelong non-disparagement agreement forbidding them from criticizing OpenAI and acknowledging the existence of the agreement. Daniel Kokotajlo, a former employee, publicly stated that he forfeited his vested equity in OpenAI in order to leave without signing the agreement. Sam Altman stated that he was unaware of the equity cancellation provision, and that OpenAI never enforced it to cancel any employee's vested equity. However, leaked documents and emails refute this claim. On May 23, 2024, OpenAI sent a memo releasing former employees from the agreement. OpenAI was sued for copyright infringement by authors Sarah Silverman, Matthew Butterick, Paul Tremblay and Mona Awad in July 2023. In September 2023, 17 authors, including George R. R. Martin, John Grisham, Jodi Picoult and Jonathan Franzen, joined the Authors Guild in filing a class action lawsuit against OpenAI, alleging that the company's technology was illegally using their copyrighted work. The New York Times also sued the company in late December 2023. In May 2024 it was revealed that OpenAI had destroyed its Books1 and Books2 training datasets, which were used in the training of GPT-3, and which the Authors Guild believed to have contained over 100,000 copyrighted books. In 2021, OpenAI developed a speech recognition tool called Whisper. OpenAI used it to transcribe more than one million hours of YouTube videos into text for training GPT-4. The automated transcription of YouTube videos raised concerns within OpenAI employees regarding potential violations of YouTube's terms of service, which prohibit the use of videos for applications independent of the platform, as well as any type of automated access to its videos. Despite these concerns, the project proceeded with notable involvement from OpenAI's president, Greg Brockman. The resulting dataset proved instrumental in training GPT-4. In February 2024, The Intercept as well as Raw Story and Alternate Media Inc. filed lawsuit against OpenAI on copyright litigation ground. The lawsuit is said to have charted a new legal strategy for digital-only publishers to sue OpenAI. On April 30, 2024, eight newspapers filed a lawsuit in the Southern District of New York against OpenAI and Microsoft, claiming illegal harvesting of their copyrighted articles. The suing publications included The Mercury News, The Denver Post, The Orange County Register, St. Paul Pioneer Press, Chicago Tribune, Orlando Sentinel, Sun Sentinel, and New York Daily News. In June 2023, a lawsuit claimed that OpenAI scraped 300 billion words online without consent and without registering as a data broker. It was filed in San Francisco, California, by sixteen anonymous plaintiffs. They also claimed that OpenAI and its partner as well as customer Microsoft continued to unlawfully collect and use personal data from millions of consumers worldwide to train artificial intelligence models. On May 22, 2024, OpenAI entered into an agreement with News Corp to integrate news content from The Wall Street Journal, the New York Post, The Times, and The Sunday Times into its AI platform. Meanwhile, other publications like The New York Times chose to sue OpenAI and Microsoft for copyright infringement over the use of their content to train AI models. In November 2024, a coalition of Canadian news outlets, including the Toronto Star, Metroland Media, Postmedia, The Globe and Mail, The Canadian Press and CBC, sued OpenAI for using their news articles to train its software without permission. In October 2024 during a New York Times interview, Suchir Balaji accused OpenAI of violating copyright law in developing its commercial LLMs which he had helped engineer. He was a likely witness in a major copyright trial against the AI company, and was one of several of its current or former employees named in court filings as potentially having documents relevant to the case. On November 26, 2024, Balaji died by suicide. His death prompted the circulation of conspiracy theories alleging that he had been deliberately silenced. California Congressman Ro Khanna endorsed calls for an investigation. On April 24, 2025, Ziff Davis sued OpenAI in Delaware federal court for copyright infringement. Ziff Davis is known for publications such as ZDNet, PCMag, CNET, IGN and Lifehacker. In April 2023, the EU's European Data Protection Board (EDPB) formed a dedicated task force on ChatGPT "to foster cooperation and to exchange information on possible enforcement actions conducted by data protection authorities" based on the "enforcement action undertaken by the Italian data protection authority against OpenAI about the ChatGPT service". In late April 2024 NOYB filed a complaint with the Austrian Datenschutzbehörde against OpenAI for violating the European General Data Protection Regulation. A text created with ChatGPT gave a false date of birth for a living person without giving the individual the option to see the personal data used in the process. A request to correct the mistake was denied. Additionally, neither the recipients of ChatGPT's work nor the sources used, could be made available, OpenAI claimed. OpenAI was criticized for lifting its ban on using ChatGPT for "military and warfare". Up until January 10, 2024, its "usage policies" included a ban on "activity that has high risk of physical harm, including", specifically, "weapons development" and "military and warfare". Its new policies prohibit "[using] our service to harm yourself or others" and to "develop or use weapons". In August 2025, the parents of a 16-year-old boy who died by suicide filed a wrongful death lawsuit against OpenAI (and CEO Sam Altman), alleging that months of conversations with ChatGPT about mental health and methods of self-harm contributed to their son's death and that safeguards were inadequate for minors. OpenAI expressed condolences and said it was strengthening protections (including updated crisis response behavior and parental controls). Coverage described it as a first-of-its-kind wrongful death case targeting the company's chatbot. The complaint was filed in California state court in San Francisco. In November 2025, the Social Media Victims Law Center and Tech Justice Law Project filed seven lawsuits against OpenAI, of which four lawsuits alleged wrongful death. The suits were filed on behalf of Zane Shamblin, 23, of Texas; Amaurie Lacey, 17, of Georgia; Joshua Enneking, 26, of Florida; and Joe Ceccanti, 48, of Oregon, who each committed suicide after prolonged ChatGPT usage. In December 2025, Stein-Erik Soelberg, who was 56 years old at the time, allegedly murdered his mother Suzanne Adams. In the months prior the paranoid, delusional man often discussed his ideas with ChatGPT. Adam's estate then sued OpenAI claiming that the company shared responsibility due to the risk of chatbot psychosis despite the fact that chatbot psychosis is not a real medical diagnosis. OpenAI responded saying they will make ChatGPT safer for users disconnected from reality. See also References Further reading External links |
======================================== |
[SOURCE: https://en.wikipedia.org/wiki/Social_network#cite_ref-74] | [TOKENS: 5247] |
Contents Social network 1800s: Martineau · Tocqueville · Marx · Spencer · Le Bon · Ward · Pareto · Tönnies · Veblen · Simmel · Durkheim · Addams · Mead · Weber · Du Bois · Mannheim · Elias A social network is a social structure consisting of a set of social actors (such as individuals or organizations), networks of dyadic ties, and other social interactions between actors. The social network perspective provides a set of methods for analyzing the structure of whole social entities along with a variety of theories explaining the patterns observed in these structures. The study of these structures uses social network analysis to identify local and global patterns, locate influential entities, and examine dynamics of networks. For instance, social network analysis has been used in studying the spread of misinformation on social media platforms or analyzing the influence of key figures in social networks. Social networks and the analysis of them is an inherently interdisciplinary academic field which emerged from social psychology, sociology, statistics, and graph theory. Georg Simmel authored early structural theories in sociology emphasizing the dynamics of triads and "web of group affiliations". Jacob Moreno is credited with developing the first sociograms in the 1930s to study interpersonal relationships. These approaches were mathematically formalized in the 1950s and theories and methods of social networks became pervasive in the social and behavioral sciences by the 1980s. Social network analysis is now one of the major paradigms in contemporary sociology, and is also employed in a number of other social and formal sciences. Together with other complex networks, it forms part of the nascent field of network science. Overview The social network is a theoretical construct useful in the social sciences to study relationships between individuals, groups, organizations, or even entire societies (social units, see differentiation). The term is used to describe a social structure determined by such interactions. The ties through which any given social unit connects represent the convergence of the various social contacts of that unit. This theoretical approach is, necessarily, relational. An axiom of the social network approach to understanding social interaction is that social phenomena should be primarily conceived and investigated through the properties of relations between and within units, instead of the properties of these units themselves. Thus, one common criticism of social network theory is that individual agency is often ignored although this may not be the case in practice (see agent-based modeling). Precisely because many different types of relations, singular or in combination, form these network configurations, network analytics are useful to a broad range of research enterprises. In social science, these fields of study include, but are not limited to anthropology, biology, communication studies, economics, geography, information science, organizational studies, social psychology, sociology, and sociolinguistics. History In the late 1890s, both Émile Durkheim and Ferdinand Tönnies foreshadowed the idea of social networks in their theories and research of social groups. Tönnies argued that social groups can exist as personal and direct social ties that either link individuals who share values and belief (Gemeinschaft, German, commonly translated as "community") or impersonal, formal, and instrumental social links (Gesellschaft, German, commonly translated as "society"). Durkheim gave a non-individualistic explanation of social facts, arguing that social phenomena arise when interacting individuals constitute a reality that can no longer be accounted for in terms of the properties of individual actors. Georg Simmel, writing at the turn of the twentieth century, pointed to the nature of networks and the effect of network size on interaction and examined the likelihood of interaction in loosely knit networks rather than groups. Major developments in the field can be seen in the 1930s by several groups in psychology, anthropology, and mathematics working independently. In psychology, in the 1930s, Jacob L. Moreno began systematic recording and analysis of social interaction in small groups, especially classrooms and work groups (see sociometry). In anthropology, the foundation for social network theory is the theoretical and ethnographic work of Bronislaw Malinowski, Alfred Radcliffe-Brown, and Claude Lévi-Strauss. A group of social anthropologists associated with Max Gluckman and the Manchester School, including John A. Barnes, J. Clyde Mitchell and Elizabeth Bott Spillius, often are credited with performing some of the first fieldwork from which network analyses were performed, investigating community networks in southern Africa, India and the United Kingdom. Concomitantly, British anthropologist S. F. Nadel codified a theory of social structure that was influential in later network analysis. In sociology, the early (1930s) work of Talcott Parsons set the stage for taking a relational approach to understanding social structure. Later, drawing upon Parsons' theory, the work of sociologist Peter Blau provides a strong impetus for analyzing the relational ties of social units with his work on social exchange theory. By the 1970s, a growing number of scholars worked to combine the different tracks and traditions. One group consisted of sociologist Harrison White and his students at the Harvard University Department of Social Relations. Also independently active in the Harvard Social Relations department at the time were Charles Tilly, who focused on networks in political and community sociology and social movements, and Stanley Milgram, who developed the "six degrees of separation" thesis. Mark Granovetter and Barry Wellman are among the former students of White who elaborated and championed the analysis of social networks. Beginning in the late 1990s, social network analysis experienced work by sociologists, political scientists, and physicists such as Duncan J. Watts, Albert-László Barabási, Peter Bearman, Nicholas A. Christakis, James H. Fowler, and others, developing and applying new models and methods to emerging data available about online social networks, as well as "digital traces" regarding face-to-face networks. Levels of analysis In general, social networks are self-organizing, emergent, and complex, such that a globally coherent pattern appears from the local interaction of the elements that make up the system. These patterns become more apparent as network size increases. However, a global network analysis of, for example, all interpersonal relationships in the world is not feasible and is likely to contain so much information as to be uninformative. Practical limitations of computing power, ethics and participant recruitment and payment also limit the scope of a social network analysis. The nuances of a local system may be lost in a large network analysis, hence the quality of information may be more important than its scale for understanding network properties. Thus, social networks are analyzed at the scale relevant to the researcher's theoretical question. Although levels of analysis are not necessarily mutually exclusive, there are three general levels into which networks may fall: micro-level, meso-level, and macro-level. At the micro-level, social network research typically begins with an individual, snowballing as social relationships are traced, or may begin with a small group of individuals in a particular social context. Dyadic level: A dyad is a social relationship between two individuals. Network research on dyads may concentrate on structure of the relationship (e.g. multiplexity, strength), social equality, and tendencies toward reciprocity/mutuality. Triadic level: Add one individual to a dyad, and you have a triad. Research at this level may concentrate on factors such as balance and transitivity, as well as social equality and tendencies toward reciprocity/mutuality. In the balance theory of Fritz Heider the triad is the key to social dynamics. The discord in a rivalrous love triangle is an example of an unbalanced triad, likely to change to a balanced triad by a change in one of the relations. The dynamics of social friendships in society has been modeled by balancing triads. The study is carried forward with the theory of signed graphs. Actor level: The smallest unit of analysis in a social network is an individual in their social setting, i.e., an "actor" or "ego." Egonetwork analysis focuses on network characteristics, such as size, relationship strength, density, centrality, prestige and roles such as isolates, liaisons, and bridges. Such analyses, are most commonly used in the fields of psychology or social psychology, ethnographic kinship analysis or other genealogical studies of relationships between individuals. Subset level: Subset levels of network research problems begin at the micro-level, but may cross over into the meso-level of analysis. Subset level research may focus on distance and reachability, cliques, cohesive subgroups, or other group actions or behavior. In general, meso-level theories begin with a population size that falls between the micro- and macro-levels. However, meso-level may also refer to analyses that are specifically designed to reveal connections between micro- and macro-levels. Meso-level networks are low density and may exhibit causal processes distinct from interpersonal micro-level networks. Organizations: Formal organizations are social groups that distribute tasks for a collective goal. Network research on organizations may focus on either intra-organizational or inter-organizational ties in terms of formal or informal relationships. Intra-organizational networks themselves often contain multiple levels of analysis, especially in larger organizations with multiple branches, franchises or semi-autonomous departments. In these cases, research is often conducted at a work group level and organization level, focusing on the interplay between the two structures. Experiments with networked groups online have documented ways to optimize group-level coordination through diverse interventions, including the addition of autonomous agents to the groups. Randomly distributed networks: Exponential random graph models of social networks became state-of-the-art methods of social network analysis in the 1980s. This framework has the capacity to represent social-structural effects commonly observed in many human social networks, including general degree-based structural effects commonly observed in many human social networks as well as reciprocity and transitivity, and at the node-level, homophily and attribute-based activity and popularity effects, as derived from explicit hypotheses about dependencies among network ties. Parameters are given in terms of the prevalence of small subgraph configurations in the network and can be interpreted as describing the combinations of local social processes from which a given network emerges. These probability models for networks on a given set of actors allow generalization beyond the restrictive dyadic independence assumption of micro-networks, allowing models to be built from theoretical structural foundations of social behavior. Scale-free networks: A scale-free network is a network whose degree distribution follows a power law, at least asymptotically. In network theory a scale-free ideal network is a random network with a degree distribution that unravels the size distribution of social groups. Specific characteristics of scale-free networks vary with the theories and analytical tools used to create them, however, in general, scale-free networks have some common characteristics. One notable characteristic in a scale-free network is the relative commonness of vertices with a degree that greatly exceeds the average. The highest-degree nodes are often called "hubs", and may serve specific purposes in their networks, although this depends greatly on the social context. Another general characteristic of scale-free networks is the clustering coefficient distribution, which decreases as the node degree increases. This distribution also follows a power law. The Barabási model of network evolution shown above is an example of a scale-free network. Rather than tracing interpersonal interactions, macro-level analyses generally trace the outcomes of interactions, such as economic or other resource transfer interactions over a large population. Large-scale networks: Large-scale network is a term somewhat synonymous with "macro-level." It is primarily used in social and behavioral sciences, and in economics. Originally, the term was used extensively in the computer sciences (see large-scale network mapping). Complex networks: Most larger social networks display features of social complexity, which involves substantial non-trivial features of network topology, with patterns of complex connections between elements that are neither purely regular nor purely random (see, complexity science, dynamical system and chaos theory), as do biological, and technological networks. Such complex network features include a heavy tail in the degree distribution, a high clustering coefficient, assortativity or disassortativity among vertices, community structure (see stochastic block model), and hierarchical structure. In the case of agency-directed networks these features also include reciprocity, triad significance profile (TSP, see network motif), and other features. In contrast, many of the mathematical models of networks that have been studied in the past, such as lattices and random graphs, do not show these features. Theoretical links Various theoretical frameworks have been imported for the use of social network analysis. The most prominent of these are Graph theory, Balance theory, Social comparison theory, and more recently, the Social identity approach. Few complete theories have been produced from social network analysis. Two that have are structural role theory and heterophily theory. The basis of Heterophily Theory was the finding in one study that more numerous weak ties can be important in seeking information and innovation, as cliques have a tendency to have more homogeneous opinions as well as share many common traits. This homophilic tendency was the reason for the members of the cliques to be attracted together in the first place. However, being similar, each member of the clique would also know more or less what the other members knew. To find new information or insights, members of the clique will have to look beyond the clique to its other friends and acquaintances. This is what Granovetter called "the strength of weak ties". Structural holes In the context of networks, social capital exists where people have an advantage because of their location in a network. Contacts in a network provide information, opportunities and perspectives that can be beneficial to the central player in the network. Most social structures tend to be characterized by dense clusters of strong connections. Information within these clusters tends to be rather homogeneous and redundant. Non-redundant information is most often obtained through contacts in different clusters. When two separate clusters possess non-redundant information, there is said to be a structural hole between them. Thus, a network that bridges structural holes will provide network benefits that are in some degree additive, rather than overlapping. An ideal network structure has a vine and cluster structure, providing access to many different clusters and structural holes. Networks rich in structural holes are a form of social capital in that they offer information benefits. The main player in a network that bridges structural holes is able to access information from diverse sources and clusters. For example, in business networks, this is beneficial to an individual's career because he is more likely to hear of job openings and opportunities if his network spans a wide range of contacts in different industries/sectors. This concept is similar to Mark Granovetter's theory of weak ties, which rests on the basis that having a broad range of contacts is most effective for job attainment. Structural holes have been widely applied in social network analysis, resulting in applications in a wide range of practical scenarios as well as machine learning-based social prediction. Research clusters Research has used network analysis to examine networks created when artists are exhibited together in museum exhibition. Such networks have been shown to affect an artist's recognition in history and historical narratives, even when controlling for individual accomplishments of the artist. Other work examines how network grouping of artists can affect an individual artist's auction performance. An artist's status has been shown to increase when associated with higher status networks, though this association has diminishing returns over an artist's career. In J.A. Barnes' day, a "community" referred to a specific geographic location and studies of community ties had to do with who talked, associated, traded, and attended church with whom. Today, however, there are extended "online" communities developed through telecommunications devices and social network services. Such devices and services require extensive and ongoing maintenance and analysis, often using network science methods. Community development studies, today, also make extensive use of such methods. Complex networks require methods specific to modelling and interpreting social complexity and complex adaptive systems, including techniques of dynamic network analysis. Mechanisms such as Dual-phase evolution explain how temporal changes in connectivity contribute to the formation of structure in social networks. The study of social networks is being used to examine the nature of interdependencies between actors and the ways in which these are related to outcomes of conflict and cooperation. Areas of study include cooperative behavior among participants in collective actions such as protests; promotion of peaceful behavior, social norms, and public goods within communities through networks of informal governance; the role of social networks in both intrastate conflict and interstate conflict; and social networking among politicians, constituents, and bureaucrats. In criminology and urban sociology, much attention has been paid to the social networks among criminal actors. For example, murders can be seen as a series of exchanges between gangs. Murders can be seen to diffuse outwards from a single source, because weaker gangs cannot afford to kill members of stronger gangs in retaliation, but must commit other violent acts to maintain their reputation for strength. Diffusion of ideas and innovations studies focus on the spread and use of ideas from one actor to another or one culture and another. This line of research seeks to explain why some become "early adopters" of ideas and innovations, and links social network structure with facilitating or impeding the spread of an innovation. A case in point is the social diffusion of linguistic innovation such as neologisms. Experiments and large-scale field trials (e.g., by Nicholas Christakis and collaborators) have shown that cascades of desirable behaviors can be induced in social groups, in settings as diverse as Honduras villages, Indian slums, or in the lab. Still other experiments have documented the experimental induction of social contagion of voting behavior, emotions, risk perception, and commercial products. In demography, the study of social networks has led to new sampling methods for estimating and reaching populations that are hard to enumerate (for example, homeless people or intravenous drug users.) For example, respondent driven sampling is a network-based sampling technique that relies on respondents to a survey recommending further respondents. The field of sociology focuses almost entirely on networks of outcomes of social interactions. More narrowly, economic sociology considers behavioral interactions of individuals and groups through social capital and social "markets". Sociologists, such as Mark Granovetter, have developed core principles about the interactions of social structure, information, ability to punish or reward, and trust that frequently recur in their analyses of political, economic and other institutions. Granovetter examines how social structures and social networks can affect economic outcomes like hiring, price, productivity and innovation and describes sociologists' contributions to analyzing the impact of social structure and networks on the economy. Analysis of social networks is increasingly incorporated into health care analytics, not only in epidemiological studies but also in models of patient communication and education, disease prevention, mental health diagnosis and treatment, and in the study of health care organizations and systems. Human ecology is an interdisciplinary and transdisciplinary study of the relationship between humans and their natural, social, and built environments. The scientific philosophy of human ecology has a diffuse history with connections to geography, sociology, psychology, anthropology, zoology, and natural ecology. In the study of literary systems, network analysis has been applied by Anheier, Gerhards and Romo, De Nooy, Senekal, and Lotker, to study various aspects of how literature functions. The basic premise is that polysystem theory, which has been around since the writings of Even-Zohar, can be integrated with network theory and the relationships between different actors in the literary network, e.g. writers, critics, publishers, literary histories, etc., can be mapped using visualization from SNA. Research studies of formal or informal organization relationships, organizational communication, economics, economic sociology, and other resource transfers. Social networks have also been used to examine how organizations interact with each other, characterizing the many informal connections that link executives together, as well as associations and connections between individual employees at different organizations. Many organizational social network studies focus on teams. Within team network studies, research assesses, for example, the predictors and outcomes of centrality and power, density and centralization of team instrumental and expressive ties, and the role of between-team networks. Intra-organizational networks have been found to affect organizational commitment, organizational identification, interpersonal citizenship behaviour. Social capital is a form of economic and cultural capital in which social networks are central, transactions are marked by reciprocity, trust, and cooperation, and market agents produce goods and services not mainly for themselves, but for a common good. Social capital is split into three dimensions: the structural, the relational and the cognitive dimension. The structural dimension describes how partners interact with each other and which specific partners meet in a social network. Also, the structural dimension of social capital indicates the level of ties among organizations. This dimension is highly connected to the relational dimension which refers to trustworthiness, norms, expectations and identifications of the bonds between partners. The relational dimension explains the nature of these ties which is mainly illustrated by the level of trust accorded to the network of organizations. The cognitive dimension analyses the extent to which organizations share common goals and objectives as a result of their ties and interactions. Social capital is a sociological concept about the value of social relations and the role of cooperation and confidence to achieve positive outcomes. The term refers to the value one can get from their social ties. For example, newly arrived immigrants can make use of their social ties to established migrants to acquire jobs they may otherwise have trouble getting (e.g., because of unfamiliarity with the local language). A positive relationship exists between social capital and the intensity of social network use. In a dynamic framework, higher activity in a network feeds into higher social capital which itself encourages more activity. This particular cluster focuses on brand-image and promotional strategy effectiveness, taking into account the impact of customer participation on sales and brand-image. This is gauged through techniques such as sentiment analysis which rely on mathematical areas of study such as data mining and analytics. This area of research produces vast numbers of commercial applications as the main goal of any study is to understand consumer behaviour and drive sales. In many organizations, members tend to focus their activities inside their own groups, which stifles creativity and restricts opportunities. A player whose network bridges structural holes has an advantage in detecting and developing rewarding opportunities. Such a player can mobilize social capital by acting as a "broker" of information between two clusters that otherwise would not have been in contact, thus providing access to new ideas, opinions and opportunities. British philosopher and political economist John Stuart Mill, writes, "it is hardly possible to overrate the value of placing human beings in contact with persons dissimilar to themselves.... Such communication [is] one of the primary sources of progress." Thus, a player with a network rich in structural holes can add value to an organization through new ideas and opportunities. This in turn, helps an individual's career development and advancement. A social capital broker also reaps control benefits of being the facilitator of information flow between contacts. Full communication with exploratory mindsets and information exchange generated by dynamically alternating positions in a social network promotes creative and deep thinking. In the case of consulting firm Eden McCallum, the founders were able to advance their careers by bridging their connections with former big three consulting firm consultants and mid-size industry firms. By bridging structural holes and mobilizing social capital, players can advance their careers by executing new opportunities between contacts. There has been research that both substantiates and refutes the benefits of information brokerage. A study of high tech Chinese firms by Zhixing Xiao found that the control benefits of structural holes are "dissonant to the dominant firm-wide spirit of cooperation and the information benefits cannot materialize due to the communal sharing values" of such organizations. However, this study only analyzed Chinese firms, which tend to have strong communal sharing values. Information and control benefits of structural holes are still valuable in firms that are not quite as inclusive and cooperative on the firm-wide level. In 2004, Ronald Burt studied 673 managers who ran the supply chain for one of America's largest electronics companies. He found that managers who often discussed issues with other groups were better paid, received more positive job evaluations and were more likely to be promoted. Thus, bridging structural holes can be beneficial to an organization, and in turn, to an individual's career. Computer networks combined with social networking software produce a new medium for social interaction. A relationship over a computerized social networking service can be characterized by context, direction, and strength. The content of a relation refers to the resource that is exchanged. In a computer-mediated communication context, social pairs exchange different kinds of information, including sending a data file or a computer program as well as providing emotional support or arranging a meeting. With the rise of electronic commerce, information exchanged may also correspond to exchanges of money, goods or services in the "real" world. Social network analysis methods have become essential to examining these types of computer mediated communication. In addition, the sheer size and the volatile nature of social media has given rise to new network metrics. A key concern with networks extracted from social media is the lack of robustness of network metrics given missing data. Based on the pattern of homophily, ties between people are most likely to occur between nodes that are most similar to each other, or within neighbourhood segregation, individuals are most likely to inhabit the same regional areas as other individuals who are like them. Therefore, social networks can be used as a tool to measure the degree of segregation or homophily within a social network. Social Networks can both be used to simulate the process of homophily but it can also serve as a measure of level of exposure of different groups to each other within a current social network of individuals in a certain area. See also References Further reading External links |
======================================== |
[SOURCE: https://en.wikipedia.org/wiki/Qur%27an] | [TOKENS: 16667] |
Contents Quran The Quran[a] (Arabic: الْقُرْآن,[b][c] lit. 'the recitation' or 'the lecture'), also romanized Qur'an or Koran,[d] is the central religious text of Islam, believed by Muslims to be a revelation directly from God (Allāh). It is organized in 114 chapters (sūrah, pl. suwar) which consist of individual verses (āyah). Besides its religious significance, it is widely regarded as the finest work in Arabic literature, and has significantly influenced the Arabic language. Muslims believe the Quran was orally revealed by God to the final Islamic prophet Muhammad through the angel Gabriel incrementally over a period of some 23 years, beginning on the Laylat al-Qadr, when Muhammad was 40, and concluding in 632, the year of his death. Muslims regard the Quran as Muhammad's most important miracle, a proof of his prophethood, and the culmination of a series of divine messages starting with those revealed to the first Islamic prophet Adam, including the holy books of the Torah, Psalms, and Gospel in Islam. The Quran is believed by Muslims to be God's own divine speech providing a complete code of conduct across all facets of life. This has led Muslim theologians to fiercely debate whether the Quran was "created or uncreated". According to tradition, several of Muhammad's companions served as scribes, recording the revelations. Shortly after Muhammad's death, the Quran was compiled on the order of the first caliph Abu Bakr (r. 632–634) by the companions, who had written down or memorized parts of it. Caliph Uthman (r. 644–656) established a standard version, now known as the Uthmanic codex, which is generally considered the archetype of the Quran known today. There are, however, variant readings, with some differences in meaning. It is the object of a modern field of academic research known as Quranic studies. The Quran describes itself as a guide for humanity (2:185). It assumes the reader is familiar with the significant narratives found in the Biblical and apocryphal texts. While it devotes considerable space to some of these narratives, it does not quote any paragraph from these sources directly,[e] but summarizes and uses different versions of some in different sections. Islamic theorists tend to explain this situation as the Quran's focus on the moral lessons to be drawn from events. This assumption and summarization entails many explanatory efforts (exegesis). In order to extrapolate the meaning of a particular Quranic verse, Muslims rely on tafsir, or commentary rather than a direct translation of the text. Furthermore, hadiths, which are valued in most Islamic schools of thought for their role in establishing Islamic laws and regulations, are oral traditions that were later written down and are believed to reflect the words and actions of Muhammad, serving as additional guidance alongside the Quran for most Muslims. During prayers, the Quran is recited only in Arabic. Someone who has memorized the entire Quran is called a hafiz. Ideally, verses are recited with a special kind of prosody reserved for this purpose called tajwid. It is a tradition for some Muslims to recite the entire Quran during the tarawih prayers throughout the month of Ramadan. Etymology and meaning The word qur'ān appears about 70 times in the Quran itself, assuming various meanings. It is a verbal noun (maṣdar) of the Arabic verb qara'a (قرأ) meaning 'he read' or 'he recited'. The Syriac equivalent is qeryānā (ܩܪܝܢܐ), which refers to 'scripture reading' or 'lesson'. While some Western scholars consider the word to be derived from the Syriac, the majority of Muslim authorities hold the origin of the word is qara'a itself. Regardless, it had become an Arabic term by Muhammad's lifetime. An important meaning of the word is the 'act of reciting', as reflected in an early Quranic passage: "It is for Us to collect it and to recite it (qur'ānahu)." In other verses, the word refers to 'an individual passage recited [by Muhammad]'. Its liturgical context is seen in a number of passages, for example: "So when al-qur'ān is recited, listen to it and keep silent." The word may also assume the meaning of a codified scripture when mentioned with other scriptures such as the Torah and Gospel. The term also has closely related synonyms that are employed throughout the Quran. Each synonym possesses its own distinct meaning, but its use may converge with that of qur'ān in certain contexts. Such terms include kitāb ('book'), āyah ('sign'), and sūrah ('scripture'); the latter two terms also denote units of revelation. In the large majority of contexts, usually with a definite article (al-), the word is referred to as the waḥy ('revelation'), that which has been "sent down" (tanzīl) at intervals. Other related words include: dhikr ('remembrance'), used to refer to the Quran in the sense of a reminder and warning; and ḥikmah ('wisdom'), sometimes referring to the revelation or part of it.[f] The Quran describes itself as 'the discernment' (al-furqān), 'the mother book' (umm al-kitāb), 'the guide' (huda), 'the wisdom' (hikmah), 'the remembrance' (dhikr), and 'the revelation' (tanzīl; 'something sent down', signifying the descent of an object from a higher place to lower place). Another term is al-kitāb ('The Book'), though it is also used in the Arabic language for other scriptures, such as the Torah and the Gospels. The term mus'haf ('written work') is often used to refer to particular Quranic manuscripts but is also used in the Quran to identify earlier revealed books. History Islamic tradition relates that Muhammad received his first revelation in 610 CE in the Cave of Hira on the Night of Power during one of his isolated retreats to the mountains. Thereafter, he received revelations over a period of 23 years. According to hadith (traditions ascribed to Muhammad)[g] and Muslim history, after Muhammad and his followers immigrated to Medina and formed an independent Muslim community, he ordered many of his companions to recite the Quran and to learn and teach the laws, which were revealed daily. It is related that some of the Quraysh who were taken prisoners at the Battle of Badr regained their freedom after they had taught some of the Muslims the simple writing of the time. Thus a group of Muslims gradually became literate. As it was initially spoken, the Quran was recorded on tablets, bones, and the wide, flat ends of date palm fronds. Most suras (also usually transliterated as Surah) were in use amongst early Muslims since they are mentioned in numerous sayings by both Sunni and Shia sources, relating Muhammad's use of the Quran as a call to Islam, the making of prayer and the manner of recitation. However, the Quran did not exist in book form at the time of Muhammad's death in 632 at age 61–62. There is agreement among scholars that Muhammad himself did not write down the revelation. Sahih al-Bukhari narrates Muhammad describing the revelations as, "Sometimes it is (revealed) like the ringing of a bell" and A'isha reported, "I saw the Prophet being inspired Divinely on a very cold day and noticed the sweat dropping from his forehead (as the Inspiration was over)."[h] Muhammad's first revelation, according to the Quran, was accompanied with a vision. The agent of revelation is mentioned as the "one mighty in power," the one who "grew clear to view when he was on the uppermost horizon. Then he drew nigh and came down till he was (distant) two bows' length or even nearer." The Islamic studies scholar Welch states in the Encyclopaedia of Islam that he believes the graphic descriptions of Muhammad's condition at these moments may be regarded as genuine, because he was severely disturbed after these revelations. According to Welch, these seizures would have been seen by those around him as convincing evidence for the superhuman origin of Muhammad's inspirations. However, Muhammad's critics accused him of being a possessed man, a soothsayer, or a magician since his experiences were similar to those claimed by such figures well known in ancient Arabia. Welch additionally states that it remains uncertain whether these experiences occurred before or after Muhammad's initial claim of prophethood. The Quran describes Muhammad as "ummi", which is traditionally interpreted as 'illiterate', but the meaning is more complex. Medieval commentators such as al-Tabari (d. 923) maintained that the term induced two meanings: first, the inability to read or write in general; second, the inexperience or ignorance of the previous books or scriptures (but they gave priority to the first meaning). Muhammad's illiteracy was taken as a sign of the genuineness of his prophethood. For example, according to Fakhr al-Din al-Razi, if Muhammad had mastered writing and reading he possibly would have been suspected of having studied the books of the ancestors. Some scholars such as W. Montgomery Watt prefer the second meaning of ummi—they take it to indicate unfamiliarity with earlier sacred texts. The final verse of the Quran was revealed on the 18th of the Islamic month of Dhu al-Hijjah in the year 10 A.H., a date that roughly corresponds to February or March 632. The verse was revealed after the Prophet finished delivering his sermon at Ghadir Khumm. According to Islamic tradition, the Quran was revealed to Muhammad in seven different ahruf (meaning letters; however, it could mean dialects, forms, styles or modes). Most Islamic scholars agree that these different ahruf are the same Quran revealed in seven different Arabic dialects and that they do not change the meaning of the Quran, the purpose of which was to make the Quran easy for recitation and memorization among the different Arab tribes. While Sunni Muslims believe in the seven ahruf, some Shia reject the idea of seven Quranic variants. A common misconception is that the seven ahruf and the qira'at are the same. Following Muhammad's death in 632, a number of his companions who memorized the Quran were killed in the Battle of al-Yamama by Musaylima. The first caliph, Abu Bakr (r. 632–634), subsequently decided to collect the book in one volume so that it could be preserved. Zayd ibn Thabit (d. 655) was the person to collect the Quran since "he used to write the Divine Inspiration for Allah's Apostle". Thus, a group of scribes, most importantly Zayd, collected the verses and produced a hand-written manuscript of the complete book. The manuscript according to Zayd remained with Abu Bakr until he died. Zayd's reaction to the task and the difficulties in collecting the Quranic material from parchments, palm-leaf stalks, thin stones (collectively known as suhuf, any written work containing divine teachings) and from men who knew it by heart is recorded in earlier narratives. In 644, Muhammad's widow Hafsa bint Umar was entrusted with the manuscript until the third caliph, Uthman (r. 644–656), requested the standard copy from her. According to historian Michael Cook, early Muslim narratives about the collection and compilation of the Quran sometimes contradict themselves: "Most ... make Uthman little more than an editor, but there are some in which he appears very much a collector, appealing to people to bring him any bit of the Quran they happen to possess." Some accounts also "suggest that in fact the material" Abu Bakr worked with "had already been assembled", which since he was the first caliph, would mean they were collected when Muhammad was still alive. Around the 650s, the Islamic expansion beyond the Arabian Peninsula and into Persia, the Levant and North Africa, as well as the use of the seven ahruf, had caused some confusion and differences in the pronunciation of the Quran, and conflict was arising between different Arab tribes due to some claiming to be more superior to other Arab tribes and non-Arabs based on dialect, which Uthman noticed. In order to preserve the sanctity of the text, he ordered a committee headed by Zayd to use Abu Bakr's copy and prepare a standard text of the Quran. Thus, within 20 years of Muhammad's death in 632, the complete Quran was committed to written form as the Uthmanic codex. That text became the model from which copies were made and promulgated throughout the urban centers of the Muslim world, and other versions are believed to have been destroyed. and the six other ahruf of the Quran fell out of use. The present form of the Quran text is accepted by Muslim scholars to be the original version compiled by Abu Bakr.[i] Qira'at which is a way and method of reciting the Quran was developed sometime afterwards. There are ten canonical recitations and they are not to be confused with ahruf. Shias recite the Quran according to the qira'at of Hafs on authority of ‘Asim, which is the prevalent qira'at in the Islamic world and believe that the Quran was gathered and compiled by Muhammad during his lifetime. It is claimed that the Shia had more than 1,000 hadiths ascribed to the Shia Imams which indicate the distortion of the Quran and according to Etan Kohlberg, this belief about Quran was common among Shiites in the early centuries of Islam. In his view, Ibn Babawayh was the first major Twelver author "to adopt a position identical to that of the Sunnis" and the change was a result of the "rise to power of the Sunni 'Abbasid caliphate," whence belief in the corruption of the Quran became untenable vis-a-vis the position of Sunni "orthodoxy". Alleged distortions have been carried out to remove any references to the rights of Ali, the Imams and their supporters and the disapproval of enemies, such as Umayyads and Abbasids. Other personal copies of the Quran might have existed including Ibn Mas'ud's and Ubay ibn Ka'b's codex, none of which exist today. Academic research Studies on the Quran rarely went beyond textual criticism.[when?] Until the early 1970s, non-Muslim scholars of Islam —while not accepting traditional explanations for divine intervention— accepted the above-mentioned traditional origin story in most details. University of Chicago professor Fred Donner states that: [T]here was a very early attempt to establish a uniform consonantal text of the Qurʾān from what was probably a wider and more varied group of related texts in early transmission.… After the creation of this standardized canonical text, earlier authoritative texts were suppressed, and all extant manuscripts—despite their numerous variants—seem to date to a time after this standard consonantal text was established. Although most variant readings of the text of the Quran have ceased to be transmitted, some still are. There has been no critical text produced on which a scholarly reconstruction of the Quranic text could be based.[j] In 1972, in a mosque in the city of Sanaa, Yemen, manuscripts "consisting of 12,000 pieces" were discovered that were later proven to be the oldest Quranic text known to exist at the time. The Sanaa manuscripts contain palimpsests, manuscript pages from which the text has been washed off to make the parchment reusable again—a practice which was common in ancient times due to the scarcity of writing material. However, the faint washed-off underlying text (scriptio inferior) is still barely visible. Studies using radiocarbon dating indicate that the parchments are dated to the period before 671 CE with a 99 percent probability. The German scholar Gerd R. Puin has been investigating these Quran fragments for years. His research team made 35,000 microfilm photographs of the manuscripts, which he dated to the early part of the 8th century. Puin has noted unconventional verse orderings, minor textual variations, and rare styles of orthography, and suggested that some of the parchments were palimpsests which had been reused. Puin believed that this implied an evolving text as opposed to a fixed one. It is also possible that the content of the Quran itself may provide data regarding the date and probably nearby geography of writing of the text. Sources based on some archaeological data give the construction date of Masjid al-Haram, an architectural work mentioned 16 times in the Quran, as 78 AH an additional finding that sheds light on the evolutionary history of the Quranic texts mentioned, which is known to continue even during the time of Hajjaj, in a similar situation that can be seen with al-Aksa, though different suggestions have been put forward -contrary to some of the hadiths of Muhammad's ascension, which indicate that these places are architectural structures- to explain.[k] These structures, -expected to be somewhere near Muhammad-[l] which were placed in cities like Mecca and Jerusalem, which are thousands of kilometers apart today, with interpretations based on narrations and miracles, were a miracle performed under a night's journey according to the outward and literal meaning of the verse.Surah Al-Isra 17:1 (See also:Bakkah) In 2015, a single folio of a very early Quran, dating back to 1370 years earlier, was discovered in the library of the University of Birmingham, England. According to the tests carried out by the Oxford University Radiocarbon Accelerator Unit, "with a probability of more than 95%, the parchment was from between 568 and 645". The manuscript is written in Hijazi script, an early form of written Arabic. This possibly was one of the earliest extant exemplars of the Quran, but as the tests allow a range of possible dates, it cannot be said with certainty which of the existing versions is the oldest. Saudi scholar Saud al-Sarhan has expressed doubt over the age of the fragments as they contain dots and chapter separators that are believed to have originated later. The Birmingham manuscript caused excitement amongst believers because of its potential overlapping with the dominant tradition over the lifetime of Muhammad c. 570 to 632 CE and used as evidence to support conventional wisdom and to refute the revisionists' views that expresses findings and views different from the traditional approach to the early history of the Quran and Islam. Contents The Quranic content is concerned with basic Islamic beliefs including the existence of God and the resurrection. Narratives of the early prophets, ethical and legal subjects, historical events of Muhammad's time, charity and prayer also appear in the Quran. The Quranic verses contain general exhortations regarding right and wrong and historical events are related to outline general moral lessons. The style of the Quran has been called "allusive", with commentaries needed to explain what is being referred to—"events are referred to, but not narrated; disagreements are debated without being explained; people and places are mentioned, but rarely named." While tafsir in Islamic sciences expresses the effort to understand the implied and implicit expressions of the Quran, fiqh refers to the efforts to expand the meaning of expressions, especially in the verses related to the provisions, as well as understanding it. Quranic studies state that, in the historical context, the content of the Quran is related to Rabbinic, Jewish-Christian, Syriac Christian and Hellenic literature, as well as pre-Islamic Arabia. Many places, subjects and mythological figures in the culture of Arabs and many nations in their historical neighbourhoods, especially Judeo-Christian stories, are included in the Quran with small allusions, references or sometimes small narratives such as jannāt ʿadn, jahannam, Seven Sleepers, Queen of Sheba etc. The stories of Yusuf and Zulaikha, Moses, Family of Amram (parents of Mary according to the Quran) and mysterious hero Dhul-Qarnayn ("the man with two horns") who built a barrier against Gog and Magog that will remain until the end of time are more detailed and longer stories. Apart from semi-historical events and characters such as King Solomon and David, about Jewish history as well as the exodus of the Israelites from Egypt, tales of the hebrew prophets accepted in Islam, such as Creation, the Flood, struggle of Abraham with Nimrod, sacrifice of his son occupy a wide place in the Quran. Some philosophers and scholars such as Mohammed Arkoun, who emphasize the mythological content of the Quran, are met with rejectionist attitudes in Islamic circles. In response to the fact that the people mentioned do not correspond to historically known figures, some modern commentators have suggested that these figures should be understood as representative figures portraying certain characters, rather than real ones. Furthermore, considering the findings of the revisionist school of Islamic studies, it is clear that the expression of certain narrative concepts in the Quran that refer to places, people, and events (such as Quraysh, Ababil, and Abu Lahab) in a single word or a few short sentences will require new interpretations and meanings that differ from the traditional narrative within this framework of understanding. (see:Quranic hermeneutics) The central theme of the Quran is monotheism. God is depicted as living, eternal, omniscient and omnipotent (see, e.g., Quran 2:20, 2:29, 2:255). He is the creator of everything, of the heavens and the earth and what is between them (see, e.g., Quran 13:16, 2:253, 50:38, etc.). All human beings are equal in their utter dependence upon God, and their well-being depends upon their acknowledging that fact and living accordingly. The Quran uses cosmological and contingency arguments in various verses without referring to the terms to prove the existence of God. Therefore, the universe is originated and needs an originator, and whatever exists must have a sufficient cause for its existence. Besides, the design of the universe is frequently referred to as a point of contemplation: "It is He who has created seven heavens in harmony. You cannot see any fault in God's creation; then look again: Can you see any flaw?" "The Qur'ān insists that Muhammad and his followers worship the same God as the Jews (29:46). The Qur'an's Allah is the same Creator God who covenanted with Abraham". Francis Edward Peters states that the Qur'an portrays Allah as both more powerful and more remote than Yahweh, and as a universal deity, unlike Yahweh who closely follows Israelites. However, Yahweh is never used for God in the Quran and Islamic texts, but, Rabb is an Arabic word to refers to God meaning Lord and the Quran cites in several places as in the Al-Fatiha; "All Praise and Gratitude is due to God, Lord of all the Universe". Even though Muslims do not doubt about the existence and unity of God, they may have adopted different attitudes that have changed and developed throughout history regarding his nature (attributes), names and relationship with creation. In contrast with pre-Islamic Arabian polytheism, as stated by Gerhard Böwering, God in Islam does not have associates and companions, nor is there any kinship between God and jinn. Pre Islamic Arabs believed in a blind, powerful, unstoppable and insensible fate over which man had no control. This was replaced with the Islamic belief of a powerful yet benevolent and merciful God's control over man's life. In the early periods of Islam, the concept of God was established as a personal deity living in the heavens. This understanding developed over time under the influence of Islamic theology, acquiring a transcendent character. However, in contrast to this transcendent and absolute conception of God established among the elite, the public and Sufis[m] maintained the traditional understanding on God. Also actions and attributes such as coming, going, sitting, satisfaction, anger and sadness etc. similar to humans used for this God in the Quran were considered mutashabihat—"no one knows its interpretation except God" (Quran 3:7)—by later scholars stating that God was free from resemblance to humans in any way.[n] Mustafa Öztürk points out the following words of Ahmad Ibn Hanbal in this regard: "Whoever says that Allah is everywhere is a heretic, an infidel. He should be invited to repent, but if he does not, be killed." This understanding changes later and gives way to the understanding that "God cannot be assigned a place and He is everywhere." In Islam, God speaks to people called prophets through a kind of revelation called wahy, or through angels. (42:51) nubuwwah (Arabic: نبوة 'prophethood') is seen as a duty imposed by God on individuals who have some characteristics such as intelligence, honesty, fortitude and justice: "Nothing is said to you that was not said to the messengers before you, that your lord has at his Command forgiveness as well as a most Grievous Penalty."[citation needed] Islam regards Abraham as a link in the chain of prophets that begins with Adam and culminates in Muhammad via Ishmael and mentioned in 35 chapters of the Quran, more often than any other biblical personage apart from Moses. Muslims regard him as a hanif, an archetype of the perfect Muslim, and revered prophet and builder of the Kaaba in Mecca. The Quran consistently refers to Islam as 'the religion of Abraham' (millat Ibrahim). In Islam, Eid-al-Adha is celebrated to commemorate Abraham's attempt to sacrifice his son by surrendering in line with his dream, (As-Saaffat; 100–107) which he accepted as the will of God. In Islam, Moses is a prominent prophet and messenger of God and the most frequently mentioned individual in the Quran, with his name being mentioned 136 times and his life being narrated and recounted more than that of any other prophet. Unlike the hundreds of references in the Quran to the stories of prophets such as Moses and Jesus, it provides very little information about Muhammad himself, his companions, or his contemporaries. The individuals to whom the expressions used in Quranic polemics belong and the contexts in which they were used are merely notes made in commentaries written in later centuries. An exception is his slave/adopted son Zayd, whose name is mentioned in the verses (Al-Aḥzāb;37) in the context of his -divorced- wife being taken into Muhammad's marriages. Probably the clearest biographical account of Muhammad in the Quran is the brief mention of his followers' settlement in Yathrib after their expulsion by the Quraysh, and of military encounters such as the Muslim victory at Badr. Stories of the prophets in the Quran often revolve around a certain pattern, according to which a prophet is sent to a group of people, who then reject or attack him, and ultimately suffer extinction as God's punishment. However, the Quran, given its paraenetic character, does not offer a full narrative; but rather offers a parabolic reference to the doom of previous generations, assuming the audience is familiar with the told stories. While belief in God and obedience to the prophets are the main emphasis in the prophetic stories, there are also non-prophetic stories in the Quran that emphasize the importance of humility and having profound-inner knowledge (hikmah) besides trusting in God. This is the main theme in the stories of Khidr, Luqman and Dhul Qarnayn. According to the later ascriptions to these stories, it is possible for those with this knowledge and divine support to teach the prophets (Khidr-Moses story Quran 18:65–82) and employ jinn (Dhul Qarnayn). Those who "spend their wealth" on people who are in need because they devoted their lives to the way of Allah and whose situation is unknown because they are ashamed to ask, will be rewarded by Allah. (Al Baqara; 272-274) In the story of Qārūn, the person who avoids searching for the afterlife with his wealth and becomes arrogant will be punished, arrogance befits only God. (Al Mutakabbir) Characters of the stories can be closed-mythical, (Khidr) demi-mythologic or combined characters, and it can also be seen that they are Islamized. While some believe he was a prophet, some researchers equate Luqman with the Alcmaeon of Croton or Aesop. Commanding ma’ruf and forbidding munkar (Ar. ٱلْأَمْرُ بِٱلْمَعْرُوفِ وَٱلنَّهْيُ عَنِ ٱلْمُنْكَرِ) is repeated or referred to in nearly 30 verses in different contexts in the Quran and is an important part of Islamist / jihadist indoctrination today, as well as Shiite teachings. Although a common translation of the phrase is "Enjoining good and forbidding evil", the words used by Islamic philosophy determining good and evil in discourses are "husn" and "qubh". The word ma’ruf literally means "known" or what is approved because of its familiarity for a certain society and its antithesis munkar means what is disapproved because it is unknown and extraneous. The Quran is one of the fundamental sources of Islamic law (sharia). Some formal religious practices receive significant attention in the Quran including the salat and fasting in the month of Ramadan. As for the manner in which the prayer is to be conducted, the Quran refers to prostration. The term chosen for charity, zakat, literally means purification implies that it is a self-purification. In fiqh, the term fard is used for clear imperative provisions based on the Quran. However, it is not possible to say that the relevant verses are understood in the same way by all segments of Islamic commentators; For example, Hanafis accept 5 daily prayers as fard. However, some religious groups such as Quranists and Shiites, who do not doubt that the Quran existing today is a religious source, infer from the same verses that it is clearly ordered to pray two or three times, not five times. About six verses address the way a woman should dress when in public; Some Muslim scholars consider this verse referring to the Hijab while others to clothings in general.[p] Research shows that the rituals in the Quran, along with laws such as qisas and tax (zakat), developed as an evolution of pre-Islamic Arabian rituals. Arabic words meaning pilgrimage (hajj), prayer (salāt) and charity (zakāt) can be seen in pre-Islamic Safaitic-Arabic inscriptions, and this continuity can be observed in many details, especially in hajj and umrah. A small number of verses in the Quran are about general rules of governance, inheritance, marriage, crime and punishment. Although the Quran does not impose a specific legal-management system, it emphasizes custom in nearly 40 verses and commands justice. (An-Nahl; 90) The practices prescribed in the Quran are considered as reflections of contextual legal understandings, as can be clearly seen in some examples such as Qisas and Diya. The following statement in the Quran is thought to be the general rule of testimony in Islamic jurisprudence, except for crime and punishment - for example, debt, shopping, etc.; O believers! When you contract a loan for a fixed period of time, commit it to writing....with justice. Call upon two of your men to witness. If two men cannot be found, then one man and two women of your choice will witness so if one of the women forgets the other may remind her.[Quran 4:11] As a different example, in the necklace story of Aisha, called Asbab al-Nuzul for surah An-Nur :11-20 four witnesses were required for the accusation of adultery. In addition, those who made accusations that did not meet the specified conditions would be punished with 80 lashes. The jurisprudence of later periods stipulates that witnesses must be men, covering all hadd crimes and people who did not have credibility and honesty in society (slaves, non-adl; sinners, infidels) could not testify against believers. In addition, the Islamic judiciary did not require proof of the issues defined as tazir.: 45 The statement in the Quran that determines the status of slaves in community is; Ma malakat aymanuhum meaning "those whom your right hands possess". The widespread use of slavery in the Islamic world continued until the last century,[s] and the unrestricted sexual use of female slaves, with a few exceptions such as they couldn't be loaned out[t] in traditional islamic jurisprudence while stated today often that sharia provides many rights to slaves and aims to eradicate slavery over time. Sharia is a collection of laws and rules created by scholars' interpretations on the Quran and hadith collections, and has been developed over the centuries, changing according to different geographies and societies. Fiqh sects are schools of understanding that try to determine the actions that people should do or avoid based on the Quran and hadiths. The place of hadiths in legislation is controversial; for example, in the Hanafi sect, in order to claim that something is obligatory, that issue must be clearly expressed in the Quran. Some of these results may also indicate exaggeration of statements, generalizations taken out of context, and imperative broadening of scope.[p] Of the few criminal cases listed as crimes in the Quran, only a few of them are punished by the classical books of sharia as determined by the verses of the Quran and are called hudud laws. How the verse Al-Ma'idah 33, which describes the crime of hirabah, should be understood is a matter of debate even today. The verse talks about the punishment of criminals by killing, hanging, having their hands and feet cut off on opposite sides, and being exiled from the earth, in response to an -abstract- crime such as "fighting against Allah and His Messenger". Expanding or narrowing the conditions and scope of this crime according to new situations and universal legal standards are issues that continue to be discussed today such as punishing in addition to rebellion against the legitimate government on "concrete sequential criminal acts" ie massacre, robbery and rape as preconditions. Although the constitutions of most Muslim-majority states contain references to sharia, its rules are largely preserved only in family law and criminal law in some. The Islamic revival of the late 20th century brought calls by Islamic movements for the full implementation of sharia, including corporal punishment such as stoning for adultery, through a variety of propaganda methods, from civic political activities to terrorism. The doctrine of the last day and eschatology (the final fate of the universe) may be considered the second great doctrine of the Quran. It is estimated that approximately one-third of the Quran is eschatological, dealing with the afterlife in the next world and with the day of judgment at the end of time. The Quran does not assert a natural immortality of the human soul, since man's existence is dependent on the will of God: when he wills, he causes man to die; and when he wills, he raises him to life again in a bodily resurrection. In the Quran belief in the afterlife is often referred in conjunction with belief in God: "Believe in God and the last day" emphasizing what is considered impossible is easy in the sight of God. A number of suras such as 44, 56, 75, 78, 81 and 101 are directly related to the afterlife and warn people to be prepared for the "imminent" day referred to in various ways. It is 'the Day of Judgment,' 'the Last Day,' 'the Day of Resurrection,' or simply 'the Hour.' Less frequently it is 'the Day of Distinction', 'the Day of the Gathering' or 'the Day of the Meeting'. While most topics known as "signs of the apocalypse" in Islamic eschatology are based on non-Quranic sources, some references in the Quran were frequently understood as apocalyptic terms, such as fitna, Dabba, and Gog and Magog. At the time of the Mongol conquests, ibn Kathir identified the latter with the historical Turks and Mongols. The apocalyptic writings frequently feature extra-Quranic figures such as the Dajjāl (corresponding to Armilos and Anti-Christ) and the Mahdī. The Dajjāl is supposed to become a cause of misguidance and causes havoc on earth, but is ultimately stopped by either the Mahdī or ʿĪsā (Jesus) who returns to earth from heaven. When the time of the apocalypse comes -which is told in a poetic form- the sun is rolled up, the stars fall down, seas are set on fire, mountains are moved people run away in fear and pregnant women miscarry.(At-Takwir 1-7) Then a square is set up and the king or lord of the day;(māliki yawmi-d-dīn)[u] comes and shows his shin; the looks are fearful, those gathered in the square are invited to prostrate;(Al-Qalam 42-43) the question asked is why the innocent female children were killed.(At-Takwir 8-9) Text and arrangement The Quran consists of 114 chapters of varying lengths, known as a sūrah. Each sūrah consists of verses, known as āyāt, which originally means a 'sign' or 'evidence' sent by God. The number of verses differs from sūrah to sūrah. An individual verse may be just a few letters or several lines. The total number of verses in the most popular Hafs Quran is 6,236;[v] however, the number varies if the bismillahs are counted separately. According to one estimate the Quran consists of 77,430 words, 18,994 unique words, 12,183 stems, 3,382 lemmas and 1,685 roots. Chapters are classified as Meccan or Medinan, depending on whether the verses were revealed before or after the migration of Muhammad to the city of Medina on traditional account. However, a sūrah classified as Medinan may contain Meccan verses in it and vice versa. Sūrah names are derived from a name or a character in the text, or from the first letters or words of the sūrah. Chapters are not arranged in chronological order, rather the chapters appear to be arranged roughly in order of decreasing size. Each sūrah except the ninth starts with the Bismillah (بِسْمِ ٱللَّٰهِ ٱلرَّحْمَٰنِ ٱلرَّحِيمِ), an Arabic phrase meaning 'In the name of God.' There are, however, still 114 occurrences of the Bismillah in the Quran, due to its presence in Quran 27:30 as the opening of Solomon's letter to the Queen of Sheba. The Muqattaʿat (Arabic: حروف مقطعات ḥurūf muqaṭṭaʿāt, 'disjoined letters, disconnected letters'; also 'mysterious letters') are combinations of between one and five Arabic letters figuring at the beginning of 29 out of the 114 chapters of the Quran just after the basmala. The letters are also known as fawātih (فواتح), or 'openers', as they form the opening verse of their respective suras. Four surahs are named for their muqatta'at: Ṭāʾ-Hāʾ, Yāʾ-Sīn, Ṣād, and Qāf. Various theories have been put forward; they were a secret communication language between Allah and Muhammad, abbreviations of various names or attributes of Allah, symbols of the versions of the Quran belonging to different companions, elements of a secret coding system, or expressions containing esoteric meanings. Some researchers associate them with hymns used in Syrian Christianity. The phrases must have been part of these hymns or abbreviations of frequently repeated introductory phrases. Some of them, such as Nun, were used in symbolic meanings. In addition of the division into chapters, there are various ways of dividing Quran into parts of approximately equal length for convenience in reading. The 30 juz' (plural ajzāʼ) can be used to read through the entire Quran in a month. A juz' is sometimes further divided into two ḥizb (plural aḥzāb), and each hizb subdivided into four rubʻ al-ahzab. The Quran is also divided into seven approximately equal parts, manzil (plural manāzil), for it to be recited in a week. A different structure is provided by semantic units resembling paragraphs and comprising roughly ten āyāt each. Such a section is called a ruku. The Quran's message is conveyed with various literary structures and devices. In the original Arabic, the suras and verses employ phonetic and thematic structures that assist the audience's efforts to recall the message of the text. Muslims, such as Dr. Mustafa Khattab in his The Clear Quran assert, according to the Quran itself, that the Quranic content and style is inimitable. The language of the Quran has been described as "rhymed prose" as it partakes of both poetry and prose; however, this description runs the risk of failing to convey the rhythmic quality of Quranic language, which is more poetic in some parts and more prose-like in others. Rhyme, while found throughout the Quran, is conspicuous in many of the earlier Meccan suras, in which relatively short verses throw the rhyming words into prominence. The effectiveness of such a form is evident for instance in Sura 81, and there can be no doubt that these passages impressed the conscience of the hearers. Frequently a change of rhyme from one set of verses to another signals a change in the subject of discussion. Later sections also preserve this form but the style is more expository. The Quranic text seems to have no beginning, middle, or end, its nonlinear structure being akin to a web or net. The textual arrangement is sometimes considered to exhibit lack of continuity, absence of any chronological or thematic order and repetitiousness.[w] Michael Sells, citing the work of the critic Norman O. Brown, acknowledges Brown's observation that the seeming disorganization of Quranic literary expression—its scattered or fragmented mode of composition in Sells's phrase—is in fact a literary device capable of delivering profound effects as if the intensity of the prophetic message were shattering the vehicle of human language in which it was being communicated. Sells also addresses the much-discussed repetitiveness of the Quran, seeing this, too, as a literary device. Another group of researchers explores the irregularities and repetitions mentioned in the Quranic text in a way that refutes the traditional claim that it was preserved by memorization alongside writing. According to them, an oral period shaped the Quran as a text and order, and mentioned repetitions and irregularities were remnants of this period. A text is self-referential when it speaks about itself and makes reference to itself. According to Stefan Wild, the Quran demonstrates this metatextuality by explaining, classifying, interpreting and justifying the words to be transmitted. Self-referentiality is evident in those passages where the Quran refers to itself as revelation (tanzil), remembrance (dhikr), news (naba'), criterion (furqan) in a self-designating manner (explicitly asserting its Divinity, "And this is a blessed Remembrance that We have sent down; so are you now denying it?"), or in the frequent appearance of the "Say" tags, when Muhammad is commanded to speak (e.g., "Say: 'God's guidance is the true guidance'", "Say: 'Would you then dispute with us concerning God?'"). According to Wild the Quran is highly self-referential. The feature is more evident in early Meccan suras. In Islam, ’i‘jāz (Arabic: اَلْإِعْجَازُ), "inimitability challenge" of the Quran in sense of feṣāḥa and belagha (both eloquence and rhetoric) is the doctrine which holds that the Quran has a miraculous quality, both in content and in form, that no human speech can match. According to this, the Quran is a miracle and its inimitability is the proof granted to Muhammad in authentication of his prophetic status. The literary quality of the Quran has been praised by Muslim scholars and by many non-Muslim scholars. The doctrine of the miraculousness of the Quran is further emphasized by Muhammad's illiteracy since the unlettered prophet could not have been suspected of composing the Quran. The Quran is widely regarded as the finest work in Arabic literature. The emergence of the Quran was an oral and aural poetic experience; the aesthetic experience of reciting and hearing the Quran is often regarded as one of the main reasons behind conversion to Islam in the early days. Pre-Islamic Arabic poetry was an element of challenge, propaganda and warfare, and those who incapacitated their opponents from doing the same in feṣāḥa and Belagha socially honored, as could be seen on Mu'allaqat poets. The etymology of the word "shā'ir; (poet)" connotes the meaning of a man of inspirational knowledge, of unseen powers. `To the early Arabs poetry was ṣihr ḥalāl and the poet was a genius who had supernatural communications with the jinn or spirits, the muses who inspired him.’ Although pre-Islamic Arabs gave poets status associated with suprahuman beings, soothsayers and prophecies were seen as persons of lower status. Contrary to later hurufic and recent scientific prophecy claims, traditional miracle statements about the Quran hadn't focused on prophecies, with a few exceptions like the Byzantine victory over the Persians in wars that continued for hundreds of years with mutual victories and defeats. The first works about the ’i‘jāz of the Quran began to appear in the 9th century in the Mu'tazila circles, which emphasized only its literary aspect, and were adopted by other religious groups. According to grammarian Al-Rummani the eloquence contained in the Quran consisted of tashbīh, istiʿāra, taǧānus, mubālaġa, concision, clarity of speech (bayān), and talāʾum. He also added other features developed by himself; the free variation of themes (taṣrīf al-maʿānī), the implication content (taḍmīn) of the expressions and the rhyming closures (fawāṣil). The most famous works on the doctrine of inimitability are two medieval books by the grammarian Al Jurjani (d. 1078 CE), Dala’il al-i'jaz ('the Arguments of Inimitability') and Asraral-balagha ('the Secrets of Eloquence'). Al Jurjani believed that Quran's eloquence must be a certain special quality in the manner of its stylistic arrangement and composition or a certain special way of joining words. Angelika Neuwirth lists the factors that led to the emergence of the doctrine of ’i‘jāz: The necessity of explaining some challenging verses in the Quran; In the context of the emergence of the theory of "proofs of prophecy" (dâ'il an-nubuwwa) in Islamic theology, proving that the Quran is a work worthy of the emphasized superior place of Muhammad in the history of the prophets, thus gaining polemical superiority over Jews and Christians; Preservation of Arab national pride in the face of confrontation with the Iranian Shu'ubiyya movement, etc. Orientalist scholars Theodor Nöldeke, Friedrich Schwally and John Wansbrough, pointing out linguistic defects, held similar opinions on the Quranic text as careless and imperfect. Significance in Islam Quran says, "We have sent down the Quran in truth, and with the truth it has come down" and frequently asserts in its text that it is divinely ordained. The Quran speaks of a written pre-text that records God's speech before it is sent down, the "preserved tablet" that is the basis of the belief in fate also, and Muslims believe that the Quran was sent down or started to be sent down on the Laylat al-Qadr. Revered by pious Muslims as "the holy of holies", whose sound moves some to "tears and ecstasy", it is the physical symbol of the faith, the text often used as a charm on occasions of birth, death, marriage. Traditionally, before starting to read the Quran, ablution is performed, one seeks refuge in Allah from the accursed Satan, and the reading begins by mentioning the names of Allah, Rahman and Rahim together known as basmala. Consequently, It must never rest beneath other books, but always on top of them, one must never drink or smoke when it is being read aloud, and it must be listened to in silence. It is a talisman against disease and disaster. According to Islam, the Quran is the word of God (Kalām Allāh). Its nature and whether it was created became a matter of fierce debate among religious scholars; and with the involvement of the political authority in the discussions, some Muslim religious scholars who stood against the political stance faced religious persecution during the caliph al-Ma'mun period and the following years. Muslims believe that the present Quranic text corresponds to that revealed to Muhammad, and according to their interpretation of Quran 15:9, it is protected from corruption ("Indeed, it is We who sent down the Quran and indeed, We will be its guardians"). Muslims consider the Quran to be a sign of the prophethood of Muhammad and the truth of the religion. For this reason, in traditional Islamic societies, great importance was given to children memorizing the Quran, and those who memorized the entire Quran were honored with the title of hafiz. Even today, "millions and millions of people refer to the Koran daily to explain their actions and to justify their aspirations"[y] or see it as the source of scientific knowledge, though this has been criticized as pseudoscientific. Muslims believe the Quran to be God's literal words, a complete code of life, the final revelation to humanity, a work of divine guidance revealed to Muhammad through the angel Gabriel. On the other hand it is believed in Muslim community that full understanding of it can only be possible with the depths obtained in the basic and religious sciences that the ulema (imams in shia) might access, as "heirs of the prophets". For this reason, direct reading of the Quran or applications based on its literal translations are considered problematic except for some groups such as Quranists thinking that the Quran is a complete and clear book; and tafsir / fiqh are brought fore to correct understandings in it. With a classical approach, scholars will discuss verses of the Quran in context called asbab al-nuzul in islamic literature, as well as language and linguistics; will pass it through filters such as muhkam and mutashabih, nasıkh and abrogated; will open the closed expressions and try to guide the believers. There is no standardization in Quran translations, and interpretations range from traditional scholastic, to literalist-salafist understandings to Esoteric-Sufist, to modern and secular exegesis according to the personal scientific depth and tendencies of scholars. Surah Al-Fatiha, the first chapter of the Quran, is recited in full in every rakat of salah and on other occasions. This sura, which consists of seven verses, is the most often recited surah of the Quran: بِسْمِ ٱللَّهِ ٱلرَّحْمَٰنِ ٱلرَّحِيمِ ٱلْحَمْدُ لِلَّهِ رَبِّ ٱلْعَٰلَمِينَ ٱلرَّحْمَٰنِ ٱلرَّحِيمِ مَٰلِكِ يَوْمِ ٱلدِّينِ إِيَّاكَ نَعْبُدُ وَإِيَّاكَ نَسْتَعِينُ ٱهْدِنَا ٱلصِّرَٰطَ ٱلْمُسْتَقِيمَ صِرَٰطَ ٱلَّذِينَ أَنْعَمْتَ عَلَيْهِمْ غَيْرِ ٱلْمَغْضُوبِ عَلَيْهِمْ وَلَا ٱلضَّآلِّينَ In the Name of Allah the Entirely Merciful, the Especially Merciful. [All] praise is [due] to Allah, Lord of the worlds — The Entirely Merciful, the Especially Merciful, Sovereign of the Day of Recompense. It is You we worship and You we ask for help Guide us to the straight path— The path of those upon whom You have bestowed favor, not of those who have evoked [Your] anger or of those who are astray. Other sections of the Quran of choice are also read in daily prayers. Sura Al-Ikhlāṣ is second in frequency of Quran recitation, for according to many early authorities, Muhammad said that Ikhlāṣ is equivalent to one-third of the whole Quran. قُلۡ هُوَ ٱللَّهُ أَحَدٌ ٱللَّهُ ٱلصَّمَدُ لَمۡ یَلِدۡ وَلَمۡ یُولَدۡ وَلَمۡ یَكُن لَّهُۥ كُفُوًا أَحَدُۢ Say, ˹O Prophet,˺ "He is God—One ˹and Indivisible˺; God—the Sustainer ˹needed by all˺. He has never had offspring, nor was He born. And there is none comparable to Him." Respect for the written text of the Quran is an important element of religious faith by many Muslims, and the Quran is treated with reverence. Based on tradition and a literal interpretation of Quran 56:79 ("none shall touch but those who are clean"), some Muslims believe that they must perform a ritual cleansing with water (wudu or ghusl) before touching a copy of the Quran, although this view is not universal. Worn-out and old copies of the Quran are wrapped in a cloth and stored indefinitely in a safe place, buried in a mosque or a Muslim cemetery, or burned and the ashes buried or scattered over water. While praying, the Quran is only recited in Arabic. In Islam, most intellectual disciplines, including Islamic theology, philosophy, mysticism and jurisprudence, have been concerned with the Quran or have their foundation in its teachings. Muslims believe that the preaching or reading of the Quran is rewarded with divine rewards variously called ajr, thawab, or hasanat. The Quran also inspired Islamic arts and specifically the so-called Quranic arts of calligraphy and illumination. The Quran is never decorated with figurative images, but many Qurans have been highly decorated with decorative patterns in the margins of the page, or between the lines or at the start of suras. Islamic verses appear in many other media, on buildings and on objects of all sizes, such as mosque lamps, metal work, pottery and single pages of calligraphy for muraqqas or albums. Interpretation Tafsir (Arabic: تفسير, romanized: tafsīr [tafˈsiːr]; English: explanation) refers to an exegesis, or commentary, of the Quran. An author of a tafsir is a mufassir (Arabic: مُفسّر; plural: Arabic: مفسّرون, romanized: mufassirūn). A Quranic tafsir attempts to provide elucidation, explanation, interpretation, context or commentary for clear understanding and conviction of God's will in Islam. In the classical approach, the exegesis of the Qur'an is built on the 7th verse of Al Imran-7; "He is the one who has sent down to you the Book. In it are verses that are sound (muhkamat), which are the foundation of the Book, and others that are assimilated (mutashabihat; resembling one another)." In classical approach verses that speak of Allah's "hand," "throne" (arsh), or His "sitting on throne" (istiwa) are considered mutashabih. These expressions are not believed to refer to a physical organ or action but are seen as metaphorical expressions symbolizing Allah's power, knowledge, or majesty. Principally, a tafsir deals with the issues of linguistics, jurisprudence, and theology. In terms of perspective and approach, tafsir can be broadly divided into two main categories, namely tafsir bi-al-ma'thur (lit. received tafsir), which is transmitted from the early days of Islam through the Islamic prophet Muhammad and his companions, and tafsir bi-al-ra'y (lit. tafsir by opinion), which is arrived through personal reflection or independent rational thinking. Knowledge of context / Asbab al-Nuzul; a scholar must understand the reasons and circumstances of a revelation in order to interpret the text correctly. If a verse is isolated without knowing the context, it can be misinterpreted. Quranic hermeneutics, and hermeneutics in general, refers to efforts to understand the Quran and Islamic texts through a more objective and secular approach, using universal criteria and moving beyond traditionally imposed forms. The Syro-Aramaic Reading of the Koran offers a new and compelling approach to understanding the implications of the Qur'an, based on linguistic and geographical-cultural knowledge that also coincides with the findings of historical research of the revisionist school. There are different characteristics and traditions for each of the tafsirs representing respective schools and doctrines, such as Sunni Islam, Shia Islam, and Sufism. There are also general distinctions between classic tafsirs compiled by authoritative figures of Muslim scholarship during the formative ages of Islam, and modern tafsir which seeks to address a wider audience, including the common people. Commentaries dealing with the zahir ('outward aspects') of the text are called tafsir, (explanation) and hermeneutic and esoteric commentaries dealing with the batin are called ta'wil ('interpretation'). Commentators with an esoteric slant believe that the ultimate meaning of the Quran is known only to God. Esoteric or Sufi interpretation relates Quranic verses to the inner or esoteric (batin) and metaphysical dimensions of existence and consciousness. According to Sands, esoteric interpretations are more suggestive than declarative, and are allusions (isharat) rather than explanations (tafsir). They indicate possibilities as much as they demonstrate the insights of writers. Shias and Sunnis as well as some Muslim philosophers believe the meaning of the Quran is not restricted to the literal aspect.: 7 In contrast, Quranic literalism, followed by Salafis and Zahiris, is the belief that the Quran should only be taken at its apparent meaning. Henry Corbin narrates a hadith that goes back to Muhammad: The Quran possesses an external appearance and a hidden depth, an exoteric meaning and an esoteric meaning. This esoteric meaning in turn conceals an esoteric meaning. So it goes on for seven esoteric meanings.: 7 According to esoteric interpreters, the inner meaning of the Quran does not eradicate or invalidate its outward meaning. Rather, it is like the soul, which gives life to the body. Corbin considers the Quran to play a part in Islamic philosophy, because Gnosiology itself goes hand in hand with prophetology.: 13 Translations Translating the Quran has always been problematic and difficult. Many argue that the Quranic text cannot be reproduced in another language or form. An Arabic word may have a range of meanings depending on the context, making an accurate translation difficult. Moreover, one of the biggest difficulties in understanding the Quran for those who do not know its language in the face of shifts in linguistic usage over the centuries is semantic translations (meanings) that include the translator's contributions to the relevant text instead of literal ones. Although the author's contributions are often bracketed and shown separately, the author's individual tendencies may also come to the fore in making sense of the main text. These studies contain reflections and even distortions caused by the region, sect, education, ideology and knowledge of the people who made them, and efforts to reach the real content are drowned in the details of volumes of commentaries. These distortions can manifest themselves in many areas of belief and practices such as hijab.[z] In fact, every new interpretation and translation of the Quran inherently involves a new semantic restructuring of it. Islamic tradition also holds that translations were made for Negus of Abyssinia and Byzantine Emperor Heraclius, as both received letters by Muhammad containing verses from the Quran. In early centuries, the permissibility of translations was not an issue, but whether one could use translations in prayer.[citation needed] The Quran has been translated into most African, Asian, and European languages. The first translator of the Quran was Salman the Persian, who translated surat al-Fatiha into Persian during the seventh century. Another translation of the Quran was completed in 884 in Alwar (Sindh, India, now Pakistan) by the orders of Abdullah bin Umar bin Abdul Aziz on the request of the Hindu Raja Mehruk. The first fully attested complete translations of the Quran were done between the 10th and 12th centuries in Persian. The Samanid king, Mansur I (961–976), ordered a group of scholars from Khorasan to translate the Tafsir al-Tabari, originally in Arabic, into Persian. Later in the 11th century, one of the students of Abu Mansur Abdullah al-Ansari wrote a complete tafsir of the Quran in Persian. In the 12th century, Najm al-Din Abu Hafs al-Nasafi translated the Quran into Persian. The manuscripts of all three books have survived and have been published several times. In 1936, translations in 102 languages were known. In 2010, the Hürriyet Daily News and Economic Review reported that the Quran was presented in 112 languages at the 18th International Quran Exhibition in Tehran. Robert of Ketton's 1143 translation of the Quran for Peter the Venerable, Lex Mahumet pseudoprophete, was the first into a Western language (Latin). Alexander Ross offered the first English version in 1649, from the French translation of L'Alcoran de Mahomet (1647) by Andre du Ryer. In 1734, George Sale produced the first scholarly translation of the Quran into English; another was produced by Richard Bell in 1937, and yet another by Arthur John Arberry in 1955. While all these translators were non-Muslims, there have been numerous translations by Muslims: popular modern English translations by Muslims include The Oxford World Classics translation by Muhammad Abdel Haleem, The Clear Quran by Mustafa Khattab, Sahih International's translation, among various others. As with translations of the Bible, the English translators have sometimes favored archaic English words and constructions over their more modern or conventional equivalents; for example, two widely read translators, Abdullah Yusuf Ali and Marmaduke Pickthall, use the plural and singular ye and thou instead of the more common you. The oldest Gurmukhi translation of the Quran Sharif has been found in village Lande of Moga district of Indian Punjab which was printed in 1911. Recitation The proper recitation of the Quran is the subject of a separate discipline named tajwid which determines in detail how the Quran should be recited, how each individual syllable is to be pronounced, the need to pay attention to the places where there should be a pause, to elisions, where the pronunciation should be long or short, where letters should be sounded together and where they should be kept separate, etc. It may be said that this discipline studies the laws and methods of the proper recitation of the Quran and covers three main areas: the proper pronunciation of consonants and vowels (the articulation of the Quranic phonemes), the rules of pause in recitation and of resumption of recitation, and the musical and melodious features of recitation. In order to avoid incorrect pronunciation, reciters follow a program of training with a qualified teacher. The two most popular texts used as references for tajwid rules are Matn al-Jazariyyah by Ibn al-Jazari and Tuhfat al-Atfal by Sulayman al-Jamzuri. The recitations of a few Egyptian reciters, like El Minshawy, Al-Hussary, Abdul Basit, Mustafa Ismail, were highly influential in the development of current styles of recitation.: 83 Southeast Asia is well known for world-class recitation, evidenced in the popularity of the woman reciters such as Maria Ulfah of Jakarta. Today, crowds fill auditoriums for public Quran recitation competitions. There are generally two types of recitation (based on pace of recitation): The variant readings of the Quran are one type of textual variant. According to Melchert (2008), the majority of disagreements have to do with vowels to supply, most of them in turn not conceivably reflecting dialectal differences and about one in eight disagreements has to do with whether to place dots above or below the line. Nasser categorizes variant readings into various subtypes, including internal vowels, long vowels, gemination (shaddah), assimilation and alternation. It is generally stated that there are small differences between readings. However, these small changes may also include differences that may lead to serious differences in Islam, ranging from the definition of God[aa] to practices such as the formal conditions of ablution. The first Quranic manuscripts lacked marks, enabling multiple possible recitations to be conveyed by the same written text. The 10th-century Muslim scholar from Baghdad, Ibn Mujāhid, is famous for establishing seven acceptable textual readings of the Quran. He studied various readings and their trustworthiness and chose seven 8th-century readers from the cities of Mecca, Medina, Kufa, Basra and Damascus. Ibn Mujahid did not explain why he chose seven readers, rather than six or ten, but this may be related to a prophetic tradition (Muhammad's saying) reporting that the Quran had been revealed in seven ahruf. Today, the most popular readings are those transmitted by Ḥafṣ (d. 796) and Warsh (d. 812) which are according to two of Ibn Mujahid's reciters, Aasim ibn Abi al-Najud (Kufa, d. 745) and Nafiʽ al-Madani (Medina, d. 785), respectively. The influential standard Quran of Cairo uses an elaborate system of modified vowel-signs and a set of additional symbols for minute details and is based on ʻAsim's recitation, the 8th-century recitation of Kufa. This edition has become the standard for modern printings of the Quran. Occasionally, an early Quran shows compatibility with a particular reading. A Syrian manuscript from the 8th century is shown to have been written according to the reading of Ibn Amir ad-Dimashqi. Another study suggests that this manuscript bears the vocalization of himsi region. According to Ibn Taymiyyah vocalization markers indicating specific vowel sounds (tashkeel) were introduced into the text of the Quran during the lifetimes of the last Sahabah. Writing and printing Before printing was widely adopted in the 19th century, the Quran was transmitted in manuscripts made by calligraphers and copyists. The earliest manuscripts were written in Ḥijāzī-typescript. The Hijazi style manuscripts nevertheless confirm that transmission of the Quran in writing began at an early stage. Probably in the ninth century, scripts began to feature thicker strokes, which are traditionally known as Kufic scripts. Toward the end of the ninth century, new scripts began to appear in copies of the Quran and replace earlier scripts. The reason for discontinuation in the use of the earlier style was that it took too long to produce and the demand for copies was increasing. Copyists would therefore choose simpler writing styles. Beginning in the 11th century, the styles of writing employed were primarily the naskh, muhaqqaq, rayḥānī and, on rarer occasions, the thuluth script. Naskh was in very widespread use. In North Africa and Iberia, the Maghribī style was popular. More distinct is the Bihari script which was used solely in the north of India. Nastaʻlīq style was also rarely used in Persian world. In the beginning, the Quran was not written with dots or tashkeel. These features were added to the text during the lifetimes of the last of the Sahabah. Since it would have been too costly for most Muslims to purchase a manuscript, copies of the Quran were held in mosques in order to make them accessible to people. These copies frequently took the form of a series of 30 parts or juzʼ. In terms of productivity, the Ottoman copyists provide the best example. This was in response to widespread demand, unpopularity of printing methods and for aesthetic reasons. Whilst the majority of Islamic scribes were men, some women also worked as scholars and copyists; one such woman who made a copy of this text was the Moroccan jurist, Amina, bint al-Hajj ʿAbd al-Latif. Wood-block printing of extracts from the Quran is on record as early as the 10th century. Arabic movable type printing was ordered by Pope Julius II (r. 1503–1512) for distribution among Middle Eastern Christians. The first complete Quran printed with movable type was produced in Venice in 1537–1538 for the Ottoman market by Paganino Paganini and Alessandro Paganini. But this Quran was not used as it contained a large number of errors. Two more editions include the Hinckelmann edition published by the pastor Abraham Hinckelmann in Hamburg in 1694, and the edition by the Italian priest Ludovico Maracci in Padua in 1698 with Latin translation and commentary. Printed copies of the Quran during this period met with strong opposition from Muslim legal scholars: printing anything in Arabic was prohibited in the Ottoman empire between 1483 and 1726—initially, even on penalty of death. The Ottoman ban on printing in Arabic script was lifted in 1726 for non-religious texts only upon the request of Ibrahim Muteferrika, who printed his first book in 1729. Except for books in Hebrew and European languages, which were unrestricted, very few books, and no religious texts, were printed in the Ottoman Empire for another century.[ab] In 1786, Catherine the Great of Russia, sponsored a printing press for "Tatar and Turkish orthography" in Saint Petersburg, with one Mullah Osman Ismail responsible for producing the Arabic types. A Quran was printed with this press in 1787, reprinted in 1790 and 1793 in Saint Petersburg, and in 1803 in Kazan.[ac] The first edition printed in Iran appeared in Tehran (1828), a translation in Turkish was printed in Cairo in 1842, and the first officially sanctioned Ottoman edition was finally printed in Constantinople between 1875 and 1877 as a two-volume set, during the First Constitutional Era. Gustav Flügel published an edition of the Quran in 1834 in Leipzig, which remained authoritative in Europe for close to a century, until Cairo's Al-Azhar University published an edition of the Quran in 1924. This edition was the result of a long preparation, as it standardized Quranic orthography, and it remains the basis of later editions. Criticism Regarding the claim of divine origin, critics[who?] refer to pre-existing sources, not only taken from the Bible, supposed to be older revelations of God, but also from heretic, Apocryphic and talmudic sources, such as the Syriac Infancy Gospel and Gospel of James.[citation needed] The Quran acknowledges that accusations of borrowing popular ancient fables were being made against Muhammad. Relationship with other literature Some non-Muslim groups such as the Baháʼí Faith and Druze view the Quran as holy. In the Baháʼí Faith, the Quran is accepted as authentic revelation from God along with the revelations of the other world religions, Islam being viewed as a stage within the divine process of progressive revelation. Bahá'u'lláh, the Prophet-Founder of the Baháʼí Faith, wrote about the Quran. Unitarian Universalists may also seek inspiration from the Quran. It has been suggested that the Quran has some narrative similarities to the Diatessaron, Protoevangelium of James, Infancy Gospel of Thomas, Gospel of Pseudo-Matthew and the Arabic Infancy Gospel. One scholar has suggested that the Diatessaron, as a gospel harmony, may have led to the conception that the Christian Gospel is one text. The Quran attributes its relationship with former books (the Torah, or Tawrāh and the Gospel, or ʾInjīl) to their unique origin, saying all of them have been revealed by Allah, through it asserts that Jews and Christians have corrupted it through falsification and alteration (Taḥrīf). According to Christoph Luxenberg (in The Syro-Aramaic Reading of the Koran) the Quran's language was similar to the Syriac language. The Quran recounts stories of many of the people and events recounted in Jewish and Christian sacred books (Tanakh, Bible) and devotional literature (Apocrypha, Midrash), although it differs in many details. Adam, Enoch, Noah, Eber, Shelah, Abraham, Lot, Ishmael, Isaac, Jacob, Joseph, Job, Jethro, David, Solomon, Elijah, Elisha, Jonah, Aaron, Moses, Zechariah, John the Baptist and Jesus are mentioned in the Quran as prophets of God (see Prophets of Islam). In fact, Moses is mentioned more in the Quran than any other individual. Jesus is mentioned more often in the Quran than Muhammad (by name—Muhammad is often alluded to as "The Prophet" or "The Apostle"), while Mary is mentioned in the Quran more than in the New Testament. After the Quran, and the general rise of Islam, the Arabic alphabet developed rapidly into an art form. The Arabic grammarian Sibawayh wrote one of the earliest books on Arabic grammar, referred to as "Al-Kitab", which relied heavily on the language in the Quran. Wadad Kadi, Professor of Near Eastern Languages and Civilizations at University of Chicago, and Mustansir Mir, Professor of Islamic studies at Youngstown State University, state that the Quran exerted a particular influence on Arabic literature's diction, themes, metaphors, motifs and symbols and added new expressions and new meanings to old, pre-Islamic words that would become ubiquitous. See also Notes The first of those to collect the Qur'an into a mushaf (codex) was Salim, the freed slave of Abu Hudhaifah. In other words, veiling is a security issue that arose according to the needs of that period. These are not taken into consideration at all and are reflected as God's command. Women have been called God's command for a thousand years. Women said the same thing to their daughters and daughters-in-law. They tell me; 'Do you cover yourself while praying?' Of course, I cover up when I'm in congregation. I am obliged not to disturb the peace. But I also pray with my head uncovered in my own home. Because the Quran's requirement for prayer is not covering up, but ablution and turning towards the qibla. This is a thousand year old issue. It's so ingrained in us. But this should definitely not be underestimated. Because people do it thinking it is God's command. But on the other hand, we should not declare a person who does not cover up as a bad woman. References Since the word of God seemed threatened with alteration, the [third] caliph ordered that five of the qurrā' from amongst the companions, (one of them being Zayd ibn Thābit who had compiled the first volume), produce other copies from the first volume which had been prepared on the orders of the first caliph and which had been kept with Ḥafṣah, the wife of the Prophet and daughter of the second caliph. The other copies, already in the hands of Muslims in other areas, were collected and sent to Medina where, on orders of the Caliph, they were burnt (or, according to some historians, were destroyed by boiling). Thus several copies were made, one being kept in Medina, one in Mecca, and one each sent to Sham (a territory now divided into Syria, Lebanon, Palestine and Jordan), Kufa and Basra. It is said that beside these five, one copy was also sent to Yemen and one to Bahrein. These copies were called the Imam copies and served as original for all future copies. The only difference of order between these copies and the first volume was that the chapters "Spirits of War" and "Immunity" were written in one place between "The Heights" and "Jonah." Further reading External links |
======================================== |
[SOURCE: https://en.wikipedia.org/wiki/Mars#cite_note-271] | [TOKENS: 11899] |
Contents Mars Mars is the fourth planet from the Sun. It is also known as the "Red Planet", for its orange-red appearance. Mars is a desert-like rocky planet with a tenuous atmosphere that is primarily carbon dioxide (CO2). At the average surface level the atmospheric pressure is a few thousandths of Earth's, atmospheric temperature ranges from −153 to 20 °C (−243 to 68 °F), and cosmic radiation is high. Mars retains some water, in the ground as well as thinly in the atmosphere, forming cirrus clouds, fog, frost, larger polar regions of permafrost and ice caps (with seasonal CO2 snow), but no bodies of liquid surface water. Its surface gravity is roughly a third of Earth's or double that of the Moon. Its diameter, 6,779 km (4,212 mi), is about half the Earth's, or twice the Moon's, and its surface area is the size of all the dry land of Earth. Fine dust is prevalent across the surface and the atmosphere, being picked up and spread at the low Martian gravity even by the weak wind of the tenuous atmosphere. The terrain of Mars roughly follows a north-south divide, the Martian dichotomy, with the northern hemisphere mainly consisting of relatively flat, low lying plains, and the southern hemisphere of cratered highlands. Geologically, the planet is fairly active with marsquakes trembling underneath the ground, but also hosts many enormous volcanoes that are extinct (the tallest is Olympus Mons, 21.9 km or 13.6 mi tall), as well as one of the largest canyons in the Solar System (Valles Marineris, 4,000 km or 2,500 mi long). Mars has two natural satellites that are small and irregular in shape: Phobos and Deimos. With a significant axial tilt of 25 degrees, Mars experiences seasons, like Earth (which has an axial tilt of 23.5 degrees). A Martian solar year is equal to 1.88 Earth years (687 Earth days), a Martian solar day (sol) is equal to 24.6 hours. Mars formed along with the other planets approximately 4.5 billion years ago. During the martian Noachian period (4.5 to 3.5 billion years ago), its surface was marked by meteor impacts, valley formation, erosion, the possible presence of water oceans and the loss of its magnetosphere. The Hesperian period (beginning 3.5 billion years ago and ending 3.3–2.9 billion years ago) was dominated by widespread volcanic activity and flooding that carved immense outflow channels. The Amazonian period, which continues to the present, is the currently dominating and remaining influence on geological processes. Because of Mars's geological history, the possibility of past or present life on Mars remains an area of active scientific investigation, with some possible traces needing further examination. Being visible with the naked eye in Earth's sky as a red wandering star, Mars has been observed throughout history, acquiring diverse associations in different cultures. In 1963 the first flight to Mars took place with Mars 1, but communication was lost en route. The first successful flyby exploration of Mars was conducted in 1965 with Mariner 4. In 1971 Mariner 9 entered orbit around Mars, being the first spacecraft to orbit any body other than the Moon, Sun or Earth; following in the same year were the first uncontrolled impact (Mars 2) and first successful landing (Mars 3) on Mars. Probes have been active on Mars continuously since 1997. At times, more than ten probes have simultaneously operated in orbit or on the surface, more than at any other planet beyond Earth. Mars is an often proposed target for future crewed exploration missions, though no such mission is currently planned. Natural history Scientists have theorized that during the Solar System's formation, Mars was created as the result of a random process of run-away accretion of material from the protoplanetary disk that orbited the Sun. Mars has many distinctive chemical features caused by its position in the Solar System. Elements with comparatively low boiling points, such as chlorine, phosphorus, and sulfur, are much more common on Mars than on Earth; these elements were probably pushed outward by the young Sun's energetic solar wind. After the formation of the planets, the inner Solar System may have been subjected to the so-called Late Heavy Bombardment. About 60% of the surface of Mars shows a record of impacts from that era, whereas much of the remaining surface is probably underlain by immense impact basins caused by those events. However, more recent modeling has disputed the existence of the Late Heavy Bombardment. There is evidence of an enormous impact basin in the Northern Hemisphere of Mars, spanning 10,600 by 8,500 kilometres (6,600 by 5,300 mi), or roughly four times the size of the Moon's South Pole–Aitken basin, which would be the largest impact basin yet discovered if confirmed. It has been hypothesized that the basin was formed when Mars was struck by a Pluto-sized body about four billion years ago. The event, thought to be the cause of the Martian hemispheric dichotomy, created the smooth Borealis basin that covers 40% of the planet. A 2023 study shows evidence, based on the orbital inclination of Deimos (a small moon of Mars), that Mars may once have had a ring system 3.5 billion years to 4 billion years ago. This ring system may have been formed from a moon, 20 times more massive than Phobos, orbiting Mars billions of years ago; and Phobos would be a remnant of that ring. Epochs: The geological history of Mars can be split into many periods, but the following are the three primary periods: Geological activity is still taking place on Mars. The Athabasca Valles is home to sheet-like lava flows created about 200 million years ago. Water flows in the grabens called the Cerberus Fossae occurred less than 20 million years ago, indicating equally recent volcanic intrusions. The Mars Reconnaissance Orbiter has captured images of avalanches. Physical characteristics Mars is approximately half the diameter of Earth or twice that of the Moon, with a surface area only slightly less than the total area of Earth's dry land. Mars is less dense than Earth, having about 15% of Earth's volume and 11% of Earth's mass, resulting in about 38% of Earth's surface gravity. Mars is the only presently known example of a desert planet, a rocky planet with a surface akin to that of Earth's deserts. The red-orange appearance of the Martian surface is caused by iron(III) oxide (nanophase Fe2O3) and the iron(III) oxide-hydroxide mineral goethite. It can look like butterscotch; other common surface colors include golden, brown, tan, and greenish, depending on the minerals present. Like Earth, Mars is differentiated into a dense metallic core overlaid by less dense rocky layers. The outermost layer is the crust, which is on average about 42–56 kilometres (26–35 mi) thick, with a minimum thickness of 6 kilometres (3.7 mi) in Isidis Planitia, and a maximum thickness of 117 kilometres (73 mi) in the southern Tharsis plateau. For comparison, Earth's crust averages 27.3 ± 4.8 km in thickness. The most abundant elements in the Martian crust are silicon, oxygen, iron, magnesium, aluminum, calcium, and potassium. Mars is confirmed to be seismically active; in 2019, it was reported that InSight had detected and recorded over 450 marsquakes and related events. Beneath the crust is a silicate mantle responsible for many of the tectonic and volcanic features on the planet's surface. The upper Martian mantle is a low-velocity zone, where the velocity of seismic waves is lower than surrounding depth intervals. The mantle appears to be rigid down to the depth of about 250 km, giving Mars a very thick lithosphere compared to Earth. Below this the mantle gradually becomes more ductile, and the seismic wave velocity starts to grow again. The Martian mantle does not appear to have a thermally insulating layer analogous to Earth's lower mantle; instead, below 1050 km in depth, it becomes mineralogically similar to Earth's transition zone. At the bottom of the mantle lies a basal liquid silicate layer approximately 150–180 km thick. The Martian mantle appears to be highly heterogenous, with dense fragments up to 4 km across, likely injected deep into the planet by colossal impacts ~4.5 billion years ago; high-frequency waves from eight marsquakes slowed as they passed these localized regions, and modeling indicates the heterogeneities are compositionally distinct debris preserved because Mars lacks plate tectonics and has a sluggishly convecting interior that prevents complete homogenization. Mars's iron and nickel core is at least partially molten, and may have a solid inner core. It is around half of Mars's radius, approximately 1650–1675 km, and is enriched in light elements such as sulfur, oxygen, carbon, and hydrogen. The temperature of the core is estimated to be 2000–2400 K, compared to 5400–6230 K for Earth's solid inner core. In 2025, based on data from the InSight lander, a group of researchers reported the detection of a solid inner core 613 kilometres (381 mi) ± 67 kilometres (42 mi) in radius. Mars is a terrestrial planet with a surface that consists of minerals containing silicon and oxygen, metals, and other elements that typically make up rock. The Martian surface is primarily composed of tholeiitic basalt, although parts are more silica-rich than typical basalt and may be similar to andesitic rocks on Earth, or silica glass. Regions of low albedo suggest concentrations of plagioclase feldspar, with northern low albedo regions displaying higher than normal concentrations of sheet silicates and high-silicon glass. Parts of the southern highlands include detectable amounts of high-calcium pyroxenes. Localized concentrations of hematite and olivine have been found. Much of the surface is deeply covered by finely grained iron(III) oxide dust. The Phoenix lander returned data showing Martian soil to be slightly alkaline and containing elements such as magnesium, sodium, potassium and chlorine. These nutrients are found in soils on Earth, and are necessary for plant growth. Experiments performed by the lander showed that the Martian soil has a basic pH of 7.7, and contains 0.6% perchlorate by weight, concentrations that are toxic to humans. Streaks are common across Mars and new ones appear frequently on steep slopes of craters, troughs, and valleys. The streaks are dark at first and get lighter with age. The streaks can start in a tiny area, then spread out for hundreds of metres. They have been seen to follow the edges of boulders and other obstacles in their path. The commonly accepted hypotheses include that they are dark underlying layers of soil revealed after avalanches of bright dust or dust devils. Several other explanations have been put forward, including those that involve water or even the growth of organisms. Environmental radiation levels on the surface are on average 0.64 millisieverts of radiation per day, and significantly less than the radiation of 1.84 millisieverts per day or 22 millirads per day during the flight to and from Mars. For comparison the radiation levels in low Earth orbit, where Earth's space stations orbit, are around 0.5 millisieverts of radiation per day. Hellas Planitia has the lowest surface radiation at about 0.342 millisieverts per day, featuring lava tubes southwest of Hadriacus Mons with potentially levels as low as 0.064 millisieverts per day, comparable to radiation levels during flights on Earth. Although Mars has no evidence of a structured global magnetic field, observations show that parts of the planet's crust have been magnetized, suggesting that alternating polarity reversals of its dipole field have occurred in the past. This paleomagnetism of magnetically susceptible minerals is similar to the alternating bands found on Earth's ocean floors. One hypothesis, published in 1999 and re-examined in October 2005 (with the help of the Mars Global Surveyor), is that these bands suggest plate tectonic activity on Mars four billion years ago, before the planetary dynamo ceased to function and the planet's magnetic field faded. Geography and features Although better remembered for mapping the Moon, Johann Heinrich von Mädler and Wilhelm Beer were the first areographers. They began by establishing that most of Mars's surface features were permanent and by more precisely determining the planet's rotation period. In 1840, Mädler combined ten years of observations and drew the first map of Mars. Features on Mars are named from a variety of sources. Albedo features are named for classical mythology. Craters larger than roughly 50 km are named for deceased scientists and writers and others who have contributed to the study of Mars. Smaller craters are named for towns and villages of the world with populations of less than 100,000. Large valleys are named for the word "Mars" or "star" in various languages; smaller valleys are named for rivers. Large albedo features retain many of the older names but are often updated to reflect new knowledge of the nature of the features. For example, Nix Olympica (the snows of Olympus) has become Olympus Mons (Mount Olympus). The surface of Mars as seen from Earth is divided into two kinds of areas, with differing albedo. The paler plains covered with dust and sand rich in reddish iron oxides were once thought of as Martian "continents" and given names like Arabia Terra (land of Arabia) or Amazonis Planitia (Amazonian plain). The dark features were thought to be seas, hence their names Mare Erythraeum, Mare Sirenum and Aurorae Sinus. The largest dark feature seen from Earth is Syrtis Major Planum. The permanent northern polar ice cap is named Planum Boreum. The southern cap is called Planum Australe. Mars's equator is defined by its rotation, but the location of its Prime Meridian was specified, as was Earth's (at Greenwich), by choice of an arbitrary point; Mädler and Beer selected a line for their first maps of Mars in 1830. After the spacecraft Mariner 9 provided extensive imagery of Mars in 1972, a small crater (later called Airy-0), located in the Sinus Meridiani ("Middle Bay" or "Meridian Bay"), was chosen by Merton E. Davies, Harold Masursky, and Gérard de Vaucouleurs for the definition of 0.0° longitude to coincide with the original selection. Because Mars has no oceans, and hence no "sea level", a zero-elevation surface had to be selected as a reference level; this is called the areoid of Mars, analogous to the terrestrial geoid. Zero altitude was defined by the height at which there is 610.5 Pa (6.105 mbar) of atmospheric pressure. This pressure corresponds to the triple point of water, and it is about 0.6% of the sea level surface pressure on Earth (0.006 atm). For mapping purposes, the United States Geological Survey divides the surface of Mars into thirty cartographic quadrangles, each named for a classical albedo feature it contains. In April 2023, The New York Times reported an updated global map of Mars based on images from the Hope spacecraft. A related, but much more detailed, global Mars map was released by NASA on 16 April 2023. The vast upland region Tharsis contains several massive volcanoes, which include the shield volcano Olympus Mons. The edifice is over 600 km (370 mi) wide. Because the mountain is so large, with complex structure at its edges, giving a definite height to it is difficult. Its local relief, from the foot of the cliffs which form its northwest margin to its peak, is over 21 km (13 mi), a little over twice the height of Mauna Kea as measured from its base on the ocean floor. The total elevation change from the plains of Amazonis Planitia, over 1,000 km (620 mi) to the northwest, to the summit approaches 26 km (16 mi), roughly three times the height of Mount Everest, which in comparison stands at just over 8.8 kilometres (5.5 mi). Consequently, Olympus Mons is either the tallest or second-tallest mountain in the Solar System; the only known mountain which might be taller is the Rheasilvia peak on the asteroid Vesta, at 20–25 km (12–16 mi). The dichotomy of Martian topography is striking: northern plains flattened by lava flows contrast with the southern highlands, pitted and cratered by ancient impacts. It is possible that, four billion years ago, the Northern Hemisphere of Mars was struck by an object one-tenth to two-thirds the size of Earth's Moon. If this is the case, the Northern Hemisphere of Mars would be the site of an impact crater 10,600 by 8,500 kilometres (6,600 by 5,300 mi) in size, or roughly the area of Europe, Asia, and Australia combined, surpassing Utopia Planitia and the Moon's South Pole–Aitken basin as the largest impact crater in the Solar System. Mars is scarred by 43,000 impact craters with a diameter of 5 kilometres (3.1 mi) or greater. The largest exposed crater is Hellas, which is 2,300 kilometres (1,400 mi) wide and 7,000 metres (23,000 ft) deep, and is a light albedo feature clearly visible from Earth. There are other notable impact features, such as Argyre, which is around 1,800 kilometres (1,100 mi) in diameter, and Isidis, which is around 1,500 kilometres (930 mi) in diameter. Due to the smaller mass and size of Mars, the probability of an object colliding with the planet is about half that of Earth. Mars is located closer to the asteroid belt, so it has an increased chance of being struck by materials from that source. Mars is more likely to be struck by short-period comets, i.e., those that lie within the orbit of Jupiter. Martian craters can[discuss] have a morphology that suggests the ground became wet after the meteor impact. The large canyon, Valles Marineris (Latin for 'Mariner Valleys, also known as Agathodaemon in the old canal maps), has a length of 4,000 kilometres (2,500 mi) and a depth of up to 7 kilometres (4.3 mi). The length of Valles Marineris is equivalent to the length of Europe and extends across one-fifth the circumference of Mars. By comparison, the Grand Canyon on Earth is only 446 kilometres (277 mi) long and nearly 2 kilometres (1.2 mi) deep. Valles Marineris was formed due to the swelling of the Tharsis area, which caused the crust in the area of Valles Marineris to collapse. In 2012, it was proposed that Valles Marineris is not just a graben, but a plate boundary where 150 kilometres (93 mi) of transverse motion has occurred, making Mars a planet with possibly a two-tectonic plate arrangement. Images from the Thermal Emission Imaging System (THEMIS) aboard NASA's Mars Odyssey orbiter have revealed seven possible cave entrances on the flanks of the volcano Arsia Mons. The caves, named after loved ones of their discoverers, are collectively known as the "seven sisters". Cave entrances measure from 100 to 252 metres (328 to 827 ft) wide and they are estimated to be at least 73 to 96 metres (240 to 315 ft) deep. Because light does not reach the floor of most of the caves, they may extend much deeper than these lower estimates and widen below the surface. "Dena" is the only exception; its floor is visible and was measured to be 130 metres (430 ft) deep. The interiors of these caverns may be protected from micrometeoroids, UV radiation, solar flares and high energy particles that bombard the planet's surface. Martian geysers (or CO2 jets) are putative sites of small gas and dust eruptions that occur in the south polar region of Mars during the spring thaw. "Dark dune spots" and "spiders" – or araneiforms – are the two most visible types of features ascribed to these eruptions. Similarly sized dust will settle from the thinner Martian atmosphere sooner than it would on Earth. For example, the dust suspended by the 2001 global dust storms on Mars only remained in the Martian atmosphere for 0.6 years, while the dust from Mount Pinatubo took about two years to settle. However, under current Martian conditions, the mass movements involved are generally much smaller than on Earth. Even the 2001 global dust storms on Mars moved only the equivalent of a very thin dust layer – about 3 μm thick if deposited with uniform thickness between 58° north and south of the equator. Dust deposition at the two rover sites has proceeded at a rate of about the thickness of a grain every 100 sols. Atmosphere Mars lost its magnetosphere 4 billion years ago, possibly because of numerous asteroid strikes, so the solar wind interacts directly with the Martian ionosphere, lowering the atmospheric density by stripping away atoms from the outer layer. Both Mars Global Surveyor and Mars Express have detected ionized atmospheric particles trailing off into space behind Mars, and this atmospheric loss is being studied by the MAVEN orbiter. Compared to Earth, the atmosphere of Mars is quite rarefied. Atmospheric pressure on the surface today ranges from a low of 30 Pa (0.0044 psi) on Olympus Mons to over 1,155 Pa (0.1675 psi) in Hellas Planitia, with a mean pressure at the surface level of 600 Pa (0.087 psi). The highest atmospheric density on Mars is equal to that found 35 kilometres (22 mi) above Earth's surface. The resulting mean surface pressure is only 0.6% of Earth's 101.3 kPa (14.69 psi). The scale height of the atmosphere is about 10.8 kilometres (6.7 mi), which is higher than Earth's 6 kilometres (3.7 mi), because the surface gravity of Mars is only about 38% of Earth's. The atmosphere of Mars consists of about 96% carbon dioxide, 1.93% argon and 1.89% nitrogen along with traces of oxygen and water. The atmosphere is quite dusty, containing particulates about 1.5 μm in diameter which give the Martian sky a tawny color when seen from the surface. It may take on a pink hue due to iron oxide particles suspended in it. Despite repeated detections of methane on Mars, there is no scientific consensus as to its origin. One suggestion is that methane exists on Mars and that its concentration fluctuates seasonally. The existence of methane could be produced by non-biological process such as serpentinization involving water, carbon dioxide, and the mineral olivine, which is known to be common on Mars, or by Martian life. Compared to Earth, its higher concentration of atmospheric CO2 and lower surface pressure may be why sound is attenuated more on Mars, where natural sources are rare apart from the wind. Using acoustic recordings collected by the Perseverance rover, researchers concluded that the speed of sound there is approximately 240 m/s for frequencies below 240 Hz, and 250 m/s for those above. Auroras have been detected on Mars. Because Mars lacks a global magnetic field, the types and distribution of auroras there differ from those on Earth; rather than being mostly restricted to polar regions as is the case on Earth, a Martian aurora can encompass the planet. In September 2017, NASA reported radiation levels on the surface of the planet Mars were temporarily doubled, and were associated with an aurora 25 times brighter than any observed earlier, due to a massive, and unexpected, solar storm in the middle of the month. Mars has seasons, alternating between its northern and southern hemispheres, similar to on Earth. Additionally the orbit of Mars has, compared to Earth's, a large eccentricity and approaches perihelion when it is summer in its southern hemisphere and winter in its northern, and aphelion when it is winter in its southern hemisphere and summer in its northern. As a result, the seasons in its southern hemisphere are more extreme and the seasons in its northern are milder than would otherwise be the case. The summer temperatures in the south can be warmer than the equivalent summer temperatures in the north by up to 30 °C (54 °F). Martian surface temperatures vary from lows of about −110 °C (−166 °F) to highs of up to 35 °C (95 °F) in equatorial summer. The wide range in temperatures is due to the thin atmosphere which cannot store much solar heat, the low atmospheric pressure (about 1% that of the atmosphere of Earth), and the low thermal inertia of Martian soil. The planet is 1.52 times as far from the Sun as Earth, resulting in just 43% of the amount of sunlight. Mars has the largest dust storms in the Solar System, reaching speeds of over 160 km/h (100 mph). These can vary from a storm over a small area, to gigantic storms that cover the entire planet. They tend to occur when Mars is closest to the Sun, and have been shown to increase global temperature. Seasons also produce dry ice covering polar ice caps. Hydrology While Mars contains water in larger amounts, most of it is dust covered water ice at the Martian polar ice caps. The volume of water ice in the south polar ice cap, if melted, would be enough to cover most of the surface of the planet with a depth of 11 metres (36 ft). Water in its liquid form cannot persist on the surface due to Mars's low atmospheric pressure, which is less than 1% that of Earth. Only at the lowest of elevations are the pressure and temperature high enough for liquid water to exist for short periods. Although little water is present in the atmosphere, there is enough to produce clouds of water ice and different cases of snow and frost, often mixed with snow of carbon dioxide dry ice. Landforms visible on Mars strongly suggest that liquid water has existed on the planet's surface. Huge linear swathes of scoured ground, known as outflow channels, cut across the surface in about 25 places. These are thought to be a record of erosion caused by the catastrophic release of water from subsurface aquifers, though some of these structures have been hypothesized to result from the action of glaciers or lava. One of the larger examples, Ma'adim Vallis, is 700 kilometres (430 mi) long, much greater than the Grand Canyon, with a width of 20 kilometres (12 mi) and a depth of 2 kilometres (1.2 mi) in places. It is thought to have been carved by flowing water early in Mars's history. The youngest of these channels is thought to have formed only a few million years ago. Elsewhere, particularly on the oldest areas of the Martian surface, finer-scale, dendritic networks of valleys are spread across significant proportions of the landscape. Features of these valleys and their distribution strongly imply that they were carved by runoff resulting from precipitation in early Mars history. Subsurface water flow and groundwater sapping may play important subsidiary roles in some networks, but precipitation was probably the root cause of the incision in almost all cases. Along craters and canyon walls, there are thousands of features that appear similar to terrestrial gullies. The gullies tend to be in the highlands of the Southern Hemisphere and face the Equator; all are poleward of 30° latitude. A number of authors have suggested that their formation process involves liquid water, probably from melting ice, although others have argued for formation mechanisms involving carbon dioxide frost or the movement of dry dust. No partially degraded gullies have formed by weathering and no superimposed impact craters have been observed, indicating that these are young features, possibly still active. Other geological features, such as deltas and alluvial fans preserved in craters, are further evidence for warmer, wetter conditions at an interval or intervals in earlier Mars history. Such conditions necessarily require the widespread presence of crater lakes across a large proportion of the surface, for which there is independent mineralogical, sedimentological and geomorphological evidence. Further evidence that liquid water once existed on the surface of Mars comes from the detection of specific minerals such as hematite and goethite, both of which sometimes form in the presence of water. The chemical signature of water vapor on Mars was first unequivocally demonstrated in 1963 by spectroscopy using an Earth-based telescope. In 2004, Opportunity detected the mineral jarosite. This forms only in the presence of acidic water, showing that water once existed on Mars. The Spirit rover found concentrated deposits of silica in 2007 that indicated wet conditions in the past, and in December 2011, the mineral gypsum, which also forms in the presence of water, was found on the surface by NASA's Mars rover Opportunity. It is estimated that the amount of water in the upper mantle of Mars, represented by hydroxyl ions contained within Martian minerals, is equal to or greater than that of Earth at 50–300 parts per million of water, which is enough to cover the entire planet to a depth of 200–1,000 metres (660–3,280 ft). On 18 March 2013, NASA reported evidence from instruments on the Curiosity rover of mineral hydration, likely hydrated calcium sulfate, in several rock samples including the broken fragments of "Tintina" rock and "Sutton Inlier" rock as well as in veins and nodules in other rocks like "Knorr" rock and "Wernicke" rock. Analysis using the rover's DAN instrument provided evidence of subsurface water, amounting to as much as 4% water content, down to a depth of 60 centimetres (24 in), during the rover's traverse from the Bradbury Landing site to the Yellowknife Bay area in the Glenelg terrain. In September 2015, NASA announced that they had found strong evidence of hydrated brine flows in recurring slope lineae, based on spectrometer readings of the darkened areas of slopes. These streaks flow downhill in Martian summer, when the temperature is above −23 °C, and freeze at lower temperatures. These observations supported earlier hypotheses, based on timing of formation and their rate of growth, that these dark streaks resulted from water flowing just below the surface. However, later work suggested that the lineae may be dry, granular flows instead, with at most a limited role for water in initiating the process. A definitive conclusion about the presence, extent, and role of liquid water on the Martian surface remains elusive. Researchers suspect much of the low northern plains of the planet were covered with an ocean hundreds of meters deep, though this theory remains controversial. In March 2015, scientists stated that such an ocean might have been the size of Earth's Arctic Ocean. This finding was derived from the ratio of protium to deuterium in the modern Martian atmosphere compared to that ratio on Earth. The amount of Martian deuterium (D/H = 9.3 ± 1.7 10−4) is five to seven times the amount on Earth (D/H = 1.56 10−4), suggesting that ancient Mars had significantly higher levels of water. Results from the Curiosity rover had previously found a high ratio of deuterium in Gale Crater, though not significantly high enough to suggest the former presence of an ocean. Other scientists caution that these results have not been confirmed, and point out that Martian climate models have not yet shown that the planet was warm enough in the past to support bodies of liquid water. Near the northern polar cap is the 81.4 kilometres (50.6 mi) wide Korolev Crater, which the Mars Express orbiter found to be filled with approximately 2,200 cubic kilometres (530 cu mi) of water ice. In November 2016, NASA reported finding a large amount of underground ice in the Utopia Planitia region. The volume of water detected has been estimated to be equivalent to the volume of water in Lake Superior (which is 12,100 cubic kilometers). During observations from 2018 through 2021, the ExoMars Trace Gas Orbiter spotted indications of water, probably subsurface ice, in the Valles Marineris canyon system. Orbital motion Mars's average distance from the Sun is roughly 230 million km (143 million mi), and its orbital period is 687 (Earth) days. The solar day (or sol) on Mars is only slightly longer than an Earth day: 24 hours, 39 minutes, and 35.244 seconds. A Martian year is equal to 1.8809 Earth years, or 1 year, 320 days, and 18.2 hours. The gravitational potential difference and thus the delta-v needed to transfer between Mars and Earth is the second lowest for Earth. The axial tilt of Mars is 25.19° relative to its orbital plane, which is similar to the axial tilt of Earth. As a result, Mars has seasons like Earth, though on Mars they are nearly twice as long because its orbital period is that much longer. In the present day, the orientation of the north pole of Mars is close to the star Deneb. Mars has a relatively pronounced orbital eccentricity of about 0.09; of the seven other planets in the Solar System, only Mercury has a larger orbital eccentricity. It is known that in the past, Mars has had a much more circular orbit. At one point, 1.35 million Earth years ago, Mars had an eccentricity of roughly 0.002, much less than that of Earth today. Mars's cycle of eccentricity is 96,000 Earth years compared to Earth's cycle of 100,000 years. Mars has its closest approach to Earth (opposition) in a synodic period of 779.94 days. It should not be confused with Mars conjunction, where the Earth and Mars are at opposite sides of the Solar System and form a straight line crossing the Sun. The average time between the successive oppositions of Mars, its synodic period, is 780 days; but the number of days between successive oppositions can range from 764 to 812. The distance at close approach varies between about 54 and 103 million km (34 and 64 million mi) due to the planets' elliptical orbits, which causes comparable variation in angular size. At their furthest Mars and Earth can be as far as 401 million km (249 million mi) apart. Mars comes into opposition from Earth every 2.1 years. The planets come into opposition near Mars's perihelion in 2003, 2018 and 2035, with the 2020 and 2033 events being particularly close to perihelic opposition. The mean apparent magnitude of Mars is +0.71 with a standard deviation of 1.05. Because the orbit of Mars is eccentric, the magnitude at opposition from the Sun can range from about −3.0 to −1.4. The minimum brightness is magnitude +1.86 when the planet is near aphelion and in conjunction with the Sun. At its brightest, Mars (along with Jupiter) is second only to Venus in apparent brightness. Mars usually appears distinctly yellow, orange, or red. When farthest away from Earth, it is more than seven times farther away than when it is closest. Mars is usually close enough for particularly good viewing once or twice at 15-year or 17-year intervals. Optical ground-based telescopes are typically limited to resolving features about 300 kilometres (190 mi) across when Earth and Mars are closest because of Earth's atmosphere. As Mars approaches opposition, it begins a period of retrograde motion, which means it will appear to move backwards in a looping curve with respect to the background stars. This retrograde motion lasts for about 72 days, and Mars reaches its peak apparent brightness in the middle of this interval. Moons Mars has two relatively small (compared to Earth's) natural moons, Phobos (about 22 km (14 mi) in diameter) and Deimos (about 12 km (7.5 mi) in diameter), which orbit at 9,376 km (5,826 mi) and 23,460 km (14,580 mi) around the planet. The origin of both moons is unclear, although a popular theory states that they were asteroids captured into Martian orbit. Both satellites were discovered in 1877 by Asaph Hall and were named after the characters Phobos (the deity of panic and fear) and Deimos (the deity of terror and dread), twins from Greek mythology who accompanied their father Ares, god of war, into battle. Mars was the Roman equivalent to Ares. In modern Greek, the planet retains its ancient name Ares (Aris: Άρης). From the surface of Mars, the motions of Phobos and Deimos appear different from that of the Earth's satellite, the Moon. Phobos rises in the west, sets in the east, and rises again in just 11 hours. Deimos, being only just outside synchronous orbit – where the orbital period would match the planet's period of rotation – rises as expected in the east, but slowly. Because the orbit of Phobos is below a synchronous altitude, tidal forces from Mars are gradually lowering its orbit. In about 50 million years, it could either crash into Mars's surface or break up into a ring structure around the planet. The origin of the two satellites is not well understood. Their low albedo and carbonaceous chondrite composition have been regarded as similar to asteroids, supporting a capture theory. The unstable orbit of Phobos would seem to point toward a relatively recent capture. But both have circular orbits near the equator, which is unusual for captured objects, and the required capture dynamics are complex. Accretion early in the history of Mars is plausible, but would not account for a composition resembling asteroids rather than Mars itself, if that is confirmed. Mars may have yet-undiscovered moons, smaller than 50 to 100 metres (160 to 330 ft) in diameter, and a dust ring is predicted to exist between Phobos and Deimos. A third possibility for their origin as satellites of Mars is the involvement of a third body or a type of impact disruption. More-recent lines of evidence for Phobos having a highly porous interior, and suggesting a composition containing mainly phyllosilicates and other minerals known from Mars, point toward an origin of Phobos from material ejected by an impact on Mars that reaccreted in Martian orbit, similar to the prevailing theory for the origin of Earth's satellite. Although the visible and near-infrared (VNIR) spectra of the moons of Mars resemble those of outer-belt asteroids, the thermal infrared spectra of Phobos are reported to be inconsistent with chondrites of any class. It is also possible that Phobos and Deimos were fragments of an older moon, formed by debris from a large impact on Mars, and then destroyed by a more recent impact upon the satellite. More recently, a study conducted by a team of researchers from multiple countries suggests that a lost moon, at least fifteen times the size of Phobos, may have existed in the past. By analyzing rocks which point to tidal processes on the planet, it is possible that these tides may have been regulated by a past moon. Human observations and exploration The history of observations of Mars is marked by oppositions of Mars when the planet is closest to Earth and hence is most easily visible, which occur every couple of years. Even more notable are the perihelic oppositions of Mars, which are distinguished because Mars is close to perihelion, making it even closer to Earth. The ancient Sumerians named Mars Nergal, the god of war and plague. During Sumerian times, Nergal was a minor deity of little significance, but, during later times, his main cult center was the city of Nineveh. In Mesopotamian texts, Mars is referred to as the "star of judgement of the fate of the dead". The existence of Mars as a wandering object in the night sky was also recorded by the ancient Egyptian astronomers and, by 1534 BCE, they were familiar with the retrograde motion of the planet. By the period of the Neo-Babylonian Empire, the Babylonian astronomers were making regular records of the positions of the planets and systematic observations of their behavior. For Mars, they knew that the planet made 37 synodic periods, or 42 circuits of the zodiac, every 79 years. They invented arithmetic methods for making minor corrections to the predicted positions of the planets. In Ancient Greece, the planet was known as Πυρόεις. Commonly, the Greek name for the planet now referred to as Mars, was Ares. It was the Romans who named the planet Mars, for their god of war, often represented by the sword and shield of the planet's namesake. In the fourth century BCE, Aristotle noted that Mars disappeared behind the Moon during an occultation, indicating that the planet was farther away. Ptolemy, a Greek living in Alexandria, attempted to address the problem of the orbital motion of Mars. Ptolemy's model and his collective work on astronomy was presented in the multi-volume collection later called the Almagest (from the Arabic for "greatest"), which became the authoritative treatise on Western astronomy for the next fourteen centuries. Literature from ancient China confirms that Mars was known by Chinese astronomers by no later than the fourth century BCE. In the East Asian cultures, Mars is traditionally referred to as the "fire star" (火星) based on the Wuxing system. In 1609 Johannes Kepler published a 10 year study of Martian orbit, using the diurnal parallax of Mars, measured by Tycho Brahe, to make a preliminary calculation of the relative distance to the planet. From Brahe's observations of Mars, Kepler deduced that the planet orbited the Sun not in a circle, but in an ellipse. Moreover, Kepler showed that Mars sped up as it approached the Sun and slowed down as it moved farther away, in a manner that later physicists would explain as a consequence of the conservation of angular momentum.: 433–437 In 1610 the first use of a telescope for astronomical observation, including Mars, was performed by Italian astronomer Galileo Galilei. With the telescope the diurnal parallax of Mars was again measured in an effort to determine the Sun-Earth distance. This was first performed by Giovanni Domenico Cassini in 1672. The early parallax measurements were hampered by the quality of the instruments. The only occultation of Mars by Venus observed was that of 13 October 1590, seen by Michael Maestlin at Heidelberg. By the 19th century, the resolution of telescopes reached a level sufficient for surface features to be identified. On 5 September 1877, a perihelic opposition to Mars occurred. The Italian astronomer Giovanni Schiaparelli used a 22-centimetre (8.7 in) telescope in Milan to help produce the first detailed map of Mars. These maps notably contained features he called canali, which, with the possible exception of the natural canyon Valles Marineris, were later shown to be an optical illusion. These canali were supposedly long, straight lines on the surface of Mars, to which he gave names of famous rivers on Earth. His term, which means "channels" or "grooves", was popularly mistranslated in English as "canals". Influenced by the observations, the orientalist Percival Lowell founded an observatory which had 30- and 45-centimetre (12- and 18-in) telescopes. The observatory was used for the exploration of Mars during the last good opportunity in 1894, and the following less favorable oppositions. He published several books on Mars and life on the planet, which had a great influence on the public. The canali were independently observed by other astronomers, like Henri Joseph Perrotin and Louis Thollon in Nice, using one of the largest telescopes of that time. The seasonal changes (consisting of the diminishing of the polar caps and the dark areas formed during Martian summers) in combination with the canals led to speculation about life on Mars, and it was a long-held belief that Mars contained vast seas and vegetation. As bigger telescopes were used, fewer long, straight canali were observed. During observations in 1909 by Antoniadi with an 84-centimetre (33 in) telescope, irregular patterns were observed, but no canali were seen. The first spacecraft from Earth to visit Mars was Mars 1 of the Soviet Union, which flew by in 1963, but contact was lost en route. NASA's Mariner 4 followed and became the first spacecraft to successfully transmit from Mars; launched on 28 November 1964, it made its closest approach to the planet on 15 July 1965. Mariner 4 detected the weak Martian radiation belt, measured at about 0.1% that of Earth, and captured the first images of another planet from deep space. Once spacecraft visited the planet during the 1960s and 1970s, many previous concepts of Mars were radically broken. After the results of the Viking life-detection experiments, the hypothesis of a dead planet was generally accepted. The data from Mariner 9 and Viking allowed better maps of Mars to be made. Until 1997 and after Viking 1 shut down in 1982, Mars was only visited by three unsuccessful probes, two flying past without contact (Phobos 1, 1988; Mars Observer, 1993), and one (Phobos 2 1989) malfunctioning in orbit before reaching its destination Phobos. In 1997 Mars Pathfinder became the first successful rover mission beyond the Moon and started together with Mars Global Surveyor (operated until late 2006) an uninterrupted active robotic presence at Mars that has lasted until today. It produced complete, extremely detailed maps of the Martian topography, magnetic field and surface minerals. Starting with these missions a range of new improved crewless spacecraft, including orbiters, landers, and rovers, have been sent to Mars, with successful missions by the NASA (United States), Jaxa (Japan), ESA, United Kingdom, ISRO (India), Roscosmos (Russia), the United Arab Emirates, and CNSA (China) to study the planet's surface, climate, and geology, uncovering the different elements of the history and dynamic of the hydrosphere of Mars and possible traces of ancient life. As of 2023[update], Mars is host to ten functioning spacecraft. Eight are in orbit: 2001 Mars Odyssey, Mars Express, Mars Reconnaissance Orbiter, MAVEN, ExoMars Trace Gas Orbiter, the Hope orbiter, and the Tianwen-1 orbiter. Another two are on the surface: the Mars Science Laboratory Curiosity rover and the Perseverance rover. Collected maps are available online at websites including Google Mars. NASA provides two online tools: Mars Trek, which provides visualizations of the planet using data from 50 years of exploration, and Experience Curiosity, which simulates traveling on Mars in 3-D with Curiosity. Planned missions to Mars include: As of February 2024[update], debris from these types of missions has reached over seven tons. Most of it consists of crashed and inactive spacecraft as well as discarded components. In April 2024, NASA selected several companies to begin studies on providing commercial services to further enable robotic science on Mars. Key areas include establishing telecommunications, payload delivery and surface imaging. Habitability and habitation During the late 19th century, it was widely accepted in the astronomical community that Mars had life-supporting qualities, including the presence of oxygen and water. However, in 1894 W. W. Campbell at Lick Observatory observed the planet and found that "if water vapor or oxygen occur in the atmosphere of Mars it is in quantities too small to be detected by spectroscopes then available". That observation contradicted many of the measurements of the time and was not widely accepted. Campbell and V. M. Slipher repeated the study in 1909 using better instruments, but with the same results. It was not until the findings were confirmed by W. S. Adams in 1925 that the myth of the Earth-like habitability of Mars was finally broken. However, even in the 1960s, articles were published on Martian biology, putting aside explanations other than life for the seasonal changes on Mars. The current understanding of planetary habitability – the ability of a world to develop environmental conditions favorable to the emergence of life – favors planets that have liquid water on their surface. Most often this requires the orbit of a planet to lie within the habitable zone, which for the Sun is estimated to extend from within the orbit of Earth to about that of Mars. During perihelion, Mars dips inside this region, but Mars's thin (low-pressure) atmosphere prevents liquid water from existing over large regions for extended periods. The past flow of liquid water demonstrates the planet's potential for habitability. Recent evidence has suggested that any water on the Martian surface may have been too salty and acidic to support regular terrestrial life. The environmental conditions on Mars are a challenge to sustaining organic life: the planet has little heat transfer across its surface, it has poor insulation against bombardment by the solar wind due to the absence of a magnetosphere and has insufficient atmospheric pressure to retain water in a liquid form (water instead sublimes to a gaseous state). Mars is nearly, or perhaps totally, geologically dead; the end of volcanic activity has apparently stopped the recycling of chemicals and minerals between the surface and interior of the planet. Evidence suggests that the planet was once significantly more habitable than it is today, but whether living organisms ever existed there remains unknown. The Viking probes of the mid-1970s carried experiments designed to detect microorganisms in Martian soil at their respective landing sites and had positive results, including a temporary increase in CO2 production on exposure to water and nutrients. This sign of life was later disputed by scientists, resulting in a continuing debate, with NASA scientist Gilbert Levin asserting that Viking may have found life. A 2014 analysis of Martian meteorite EETA79001 found chlorate, perchlorate, and nitrate ions in sufficiently high concentrations to suggest that they are widespread on Mars. UV and X-ray radiation would turn chlorate and perchlorate ions into other, highly reactive oxychlorines, indicating that any organic molecules would have to be buried under the surface to survive. Small quantities of methane and formaldehyde detected by Mars orbiters are both claimed to be possible evidence for life, as these chemical compounds would quickly break down in the Martian atmosphere. Alternatively, these compounds may instead be replenished by volcanic or other geological means, such as serpentinite. Impact glass, formed by the impact of meteors, which on Earth can preserve signs of life, has also been found on the surface of the impact craters on Mars. Likewise, the glass in impact craters on Mars could have preserved signs of life, if life existed at the site. The Cheyava Falls rock discovered on Mars in June 2024 has been designated by NASA as a "potential biosignature" and was core sampled by the Perseverance rover for possible return to Earth and further examination. Although highly intriguing, no definitive final determination on a biological or abiotic origin of this rock can be made with the data currently available. Several plans for a human mission to Mars have been proposed, but none have come to fruition. The NASA Authorization Act of 2017 directed NASA to study the feasibility of a crewed Mars mission in the early 2030s; the resulting report concluded that this would be unfeasible. In addition, in 2021, China was planning to send a crewed Mars mission in 2033. Privately held companies such as SpaceX have also proposed plans to send humans to Mars, with the eventual goal to settle on the planet. As of 2024, SpaceX has proceeded with the development of the Starship launch vehicle with the goal of Mars colonization. In plans shared with the company in April 2024, Elon Musk envisions the beginning of a Mars colony within the next twenty years. This would be enabled by the planned mass manufacturing of Starship and initially sustained by resupply from Earth, and in situ resource utilization on Mars, until the Mars colony reaches full self sustainability. Any future human mission to Mars will likely take place within the optimal Mars launch window, which occurs every 26 months. The moon Phobos has been proposed as an anchor point for a space elevator. Besides national space agencies and space companies, groups such as the Mars Society and The Planetary Society advocate for human missions to Mars. In culture Mars is named after the Roman god of war (Greek Ares), but was also associated with the demi-god Heracles (Roman Hercules) by ancient Greek astronomers, as detailed by Aristotle. This association between Mars and war dates back at least to Babylonian astronomy, in which the planet was named for the god Nergal, deity of war and destruction. It persisted into modern times, as exemplified by Gustav Holst's orchestral suite The Planets, whose famous first movement labels Mars "The Bringer of War". The planet's symbol, a circle with a spear pointing out to the upper right, is also used as a symbol for the male gender. The symbol dates from at least the 11th century, though a possible predecessor has been found in the Greek Oxyrhynchus Papyri. The idea that Mars was populated by intelligent Martians became widespread in the late 19th century. Schiaparelli's "canali" observations combined with Percival Lowell's books on the subject put forward the standard notion of a planet that was a drying, cooling, dying world with ancient civilizations constructing irrigation works. Many other observations and proclamations by notable personalities added to what has been termed "Mars Fever". In the present day, high-resolution mapping of the surface of Mars has revealed no artifacts of habitation, but pseudoscientific speculation about intelligent life on Mars still continues. Reminiscent of the canali observations, these speculations are based on small scale features perceived in the spacecraft images, such as "pyramids" and the "Face on Mars". In his book Cosmos, planetary astronomer Carl Sagan wrote: "Mars has become a kind of mythic arena onto which we have projected our Earthly hopes and fears." The depiction of Mars in fiction has been stimulated by its dramatic red color and by nineteenth-century scientific speculations that its surface conditions might support not just life but intelligent life. This gave way to many science fiction stories involving these concepts, such as H. G. Wells's The War of the Worlds, in which Martians seek to escape their dying planet by invading Earth; Ray Bradbury's The Martian Chronicles, in which human explorers accidentally destroy a Martian civilization; as well as Edgar Rice Burroughs's series Barsoom, C. S. Lewis's novel Out of the Silent Planet (1938), and a number of Robert A. Heinlein stories before the mid-sixties. Since then, depictions of Martians have also extended to animation. A comic figure of an intelligent Martian, Marvin the Martian, appeared in Haredevil Hare (1948) as a character in the Looney Tunes animated cartoons of Warner Brothers, and has continued as part of popular culture to the present. After the Mariner and Viking spacecraft had returned pictures of Mars as a lifeless and canal-less world, these ideas about Mars were abandoned; for many science-fiction authors, the new discoveries initially seemed like a constraint, but eventually the post-Viking knowledge of Mars became itself a source of inspiration for works like Kim Stanley Robinson's Mars trilogy. See also Notes References Further reading External links Solar System → Local Interstellar Cloud → Local Bubble → Gould Belt → Orion Arm → Milky Way → Milky Way subgroup → Local Group → Local Sheet → Local Volume → Virgo Supercluster → Laniakea Supercluster → Pisces–Cetus Supercluster Complex → Local Hole → Observable universe → UniverseEach arrow (→) may be read as "within" or "part of". |
======================================== |
[SOURCE: https://en.wikipedia.org/wiki/Animal#cite_ref-Heino_18-0] | [TOKENS: 6011] |
Contents Animal Animals are multicellular, eukaryotic organisms belonging to the biological kingdom Animalia (/ˌænɪˈmeɪliə/). With few exceptions, animals consume organic material, breathe oxygen, have myocytes and are able to move, can reproduce sexually, and grow from a hollow sphere of cells, the blastula, during embryonic development. Animals form a clade, meaning that they arose from a single common ancestor. Over 1.5 million living animal species have been described, of which around 1.05 million are insects, over 85,000 are molluscs, and around 65,000 are vertebrates. It has been estimated there are as many as 7.77 million animal species on Earth. Animal body lengths range from 8.5 μm (0.00033 in) to 33.6 m (110 ft). They have complex ecologies and interactions with each other and their environments, forming intricate food webs. The scientific study of animals is known as zoology, and the study of animal behaviour is known as ethology. The animal kingdom is divided into five major clades, namely Porifera, Ctenophora, Placozoa, Cnidaria and Bilateria. Most living animal species belong to the clade Bilateria, a highly proliferative clade whose members have a bilaterally symmetric and significantly cephalised body plan, and the vast majority of bilaterians belong to two large clades: the protostomes, which includes organisms such as arthropods, molluscs, flatworms, annelids and nematodes; and the deuterostomes, which include echinoderms, hemichordates and chordates, the latter of which contains the vertebrates. The much smaller basal phylum Xenacoelomorpha have an uncertain position within Bilateria. Animals first appeared in the fossil record in the late Cryogenian period and diversified in the subsequent Ediacaran period in what is known as the Avalon explosion. Nearly all modern animal phyla first appeared in the fossil record as marine species during the Cambrian explosion, which began around 539 million years ago (Mya), and most classes during the Ordovician radiation 485.4 Mya. Common to all living animals, 6,331 groups of genes have been identified that may have arisen from a single common ancestor that lived about 650 Mya during the Cryogenian period. Historically, Aristotle divided animals into those with blood and those without. Carl Linnaeus created the first hierarchical biological classification for animals in 1758 with his Systema Naturae, which Jean-Baptiste Lamarck expanded into 14 phyla by 1809. In 1874, Ernst Haeckel divided the animal kingdom into the multicellular Metazoa (now synonymous with Animalia) and the Protozoa, single-celled organisms no longer considered animals. In modern times, the biological classification of animals relies on advanced techniques, such as molecular phylogenetics, which are effective at demonstrating the evolutionary relationships between taxa. Humans make use of many other animal species for food (including meat, eggs, and dairy products), for materials (such as leather, fur, and wool), as pets and as working animals for transportation, and services. Dogs, the first domesticated animal, have been used in hunting, in security and in warfare, as have horses, pigeons and birds of prey; while other terrestrial and aquatic animals are hunted for sports, trophies or profits. Non-human animals are also an important cultural element of human evolution, having appeared in cave arts and totems since the earliest times, and are frequently featured in mythology, religion, arts, literature, heraldry, politics, and sports. Etymology The word animal comes from the Latin noun animal of the same meaning, which is itself derived from Latin animalis 'having breath or soul'. The biological definition includes all members of the kingdom Animalia. In colloquial usage, the term animal is often used to refer only to nonhuman animals. The term metazoa is derived from Ancient Greek μετα meta 'after' (in biology, the prefix meta- stands for 'later') and ζῷᾰ zōia 'animals', plural of ζῷον zōion 'animal'. A metazoan is any member of the group Metazoa. Characteristics Animals have several characteristics that they share with other living things. Animals are eukaryotic, multicellular, and aerobic, as are plants and fungi. Unlike plants and algae, which produce their own food, animals cannot produce their own food, a feature they share with fungi. Animals ingest organic material and digest it internally. Animals have structural characteristics that set them apart from all other living things: Typically, there is an internal digestive chamber with either one opening (in Ctenophora, Cnidaria, and flatworms) or two openings (in most bilaterians). Animal development is controlled by Hox genes, which signal the times and places to develop structures such as body segments and limbs. During development, the animal extracellular matrix forms a relatively flexible framework upon which cells can move about and be reorganised into specialised tissues and organs, making the formation of complex structures possible, and allowing cells to be differentiated. The extracellular matrix may be calcified, forming structures such as shells, bones, and spicules. In contrast, the cells of other multicellular organisms (primarily algae, plants, and fungi) are held in place by cell walls, and so develop by progressive growth. Nearly all animals make use of some form of sexual reproduction. They produce haploid gametes by meiosis; the smaller, motile gametes are spermatozoa and the larger, non-motile gametes are ova. These fuse to form zygotes, which develop via mitosis into a hollow sphere, called a blastula. In sponges, blastula larvae swim to a new location, attach to the seabed, and develop into a new sponge. In most other groups, the blastula undergoes more complicated rearrangement. It first invaginates to form a gastrula with a digestive chamber and two separate germ layers, an external ectoderm and an internal endoderm. In most cases, a third germ layer, the mesoderm, also develops between them. These germ layers then differentiate to form tissues and organs. Repeated instances of mating with a close relative during sexual reproduction generally leads to inbreeding depression within a population due to the increased prevalence of harmful recessive traits. Animals have evolved numerous mechanisms for avoiding close inbreeding. Some animals are capable of asexual reproduction, which often results in a genetic clone of the parent. This may take place through fragmentation; budding, such as in Hydra and other cnidarians; or parthenogenesis, where fertile eggs are produced without mating, such as in aphids. Ecology Animals are categorised into ecological groups depending on their trophic levels and how they consume organic material. Such groupings include carnivores (further divided into subcategories such as piscivores, insectivores, ovivores, etc.), herbivores (subcategorised into folivores, graminivores, frugivores, granivores, nectarivores, algivores, etc.), omnivores, fungivores, scavengers/detritivores, and parasites. Interactions between animals of each biome form complex food webs within that ecosystem. In carnivorous or omnivorous species, predation is a consumer–resource interaction where the predator feeds on another organism, its prey, who often evolves anti-predator adaptations to avoid being fed upon. Selective pressures imposed on one another lead to an evolutionary arms race between predator and prey, resulting in various antagonistic/competitive coevolutions. Almost all multicellular predators are animals. Some consumers use multiple methods; for example, in parasitoid wasps, the larvae feed on the hosts' living tissues, killing them in the process, but the adults primarily consume nectar from flowers. Other animals may have very specific feeding behaviours, such as hawksbill sea turtles which mainly eat sponges. Most animals rely on biomass and bioenergy produced by plants and phytoplanktons (collectively called producers) through photosynthesis. Herbivores, as primary consumers, eat the plant material directly to digest and absorb the nutrients, while carnivores and other animals on higher trophic levels indirectly acquire the nutrients by eating the herbivores or other animals that have eaten the herbivores. Animals oxidise carbohydrates, lipids, proteins and other biomolecules in cellular respiration, which allows the animal to grow and to sustain basal metabolism and fuel other biological processes such as locomotion. Some benthic animals living close to hydrothermal vents and cold seeps on the dark sea floor consume organic matter produced through chemosynthesis (via oxidising inorganic compounds such as hydrogen sulfide) by archaea and bacteria. Animals originated in the ocean; all extant animal phyla, except for Micrognathozoa and Onychophora, feature at least some marine species. However, several lineages of arthropods begun to colonise land around the same time as land plants, probably between 510 and 471 million years ago, during the Late Cambrian or Early Ordovician. Vertebrates such as the lobe-finned fish Tiktaalik started to move on to land in the late Devonian, about 375 million years ago. Other notable animal groups that colonized land environments are Mollusca, Platyhelmintha, Annelida, Tardigrada, Onychophora, Rotifera, Nematoda. Animals occupy virtually all of earth's habitats and microhabitats, with faunas adapted to salt water, hydrothermal vents, fresh water, hot springs, swamps, forests, pastures, deserts, air, and the interiors of other organisms. Animals are however not particularly heat tolerant; very few of them can survive at constant temperatures above 50 °C (122 °F) or in the most extreme cold deserts of continental Antarctica. The collective global geomorphic influence of animals on the processes shaping the Earth's surface remains largely understudied, with most studies limited to individual species and well-known exemplars. Diversity The blue whale (Balaenoptera musculus) is the largest animal that has ever lived, weighing up to 190 tonnes and measuring up to 33.6 metres (110 ft) long. The largest extant terrestrial animal is the African bush elephant (Loxodonta africana), weighing up to 12.25 tonnes and measuring up to 10.67 metres (35.0 ft) long. The largest terrestrial animals that ever lived were titanosaur sauropod dinosaurs such as Argentinosaurus, which may have weighed as much as 73 tonnes, and Supersaurus which may have reached 39 metres. Several animals are microscopic; some Myxozoa (obligate parasites within the Cnidaria) never grow larger than 20 μm, and one of the smallest species (Myxobolus shekel) is no more than 8.5 μm when fully grown. The following table lists estimated numbers of described extant species for the major animal phyla, along with their principal habitats (terrestrial, fresh water, and marine), and free-living or parasitic ways of life. Species estimates shown here are based on numbers described scientifically; much larger estimates have been calculated based on various means of prediction, and these can vary wildly. For instance, around 25,000–27,000 species of nematodes have been described, while published estimates of the total number of nematode species include 10,000–20,000; 500,000; 10 million; and 100 million. Using patterns within the taxonomic hierarchy, the total number of animal species—including those not yet described—was calculated to be about 7.77 million in 2011.[a] 3,000–6,500 4,000–25,000 Evolutionary origin Evidence of animals is found as long ago as the Cryogenian period. 24-Isopropylcholestane (24-ipc) has been found in rocks from roughly 650 million years ago; it is only produced by sponges and pelagophyte algae. Its likely origin is from sponges based on molecular clock estimates for the origin of 24-ipc production in both groups. Analyses of pelagophyte algae consistently recover a Phanerozoic origin, while analyses of sponges recover a Neoproterozoic origin, consistent with the appearance of 24-ipc in the fossil record. The first body fossils of animals appear in the Ediacaran, represented by forms such as Charnia and Spriggina. It had long been doubted whether these fossils truly represented animals, but the discovery of the animal lipid cholesterol in fossils of Dickinsonia establishes their nature. Animals are thought to have originated under low-oxygen conditions, suggesting that they were capable of living entirely by anaerobic respiration, but as they became specialised for aerobic metabolism they became fully dependent on oxygen in their environments. Many animal phyla first appear in the fossil record during the Cambrian explosion, starting about 539 million years ago, in beds such as the Burgess Shale. Extant phyla in these rocks include molluscs, brachiopods, onychophorans, tardigrades, arthropods, echinoderms and hemichordates, along with numerous now-extinct forms such as the predatory Anomalocaris. The apparent suddenness of the event may however be an artefact of the fossil record, rather than showing that all these animals appeared simultaneously. That view is supported by the discovery of Auroralumina attenboroughii, the earliest known Ediacaran crown-group cnidarian (557–562 mya, some 20 million years before the Cambrian explosion) from Charnwood Forest, England. It is thought to be one of the earliest predators, catching small prey with its nematocysts as modern cnidarians do. Some palaeontologists have suggested that animals appeared much earlier than the Cambrian explosion, possibly as early as 1 billion years ago. Early fossils that might represent animals appear for example in the 665-million-year-old rocks of the Trezona Formation of South Australia. These fossils are interpreted as most probably being early sponges. Trace fossils such as tracks and burrows found in the Tonian period (from 1 gya) may indicate the presence of triploblastic worm-like animals, roughly as large (about 5 mm wide) and complex as earthworms. However, similar tracks are produced by the giant single-celled protist Gromia sphaerica, so the Tonian trace fossils may not indicate early animal evolution. Around the same time, the layered mats of microorganisms called stromatolites decreased in diversity, perhaps due to grazing by newly evolved animals. Objects such as sediment-filled tubes that resemble trace fossils of the burrows of wormlike animals have been found in 1.2 gya rocks in North America, in 1.5 gya rocks in Australia and North America, and in 1.7 gya rocks in Australia. Their interpretation as having an animal origin is disputed, as they might be water-escape or other structures. Phylogeny Animals are monophyletic, meaning they are derived from a common ancestor. Animals are the sister group to the choanoflagellates, with which they form the Choanozoa. Ros-Rocher and colleagues (2021) trace the origins of animals to unicellular ancestors, providing the external phylogeny shown in the cladogram. Uncertainty of relationships is indicated with dashed lines. The animal clade had certainly originated by 650 mya, and may have come into being as much as 800 mya, based on molecular clock evidence for different phyla. Holomycota (inc. fungi) Ichthyosporea Pluriformea Filasterea The relationships at the base of the animal tree have been debated. Other than Ctenophora, the Bilateria and Cnidaria are the only groups with symmetry, and other evidence shows they are closely related. In addition to sponges, Placozoa has no symmetry and was often considered a "missing link" between protists and multicellular animals. The presence of hox genes in Placozoa shows that they were once more complex. The Porifera (sponges) have long been assumed to be sister to the rest of the animals, but there is evidence that the Ctenophora may be in that position. Molecular phylogenetics has supported both the sponge-sister and ctenophore-sister hypotheses. In 2017, Roberto Feuda and colleagues, using amino acid differences, presented both, with the following cladogram for the sponge-sister view that they supported (their ctenophore-sister tree simply interchanging the places of ctenophores and sponges): Porifera Ctenophora Placozoa Cnidaria Bilateria Conversely, a 2023 study by Darrin Schultz and colleagues uses ancient gene linkages to construct the following ctenophore-sister phylogeny: Ctenophora Porifera Placozoa Cnidaria Bilateria Sponges are physically very distinct from other animals, and were long thought to have diverged first, representing the oldest animal phylum and forming a sister clade to all other animals. Despite their morphological dissimilarity with all other animals, genetic evidence suggests sponges may be more closely related to other animals than the comb jellies are. Sponges lack the complex organisation found in most other animal phyla; their cells are differentiated, but in most cases not organised into distinct tissues, unlike all other animals. They typically feed by drawing in water through pores, filtering out small particles of food. The Ctenophora and Cnidaria are radially symmetric and have digestive chambers with a single opening, which serves as both mouth and anus. Animals in both phyla have distinct tissues, but these are not organised into discrete organs. They are diploblastic, having only two main germ layers, ectoderm and endoderm. The tiny placozoans have no permanent digestive chamber and no symmetry; they superficially resemble amoebae. Their phylogeny is poorly defined, and under active research. The remaining animals, the great majority—comprising some 29 phyla and over a million species—form the Bilateria clade, which have a bilaterally symmetric body plan. The Bilateria are triploblastic, with three well-developed germ layers, and their tissues form distinct organs. The digestive chamber has two openings, a mouth and an anus, and in the Nephrozoa there is an internal body cavity, a coelom or pseudocoelom. These animals have a head end (anterior) and a tail end (posterior), a back (dorsal) surface and a belly (ventral) surface, and a left and a right side. A modern consensus phylogenetic tree for the Bilateria is shown below. Xenacoelomorpha Ambulacraria Chordata Ecdysozoa Spiralia Having a front end means that this part of the body encounters stimuli, such as food, favouring cephalisation, the development of a head with sense organs and a mouth. Many bilaterians have a combination of circular muscles that constrict the body, making it longer, and an opposing set of longitudinal muscles, that shorten the body; these enable soft-bodied animals with a hydrostatic skeleton to move by peristalsis. They also have a gut that extends through the basically cylindrical body from mouth to anus. Many bilaterian phyla have primary larvae which swim with cilia and have an apical organ containing sensory cells. However, over evolutionary time, descendant spaces have evolved which have lost one or more of each of these characteristics. For example, adult echinoderms are radially symmetric (unlike their larvae), while some parasitic worms have extremely simplified body structures. Genetic studies have considerably changed zoologists' understanding of the relationships within the Bilateria. Most appear to belong to two major lineages, the protostomes and the deuterostomes. It is often suggested that the basalmost bilaterians are the Xenacoelomorpha, with all other bilaterians belonging to the subclade Nephrozoa. However, this suggestion has been contested, with other studies finding that xenacoelomorphs are more closely related to Ambulacraria than to other bilaterians. Protostomes and deuterostomes differ in several ways. Early in development, deuterostome embryos undergo radial cleavage during cell division, while many protostomes (the Spiralia) undergo spiral cleavage. Animals from both groups possess a complete digestive tract, but in protostomes the first opening of the embryonic gut develops into the mouth, and the anus forms secondarily. In deuterostomes, the anus forms first while the mouth develops secondarily. Most protostomes have schizocoelous development, where cells simply fill in the interior of the gastrula to form the mesoderm. In deuterostomes, the mesoderm forms by enterocoelic pouching, through invagination of the endoderm. The main deuterostome taxa are the Ambulacraria and the Chordata. Ambulacraria are exclusively marine and include acorn worms, starfish, sea urchins, and sea cucumbers. The chordates are dominated by the vertebrates (animals with backbones), which consist of fishes, amphibians, reptiles, birds, and mammals. The protostomes include the Ecdysozoa, named after their shared trait of ecdysis, growth by moulting, Among the largest ecdysozoan phyla are the arthropods and the nematodes. The rest of the protostomes are in the Spiralia, named for their pattern of developing by spiral cleavage in the early embryo. Major spiralian phyla include the annelids and molluscs. History of classification In the classical era, Aristotle divided animals,[d] based on his own observations, into those with blood (roughly, the vertebrates) and those without. The animals were then arranged on a scale from man (with blood, two legs, rational soul) down through the live-bearing tetrapods (with blood, four legs, sensitive soul) and other groups such as crustaceans (no blood, many legs, sensitive soul) down to spontaneously generating creatures like sponges (no blood, no legs, vegetable soul). Aristotle was uncertain whether sponges were animals, which in his system ought to have sensation, appetite, and locomotion, or plants, which did not: he knew that sponges could sense touch and would contract if about to be pulled off their rocks, but that they were rooted like plants and never moved about. In 1758, Carl Linnaeus created the first hierarchical classification in his Systema Naturae. In his original scheme, the animals were one of three kingdoms, divided into the classes of Vermes, Insecta, Pisces, Amphibia, Aves, and Mammalia. Since then, the last four have all been subsumed into a single phylum, the Chordata, while his Insecta (which included the crustaceans and arachnids) and Vermes have been renamed or broken up. The process was begun in 1793 by Jean-Baptiste de Lamarck, who called the Vermes une espèce de chaos ('a chaotic mess')[e] and split the group into three new phyla: worms, echinoderms, and polyps (which contained corals and jellyfish). By 1809, in his Philosophie Zoologique, Lamarck had created nine phyla apart from vertebrates (where he still had four phyla: mammals, birds, reptiles, and fish) and molluscs, namely cirripedes, annelids, crustaceans, arachnids, insects, worms, radiates, polyps, and infusorians. In his 1817 Le Règne Animal, Georges Cuvier used comparative anatomy to group the animals into four embranchements ('branches' with different body plans, roughly corresponding to phyla), namely vertebrates, molluscs, articulated animals (arthropods and annelids), and zoophytes (radiata) (echinoderms, cnidaria and other forms). This division into four was followed by the embryologist Karl Ernst von Baer in 1828, the zoologist Louis Agassiz in 1857, and the comparative anatomist Richard Owen in 1860. In 1874, Ernst Haeckel divided the animal kingdom into two subkingdoms: Metazoa (multicellular animals, with five phyla: coelenterates, echinoderms, articulates, molluscs, and vertebrates) and Protozoa (single-celled animals), including a sixth animal phylum, sponges. The protozoa were later moved to the former kingdom Protista, leaving only the Metazoa as a synonym of Animalia. In human culture The human population exploits a large number of other animal species for food, both of domesticated livestock species in animal husbandry and, mainly at sea, by hunting wild species. Marine fish of many species are caught commercially for food. A smaller number of species are farmed commercially. Humans and their livestock make up more than 90% of the biomass of all terrestrial vertebrates, and almost as much as all insects combined. Invertebrates including cephalopods, crustaceans, insects—principally bees and silkworms—and bivalve or gastropod molluscs are hunted or farmed for food, fibres. Chickens, cattle, sheep, pigs, and other animals are raised as livestock for meat across the world. Animal fibres such as wool and silk are used to make textiles, while animal sinews have been used as lashings and bindings, and leather is widely used to make shoes and other items. Animals have been hunted and farmed for their fur to make items such as coats and hats. Dyestuffs including carmine (cochineal), shellac, and kermes have been made from the bodies of insects. Working animals including cattle and horses have been used for work and transport from the first days of agriculture. Animals such as the fruit fly Drosophila melanogaster serve a major role in science as experimental models. Animals have been used to create vaccines since their discovery in the 18th century. Some medicines such as the cancer drug trabectedin are based on toxins or other molecules of animal origin. People have used hunting dogs to help chase down and retrieve animals, and birds of prey to catch birds and mammals, while tethered cormorants have been used to catch fish. Poison dart frogs have been used to poison the tips of blowpipe darts. A wide variety of animals are kept as pets, from invertebrates such as tarantulas, octopuses, and praying mantises, reptiles such as snakes and chameleons, and birds including canaries, parakeets, and parrots all finding a place. However, the most kept pet species are mammals, namely dogs, cats, and rabbits. There is a tension between the role of animals as companions to humans, and their existence as individuals with rights of their own. A wide variety of terrestrial and aquatic animals are hunted for sport. The signs of the Western and Chinese zodiacs are based on animals. In China and Japan, the butterfly has been seen as the personification of a person's soul, and in classical representation the butterfly is also the symbol of the soul. Animals have been the subjects of art from the earliest times, both historical, as in ancient Egypt, and prehistoric, as in the cave paintings at Lascaux. Major animal paintings include Albrecht Dürer's 1515 The Rhinoceros, and George Stubbs's c. 1762 horse portrait Whistlejacket. Insects, birds and mammals play roles in literature and film, such as in giant bug movies. Animals including insects and mammals feature in mythology and religion. The scarab beetle was sacred in ancient Egypt, and the cow is sacred in Hinduism. Among other mammals, deer, horses, lions, bats, bears, and wolves are the subjects of myths and worship. See also Notes References External links |
======================================== |
[SOURCE: https://en.wikipedia.org/wiki/Bil%27in,_Gaza] | [TOKENS: 474] |
Contents Bil'in, Gaza Bil'in was a Palestinian Arab village in the Gaza Subdistrict. It was depopulated by the Israel Defense Forces during the 1948 Arab–Israeli War on July 8, 1948, under Operation An-Far. It was located 39 km northeast of Gaza and the village contained two wells which supplied it with drinking water. Etymology Biʽīln derives from Canaanite/Hebrew ba'lin (ba'als). History Bil'in did not appear in records from the 16th century. It was a small village in the 19th century, settled by families that were once nomads in the Shephelah and had origins in Bayt Jibrin and Iraq al-Manshiyya. In 1838 it was noted as a Muslim village, Ba'lin, in the Gaza district, being smaller than nearby Barqusya. In 1863 Victor Guérin noted it as a small village on a mound. In 1882, the PEF's Survey of Western Palestine described it as being a small adobe village, "with no traces of antiquity". In the 1922 census of Palestine conducted by the British Mandate authorities, Bil'in had a population of 101 Muslims, increasing by the 1931 census to 127, still all Muslim, in 32 houses. In the 1945 statistics, the village together with Ard el Ishra had a population of 180 Muslims, and the land area was 8,036 dunams, according to an official land and population survey. Of this, 143 dunams were for plantations and irrigable land, 6,972 for cereals, while 6 dunams were built-up areas. Bi'lin had an elementary school which was founded in 1937 and a shrine for al-Shaykh Ya'qub. Qedma started using some of the village land after 1948. In 1992 the village site was described: "All that remains is the rubble of a few houses, with wild herbs and thorns growing on the site, along with some trees and cactus plants. The site is surrounded by barbed wire. Parts of the surrounding land are planted with mango trees and grapes, while others serve as pastures." References Bibliography External links |
======================================== |
[SOURCE: https://en.wikipedia.org/wiki/Gali_Baharav-Miara] | [TOKENS: 1138] |
Contents Gali Baharav-Miara Gali Baharav-Miara (Hebrew: גלי בהרב-מיארה; born 18 September 1959) is an Israeli lawyer who serves as the current Attorney General of Israel. Prior to her legal career she served in Unit 8200. In her legal career she served as the Tel Aviv District Attorney for Civil Affairs, and as a consultant to the law firm Tadmor, Levy, & Co. She is the first female Attorney General of Israel. Biography She was born to Shulamit Baharav (née Davidovich), a painter, and to Emanuel Baharav, who served in the Palmach and fought in the War of Independence. She holds both a bachelor's degree (1984, with honors) and a master's degree (1990) in law from Tel Aviv University, where she served as a teaching assistant and adjunct professor. In 1985, she joined the Tel Aviv District Attorney's Office, where she served for about thirty years in a variety of positions, mainly in the fields of civil and administrative law. In 2002, she was appointed director of the administrative department of the District Attorney's Office, and between 2007 and 2015 she served as Tel Aviv District Attorney for Civil Affairs. In tort cases filed by Palestinians living in the West Bank against the State of Israel, she has led a policy of opposition to the lawsuits. In 2014, she was among the candidates for the position of Director General of the Ministry of Justice, but eventually was not chosen for the position. She was chosen as one of the 50 most influential women by Globes for 2022 and 2023. After retiring from the State Attorney's Office After retiring from the State Attorney's Office, she joined the law firm Tadmor & Co. as a consultant. Since 2015, she has been a member of the Civil Service Commission's search committee, giving her opinion to the government on senior appointments whose appointments require government approval. She also served as a member of the Advisory Committee for the Administrative Courts Law and the Advisory Committee on Civil Procedure, chaired the Committee for the Examination of Appointments in Municipal Corporations in the Ministry of the Interior, chaired the Public Committee for the Examination of Private Investigations, and a member of the Administrative Courts Council. In 2018, at the request of the State Attorney's Office, she wrote an opinion defending Benny Gantz and Amir Eshel from a tort lawsuit filed in a Dutch court by a Palestinian whose family members were killed in the 2014 Gaza War. Her opinion was accepted and the claim was dismissed. In May 2019, along with former senior members of the State Attorney's Office and the office of the Attorney General of Israel, she signed a statement against initiatives to grant the Knesset an override of the Supreme Court and the extension of the Immunity Law. Attorney General of Israel In November 2021, Baharav-Miara was named by Justice Minister Gideon Saar as one of the three candidates for the position of Attorney General, and was considered his preferred candidate. Her candidacy was also supported by Prime Minister Naftali Bennett. On 7 February 2022, the government unanimously approved her appointment. Baharav-Miara is the first woman to serve in that position. Justice minister Yariv Levin initiated the process of firing her in March 2025. The cabinet unanimously voted on 23 March to dismiss Baharav-Miara from office in a vote of no confidence. The cabinet approved a "new method" in June 2025, allowing for Baharav-Miara's dismissal by a committee of cabinet members which included chair Amichai Chikli, Bezalel Smotrich, Itamar Ben-Gvir, Gila Gamliel and Michael Malchieli, rather than a committee of "legal professionals and public figures." Noam Sohlberg of the High Court of Justice called for a return to the process of dismissal described by the Shamgar Commission, which was implemented in 2000 and called for a professional committee. The government was previously unable to find members who were willing to sit on the committee. The government fired Baharav-Miara on 4 August 2025. The High Court of Justice froze her firing the same day. Communications Minister Shlomo Karhi issued a letter in which he told ministry workers to "disregard orders" issued by the Attorney General, while the State Attorney called her dismissal "highly irregular." The High Court will hear her case on 3 September, with an expanded panel of nine judges, rather than six, which will include the most senior judges. The court issued an injunction on 10 August 2025, which "suspended" the order to fire Baharav-Miara. The judges unanimously recommended on 1 September that the government should cancel its decision to fire Baharav-Miara. On 14 December the High Court unanimously annulled the firing, asserting that the government had failed to provide legal justification for its decision to change the way an AG is fired, and determining that the new system it designed was fundamentally flawed. Personal life Baharav-Miara was married to Tzion Miara, who served in senior positions in the security services. In 2002 he was diagnosed with ALS and lived with the disease for many years, before dying in June 2025. The couple has three sons and lived in Tel Aviv. References |
======================================== |
[SOURCE: https://en.wikipedia.org/wiki/Category:Construction_and_management_simulation_games] | [TOKENS: 59] |
Category:Construction and management simulation games Subcategories This category has the following 9 subcategories, out of 9 total. Pages in category "Construction and management simulation games" The following 78 pages are in this category, out of 78 total. This list may not reflect recent changes. |
======================================== |
[SOURCE: https://en.wikipedia.org/wiki/Mail_transfer_agent] | [TOKENS: 563] |
Contents Message transfer agent Within the Internet email system, a message transfer agent (MTA), mail transfer agent, or mail relay is software that transfers electronic mail messages from one computer to another using the Simple Mail Transfer Protocol. In some contexts, the alternative names mail server, mail exchanger, or MX host are used to describe an MTA. Messages exchanged across networks are passed between mail servers, including any attached data files (such as images, multimedia, or documents). These servers often keep mailboxes for email. Access to this email by end users is typically either by webmail or an email client. Operation A message transfer agent receives mail from either another MTA, a mail submission agent (MSA), or a mail user agent (MUA). The transmission details are specified by the Simple Mail Transfer Protocol (SMTP). When a recipient mailbox of a message is not hosted locally, the message is relayed, that is, forwarded to another MTA. Every time an MTA receives an email message, it adds a Received trace header field to the top of the header of the message, thereby building a sequential record of MTAs handling the message. The process of choosing a target MTA for the next hop is also described in SMTP, but can usually be overridden by configuring the MTA software with specific routes. An MTA works in the background, while the user usually interacts directly with a mail user agent. One may distinguish initial submission as first passing through an MSA—port 465 (or, for legacy reasons, optionally port 587) is used for communication between an MUA and an MSA, while port 25 is used for communication between MTAs, or from an MSA to an MTA. this distinction is clarified in RFC 8314. For recipients hosted locally, the final delivery of email to a recipient mailbox is the task of a message delivery agent (MDA). For this purpose the MTA transfers the message to the message handling service component of the message delivery agent (MDA). Upon final delivery, the Return-Path field is added to the envelope to record the return path. Transfer versus access A relay or filtering server will typically store email only briefly, but other systems keep full mailboxes for email - in which case they usually support some means for end users to access their email via a Mail User Agent (MUA), or email client. Common protocols for this are: Submission of new email from a mail client is via SMTP, typically on port 587 or 465, and is now generally restricted to servers the user has an account with-such as their Internet service provider. This is for policy, not technical, reasons so that providers have some means of holding their users accountable for the generation of spam and other forms of email abuse. See also References |
======================================== |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.