text
stringlengths
0
473k
[SOURCE: https://en.wikipedia.org/wiki/Minecraft#cite_ref-IGFwins_323-0] | [TOKENS: 12858]
Contents Minecraft Minecraft is a sandbox game developed and published by Mojang Studios. Following its initial public alpha release in 2009, it was formally released in 2011 for personal computers. The game has since been ported to numerous platforms, including mobile devices and various video game consoles. In Minecraft, players explore a procedurally generated world with virtually infinite terrain made up of voxels (cubes). They can discover and extract raw materials, craft tools and items, build structures, fight hostile mobs, and cooperate with or compete against other players in multiplayer. The game's large community offers a wide variety of user-generated content, such as modifications, servers, player skins, texture packs, and custom maps, which add new game mechanics and possibilities. Originally created by Markus "Notch" Persson using the Java programming language, Jens "Jeb" Bergensten was handed control over the game's development following its full release. In 2014, Mojang and the Minecraft intellectual property were purchased by Microsoft for US$2.5 billion; Xbox Game Studios hold the publishing rights for the Bedrock Edition, the unified cross-platform version which evolved from the Pocket Edition codebase[i] and replaced the legacy console versions. Bedrock is updated concurrently with Mojang's original Java Edition, although with numerous, generally small, differences. Minecraft is the best-selling video game in history with over 350 million copies sold. It has received critical acclaim, winning several awards and being cited as one of the greatest video games of all time. Social media, parodies, adaptations, merchandise, and the annual Minecon conventions have played prominent roles in popularizing it. The wider Minecraft franchise includes several spin-off games, such as Minecraft: Story Mode, Minecraft Dungeons, and Minecraft Legends. A film adaptation, titled A Minecraft Movie, was released in 2025 and became the second highest-grossing video game film of all time. Gameplay Minecraft is a 3D sandbox video game that has no required goals to accomplish, giving players a large amount of freedom in choosing how to play the game. The game features an optional achievement system. Gameplay is in the first-person perspective by default, but players have the option of third-person perspectives. The game world is composed of rough 3D objects—mainly cubes, referred to as blocks—representing various materials, such as dirt, stone, ores, tree trunks, water, and lava. The core gameplay revolves around picking up and placing these objects. These blocks are arranged in a voxel grid, while players can move freely around the world. Players can break, or mine, blocks and then place them elsewhere, enabling them to build things. Very few blocks are affected by gravity, instead maintaining their voxel position in the air. Players can also craft a wide variety of items, such as armor, which mitigates damage from attacks; weapons (such as swords or bows and arrows), which allow monsters and animals to be killed more easily; and tools (such as pickaxes or shovels), which break certain types of blocks more quickly. Some items have multiple tiers depending on the material used to craft them, with higher-tier items being more effective and durable. They may also freely craft helpful blocks—such as furnaces which can cook food and smelt ores, and torches that produce light—or exchange items with villagers (NPC) through trading emeralds for different goods and vice versa. The game has an inventory system, allowing players to carry a limited number of items. The in-game time system follows a day and night cycle, with one full cycle lasting for 20 real-time minutes. The game also contains a material called redstone, which can be used to make primitive mechanical devices, electrical circuits, and logic gates, allowing for the construction of many complex systems. New players are given a randomly selected default character skin out of nine possibilities, including Steve or Alex, but are able to create and upload their own skins. Players encounter various mobs (short for mobile entities) including animals, villagers, and hostile creatures. Passive mobs, such as cows, pigs, and chickens, spawn during the daytime and can be hunted for food and crafting materials, while hostile mobs—including large spiders, witches, skeletons, and zombies—spawn during nighttime or in dark places such as caves. Some hostile mobs, such as zombies and skeletons, burn under the sun if they have no headgear and are not standing in water. Other creatures unique to Minecraft include the creeper (an exploding creature that sneaks up on the player) and the enderman (a creature with the ability to teleport as well as pick up and place blocks). There are also variants of mobs that spawn in different conditions; for example, zombies have husk and drowned variants that spawn in deserts and oceans, respectively. The Minecraft environment is procedurally generated as players explore it using a map seed that is randomly chosen at the time of world creation (or manually specified by the player). Divided into biomes representing different environments with unique resources and structures, worlds are designed to be effectively infinite in traditional gameplay, though technical limits on the player have existed throughout development, both intentionally and not. Implementation of horizontally infinite generation initially resulted in a glitch termed the "Far Lands" at over 12 million blocks away from the world center, where terrain generated as wall-like, fissured patterns. The Far Lands and associated glitches were considered the effective edge of the world until they were resolved, with the current horizontal limit instead being a special impassable barrier called the world border, located 30 million blocks away. Vertical space is comparatively limited, with an unbreakable bedrock layer at the bottom and a building limit several hundred blocks into the sky. Minecraft features three independent dimensions accessible through portals and providing alternate game environments. The Overworld is the starting dimension and represents the real world, with a terrestrial surface setting including plains, mountains, forests, oceans, caves, and small sources of lava. The Nether is a hell-like underworld dimension accessed via an obsidian portal and composed mainly of lava. Mobs that populate the Nether include shrieking, fireball-shooting ghasts, alongside anthropomorphic pigs called piglins and their zombified counterparts. Piglins in particular have a bartering system, where players can give them gold ingots and receive items in return. Structures known as Nether Fortresses generate in the Nether, containing mobs such as wither skeletons and blazes, which can drop blaze rods needed to access the End dimension. The player can also choose to build an optional boss mob known as the Wither, using skulls obtained from wither skeletons and soul sand. The End can be reached through an end portal, consisting of twelve end portal frames. End portals are found in underground structures in the Overworld known as strongholds. To find strongholds, players must craft eyes of ender using an ender pearl and blaze powder. Eyes of ender can then be thrown, traveling in the direction of the stronghold. Once the player reaches the stronghold, they can place eyes of ender into each portal frame to activate the end portal. The dimension consists of islands floating in a dark, bottomless void. A boss enemy called the Ender Dragon guards the largest, central island. Killing the dragon opens access to an exit portal, which, when entered, cues the game's ending credits and the End Poem, a roughly 1,500-word work written by Irish novelist Julian Gough, which takes about nine minutes to scroll past, is the game's only narrative text, and the only text of significant length directed at the player.: 10–12 At the conclusion of the credits, the player is teleported back to their respawn point and may continue the game indefinitely. In Survival mode, players have to gather natural resources such as wood and stone found in the environment in order to craft certain blocks and items. Depending on the difficulty, monsters spawn in darker areas outside a certain radius of the character, requiring players to build a shelter in order to survive at night. The mode also has a health bar which is depleted by attacks from mobs, falls, drowning, falling into lava, suffocation, starvation, and other events. Players also have a hunger bar, which must be periodically refilled by eating food in-game unless the player is playing on peaceful difficulty. If the hunger bar is empty, the player starves. Health replenishes when players have a full hunger bar or continuously on peaceful. Upon losing all health, players die. The items in the players' inventories are dropped unless the game is reconfigured not to do so. Players then re-spawn at their spawn point, which by default is where players first spawn in the game and can be changed by sleeping in a bed or using a respawn anchor. Dropped items can be recovered if players can reach them before they despawn after 5 minutes. Players may acquire experience points (commonly referred to as "xp" or "exp") by killing mobs and other players, mining, smelting ores, animal breeding, and cooking food. Experience can then be spent on enchanting tools, armor and weapons. Enchanted items are generally more powerful, last longer, or have other special effects. The game features two more game modes based on Survival, known as Hardcore mode and Adventure mode. Hardcore mode plays identically to Survival mode, but with the game's difficulty setting locked to "Hard" and with permadeath, forcing them to delete the world or explore it as a spectator after dying. Adventure mode was added to the game in a post-launch update, and prevents the player from directly modifying the game's world. It was designed primarily for use in custom maps, allowing map designers to let players experience it as intended. In Creative mode, players have access to an infinite number of all resources and items in the game through the inventory menu and can place or mine them instantly. Players can toggle the ability to fly freely around the game world at will, and their characters usually do not take any damage nor are affected by hunger. The game mode helps players focus on building and creating projects of any size without disturbance. Multiplayer in Minecraft enables multiple players to interact and communicate with each other on a single world. It is available through direct game-to-game multiplayer, local area network (LAN) play, local split screen (console-only), and servers (player-hosted and business-hosted). Players can run their own server by making a realm, using a host provider, hosting one themselves or connect directly to another player's game via Xbox Live, PlayStation Network or Nintendo Switch Online. Single-player worlds have LAN support, allowing players to join a world on locally interconnected computers without a server setup. Minecraft multiplayer servers are guided by server operators, who have access to server commands such as setting the time of day and teleporting players. Operators can also set up restrictions concerning which usernames or IP addresses are allowed or disallowed to enter the server. Multiplayer servers have a wide range of activities, with some servers having their own unique rules and customs. The largest and most popular server is Hypixel, which has been visited by over 14 million unique players. Player versus player combat (PvP) can be enabled to allow fighting between players. In 2013, Mojang announced Minecraft Realms, a server hosting service intended to enable players to run server multiplayer games easily and safely without having to set up their own. Unlike a standard server, only invited players can join Realms servers, and these servers do not use server addresses. Minecraft: Java Edition Realms server owners can invite up to twenty people to play on their server, with up to ten players online at a time. Minecraft Realms server owners can invite up to 3,000 people to play on their server, with up to ten players online at one time. The Minecraft: Java Edition Realms servers do not support user-made plugins, but players can play custom Minecraft maps. Minecraft Bedrock Realms servers support user-made add-ons, resource packs, behavior packs, and custom Minecraft maps. At Electronic Entertainment Expo 2016, support for cross-platform play between Windows 10, iOS, and Android platforms was added through Realms starting in June 2016, with Xbox One and Nintendo Switch support to come later in 2017, and support for virtual reality devices. On 31 July 2017, Mojang released the beta version of the update allowing cross-platform play. Nintendo Switch support for Realms was released in July 2018. The modding community consists of fans, users and third-party programmers. Using a variety of application program interfaces that have arisen over time, they have produced a wide variety of downloadable content for Minecraft, such as modifications, texture packs and custom maps. Modifications of the Minecraft code, called mods, add a variety of gameplay changes, ranging from new blocks, items, and mobs to entire arrays of mechanisms. The modding community is responsible for a substantial supply of mods from ones that enhance gameplay, such as mini-maps, waypoints, and durability counters, to ones that add to the game elements from other video games and media. While a variety of mod frameworks were independently developed by reverse engineering the code, Mojang has also enhanced vanilla Minecraft with official frameworks for modification, allowing the production of community-created resource packs, which alter certain game elements including textures and sounds. Players can also create their own "maps" (custom world save files) that often contain specific rules, challenges, puzzles and quests, and share them for others to play. Mojang added an adventure mode in August 2012 and "command blocks" in October 2012, which were created specially for custom maps in Java Edition. Data packs, introduced in version 1.13 of the Java Edition, allow further customization, including the ability to add new achievements, dimensions, functions, loot tables, predicates, recipes, structures, tags, and world generation. The Xbox 360 Edition supported downloadable content, which was available to purchase via the Xbox Games Store; these content packs usually contained additional character skins. It later received support for texture packs in its twelfth title update while introducing "mash-up packs", which combined texture packs with skin packs and changes to the game's sounds, music and user interface. The first mash-up pack (and by extension, the first texture pack) for the Xbox 360 Edition was released on 4 September 2013, and was themed after the Mass Effect franchise. Unlike Java Edition, however, the Xbox 360 Edition did not support player-made mods or custom maps. A cross-promotional resource pack based on the Super Mario franchise by Nintendo was released exclusively for the Wii U Edition worldwide on 17 May 2016, and later bundled free with the Nintendo Switch Edition at launch. Another based on Fallout was released on consoles that December, and for Windows and Mobile in April 2017. In April 2018, malware was discovered in several downloadable user-made Minecraft skins for use with the Java Edition of the game. Avast stated that nearly 50,000 accounts were infected, and when activated, the malware would attempt to reformat the user's hard drive. Mojang promptly patched the issue, and released a statement stating that "the code would not be run or read by the game itself", and would run only when the image containing the skin itself was opened. In June 2017, Mojang released the "1.1 Discovery Update" to the Pocket Edition of the game, which later became the Bedrock Edition. The update introduced the "Marketplace", a catalogue of purchasable user-generated content intended to give Minecraft creators "another way to make a living from the game". Various skins, maps, texture packs and add-ons from different creators can be bought with "Minecoins", a digital currency that is purchased with real money. Additionally, users can access specific content with a subscription service titled "Marketplace Pass". Alongside content from independent creators, the Marketplace also houses items published by Mojang and Microsoft themselves, as well as official collaborations between Minecraft and other intellectual properties. By 2022, the Marketplace had over 1.7 billion content downloads, generating over $500 million in revenue. Development Before creating Minecraft, Markus "Notch" Persson was a game developer at King, where he worked until March 2009. At King, he primarily developed browser games and learned several programming languages. During his free time, he prototyped his own games, often drawing inspiration from other titles, and was an active participant on the TIGSource forums for independent developers. One such project was "RubyDung", a base-building game inspired by Dwarf Fortress, but with an isometric, three-dimensional perspective similar to RollerCoaster Tycoon. Among the features in RubyDung that he explored was a first-person view similar to Dungeon Keeper, though he ultimately discarded this idea, feeling the graphics were too pixelated at the time. Around March 2009, Persson left King and joined jAlbum, while continuing to work on his prototypes. Infiniminer, a block-based open-ended mining game first released in April 2009, inspired Persson's vision for RubyDung's future direction. Infiniminer heavily influenced the visual style of gameplay, including bringing back the first-person mode, the "blocky" visual style and the block-building fundamentals. However, unlike Infiniminer, Persson wanted Minecraft to have RPG elements. The first public alpha build of Minecraft was released on 17 May 2009 on TIGSource. Over the years, Persson regularly released test builds that added new features, including tools, mobs, and entire new dimensions. In 2011, partly due to the game's rising popularity, Persson decided to release a full 1.0 version—a second part of the "Adventure Update"—on 18 November 2011. Shortly after, Persson stepped down from development, handing the project's lead to Jens "Jeb" Bergensten. On 15 September 2014, Microsoft, the developer behind the Microsoft Windows operating system and Xbox video game console, announced a $2.5 billion acquisition of Mojang, which included the Minecraft intellectual property. Persson had suggested the deal on Twitter, asking a corporation to buy his stake in the game after receiving criticism for enforcing terms in the game's end-user license agreement (EULA), which had been in place for the past three years. According to Persson, Mojang CEO Carl Manneh received a call from a Microsoft executive shortly after the tweet, asking if Persson was serious about a deal. Mojang was also approached by other companies including Activision Blizzard and Electronic Arts. The deal with Microsoft was arbitrated on 6 November 2014 and led to Persson becoming one of Forbes' "World's Billionaires". After 2014, Minecraft's primary versions received usually annual major updates—free to players who have purchased the game— each primarily centered around a specific theme. For instance, version 1.13, the Update Aquatic, focused on ocean-related features, while version 1.16, the Nether Update, introduced significant changes to the Nether dimension. However, in late 2024, Mojang announced a shift in their update strategy; rather than releasing large updates annually, they opted for a more frequent release schedule with smaller, incremental updates, stating, "We know that you want new Minecraft content more often." The Bedrock Edition has also received regular updates, now matching the themes of the Java Edition updates. Other versions of the game, such as various console editions and the Pocket Edition, were either merged into Bedrock or discontinued and have not received further updates. On 7 May 2019, coinciding with Minecraft's 10th anniversary, a JavaScript recreation of an old 2009 Java Edition build named Minecraft Classic was made available to play online for free. On 16 April 2020, a Bedrock Edition-exclusive beta version of Minecraft, called Minecraft RTX, was released by Nvidia. It introduced physically-based rendering, real-time path tracing, and DLSS for RTX-enabled GPUs. The public release was made available on 8 December 2020. Path tracing can only be enabled in supported worlds, which can be downloaded for free via the in-game Minecraft Marketplace, with a texture pack from Nvidia's website, or with compatible third-party texture packs. It cannot be enabled by default with any texture pack on any world. Initially, Minecraft RTX was affected by many bugs, display errors, and instability issues. On 22 March 2025, a new visual mode called Vibrant Visuals, an optional graphical overhaul similar to Minecraft RTX, was announced. It promises modern rendering features—such as dynamic shadows, screen space reflections, volumetric fog, and bloom—without the need of RTX-capable hardware. Vibrant Visuals was released as a part of the Chase the Skies update on 17 June 2025 for Bedrock Edition and is planned to release on Java Edition at a later date. Development began for the original edition of Minecraft—then known as Cave Game, and now known as the Java Edition—in May 2009,[k] and ended on 13 May, when Persson released a test video on YouTube of an early version of the game, dubbed the "Cave game tech test" or the "Cave game tech demo". The game was named Minecraft: Order of the Stone the next day, after a suggestion made by a player. "Order of the Stone" came from the webcomic The Order of the Stick, and "Minecraft" was chosen "because it's a good name". The title was later shortened to just Minecraft, omitting the subtitle. Persson completed the game's base programming over a weekend in May 2009, and private testing began on TigIRC on 16 May. The first public release followed on 17 May 2009 as a developmental version shared on the TIGSource forums. Based on feedback from forum users, Persson continued updating the game. This initial public build later became known as Classic. Further developmental phases—dubbed Survival Test, Indev, and Infdev—were released throughout 2009 and 2010. The first major update, known as Alpha, was released on 30 June 2010. At the time, Persson was still working a day job at jAlbum but later resigned to focus on Minecraft full-time as sales of the alpha version surged. Updates were distributed automatically, introducing new blocks, items, mobs, and changes to game mechanics such as water flow. With revenue generated from the game, Persson founded Mojang, a video game studio, alongside former colleagues Jakob Porser and Carl Manneh. On 11 December 2010, Persson announced that Minecraft would enter its beta phase on 20 December. He assured players that bug fixes and all pre-release updates would remain free. As development progressed, Mojang expanded, hiring additional employees to work on the project. The game officially exited beta and launched in full on 18 November 2011. On 1 December 2011, Jens "Jeb" Bergensten took full creative control over Minecraft, replacing Persson as lead designer. On 28 February 2012, Mojang announced the hiring of the developers behind Bukkit, a popular developer API for Minecraft servers, to improve Minecraft's support of server modifications. This move included Mojang taking apparent ownership of the CraftBukkit server mod, though this apparent acquisition later became controversial, and its legitimacy was questioned due to CraftBukkit's open-source nature and licensing under the GNU General Public License and Lesser General Public License. In August 2011, Minecraft: Pocket Edition was released as an early alpha for the Xperia Play via the Android Market, later expanding to other Android devices on 8 October 2011. The iOS version followed on 17 November 2011. A port was made available for Windows Phones shortly after Microsoft acquired Mojang. Unlike Java Edition, Pocket Edition initially focused on Minecraft's creative building and basic survival elements but lacked many features of the PC version. Bergensten confirmed on Twitter that the Pocket Edition was written in C++ rather than Java, as iOS does not support Java. On 10 December 2014, a port of Pocket Edition was released for Windows Phone 8.1. In July 2015, a port of the Pocket Edition to Windows 10 was released as the Windows 10 Edition, with full crossplay to other Pocket versions. In January 2017, Microsoft announced that it would no longer maintain the Windows Phone versions of Pocket Edition. On 20 September 2017, with the "Better Together Update", the Pocket Edition was ported to the Xbox One, and was renamed to the Bedrock Edition. The console versions of Minecraft debuted with the Xbox 360 edition, developed by 4J Studios and released on 9 May 2012. Announced as part of the Xbox Live Arcade NEXT promotion, this version introduced a redesigned crafting system, a new control interface, in-game tutorials, split-screen multiplayer, and online play via Xbox Live. Unlike the PC version, its worlds were finite, bordered by invisible walls. Initially, the Xbox 360 version resembled outdated PC versions but received updates to bring it closer to Java Edition before eventually being discontinued. The Xbox One version launched on 5 September 2014, featuring larger worlds and support for more players. Minecraft expanded to PlayStation platforms with PlayStation 3 and PlayStation 4 editions released on 17 December 2013 and 4 September 2014, respectively. Originally planned as a PS4 launch title, it was delayed before its eventual release. A PlayStation Vita version followed in October 2014. Like the Xbox versions, the PlayStation editions were developed by 4J Studios. Nintendo platforms received Minecraft: Wii U Edition on 17 December 2015, with a physical release in North America on 17 June 2016 and in Europe on 30 June. The Nintendo Switch version launched via the eShop on 11 May 2017. During a Nintendo Direct presentation on 13 September 2017, Nintendo announced that Minecraft: New Nintendo 3DS Edition, based on the Pocket Edition, would be available for download immediately after the livestream, and a physical copy available on a later date. The game is compatible only with the New Nintendo 3DS or New Nintendo 2DS XL systems and does not work with the original 3DS or 2DS systems. On 20 September 2017, the Better Together Update introduced Bedrock Edition across Xbox One, Windows 10, VR, and mobile platforms, enabling cross-play between these versions. Bedrock Edition later expanded to Nintendo Switch and PlayStation 4, with the latter receiving the update in December 2019, allowing cross-platform play for users with a free Xbox Live account. The Bedrock Edition released a native version for PlayStation 5 on 22 October 2024, while the Xbox Series X/S version launched on 17 June 2025. On 18 December 2018, the PlayStation 3, PlayStation Vita, Xbox 360, and Wii U versions of Minecraft received their final update and would later become known as "Legacy Console Editions". On 15 January 2019, the New Nintendo 3DS version of Minecraft received its final update, effectively becoming discontinued as well. An educational version of Minecraft, designed for use in schools, launched on 1 November 2016. It is available on Android, ChromeOS, iPadOS, iOS, MacOS, and Windows. On 20 August 2018, Mojang announced that it would bring Education Edition to iPadOS in Autumn 2018. It was released to the App Store on 6 September 2018. On 27 March 2019, it was announced that it would be operated by JD.com in China. On 26 June 2020, a public beta for the Education Edition was made available to Google Play Store compatible Chromebooks. The full game was released to the Google Play Store for Chromebooks on 7 August 2020. On 20 May 2016, China Edition (also known as My World) was announced as a localized edition for China, where it was released under a licensing agreement between NetEase and Mojang. The PC edition was released for public testing on 8 August 2017. The iOS version was released on 15 September 2017, and the Android version was released on 12 October 2017. The PC edition is based on the original Java Edition, while the iOS and Android mobile versions are based on the Bedrock Edition. The edition is free-to-play and had over 700 million registered accounts by September 2023. This version of Bedrock Edition is exclusive to Microsoft's Windows 10 and Windows 11 operating systems. The beta release for Windows 10 launched on the Windows Store on 29 July 2015. After nearly a year and a half in beta, Microsoft fully released the version on 19 December 2016. Called the "Ender Update", this release implemented new features to this version of Minecraft like world templates and add-on packs. On 7 June 2022, the Java and Bedrock Editions of Minecraft were merged into a single bundle for purchase on Windows; those who owned one version would automatically gain access to the other version. Both game versions would otherwise remain separate. Around 2011, prior to Minecraft's full release, Mojang collaborated with The Lego Group to create a Lego brick-based Minecraft game called Brickcraft. This would have modified the base Minecraft game to use Lego bricks, which meant adapting the basic 1×1 block to account for larger pieces typically used in Lego sets. Persson worked on an early version called "Project Rex Kwon Do", named after the character of the same name from the film Napoleon Dynamite. Although Lego approved the project and Mojang assigned two developers for six months, it was canceled due to the Lego Group's demands, according to Mojang's Daniel Kaplan. Lego considered buying Mojang to complete the game, but when Microsoft offered over $2 billion for the company, Lego stepped back, unsure of Minecraft's potential. On 26 June 2025, a build of Brickcraft dated 28 June 2012 was published on a community archive website Omniarchive. Initially, Markus Persson planned to support the Oculus Rift with a Minecraft port. However, after Facebook acquired Oculus in 2013, he abruptly canceled the plans, stating, "Facebook creeps me out." In 2016, a community-made mod, Minecraft VR, added VR support for Java Edition, followed by Vivecraft for HTC Vive. Later that year, Microsoft introduced official Oculus Rift support for Windows 10 Edition, leading to the discontinuation of the Minecraft VR mod due to trademark complaints. Vivecraft was endorsed by Minecraft VR contributors for its Rift support. Also available is a Gear VR version, titled Minecraft: Gear VR Edition. Windows Mixed Reality support was added in 2017. On 7 September 2020, Mojang Studios announced that the PlayStation 4 Bedrock version would receive PlayStation VR support later that month. In September 2024, the Minecraft team announced they would no longer support PlayStation VR, which received its final update in March 2025. Music and sound design Minecraft's music and sound effects were produced by German musician Daniel Rosenfeld, better known as C418. To create the sound effects for the game, Rosenfeld made extensive use of Foley techniques. On learning the processes for the game, he remarked, "Foley's an interesting thing, and I had to learn its subtleties. Early on, I wasn't that knowledgeable about it. It's a whole trial-and-error process. You just make a sound and eventually you go, 'Oh my God, that's it! Get the microphone!' There's no set way of doing anything at all." He reminisced on creating the in-game sound for grass blocks, stating "It turns out that to make grass sounds you don't actually walk on grass and record it, because grass sounds like nothing. What you want to do is get a VHS, break it apart, and just lightly touch the tape." According to Rosenfeld, his favorite sound to design for the game was the hisses of spiders. He elaborates, "I like the spiders. Recording that was a whole day of me researching what a spider sounds like. Turns out, there are spiders that make little screeching sounds, so I think I got this recording of a fire hose, put it in a sampler, and just pitched it around until it sounded like a weird spider was talking to you." Many of the sound design decisions by Rosenfeld were done accidentally or spontaneously. The creeper notably lacks any specific noises apart from a loud fuse-like sound when about to explode; Rosenfeld later recalled "That was just a complete accident by Markus and me [sic]. We just put in a placeholder sound of burning a matchstick. It seemed to work hilariously well, so we kept it." On other sounds, such as those of the zombie, Rosenfeld remarked, "I actually never wanted the zombies so scary. I intentionally made them sound comical. It's nice to hear that they work so well [...]." Rosenfeld remarked that the sound engine was "terrible" to work with, remembering "If you had two song files at once, it [the game engine] would actually crash. There were so many more weird glitches like that the guys never really fixed because they were too busy with the actual game and not the sound engine." The background music in Minecraft consists of instrumental ambient music. To compose the music of Minecraft, Rosenfeld used the package from Ableton Live, along with several additional plug-ins. Speaking on them, Rosenfeld said "They can be pretty much everything from an effect to an entire orchestra. Additionally, I've got some synthesizers that are attached to the computer. Like a Moog Voyager, Dave Smith Prophet 08 and a Virus TI." On 4 March 2011, Rosenfeld released a soundtrack titled Minecraft – Volume Alpha; it includes most of the tracks featured in Minecraft, as well as other music not featured in the game. Kirk Hamilton of Kotaku chose the music in Minecraft as one of the best video game soundtracks of 2011. On 9 November 2013, Rosenfeld released the second official soundtrack, titled Minecraft – Volume Beta, which included the music that was added in a 2013 "Music Update" for the game. A physical release of Volume Alpha, consisting of CDs, black vinyl, and limited-edition transparent green vinyl LPs, was issued by indie electronic label Ghostly International on 21 August 2015. On 14 August 2020, Ghostly released Volume Beta on CD and vinyl, with alternate color LPs and lenticular cover pressings released in limited quantities. The final update Rosenfeld worked on was 2018's 1.13 Update Aquatic. His music remained the only music in the game until 2020's "Nether Update", introducing pieces from Lena Raine. Since then, other composers have made contributions, including Kumi Tanioka, Samuel Åberg, Aaron Cherof, and Amos Roddy, with Raine remaining as the new primary composer. Ownership of all music besides Rosenfeld's independently released albums has been retained by Microsoft, with their label publishing all of the other artists' releases. Gareth Coker also composed some of the music for the game's mini games from the Legacy Console editions. Rosenfeld had stated his intent to create a third album of music for the game in a 2015 interview with Fact, and confirmed its existence in a 2017 tweet, stating that his work on the record as of then had tallied up to be longer than the previous two albums combined, which in total clocks in at over 3 hours and 18 minutes. However, due to licensing issues with Microsoft, the third volume has since not seen release. On 8 January 2021, Rosenfeld was asked in an interview with Anthony Fantano whether or not there was still a third volume of his music intended for release. Rosenfeld responded, saying, "I have something—I consider it finished—but things have become complicated, especially as Minecraft is now a big property, so I don't know." Reception Minecraft has received critical acclaim, with praise for the creative freedom it grants players in-game, as well as the ease of enabling emergent gameplay. Critics have expressed enjoyment in Minecraft's complex crafting system, commenting that it is an important aspect of the game's open-ended gameplay. Most publications were impressed by the game's "blocky" graphics, with IGN describing them as "instantly memorable". Reviewers also liked the game's adventure elements, noting that the game creates a good balance between exploring and building. The game's multiplayer feature has been generally received favorably, with IGN commenting that "adventuring is always better with friends". Jaz McDougall of PC Gamer said Minecraft is "intuitively interesting and contagiously fun, with an unparalleled scope for creativity and memorable experiences". It has been regarded as having introduced millions of children to the digital world, insofar as its basic game mechanics are logically analogous to computer commands. IGN was disappointed about the troublesome steps needed to set up multiplayer servers, calling it a "hassle". Critics also said that visual glitches occur periodically. Despite its release out of beta in 2011, GameSpot said the game had an "unfinished feel", adding that some game elements seem "incomplete or thrown together in haste". A review of the alpha version, by Scott Munro of the Daily Record, called it "already something special" and urged readers to buy it. Jim Rossignol of Rock Paper Shotgun also recommended the alpha of the game, calling it "a kind of generative 8-bit Lego Stalker". On 17 September 2010, gaming webcomic Penny Arcade began a series of comics and news posts about the addictiveness of the game. The Xbox 360 version was generally received positively by critics, but did not receive as much praise as the PC version. Although reviewers were disappointed by the lack of features such as mod support and content from the PC version, they acclaimed the port's addition of a tutorial and in-game tips and crafting recipes, saying that they make the game more user-friendly. The Xbox One Edition was one of the best received ports, being praised for its relatively large worlds. The PlayStation 3 Edition also received generally favorable reviews, being compared to the Xbox 360 Edition and praised for its well-adapted controls. The PlayStation 4 edition was the best received port to date, being praised for having 36 times larger worlds than the PlayStation 3 edition and described as nearly identical to the Xbox One edition. The PlayStation Vita Edition received generally positive reviews from critics but was noted for its technical limitations. The Wii U version received generally positive reviews from critics but was noted for a lack of GamePad integration. The 3DS version received mixed reviews, being criticized for its high price, technical issues, and lack of cross-platform play. The Nintendo Switch Edition received fairly positive reviews from critics, being praised, like other modern ports, for its relatively larger worlds. Minecraft: Pocket Edition initially received mixed reviews from critics. Although reviewers appreciated the game's intuitive controls, they were disappointed by the lack of content. The inability to collect resources and craft items, as well as the limited types of blocks and lack of hostile mobs, were especially criticized. After updates added more content, Pocket Edition started receiving more positive reviews. Reviewers complimented the controls and the graphics, but still noted a lack of content. Minecraft surpassed over a million purchases less than a month after entering its beta phase in early 2011. At the same time, the game had no publisher backing and has never been commercially advertised except through word of mouth, and various unpaid references in popular media such as the Penny Arcade webcomic. By April 2011, Persson estimated that Minecraft had made €23 million (US$33 million) in revenue, with 800,000 sales of the alpha version of the game, and over 1 million sales of the beta version. In November 2011, prior to the game's full release, Minecraft beta surpassed 16 million registered users and 4 million purchases. By March 2012, Minecraft had become the 6th best-selling PC game of all time. As of 10 October 2014[update], the game had sold 17 million copies on PC, becoming the best-selling PC game of all time. On 25 February 2014, the game reached 100 million registered users. By May 2019, 180 million copies had been sold across all platforms, making it the single best-selling video game of all time. The free-to-play Minecraft China version had over 700 million registered accounts by September 2023. By 2023, the game had sold over 300 million copies. As of April 2025, Minecraft has sold over 350 million copies. The Xbox 360 version of Minecraft became profitable within the first day of the game's release in 2012, when the game broke the Xbox Live sales records with 400,000 players online. Within a week of being on the Xbox Live Marketplace, Minecraft sold a million copies. GameSpot announced in December 2012 that Minecraft sold over 4.48 million copies since the game debuted on Xbox Live Arcade in May 2012. In 2012, Minecraft was the most purchased title on Xbox Live Arcade; it was also the fourth most played title on Xbox Live based on average unique users per day. As of 4 April 2014[update], the Xbox 360 version has sold 12 million copies. In addition, Minecraft: Pocket Edition has reached a figure of 21 million in sales. The PlayStation 3 Edition sold one million copies in five weeks. The release of the game's PlayStation Vita version boosted Minecraft sales by 79%, outselling both PS3 and PS4 debut releases and becoming the largest Minecraft launch on a PlayStation console. The PS Vita version sold 100,000 digital copies in Japan within the first two months of release, according to an announcement by SCE Japan Asia. By January 2015, 500,000 digital copies of Minecraft were sold in Japan across all PlayStation platforms, with a surge in primary school children purchasing the PS Vita version. As of 2022, the Vita version has sold over 1.65 million physical copies in Japan, making it the best-selling Vita game in the country. Minecraft helped improve Microsoft's total first-party revenue by $63 million for the 2015 second quarter. The game, including all of its versions, had over 112 million monthly active players by September 2019. On its 11th anniversary in May 2020, the company announced that Minecraft had reached over 200 million copies sold across platforms with over 126 million monthly active players. By April 2021, the number of active monthly users had climbed to 140 million. In July 2010, PC Gamer listed Minecraft as the fourth-best game to play at work. In December of that year, Good Game selected Minecraft as their choice for Best Downloadable Game of 2010, Gamasutra named it the eighth best game of the year as well as the eighth best indie game of the year, and Rock, Paper, Shotgun named it the "game of the year". Indie DB awarded the game the 2010 Indie of the Year award as chosen by voters, in addition to two out of five Editor's Choice awards for Most Innovative and Best Singleplayer Indie. It was also awarded Game of the Year by PC Gamer UK. The game was nominated for the Seumas McNally Grand Prize, Technical Excellence, and Excellence in Design awards at the March 2011 Independent Games Festival and won the Grand Prize and the community-voted Audience Award. At Game Developers Choice Awards 2011, Minecraft won awards in the categories for Best Debut Game, Best Downloadable Game and Innovation Award, winning every award for which it was nominated. It also won GameCity's video game arts award. On 5 May 2011, Minecraft was selected as one of the 80 games that would be displayed at the Smithsonian American Art Museum as part of The Art of Video Games exhibit that opened on 16 March 2012. At the 2011 Spike Video Game Awards, Minecraft won the award for Best Independent Game and was nominated in the Best PC Game category. In 2012, at the British Academy Video Games Awards, Minecraft was nominated in the GAME Award of 2011 category and Persson received The Special Award. In 2012, Minecraft XBLA was awarded a Golden Joystick Award in the Best Downloadable Game category, and a TIGA Games Industry Award in the Best Arcade Game category. In 2013, it was nominated as the family game of the year at the British Academy Video Games Awards. During the 16th Annual D.I.C.E. Awards, the Academy of Interactive Arts & Sciences nominated the Xbox 360 version of Minecraft for "Strategy/Simulation Game of the Year". Minecraft Console Edition won the award for TIGA Game Of The Year in 2014. In 2015, the game placed 6th on USgamer's The 15 Best Games Since 2000 list. In 2016, Minecraft placed 6th on Time's The 50 Best Video Games of All Time list. Minecraft was nominated for the 2013 Kids' Choice Awards for Favorite App, but lost to Temple Run. It was nominated for the 2014 Kids' Choice Awards for Favorite Video Game, but lost to Just Dance 2014. The game later won the award for the Most Addicting Game at the 2015 Kids' Choice Awards. In addition, the Java Edition was nominated for "Favorite Video Game" at the 2018 Kids' Choice Awards, while the game itself won the "Still Playing" award at the 2019 Golden Joystick Awards, as well as the "Favorite Video Game" award at the 2020 Kids' Choice Awards. Minecraft also won "Stream Game of the Year" at inaugural Streamer Awards in 2021. The game later garnered a Nickelodeon Kids' Choice Award nomination for Favorite Video Game in 2021, and won the same category in 2022 and 2023. At the Golden Joystick Awards 2025, it won the Still Playing Award - PC and Console. Minecraft has been subject to several notable controversies. In June 2014, Mojang announced that it would begin enforcing the portion of Minecraft's end-user license agreement (EULA) which prohibits servers from giving in-game advantages to players in exchange for donations or payments. Spokesperson Owen Hill stated that servers could still require players to pay a fee to access the server and could sell in-game cosmetic items. The change was supported by Persson, citing emails he received from parents of children who had spent hundreds of dollars on servers. The Minecraft community and server owners protested, arguing that the EULA's terms were more broad than Mojang was claiming, that the crackdown would force smaller servers to shut down for financial reasons, and that Mojang was suppressing competition for its own Minecraft Realms subscription service. The controversy contributed to Notch's decision to sell Mojang. In 2020, Mojang announced an eventual change to the Java Edition to require a login from a Microsoft account rather than a Mojang account, the latter of which would be sunsetted. This also required Java Edition players to create Xbox network Gamertags. Mojang defended the move to Microsoft accounts by saying that improved security could be offered, including two-factor authentication, blocking cyberbullies in chat, and improved parental controls. The community responded with intense backlash, citing various technical difficulties encountered in the process and how account migration would be mandatory, even for those who do not play on servers. As of 10 March 2022, Microsoft required that all players migrate in order to maintain access the Java Edition of Minecraft. Mojang announced a deadline of 19 September 2023 for account migration, after which all legacy Mojang accounts became inaccessible and unable to be migrated. In June 2022, Mojang added a player-reporting feature in Java Edition. Players could report other players on multiplayer servers for sending messages prohibited by the Xbox Live Code of Conduct; report categories included profane language,[l] substance abuse, hate speech, threats of violence, and nudity. If a player was found to be in violation of Xbox Community Standards, they would be banned from all servers for a specific period of time or permanently. The update containing the report feature (1.19.1) was released on 27 July 2022. Mojang received substantial backlash and protest from community members, one of the most common complaints being that banned players would be forbidden from joining any server, even private ones. Others took issue to what they saw as Microsoft increasing control over its player base and exercising censorship, leading some to start a hashtag #saveminecraft and dub the version "1.19.84", a reference to the dystopian novel Nineteen Eighty-Four. The "Mob Vote" was an online event organized by Mojang in which the Minecraft community voted between three original mob concepts; initially, the winning mob was to be implemented in a future update, while the losing mobs were scrapped, though after the first mob vote this was changed, and losing mobs would now have a chance to come to the game in the future. The first Mob Vote was held during Minecon Earth 2017 and became an annual event starting with Minecraft Live 2020. The Mob Vote was often criticized for forcing players to choose one mob instead of implementing all three, causing divisions and flaming within the community, and potentially allowing internet bots and Minecraft content creators with large fanbases to conduct vote brigading. The Mob Vote was also blamed for a perceived lack of new content added to Minecraft since Microsoft's acquisition of Mojang in 2014. The 2023 Mob Vote featured three passive mobs—the crab, the penguin, and the armadillo—with voting scheduled to start on 13 October. In response, a Change.org petition was created on 6 October, demanding that Mojang eliminate the Mob Vote and instead implement all three mobs going forward. The petition received approximately 445,000 signatures by 13 October and was joined by calls to boycott the Mob Vote, as well as a partially tongue-in-cheek "revolutionary" propaganda campaign in which sympathizers created anti-Mojang and pro-boycott posters in the vein of real 20th century propaganda posters. Mojang did not release an official response to the boycott, and the Mob Vote otherwise proceeded normally, with the armadillo winning the vote. In September 2024, as part of a blog post detailing their future plans for Minecraft's development, Mojang announced the Mob Vote would be retired. Cultural impact In September 2019, The Guardian classified Minecraft as the best video game of the 21st century to date, and in November 2019, Polygon called it the "most important game of the decade" in its 2010s "decade in review". In June 2020, Minecraft was inducted into the World Video Game Hall of Fame. Minecraft is recognized as one of the first successful games to use an early access model to draw in sales prior to its full release version to help fund development. As Minecraft helped to bolster indie game development in the early 2010s, it also helped to popularize the use of the early access model in indie game development. Social media sites such as YouTube, Facebook, and Reddit have played a significant role in popularizing Minecraft. Research conducted by the Annenberg School for Communication at the University of Pennsylvania showed that one-third of Minecraft players learned about the game via Internet videos. In 2010, Minecraft-related videos began to gain influence on YouTube, often made by commentators. The videos usually contain screen-capture footage of the game and voice-overs. Common coverage in the videos includes creations made by players, walkthroughs of various tasks, and parodies of works in popular culture. By May 2012, over four million Minecraft-related YouTube videos had been uploaded. The game would go on to be a prominent fixture within YouTube's gaming scene during the entire 2010s; in 2014, it was the second-most searched term on the entire platform. By 2018, it was still YouTube's biggest game globally. Some popular commentators have received employment at Machinima, a now-defunct gaming video company that owned a highly watched entertainment channel on YouTube. The Yogscast is a British company that regularly produces Minecraft videos; their YouTube channel has attained billions of views, and their panel at Minecon 2011 had the highest attendance. Another well-known YouTube personality is Jordan Maron, known online as CaptainSparklez, who has also created many Minecraft music parodies, including "Revenge", a parody of Usher's "DJ Got Us Fallin' in Love". Minecraft's popularity on YouTube was described by Polygon as quietly dominant, although in 2019, thanks in part to PewDiePie's playthroughs of the game, Minecraft experienced a visible uptick in popularity on the platform. Longer-running series include Far Lands or Bust, dedicated to reaching the obsolete "Far Lands" glitch by foot on an older version of the game. YouTube announced that on 14 December 2021 that the total amount of Minecraft-related views on the website had exceeded one trillion. Minecraft has been referenced by other video games, such as Torchlight II, Team Fortress 2, Borderlands 2, Choplifter HD, Super Meat Boy, The Elder Scrolls V: Skyrim, The Binding of Isaac, The Stanley Parable, and FTL: Faster Than Light. Minecraft is officially represented in downloadable content for the crossover fighter Super Smash Bros. Ultimate, with Steve as a playable character with a moveset including references to building, crafting, and redstone, alongside an Overworld-themed stage. It was also referenced by electronic music artist Deadmau5 in his performances. The game is also referenced heavily in "Informative Murder Porn", the second episode of the seventeenth season of the animated television series South Park. In 2025, A Minecraft Movie was released. It made $313 million in the box office in the first week, a record-breaking opening for a video game adaptation. Minecraft has been noted as a cultural touchstone for Generation Z, as many of the generation's members played the game at a young age. The possible applications of Minecraft have been discussed extensively, especially in the fields of computer-aided design (CAD) and education. In a panel at Minecon 2011, a Swedish developer discussed the possibility of using the game to redesign public buildings and parks, stating that rendering using Minecraft was much more user-friendly for the community, making it easier to envision the functionality of new buildings and parks. In 2012, a member of the Human Dynamics group at the MIT Media Lab, Cody Sumter, said: "Notch hasn't just built a game. He's tricked 40 million people into learning to use a CAD program." Various software has been developed to allow virtual designs to be printed using professional 3D printers or personal printers such as MakerBot and RepRap. In September 2012, Mojang began the Block by Block project in cooperation with UN Habitat to create real-world environments in Minecraft. The project allows young people who live in those environments to participate in designing the changes they would like to see. Using Minecraft, the community has helped reconstruct the areas of concern, and citizens are invited to enter the Minecraft servers and modify their own neighborhood. Carl Manneh, Mojang's managing director, called the game "the perfect tool to facilitate this process", adding "The three-year partnership will support UN-Habitat's Sustainable Urban Development Network to upgrade 300 public spaces by 2016." Mojang signed Minecraft building community, FyreUK, to help render the environments into Minecraft. The first pilot project began in Kibera, one of Nairobi's informal settlements and is in the planning phase. The Block by Block project is based on an earlier initiative started in October 2011, Mina Kvarter (My Block), which gave young people in Swedish communities a tool to visualize how they wanted to change their part of town. According to Manneh, the project was a helpful way to visualize urban planning ideas without necessarily having a training in architecture. The ideas presented by the citizens were a template for political decisions. In April 2014, the Danish Geodata Agency generated all of Denmark in fullscale in Minecraft based on their own geodata. This is possible because Denmark is one of the flattest countries with the highest point at 171 meters (ranking as the country with the 30th smallest elevation span), where the limit in default Minecraft was around 192 meters above in-game sea level when the project was completed. Taking advantage of the game's accessibility where other websites are censored, the non-governmental organization Reporters Without Borders has used an open Minecraft server to create the Uncensored Library, a repository within the game of journalism by authors from countries (including Egypt, Mexico, Russia, Saudi Arabia and Vietnam) who have been censored and arrested, such as Jamal Khashoggi. The neoclassical virtual building was created over about 250 hours by an international team of 24 people. Despite its unpredictable nature, Minecraft speedrunning, where players time themselves from spawning into a new world to reaching The End and defeating the Ender Dragon boss, is popular. Some speedrunners use a combination of mods, external programs, and debug menus, while other runners play the game in a more vanilla or more consistency-oriented way. Minecraft has been used in educational settings through initiatives such as MinecraftEdu, founded in 2011 to make the game affordable and accessible for schools in collaboration with Mojang. MinecraftEdu provided features allowing teachers to monitor student progress, including screenshot submissions as evidence of lesson completion, and by 2012 reported that approximately 250,000 students worldwide had access to the platform. Mojang also developed Minecraft: Education Edition with pre-built lesson plans for up to 30 students in a closed environment. Educators have used Minecraft to teach subjects such as history, language arts, and science through custom-built environments, including reconstructions of historical landmarks and large-scale models of biological structures such as animal cells. The introduction of redstone blocks enabled the construction of functional virtual machines such as a hard drive and an 8-bit computer. Mods have been created to use these mechanics for teaching programming. In 2014, the British Museum announced a project to reproduce its building and exhibits in Minecraft in collaboration with the public. Microsoft and Code.org have offered Minecraft-based tutorials and activities designed to teach programming, reporting by 2018 that more than 85 million children had used their resources. In 2025, the Musée de Minéralogie in Paris held a temporary exhibition titled "Minerals in Minecraft." Following the initial surge in popularity of Minecraft in 2010, other video games were criticised for having various similarities to Minecraft, and some were described as being "clones", often due to a direct inspiration from Minecraft, or a superficial similarity. Examples include Ace of Spades, CastleMiner, CraftWorld, FortressCraft, Terraria, BlockWorld 3D, Total Miner, and Luanti (formerly Minetest). David Frampton, designer of The Blockheads, reported that one failure of his 2D game was the "low resolution pixel art" that too closely resembled the art in Minecraft, which resulted in "some resistance" from fans. A homebrew adaptation of the alpha version of Minecraft for the Nintendo DS, titled DScraft, has been released; it has been noted for its similarity to the original game considering the technical limitations of the system. In response to Microsoft's acquisition of Mojang and their Minecraft IP, various developers announced further clone titles developed specifically for Nintendo's consoles, as they were the only major platforms not to officially receive Minecraft at the time. These clone titles include UCraft (Nexis Games), Cube Life: Island Survival (Cypronia), Discovery (Noowanda), Battleminer (Wobbly Tooth Games), Cube Creator 3D (Big John Games), and Stone Shire (Finger Gun Games). Despite this, the fears of fans were unfounded, with official Minecraft releases on Nintendo consoles eventually resuming. Markus Persson made another similar game, Minicraft, for a Ludum Dare competition in 2011. In 2025, Persson announced through a poll on his X account that he was considering developing a spiritual successor to Minecraft. He later clarified that he was "100% serious", and that he had "basically announced Minecraft 2". Within days, however, Persson cancelled the plans after speaking to his team. In November 2024, artificial intelligence companies Decart and Etched released Oasis, an artificially generated version of Minecraft, as a proof of concept. Every in-game element is completely AI-generated in real time and the model does not store world data, leading to "hallucinations" such as items and blocks appearing that were not there before. In January 2026, indie game developer Unomelon announced that their voxel sandbox game Allumeria would be playable in Steam Next Fest that year. On 10 February, Mojang issued a DMCA takedown of Allumeria on Steam through Valve, alleging the game was infringing on Minecraft's copyright. Some reports suggested that the takedown may have used an automatic AI copyright claiming service. The DMCA was later withdrawn. Minecon was an annual official fan convention dedicated to Minecraft. The first full Minecon was held in November 2011 at the Mandalay Bay Hotel and Casino in Las Vegas. The event included the official launch of Minecraft; keynote speeches, including one by Persson; building and costume contests; Minecraft-themed breakout classes; exhibits by leading gaming and Minecraft-related companies; commemorative merchandise; and autograph and picture times with Mojang employees and well-known contributors from the Minecraft community. In 2016, Minecon was held in-person for the last time, with the following years featuring annual "Minecon Earth" livestreams on minecraft.net and YouTube instead. These livestreams, later rebranded to "Minecraft Live", included the mob/biome votes, and announcements of new game updates. In 2025, "Minecraft Live" became a biannual event as part of Minecraft's changing update schedule.[citation needed] Notes References External links
========================================
[SOURCE: https://en.wikipedia.org/wiki/Black_hole#Spin_and_angular_momentum] | [TOKENS: 13839]
Contents Black hole A black hole is an astronomical body so compact that its gravity prevents anything, including light, from escaping. Albert Einstein's theory of general relativity predicts that a sufficiently compact mass will form a black hole. The boundary of no escape is called the event horizon. In general relativity, a black hole's event horizon seals an object's fate but produces no locally detectable change when crossed. General relativity also predicts that every black hole should have a central singularity, where the curvature of spacetime is infinite. In many ways, a black hole acts like an ideal black body, as it reflects no light. Quantum field theory in curved spacetime predicts that event horizons emit Hawking radiation, with the same spectrum as a black body of a temperature inversely proportional to its mass. This temperature is of the order of billionths of a kelvin for stellar black holes, making it essentially impossible to observe directly. Objects whose gravitational fields are too strong for light to escape were first considered in the 18th century by John Michell and Pierre-Simon Laplace. In 1916, Karl Schwarzschild found the first modern solution of general relativity that would characterise a black hole. Due to his influential research, the Schwarzschild metric is named after him. David Finkelstein, in 1958, first interpreted Schwarzschild's model as a region of space from which nothing can escape. Black holes were long considered a mathematical curiosity; it was not until the 1960s that theoretical work showed they were a generic prediction of general relativity. The first black hole known was Cygnus X-1, identified by several researchers independently in 1971. Black holes typically form when massive stars collapse at the end of their life cycle. After a black hole has formed, it can grow by absorbing mass from its surroundings. Supermassive black holes of millions of solar masses may form by absorbing other stars and merging with other black holes, or via direct collapse of gas clouds. There is consensus that supermassive black holes exist in the centres of most galaxies. The presence of a black hole can be inferred through its interaction with other matter and with electromagnetic radiation such as visible light. Matter falling toward a black hole can form an accretion disk of infalling plasma, heated by friction and emitting light. In extreme cases, this creates a quasar, some of the brightest objects in the universe. Merging black holes can also be detected by observation of the gravitational waves they emit. If other stars are orbiting a black hole, their orbits can be used to determine the black hole's mass and location. Such observations can be used to exclude possible alternatives such as neutron stars. In this way, astronomers have identified numerous stellar black hole candidates in binary systems and established that the radio source known as Sagittarius A*, at the core of the Milky Way galaxy, contains a supermassive black hole of about 4.3 million solar masses. History The idea of a body so massive that even light could not escape was first proposed in the late 18th century by English astronomer and clergyman John Michell and independently by French scientist Pierre-Simon Laplace. Both scholars proposed very large stars in contrast to the modern concept of an extremely dense object. Michell's idea, in a short part of a letter published in 1784, calculated that a star with the same density but 500 times the radius of the sun would not let any emitted light escape; the surface escape velocity would exceed the speed of light.: 122 Michell correctly hypothesized that such supermassive but non-radiating bodies might be detectable through their gravitational effects on nearby visible bodies. In 1796, Laplace mentioned that a star could be invisible if it were sufficiently large while speculating on the origin of the Solar System in his book Exposition du Système du Monde. Franz Xaver von Zach asked Laplace for a mathematical analysis, which Laplace provided and published in a journal edited by von Zach. In 1905, Albert Einstein showed that the laws of electromagnetism would be invariant under a Lorentz transformation: they would be identical for observers travelling at different velocities relative to each other. This discovery became known as the principle of special relativity. Although the laws of mechanics had already been shown to be invariant, gravity remained yet to be included.: 19 In 1907, Einstein published a paper proposing his equivalence principle, the hypothesis that inertial mass and gravitational mass have a common cause. Using the principle, Einstein predicted the redshift and half of the lensing effect of gravity on light; the full prediction of gravitational lensing required development of general relativity.: 19 By 1915, Einstein refined these ideas into his general theory of relativity, which explained how matter affects spacetime, which in turn affects the motion of other matter. This formed the basis for black hole physics. Only a few months after Einstein published the field equations describing general relativity, astrophysicist Karl Schwarzschild set out to apply the idea to stars. He assumed spherical symmetry with no spin and found a solution to Einstein's equations.: 124 A few months after Schwarzschild, Johannes Droste, a student of Hendrik Lorentz, independently gave the same solution. At a certain radius from the center of the mass, the Schwarzschild solution became singular, meaning that some of the terms in the Einstein equations became infinite. The nature of this radius, which later became known as the Schwarzschild radius, was not understood at the time. Many physicists of the early 20th century were skeptical of the existence of black holes. In a 1926 popular science book, Arthur Eddington critiqued the idea of a star with mass compressed to its Schwarzschild radius as a flaw in the then-poorly-understood theory of general relativity.: 134 In 1939, Einstein himself used his theory of general relativity in an attempt to prove that black holes were impossible. His work relied on increasing pressure or increasing centrifugal force balancing the force of gravity so that the object would not collapse beyond its Schwarzschild radius. He missed the possibility that implosion would drive the system below this critical value.: 135 By the 1920s, astronomers had classified a number of white dwarf stars as too cool and dense to be explained by the gradual cooling of ordinary stars. In 1926, Ralph Fowler showed that quantum-mechanical degeneracy pressure was larger than thermal pressure at these densities.: 145 In 1931, Subrahmanyan Chandrasekhar calculated that a non-rotating body of electron-degenerate matter below a certain limiting mass is stable, and by 1934 he showed that this explained the catalog of white dwarf stars.: 151 When Chandrasekhar announced his results, Eddington pointed out that stars above this limit would radiate until they were sufficiently dense to prevent light from exiting, a conclusion he considered absurd. Eddington and, later, Lev Landau argued that some yet unknown mechanism would stop the collapse. In the 1930s, Fritz Zwicky and Walter Baade studied stellar novae, focusing on exceptionally bright ones they called supernovae. Zwicky promoted the idea that supernovae produced stars with the density of atomic nuclei—neutron stars—but this idea was largely ignored.: 171 In 1939, based on Chandrasekhar's reasoning, J. Robert Oppenheimer and George Volkoff predicted that neutron stars below a certain mass limit, later called the Tolman–Oppenheimer–Volkoff limit, would be stable due to neutron degeneracy pressure. Above that limit, they reasoned that either their model would not apply or that gravitational contraction would not stop.: 380 John Archibald Wheeler and two of his students resolved questions about the model behind the Tolman–Oppenheimer–Volkoff (TOV) limit. Harrison and Wheeler developed the equations of state relating density to pressure for cold matter all the way through electron degeneracy and neutron degeneracy. Masami Wakano and Wheeler then used the equations to compute the equilibrium curve for stars, relating mass to circumference. They found no additional features that would invalidate the TOV limit. This meant that the only thing that could prevent black holes from forming was a dynamic process ejecting sufficient mass from a star as it cooled.: 205 The modern concept of black holes was formulated by Robert Oppenheimer and his student Hartland Snyder in 1939.: 80 In the paper, Oppenheimer and Snyder solved Einstein's equations of general relativity for an idealized imploding star, in a model later called the Oppenheimer–Snyder model, then described the results from far outside the star. The implosion starts as one might expect: the star material rapidly collapses inward. However, as the density of the star increases, gravitational time dilation increases and the collapse, viewed from afar, seems to slow down further and further until the star reaches its Schwarzschild radius, where it appears frozen in time.: 217 In 1958, David Finkelstein identified the Schwarzschild surface as an event horizon, calling it "a perfect unidirectional membrane: causal influences can cross it in only one direction". In this sense, events that occur inside of the black hole cannot affect events that occur outside of the black hole. Finkelstein created a new reference frame to include the point of view of infalling observers.: 103 Finkelstein's new frame of reference allowed events at the surface of an imploding star to be related to events far away. By 1962 the two points of view were reconciled, convincing many skeptics that implosion into a black hole made physical sense.: 226 The era from the mid-1960s to the mid-1970s was the "golden age of black hole research", when general relativity and black holes became mainstream subjects of research.: 258 In this period, more general black hole solutions were found. In 1963, Roy Kerr found the exact solution for a rotating black hole. Two years later, Ezra Newman found the cylindrically symmetric solution for a black hole that is both rotating and electrically charged. In 1967, Werner Israel found that the Schwarzschild solution was the only possible solution for a nonspinning, uncharged black hole, meaning that a Schwarzschild black hole would be defined by its mass alone. Similar identities were later found for Reissner-Nordstrom and Kerr black holes, defined only by their mass and their charge or spin respectively. Together, these findings became known as the no-hair theorem, which states that a stationary black hole is completely described by the three parameters of the Kerr–Newman metric: mass, angular momentum, and electric charge. At first, it was suspected that the strange mathematical singularities found in each of the black hole solutions only appeared due to the assumption that a black hole would be perfectly spherically symmetric, and therefore the singularities would not appear in generic situations where black holes would not necessarily be symmetric. This view was held in particular by Vladimir Belinski, Isaak Khalatnikov, and Evgeny Lifshitz, who tried to prove that no singularities appear in generic solutions, although they would later reverse their positions. However, in 1965, Roger Penrose proved that general relativity without quantum mechanics requires that singularities appear in all black holes. Astronomical observations also made great strides during this era. In 1967, Antony Hewish and Jocelyn Bell Burnell discovered pulsars and by 1969, these were shown to be rapidly rotating neutron stars. Until that time, neutron stars, like black holes, were regarded as just theoretical curiosities, but the discovery of pulsars showed their physical relevance and spurred a further interest in all types of compact objects that might be formed by gravitational collapse. Based on observations in Greenwich and Toronto in the early 1970s, Cygnus X-1, a galactic X-ray source discovered in 1964, became the first astronomical object commonly accepted to be a black hole. Work by James Bardeen, Jacob Bekenstein, Carter, and Hawking in the early 1970s led to the formulation of black hole thermodynamics. These laws describe the behaviour of a black hole in close analogy to the laws of thermodynamics by relating mass to energy, area to entropy, and surface gravity to temperature. The analogy was completed: 442 when Hawking, in 1974, showed that quantum field theory implies that black holes should radiate like a black body with a temperature proportional to the surface gravity of the black hole, predicting the effect now known as Hawking radiation. While Cygnus X-1, a stellar-mass black hole, was generally accepted by the scientific community as a black hole by the end of 1973, it would be decades before a supermassive black hole would gain the same broad recognition. Although, as early as the 1960s, physicists such as Donald Lynden-Bell and Martin Rees had suggested that powerful quasars in the center of galaxies were powered by accreting supermassive black holes, little observational proof existed at the time. However, the Hubble Space Telescope, launched decades later, found that supermassive black holes were not only present in these active galactic nuclei, but that supermassive black holes in the center of galaxies were ubiquitous: Almost every galaxy had a supermassive black hole at its center, many of which were quiescent. In 1999, David Merritt proposed the M–sigma relation, which related the dispersion of the velocity of matter in the center bulge of a galaxy to the mass of the supermassive black hole at its core. Subsequent studies confirmed this correlation. Around the same time, based on telescope observations of the velocities of stars at the center of the Milky Way galaxy, independent work groups led by Andrea Ghez and Reinhard Genzel concluded that the compact radio source in the center of the galaxy, Sagittarius A*, was likely a supermassive black hole. On 11 February 2016, the LIGO Scientific Collaboration and Virgo Collaboration announced the first direct detection of gravitational waves, named GW150914, representing the first observation of a black hole merger. At the time of the merger, the black holes were approximately 1.4 billion light-years away from Earth and had masses of 30 and 35 solar masses.: 6 In 2017, Rainer Weiss, Kip Thorne, and Barry Barish, who had spearheaded the project, were awarded the Nobel Prize in Physics for their work. Since the initial discovery in 2015, hundreds more gravitational waves have been observed by LIGO and another interferometer, Virgo. On 10 April 2019, the first direct image of a black hole and its vicinity was published, following observations made by the Event Horizon Telescope (EHT) in 2017 of the supermassive black hole in Messier 87's galactic centre. In 2022, the Event Horizon Telescope collaboration released an image of the black hole in the center of the Milky Way galaxy, Sagittarius A*; The data had been collected in 2017. In 2020, the Nobel Prize in Physics was awarded for work on black holes. Andrea Ghez and Reinhard Genzel shared one-half for their discovery that Sagittarius A* is a supermassive black hole. Penrose received the other half for his work showing that the mathematics of general relativity requires the formation of black holes. Cosmologists lamented that Hawking's extensive theoretical work on black holes would not be honored since he died in 2018. In December 1967, a student reportedly suggested the phrase black hole at a lecture by John Wheeler; Wheeler adopted the term for its brevity and "advertising value", and Wheeler's stature in the field ensured it quickly caught on, leading some to credit Wheeler with coining the phrase. However, the term was used by others around that time. Science writer Marcia Bartusiak traces the term black hole to physicist Robert H. Dicke, who in the early 1960s reportedly compared the phenomenon to the Black Hole of Calcutta, notorious as a prison where people entered but never left alive. The term was used in print by Life and Science News magazines in 1963, and by science journalist Ann Ewing in her article "'Black Holes' in Space", dated 18 January 1964, which was a report on a meeting of the American Association for the Advancement of Science held in Cleveland, Ohio. Definition A black hole is generally defined as a region of spacetime from which no information-carrying signals or objects can escape. However, verifying an object as a black hole by this definition would require waiting for an infinite time and at an infinite distance from the black hole to verify that indeed, nothing has escaped, and thus cannot be used to identify a physical black hole. Broadly, physicists do not have a precisely-agreed-upon definition of a black hole. Among astrophysicists, a black hole is a compact object with a mass larger than four solar masses. A black hole may also be defined as a reservoir of information: 142 or a region where space is falling inwards faster than the speed of light. Properties The no-hair theorem postulates that, once it achieves a stable condition after formation, a black hole has only three independent physical properties: mass, electric charge, and angular momentum; the black hole is otherwise featureless. If the conjecture is true, any two black holes that share the same values for these properties, or parameters, are indistinguishable from one another. The degree to which the conjecture is true for real black holes is currently an unsolved problem. The simplest static black holes have mass but neither electric charge nor angular momentum. According to Birkhoff's theorem, these Schwarzschild black holes are the only vacuum solution that is spherically symmetric. Solutions describing more general black holes also exist. Non-rotating charged black holes are described by the Reissner–Nordström metric, while the Kerr metric describes a non-charged rotating black hole. The most general stationary black hole solution known is the Kerr–Newman metric, which describes a black hole with both charge and angular momentum. The simplest static black holes have mass but neither electric charge nor angular momentum. Contrary to the popular notion of a black hole "sucking in everything" in its surroundings, from far away, the external gravitational field of a black hole is identical to that of any other body of the same mass. While a black hole can theoretically have any positive mass, the charge and angular momentum are constrained by the mass. The total electric charge Q and the total angular momentum J are expected to satisfy the inequality Q 2 4 π ϵ 0 + c 2 J 2 G M 2 ≤ G M 2 {\displaystyle {\frac {Q^{2}}{4\pi \epsilon _{0}}}+{\frac {c^{2}J^{2}}{GM^{2}}}\leq GM^{2}} for a black hole of mass M. Black holes with the maximum possible charge or spin satisfying this inequality are called extremal black holes. Solutions of Einstein's equations that violate this inequality exist, but they do not possess an event horizon. These are so-called naked singularities that can be observed from the outside. Because these singularities make the universe inherently unpredictable, many physicists believe they could not exist. The weak cosmic censorship hypothesis, proposed by Sir Roger Penrose, rules out the formation of such singularities, when they are created through the gravitational collapse of realistic matter. However, this theory has not yet been proven, and some physicists believe that naked singularities could exist. It is also unknown whether black holes could even become extremal, forming naked singularities, since natural processes counteract increasing spin and charge when a black hole becomes near-extremal. The total mass of a black hole can be estimated by analyzing the motion of objects near the black hole, such as stars or gas. All black holes spin, often fast—One supermassive black hole, GRS 1915+105 has been estimated to spin at over 1,000 revolutions per second. The Milky Way's central black hole Sagittarius A* rotates at about 90% of the maximum rate. The spin rate can be inferred from measurements of atomic spectral lines in the X-ray range. As gas near the black hole plunges inward, high energy X-ray emission from electron-positron pairs illuminates the gas further out, appearing red-shifted due to relativistic effects. Depending on the spin of the black hole, this plunge happens at different radii from the hole, with different degrees of redshift. Astronomers can use the gap between the x-ray emission of the outer disk and the redshifted emission from plunging material to determine the spin of the black hole. A newer way to estimate spin is based on the temperature of gasses accreting onto the black hole. The method requires an independent measurement of the black hole mass and inclination angle of the accretion disk followed by computer modeling. Gravitational waves from coalescing binary black holes can also provide the spin of both progenitor black holes and the merged hole, but such events are rare. A spinning black hole has angular momentum. The supermassive black hole in the center of the Messier 87 (M87) galaxy appears to have an angular momentum very close to the maximum theoretical value. That uncharged limit is J ≤ G M 2 c , {\displaystyle J\leq {\frac {GM^{2}}{c}},} allowing definition of a dimensionless spin magnitude such that 0 ≤ c J G M 2 ≤ 1. {\displaystyle 0\leq {\frac {cJ}{GM^{2}}}\leq 1.} Most black holes are believed to have an approximately neutral charge. For example, Michal Zajaček, Arman Tursunov, Andreas Eckart, and Silke Britzen found the electric charge of Sagittarius A* to be at least ten orders of magnitude below the theoretical maximum. A charged black hole repels other like charges just like any other charged object. If a black hole were to become charged, particles with an opposite sign of charge would be pulled in by the extra electromagnetic force, while particles with the same sign of charge would be repelled, neutralizing the black hole. This effect may not be as strong if the black hole is also spinning. The presence of charge can reduce the diameter of the black hole by up to 38%. The charge Q for a nonspinning black hole is bounded by Q ≤ G M , {\displaystyle Q\leq {\sqrt {G}}M,} where G is the gravitational constant and M is the black hole's mass. Classification Black holes can have a wide range of masses. The minimum mass of a black hole formed by stellar gravitational collapse is governed by the maximum mass of a neutron star and is believed to be approximately two-to-four solar masses. However, theoretical primordial black holes, believed to have formed soon after the Big Bang, could be far smaller, with masses as little as 10−5 grams at formation. These very small black holes are sometimes called micro black holes. Black holes formed by stellar collapse are called stellar black holes. Estimates of their maximum mass at formation vary, but generally range from 10 to 100 solar masses, with higher estimates for black holes progenated by low-metallicity stars. The mass of a black hole formed via a supernova has a lower bound: If the progenitor star is too small, the collapse may be stopped by the degeneracy pressure of the star's constituents, allowing the condensation of matter into an exotic denser state. Degeneracy pressure occurs from the Pauli exclusion principle—Particles will resist being in the same place as each other. Smaller progenitor stars, with masses less than about 8 M☉, will be held together by the degeneracy pressure of electrons and will become a white dwarf. For more massive progenitor stars, electron degeneracy pressure is no longer strong enough to resist the force of gravity and the star will be held together by neutron degeneracy pressure, which can occur at much higher densities, forming a neutron star. If the star is still too massive, even neutron degeneracy pressure will not be able to resist the force of gravity and the star will collapse into a black hole.: 5.8 Stellar black holes can also gain mass via accretion of nearby matter, often from a companion object such as a star. Black holes that are larger than stellar black holes but smaller than supermassive black holes are called intermediate-mass black holes, with masses of approximately 102 to 105 solar masses. These black holes seem to be rarer than their stellar and supermassive counterparts, with relatively few candidates having been observed. Physicists have speculated that such black holes may form from collisions in globular and star clusters or at the center of low-mass galaxies. They may also form as the result of mergers of smaller black holes, with several LIGO observations finding merged black holes within the 110-350 solar mass range. The black holes with the largest masses are called supermassive black holes, with masses more than 106 times that of the Sun. These black holes are believed to exist at the centers of almost every large galaxy, including the Milky Way. Some scientists have proposed a subcategory of even larger black holes, called ultramassive black holes, with masses greater than 109-1010 solar masses. Theoretical models predict that the accretion disc that feeds black holes will be unstable once a black hole reaches 50-100 billion times the mass of the Sun, setting a rough upper limit to black hole mass. Structure While black holes are conceptually invisible sinks of all matter and light, in astronomical settings, their enormous gravity alters the motion of surrounding objects and pulls nearby gas inwards at near-light speed, making the area around black holes the brightest objects in the universe. Some black holes have relativistic jets—thin streams of plasma travelling away from the black hole at more than one-tenth of the speed of light. A small faction of the matter falling towards the black hole gets accelerated away along the hole rotation axis. These jets can extend as far as millions of parsecs from the black hole itself. Black holes of any mass can have jets. However, they are typically observed around spinning black holes with strongly-magnetized accretion disks. Relativistic jets were more common in the early universe, when galaxies and their corresponding supermassive black holes were rapidly gaining mass. All black holes with jets also have an accretion disk, but the jets are usually brighter than the disk. Quasars, typically found in other galaxies, are believed to be supermassive black holes with jets; microquasars are believed to be stellar-mass objects with jets, typically observed in the Milky Way. The mechanism of formation of jets is not yet known, but several options have been proposed. One method proposed to fuel these jets is the Blandford-Znajek process, which suggests that the dragging of magnetic field lines by a black hole's rotation could launch jets of matter into space. The Penrose process, which involves extraction of a black hole's rotational energy, has also been proposed as a potential mechanism of jet propulsion. Due to conservation of angular momentum, gas falling into the gravitational well created by a massive object will typically form a disk-like structure around the object.: 242 As the disk's angular momentum is transferred outward due to internal processes, its matter falls farther inward, converting its gravitational energy into heat and releasing a large flux of x-rays. The temperature of these disks can range from thousands to millions of Kelvin, and temperatures can differ throughout a single accretion disk. Accretion disks can also emit in other parts of the electromagnetic spectrum, depending on the disk's turbulence and magnetization and the black hole's mass and angular momentum. Accretion disks can be defined as geometrically thin or geometrically thick. Geometrically thin disks are mostly confined to the black hole's equatorial plane and have a well-defined edge at the innermost stable circular orbit (ISCO), while geometrically thick disks are supported by internal pressure and temperature and can extend inside the ISCO. Disks with high rates of electron scattering and absorption, appearing bright and opaque, are called optically thick; optically thin disks are more translucent and produce fainter images when viewed from afar. Accretion disks of black holes accreting beyond the Eddington limit are often referred to as polish donuts due to their thick, toroidal shape that resembles that of a donut. Quasar accretion disks are expected to usually appear blue in color. The disk for a stellar black hole, on the other hand, would likely look orange, yellow, or red, with its inner regions being the brightest. Theoretical research suggests that the hotter a disk is, the bluer it should be, although this is not always supported by observations of real astronomical objects. Accretion disk colors may also be altered by the Doppler effect, with the part of the disk travelling towards an observer appearing bluer and brighter and the part of the disk travelling away from the observer appearing redder and dimmer. In Newtonian gravity, test particles can stably orbit at arbitrary distances from a central object. In general relativity, however, there exists a smallest possible radius for which a massive particle can orbit stably. Any infinitesimal inward perturbations to this orbit will lead to the particle spiraling into the black hole, and any outward perturbations will, depending on the energy, cause the particle to spiral in, move to a stable orbit further from the black hole, or escape to infinity. This orbit is called the innermost stable circular orbit, or ISCO. The location of the ISCO depends on the spin of the black hole and the spin of the particle itself. In the case of a Schwarzschild black hole (spin zero) and a particle without spin, the location of the ISCO is: r I S C O = 3 r s = 6 G M c 2 , {\displaystyle r_{\rm {ISCO}}=3\,r_{\text{s}}={\frac {6\,GM}{c^{2}}},} where r I S C O {\displaystyle r_{\rm {_{ISCO}}}} is the radius of the ISCO, r s {\displaystyle r_{\text{s}}} is the Schwarzschild radius of the black hole, G {\displaystyle G} is the gravitational constant, and c {\displaystyle c} is the speed of light. The radius of this orbit changes slightly based on particle spin. For charged black holes, the ISCO moves inwards. For spinning black holes, the ISCO is moved inwards for particles orbiting in the same direction that the black hole is spinning (prograde) and outwards for particles orbiting in the opposite direction (retrograde). For example, the ISCO for a particle orbiting retrograde can be as far out as about 9 r s {\displaystyle 9r_{\text{s}}} , while the ISCO for a particle orbiting prograde can be as close as at the event horizon itself. The photon sphere is a spherical boundary for which photons moving on tangents to that sphere are bent completely around the black hole, possibly orbiting multiple times. Light rays with impact parameters less than the radius of the photon sphere enter the black hole. For Schwarzschild black holes, the photon sphere has a radius 1.5 times the Schwarzschild radius; the radius for non-Schwarzschild black holes is at least 1.5 times the radius of the event horizon. When viewed from a great distance, the photon sphere creates an observable black hole shadow. Since no light emerges from within the black hole, this shadow is the limit for possible observations.: 152 The shadow of colliding black holes should have characteristic warped shapes, allowing scientists to detect black holes that are about to merge. While light can still escape from the photon sphere, any light that crosses the photon sphere on an inbound trajectory will be captured by the black hole. Therefore, any light that reaches an outside observer from the photon sphere must have been emitted by objects between the photon sphere and the event horizon. Light emitted towards the photon sphere may also curve around the black hole and return to the emitter. For a rotating, uncharged black hole, the radius of the photon sphere depends on the spin parameter and whether the photon is orbiting prograde or retrograde. For a photon orbiting prograde, the photon sphere will be 1-3 Schwarzschild radii from the center of the black hole, while for a photon orbiting retrograde, the photon sphere will be between 3-5 Schwarzschild radii from the center of the black hole. The exact location of the photon sphere depends on the magnitude of the black hole's rotation. For a charged, nonrotating black hole, there will only be one photon sphere, and the radius of the photon sphere will decrease for increasing black hole charge. For non-extremal, charged, rotating black holes, there will always be two photon spheres, with the exact radii depending on the parameters of the black hole. Near a rotating black hole, spacetime rotates similar to a vortex. The rotating spacetime will drag any matter and light into rotation around the spinning black hole. This effect of general relativity, called frame dragging, gets stronger closer to the spinning mass. The region of spacetime in which it is impossible to stay still is called the ergosphere. The ergosphere of a black hole is a volume bounded by the black hole's event horizon and the ergosurface, which coincides with the event horizon at the poles but bulges out from it around the equator. Matter and radiation can escape from the ergosphere. Through the Penrose process, objects can emerge from the ergosphere with more energy than they entered with. The extra energy is taken from the rotational energy of the black hole, slowing down the rotation of the black hole.: 268 A variation of the Penrose process in the presence of strong magnetic fields, the Blandford–Znajek process, is considered a likely mechanism for the enormous luminosity and relativistic jets of quasars and other active galactic nuclei. The observable region of spacetime around a black hole closest to its event horizon is called the plunging region. In this area it is no longer possible for free falling matter to follow circular orbits or stop a final descent into the black hole. Instead, it will rapidly plunge toward the black hole at close to the speed of light, growing increasingly hot and producing a characteristic, detectable thermal emission. However, light and radiation emitted from this region can still escape from the black hole's gravitational pull. For a nonspinning, uncharged black hole, the radius of the event horizon, or Schwarzschild radius, is proportional to the mass, M, through r s = 2 G M c 2 ≈ 2.95 M M ⊙ k m , {\displaystyle r_{\mathrm {s} }={\frac {2GM}{c^{2}}}\approx 2.95\,{\frac {M}{M_{\odot }}}~\mathrm {km,} } where rs is the Schwarzschild radius and M☉ is the mass of the Sun.: 124 For a black hole with nonzero spin or electric charge, the radius is smaller,[Note 1] until an extremal black hole could have an event horizon close to r + = G M c 2 , {\displaystyle r_{\mathrm {+} }={\frac {GM}{c^{2}}},} half the radius of a nonspinning, uncharged black hole of the same mass. Since the volume within the Schwarzschild radius increase with the cube of the radius, average density of a black hole inside its Schwarzschild radius is inversely proportional to the square of its mass: supermassive black holes are much less dense than stellar black holes. The average density of a 108 M☉ black hole is comparable to that of water. The defining feature of a black hole is the existence of an event horizon, a boundary in spacetime through which matter and light can pass only inward towards the center of the black hole. Nothing, not even light, can escape from inside the event horizon. The event horizon is referred to as such because if an event occurs within the boundary, information from that event cannot reach or affect an outside observer, making it impossible to determine whether such an event occurred.: 179 For non-rotating black holes, the geometry of the event horizon is precisely spherical, while for rotating black holes, the event horizon is oblate. To a distant observer, a clock near a black hole would appear to tick more slowly than one further from the black hole.: 217 This effect, known as gravitational time dilation, would also cause an object falling into a black hole to appear to slow as it approached the event horizon, never quite reaching the horizon from the perspective of an outside observer.: 218 All processes on this object would appear to slow down, and any light emitted by the object to appear redder and dimmer, an effect known as gravitational redshift. An object falling from half of a Schwarzschild radius above the event horizon would fade away until it could no longer be seen, disappearing from view within one hundredth of a second. It would also appear to flatten onto the black hole, joining all other material that had ever fallen into the hole. On the other hand, an observer falling into a black hole would not notice any of these effects as they cross the event horizon. Their own clocks appear to them to tick normally, and they cross the event horizon after a finite time without noting any singular behaviour. In general relativity, it is impossible to determine the location of the event horizon from local observations, due to Einstein's equivalence principle.: 222 Black holes that are rotating and/or charged have an inner horizon, often called the Cauchy horizon, inside of the black hole. The inner horizon is divided up into two segments: an ingoing section and an outgoing section. At the ingoing section of the Cauchy horizon, radiation and matter that fall into the black hole would build up at the horizon, causing the curvature of spacetime to go to infinity. This would cause an observer falling in to experience tidal forces. This phenomenon is often called mass inflation, since it is associated with a parameter dictating the black hole's internal mass growing exponentially, and the buildup of tidal forces is called the mass-inflation singularity or Cauchy horizon singularity. Some physicists have argued that in realistic black holes, accretion and Hawking radiation would stop mass inflation from occurring. At the outgoing section of the inner horizon, infalling radiation would backscatter off of the black hole's spacetime curvature and travel outward, building up at the outgoing Cauchy horizon. This would cause an infalling observer to experience a gravitational shock wave and tidal forces as the spacetime curvature at the horizon grew to infinity. This buildup of tidal forces is called the shock singularity. Both of these singularities are weak, meaning that an object crossing them would only be deformed a finite amount by tidal forces, even though the spacetime curvature would still be infinite at the singularity. This is as opposed to a strong singularity, where an object hitting the singularity would be stretched and squeezed by an infinite amount. They are also null singularities, meaning that a photon could travel parallel to the them without ever being intercepted. Ignoring quantum effects, every black hole has a singularity inside, points where the curvature of spacetime becomes infinite, and geodesics terminate within a finite proper time.: 205 For a non-rotating black hole, this region takes the shape of a single point; for a rotating black hole it is smeared out to form a ring singularity that lies in the plane of rotation.: 264 In both cases, the singular region has zero volume. All of the mass of the black hole ends up in the singularity.: 252 Since the singularity has nonzero mass in an infinitely small space, it can be thought of as having infinite density. Observers falling into a Schwarzschild black hole (i.e., non-rotating and not charged) cannot avoid being carried into the singularity once they cross the event horizon. As they fall further into the black hole, they will be torn apart by the growing tidal forces in a process sometimes referred to as spaghettification or the noodle effect. Eventually, they will reach the singularity and be crushed into an infinitely small point.: 182 However any perturbations, such as those caused by matter or radiation falling in, would cause space to oscillate chaotically near the singularity. Any matter falling in would experience intense tidal forces rapidly changing in direction, all while being compressed into an increasingly small volume. Alternative forms of general relativity, including addition of some quatum effects, can lead to regular, or nonsingular, black holes without singularities. For example, the fuzzball model, based on string theory, states that black holes are actually made up of quantum microstates and need not have a singularity or an event horizon. The theory of loop quantum gravity proposes that the curvature and density at the center of a black hole is large, but not infinite. Formation Black holes are formed by gravitational collapse of massive stars, either by direct collapse or during a supernova explosion in a process called fallback. Black holes can result from the merger of two neutron stars or a neutron star and a black hole. Other more speculative mechanisms include primordial black holes created from density fluctuations in the early universe, the collapse of dark stars, a hypothetical object powered by annihilation of dark matter, or from hypothetical self-interacting dark matter. Gravitational collapse occurs when an object's internal pressure is insufficient to resist the object's own gravity. At the end of a star's life, it will run out of hydrogen to fuse, and will start fusing more and more massive elements, until it gets to iron. Since the fusion of elements heavier than iron would require more energy than it would release, nuclear fusion ceases. If the iron core of the star is too massive, the star will no longer be able to support itself and will undergo gravitational collapse. While most of the energy released during gravitational collapse is emitted very quickly, an outside observer does not actually see the end of this process. Even though the collapse takes a finite amount of time from the reference frame of infalling matter, a distant observer would see the infalling material slow and halt just above the event horizon, due to gravitational time dilation. Light from the collapsing material takes longer and longer to reach the observer, with the delay growing to infinity as the emitting material reaches the event horizon. Thus the external observer never sees the formation of the event horizon; instead, the collapsing material seems to become dimmer and increasingly red-shifted, eventually fading away. Observations of quasars at redshift z ∼ 7 {\displaystyle z\sim 7} , less than a billion years after the Big Bang, has led to investigations of other ways to form black holes. The accretion process to build supermassive black holes has a limiting rate of mass accumulation and a billion years is not enough time to reach quasar status. One suggestion is direct collapse of nearly pure hydrogen gas (low metalicity) clouds characteristic of the young universe, forming a supermassive star which collapses into a black hole. It has been suggested that seed black holes with typical masses of ~105 M☉ could have formed in this way which then could grow to ~109 M☉. However, the very large amount of gas required for direct collapse is not typically stable to fragmentation to form multiple stars. Thus another approach suggests massive star formation followed by collisions that seed massive black holes which ultimately merge to create a quasar.: 85 A neutron star in a common envelope with a regular star can accrete sufficient material to collapse to a black hole or two neutron stars can merge. These avenues for the formation of black holes are considered relatively rare. In the current epoch of the universe, conditions needed to form black holes are rare and are mostly only found in stars. However, in the early universe, conditions may have allowed for black hole formations via other means. Fluctuations of spacetime soon after the Big Bang may have formed areas that were denser then their surroundings. Initially, these regions would not have been compact enough to form a black hole, but eventually, the curvature of spacetime in the regions become large enough to cause them to collapse into a black hole. Different models for the early universe vary widely in their predictions of the scale of these fluctuations. Various models predict the creation of primordial black holes ranging from a Planck mass (~2.2×10−8 kg) to hundreds of thousands of solar masses. Primordial black holes with masses less than 1015 g would have evaporated by now due to Hawking radiation. Despite the early universe being extremely dense, it did not re-collapse into a black hole during the Big Bang, since the universe was expanding rapidly and did not have the gravitational differential necessary for black hole formation. Models for the gravitational collapse of objects of relatively constant size, such as stars, do not necessarily apply in the same way to rapidly expanding space such as the Big Bang. In principle, black holes could be formed in high-energy particle collisions that achieve sufficient density, although no such events have been detected. These hypothetical micro black holes, which could form from the collision of cosmic rays and Earth's atmosphere or in particle accelerators like the Large Hadron Collider, would not be able to aggregate additional mass. Instead, they would evaporate in about 10−25 seconds, posing no threat to the Earth. Evolution Black holes can also merge with other objects such as stars or even other black holes. This is thought to have been important, especially in the early growth of supermassive black holes, which could have formed from the aggregation of many smaller objects. The process has also been proposed as the origin of some intermediate-mass black holes. Mergers of supermassive black holes may take a long time: As a binary of supermassive black holes approach each other, most nearby stars are ejected, leaving little for the remaining black holes to gravitationally interact with that would allow them to get closer to each other. This phenomenon has been called the final parsec problem, as the distance at which this happens is usually around one parsec. When a black hole accretes matter, the gas in the inner accretion disk orbits at very high speeds because of its proximity to the black hole. The resulting friction heats the inner disk to temperatures at which it emits vast amounts of electromagnetic radiation (mainly X-rays) detectable by telescopes. By the time the matter of the disk reaches the ISCO, between 5.7% and 42% of its mass will have been converted to energy, depending on the black hole's spin. About 90% of this energy is released within about 20 black hole radii. In many cases, accretion disks are accompanied by relativistic jets that are emitted along the black hole's poles, which carry away much of the energy. The mechanism for the creation of these jets is currently not well understood, in part due to insufficient data. Many of the universe's most energetic phenomena have been attributed to the accretion of matter on black holes. Active galactic nuclei and quasars are believed to be the accretion disks of supermassive black holes. X-ray binaries are generally accepted to be binary systems in which one of the two objects is a compact object accreting matter from its companion. Ultraluminous X-ray sources may be the accretion disks of intermediate-mass black holes. At a certain rate of accretion, the outward radiation pressure will become as strong as the inward gravitational force, and the black hole should unable to accrete any faster. This limit is called the Eddington limit. However, many black holes accrete beyond this rate due to their non-spherical geometry or instabilities in the accretion disk. Accretion beyond the limit is called Super-Eddington accretion and may have been commonplace in the early universe. Stars have been observed to get torn apart by tidal forces in the immediate vicinity of supermassive black holes in galaxy nuclei, in what is known as a tidal disruption event (TDE). Some of the material from the disrupted star forms an accretion disk around the black hole, which emits observable electromagnetic radiation. The correlation between the masses of supermassive black holes in the centres of galaxies with the velocity dispersion and mass of stars in their host bulges suggests that the formation of galaxies and the formation of their central black holes are related. Black hole winds from rapid accretion, particularly when the galaxy itself is still accreting matter, can compress gas nearby, accelerating star formation. However, if the winds become too strong, the black hole may blow nearly all of the gas out of the galaxy, quenching star formation. Black hole jets may also energize nearby cavities of plasma and eject low-entropy gas from out of the galactic core, causing gas in galactic centers to be hotter than expected. If Hawking's theory of black hole radiation is correct, then black holes are expected to shrink and evaporate over time as they lose mass by the emission of photons and other particles. The temperature of this thermal spectrum (Hawking temperature) is proportional to the surface gravity of the black hole, which is inversely proportional to the mass. Hence, large black holes emit less radiation than small black holes.: Ch. 9.6 A stellar black hole of 1 M☉ has a Hawking temperature of 62 nanokelvins. This is far less than the 2.7 K temperature of the cosmic microwave background radiation. Stellar-mass or larger black holes receive more mass from the cosmic microwave background than they emit through Hawking radiation and thus will grow instead of shrinking. To have a Hawking temperature larger than 2.7 K (and be able to evaporate), a black hole would need a mass less than the Moon. Such a black hole would have a diameter of less than a tenth of a millimetre. The Hawking radiation for an astrophysical black hole is predicted to be very weak and would thus be exceedingly difficult to detect from Earth. A possible exception is the burst of gamma rays emitted in the last stage of the evaporation of primordial black holes. Searches for such flashes have proven unsuccessful and provide stringent limits on the possibility of existence of low mass primordial black holes, with modern research predicting that primordial black holes must make up less than a fraction of 10−7 of the universe's total mass. NASA's Fermi Gamma-ray Space Telescope, launched in 2008, has searched for these flashes, but has not yet found any. The properties of a black hole are constrained and interrelated by the theories that predict these properties. When based on general relativity, these relationships are called the laws of black hole mechanics. For a black hole that is not still forming or accreting matter, the zeroth law of black hole mechanics states the black hole's surface gravity is constant across the event horizon. The first law relates changes in the black hole's surface area, angular momentum, and charge to changes in its energy. The second law says the surface area of a black hole never decreases on its own. Finally, the third law says that the surface gravity of a black hole is never zero. These laws are mathematical analogs of the laws of thermodynamics. They are not equivalent, however, because, according to general relativity without quantum mechanics, a black hole can never emit radiation, and thus its temperature must always be zero.: 11 Quantum mechanics predicts that a black hole will continuously emit thermal Hawking radiation, and therefore must always have a nonzero temperature. It also predicts that all black holes have entropy which scales with their surface area. When quantum mechanics is accounted for, the laws of black hole mechanics become equivalent to the classical laws of thermodynamics. However, these conclusions are derived without a complete theory of quantum gravity, although many potential theories do predict black holes having entropy and temperature. Thus, the true quantum nature of black hole thermodynamics continues to be debated.: 29 Observational evidence Millions of black holes with around 30 solar masses derived from stellar collapse are expected to exist in the Milky Way. Even a dwarf galaxy like Draco should have hundreds. Only a few of these have been detected. By nature, black holes do not themselves emit any electromagnetic radiation other than the hypothetical Hawking radiation, so astrophysicists searching for black holes must generally rely on indirect observations. The defining characteristic of a black hole is its event horizon. The horizon itself cannot be imaged, so all other possible explanations for these indirect observations must be considered and eliminated before concluding that a black hole has been observed.: 11 The Event Horizon Telescope (EHT) is a global system of radio telescopes capable of directly observing a black hole shadow. The angular resolution of a telescope is based on its aperture and the wavelengths it is observing. Because the angular diameters of Sagittarius A* and Messier 87* in the sky are very small, a single telescope would need to be about the size of the Earth to clearly distinguish their horizons using radio wavelengths. By combining data from several different radio telescopes around the world, the Event Horizon Telescope creates an effective aperture the diameter size of the Earth. The EHT team used imaging algorithms to compute the most probable image from the data in its observations of Sagittarius A* and M87*. Gravitational-wave interferometry can be used to detect merging black holes and other compact objects. In this method, a laser beam is split down two long arms of a tunnel. The laser beams reflect off of mirrors in the tunnels and converge at the intersection of the arms, cancelling each other out. However, when a gravitational wave passes, it warps spacetime, changing the lengths of the arms themselves. Since each laser beam is now travelling a slightly different distance, they do not cancel out and produce a recognizable signal. Analysis of the signal can give scientists information about what caused the gravitational waves. Since gravitational waves are very weak, gravitational-wave observatories such as LIGO must have arms several kilometers long and carefully control for noise from Earth to be able to detect these gravitational waves. Since the first measurements in 2016, multiple gravitational waves from black holes have been detected and analyzed. The proper motions of stars near the centre of the Milky Way provide strong observational evidence that these stars are orbiting a supermassive black hole. Since 1995, astronomers have tracked the motions of 90 stars orbiting an invisible object coincident with the radio source Sagittarius A*. In 1998, by fitting the motions of the stars to Keplerian orbits, the astronomers were able to infer that Sagittarius A* must be a 2.6×106 M☉ object must be contained within a radius of 0.02 light-years. Since then, one of the stars—called S2—has completed a full orbit. From the orbital data, astronomers were able to refine the calculations of the mass of Sagittarius A* to 4.3×106 M☉, with a radius of less than 0.002 light-years. This upper limit radius is larger than the Schwarzschild radius for the estimated mass, so the combination does not prove Sagittarius A* is a black hole. Nevertheless, these observations strongly suggest that the central object is a supermassive black hole as there are no other plausible scenarios for confining so much invisible mass into such a small volume. Additionally, there is some observational evidence that this object might possess an event horizon, a feature unique to black holes. The Event Horizon Telescope image of Sagittarius A*, released in 2022, provided further confirmation that it is indeed a black hole. X-ray binaries are binary systems that emit a majority of their radiation in the X-ray part of the electromagnetic spectrum. These X-ray emissions result when a compact object accretes matter from an ordinary star. The presence of an ordinary star in such a system provides an opportunity for studying the central object and to determine if it might be a black hole. By measuring the orbital period of the binary, the distance to the binary from Earth, and the mass of the companion star, scientists can estimate the mass of the compact object. The Tolman-Oppenheimer-Volkoff limit (TOV limit) dictates the largest mass a nonrotating neutron star can be, and is estimated to be about two solar masses. While a rotating neutron star can be slightly more massive, if the compact object is much more massive than the TOV limit, it cannot be a neutron star and is generally expected to be a black hole. The first strong candidate for a black hole, Cygnus X-1, was discovered in this way by Charles Thomas Bolton, Louise Webster, and Paul Murdin in 1972. Observations of rotation broadening of the optical star reported in 1986 lead to a compact object mass estimate of 16 solar masses, with 7 solar masses as the lower bound. In 2011, this estimate was updated to 14.1±1.0 M☉ for the black hole and 19.2±1.9 M☉ for the optical stellar companion. X-ray binaries can be categorized as either low-mass or high-mass; This classification is based on the mass of the companion star, not the compact object itself. In a class of X-ray binaries called soft X-ray transients, the companion star is of relatively low mass, allowing for more accurate estimates of the black hole mass. These systems actively emit X-rays for only several months once every 10–50 years. During the period of low X-ray emission, called quiescence, the accretion disk is extremely faint, allowing detailed observation of the companion star. Numerous black hole candidates have been measured by this method. Black holes are also sometimes found in binaries with other compact objects, such as white dwarfs, neutron stars, and other black holes. The centre of nearly every galaxy contains a supermassive black hole. The close observational correlation between the mass of this hole and the velocity dispersion of the host galaxy's bulge, known as the M–sigma relation, strongly suggests a connection between the formation of the black hole and that of the galaxy itself. Astronomers use the term active galaxy to describe galaxies with unusual characteristics, such as unusual spectral line emission and very strong radio emission. Theoretical and observational studies have shown that the high levels of activity in the centers of these galaxies, regions called active galactic nuclei (AGN), may be explained by accretion onto supermassive black holes. These AGN consist of a central black hole that may be millions or billions of times more massive than the Sun, a disk of interstellar gas and dust called an accretion disk, and two jets perpendicular to the accretion disk. Although supermassive black holes are expected to be found in most AGN, only some galaxies' nuclei have been more carefully studied in attempts to both identify and measure the actual masses of the central supermassive black hole candidates. Some of the most notable galaxies with supermassive black hole candidates include the Andromeda Galaxy, Messier 32, Messier 87, the Sombrero Galaxy, and the Milky Way itself. Another way black holes can be detected is through observation of effects caused by their strong gravitational field. One such effect is gravitational lensing: The deformation of spacetime around a massive object causes light rays to be deflected, making objects behind them appear distorted. When the lensing object is a black hole, this effect can be strong enough to create multiple images of a star or other luminous source. However, the distance between the lensed images may be too small for contemporary telescopes to resolve—this phenomenon is called microlensing. Instead of seeing two images of a lensed star, astronomers see the star brighten slightly as the black hole moves towards the line of sight between the star and Earth and then return to its normal luminosity as the black hole moves away. The turn of the millennium saw the first 3 candidate detections of black holes in this way, and in January 2022, astronomers reported the first confirmed detection of a microlensing event from an isolated black hole. This was also the first determination of an isolated black hole mass, 7.1±1.3 M☉. Alternatives While there is a strong case for supermassive black holes, the model for stellar-mass black holes assumes of an upper limit for the mass of a neutron star: objects observed to have more mass are assumed to be black holes. However, the properties of extremely dense matter are poorly understood. New exotic phases of matter could allow other kinds of massive objects. Quark stars would be made up of quark matter and supported by quark degeneracy pressure, a form of degeneracy pressure even stronger than neutron degeneracy pressure. This would halt gravitational collapse at a higher mass than for a neutron star. Even stronger stars called electroweak stars would convert quarks in their cores into leptons, providing additional pressure to stop the star from collapsing. If, as some extensions of the Standard Model posit, quarks and leptons are made up of the even-smaller fundamental particles called preons, a very compact star could be supported by preon degeneracy pressure. While none of these hypothetical models can explain all of the observations of stellar black hole candidates, a Q star is the only alternative which could significantly exceed the mass limit for neutron stars and thus provide an alternative for supermassive black holes.: 12 A few theoretical objects have been conjectured to match observations of astronomical black hole candidates identically or near-identically, but which function via a different mechanism. A dark energy star would convert infalling matter into vacuum energy; This vacuum energy would be much larger than the vacuum energy of outside space, exerting outwards pressure and preventing a singularity from forming. A black star would be gravitationally collapsing slowly enough that quantum effects would keep it just on the cusp of fully collapsing into a black hole. A gravastar would consist of a very thin shell and a dark-energy interior providing outward pressure to stop the collapse into a black hole or formation of a singularity; It could even have another gravastar inside, called a 'nestar'. Open questions According to the no-hair theorem, a black hole is defined by only three parameters: its mass, charge, and angular momentum. This seems to mean that all other information about the matter that went into forming the black hole is lost, as there is no way to determine anything about the black hole from outside other than those three parameters. When black holes were thought to persist forever, this information loss was not problematic, as the information can be thought of as existing inside the black hole. However, black holes slowly evaporate by emitting Hawking radiation. This radiation does not appear to carry any additional information about the matter that formed the black hole, meaning that this information is seemingly gone forever. This is called the black hole information paradox. Theoretical studies analyzing the paradox have led to both further paradoxes and new ideas about the intersection of quantum mechanics and general relativity. While there is no consensus on the resolution of the paradox, work on the problem is expected to be important for a theory of quantum gravity.: 126 Observations of faraway galaxies have found that ultraluminous quasars, powered by supermassive black holes, existed in the early universe as far as redshift z ≥ 7 {\displaystyle z\geq 7} . These black holes have been assumed to be the products of the gravitational collapse of large population III stars. However, these stellar remnants were not massive enough to produce the quasars observed at early times without accreting beyond the Eddington limit, the theoretical maximum rate of black hole accretion. Physicists have suggested a variety of different mechanisms by which these supermassive black holes may have formed. It has been proposed that smaller black holes may have also undergone mergers to produce the observed supermassive black holes. It is also possible that they were seeded by direct-collapse black holes, in which a large cloud of hot gas avoids fragmentation that would lead to multiple stars, due to low angular momentum or heating from a nearby galaxy. Given the right circumstances, a single supermassive star forms and collapses directly into a black hole without undergoing typical stellar evolution. Additionally, these supermassive black holes in the early universe may be high-mass primordial black holes, which could have accreted further matter in the centers of galaxies. Finally, certain mechanisms allow black holes to grow faster than the theoretical Eddington limit, such as dense gas in the accretion disk limiting outward radiation pressure that prevents the black hole from accreting. However, the formation of bipolar jets prevent super-Eddington rates. In fiction Black holes have been portrayed in science fiction in a variety of ways. Even before the advent of the term itself, objects with characteristics of black holes appeared in stories such as the 1928 novel The Skylark of Space with its "black Sun" and the "hole in space" in the 1935 short story Starship Invincible. As black holes grew to public recognition in the 1960s and 1970s, they began to be featured in films as well as novels, such as Disney's The Black Hole. Black holes have also been used in works of the 21st century, such as Christopher Nolan's science fiction epic Interstellar. Authors and screenwriters have exploited the relativistic effects of black holes, particularly gravitational time dilation. For example, Interstellar features a black hole planet with a time dilation factor of over 60,000:1, while the 1977 novel Gateway depicts a spaceship approaching but never crossing the event horizon of a black hole from the perspective of an outside observer due to time dilation effects. Black holes have also been appropriated as wormholes or other methods of faster-than-light travel, such as in the 1974 novel The Forever War, where a network of black holes is used for interstellar travel. Additionally, black holes can feature as hazards to spacefarers and planets: A black hole threatens a deep-space outpost in 1978 short story The Black Hole Passes, and a binary black hole dangerously alters the orbit of a planet in the 2018 Netflix reboot of Lost in Space. Notes References Further reading External links
========================================
[SOURCE: https://en.wikipedia.org/wiki/Meta_Platforms#Company_governance] | [TOKENS: 8626]
Contents Meta Platforms Meta Platforms, Inc. (doing business as Meta) is an American multinational technology company headquartered in Menlo Park, California. Meta owns and operates several prominent social media platforms and communication services, including Facebook, Instagram, WhatsApp, Messenger, Threads and Manus. The company also operates an advertising network for its own sites and third parties; as of 2023[update], advertising accounted for 97.8 percent of its total revenue. Meta has been described as a part of Big Tech, which refers to the largest six tech companies in the United States, Alphabet (Google), Amazon, Apple, Meta (Facebook), Microsoft, and Nvidia, which are also the largest companies in the world by market capitalization. The company was originally established in 2004 as TheFacebook, Inc., and was renamed Facebook, Inc. in 2005. In 2021, it rebranded as Meta Platforms, Inc. to reflect a strategic shift toward developing the metaverse—an interconnected digital ecosystem spanning virtual and augmented reality technologies. In 2023, Meta was ranked 31st on the Forbes Global 2000 list of the world's largest public companies. As of 2022, it was the world's third-largest spender on research and development, with R&D expenses totaling US$35.3 billion. History Facebook filed for an initial public offering (IPO) on January 1, 2012. The preliminary prospectus stated that the company sought to raise $5 billion, had 845 million monthly active users, and a website accruing 2.7 billion likes and comments daily. After the IPO, Zuckerberg would retain 22% of the total shares and 57% of the total voting power in Facebook. Underwriters valued the shares at $38 each, valuing the company at $104 billion, the largest valuation yet for a newly public company. On May 16, one day before the IPO, Facebook announced it would sell 25% more shares than originally planned due to high demand. The IPO raised $16 billion, making it the third-largest in US history (slightly ahead of AT&T Mobility and behind only General Motors and Visa). The stock price left the company with a higher market capitalization than all but a few U.S. corporations—surpassing heavyweights such as Amazon, McDonald's, Disney, and Kraft Foods—and made Zuckerberg's stock worth $19 billion. The New York Times stated that the offering overcame questions about Facebook's difficulties in attracting advertisers to transform the company into a "must-own stock". Jimmy Lee of JPMorgan Chase described it as "the next great blue-chip". Writers at TechCrunch, on the other hand, expressed skepticism, stating, "That's a big multiple to live up to, and Facebook will likely need to add bold new revenue streams to justify the mammoth valuation." Trading in the stock, which began on May 18, was delayed that day due to technical problems with the Nasdaq exchange. The stock struggled to stay above the IPO price for most of the day, forcing underwriters to buy back shares to support the price. At the closing bell, shares were valued at $38.23, only $0.23 above the IPO price and down $3.82 from the opening bell value. The opening was widely described by the financial press as a disappointment. The stock set a new record for trading volume of an IPO. On May 25, 2012, the stock ended its first full week of trading at $31.91, a 16.5% decline. On May 22, 2012, regulators from Wall Street's Financial Industry Regulatory Authority announced that they had begun to investigate whether banks underwriting Facebook had improperly shared information only with select clients rather than the general public. Massachusetts Secretary of State William F. Galvin subpoenaed Morgan Stanley over the same issue. The allegations sparked "fury" among some investors and led to the immediate filing of several lawsuits, one of them a class action suit claiming more than $2.5 billion in losses due to the IPO. Bloomberg estimated that retail investors may have lost approximately $630 million on Facebook stock since its debut. S&P Global Ratings added Facebook to its S&P 500 index on December 21, 2013. On May 2, 2014, Zuckerberg announced that the company would be changing its internal motto from "Move fast and break things" to "Move fast with stable infrastructure". The earlier motto had been described as Zuckerberg's "prime directive to his developers and team" in a 2009 interview in Business Insider, in which he also said, "Unless you are breaking stuff, you are not moving fast enough." In November 2016, Facebook announced the Microsoft Windows client of gaming service Facebook Gameroom, formerly Facebook Games Arcade, at the Unity Technologies developers conference. The client allows Facebook users to play "native" games in addition to its web games. The service was closed in June 2021. Lasso was a short-video sharing app from Facebook similar to TikTok that was launched on iOS and Android in 2018 and was aimed at teenagers. On July 2, 2020, Facebook announced that Lasso would be shutting down on July 10. In 2018, the Oculus lead Jason Rubin sent his 50-page vision document titled "The Metaverse" to Facebook's leadership. In the document, Rubin acknowledged that Facebook's virtual reality business had not caught on as expected, despite the hundreds of millions of dollars spent on content for early adopters. He also urged the company to execute fast and invest heavily in the vision, to shut out HTC, Apple, Google and other competitors in the VR space. Regarding other players' participation in the metaverse vision, he called for the company to build the "metaverse" to prevent their competitors from "being in the VR business in a meaningful way at all". In May 2019, Facebook founded Libra Networks, reportedly to develop their own stablecoin cryptocurrency. Later, it was reported that Libra was being supported by financial companies such as Visa, Mastercard, PayPal and Uber. The consortium of companies was expected to pool in $10 million each to fund the launch of the cryptocurrency coin named Libra. Depending on when it would receive approval from the Swiss Financial Market Supervisory authority to operate as a payments service, the Libra Association had planned to launch a limited format cryptocurrency in 2021. Libra was renamed Diem, before being shut down and sold in January 2022 after backlash from Swiss government regulators and the public. During the COVID-19 pandemic, the use of online services, including Facebook, grew globally. Zuckerberg predicted this would be a "permanent acceleration" that would continue after the pandemic. Facebook hired aggressively, growing from 48,268 employees in March 2020 to more than 87,000 by September 2022. Following a period of intense scrutiny and damaging whistleblower leaks, news started to emerge on October 21, 2021 about Facebook's plan to rebrand the company and change its name. In the Q3 2021 earnings call on October 25, Mark Zuckerberg discussed the ongoing criticism of the company's social services and the way it operates, and pointed to the pivoting efforts to building the metaverse – without mentioning the rebranding and the name change. The metaverse vision and the name change from Facebook, Inc. to Meta Platforms was introduced at Facebook Connect on October 28, 2021. Based on Facebook's PR campaign, the name change reflects the company's shifting long term focus of building the metaverse, a digital extension of the physical world by social media, virtual reality and augmented reality features. "Meta" had been registered as a trademark in the United States in 2018 (after an initial filing in 2015) for marketing, advertising, and computer services, by a Canadian company that provided big data analysis of scientific literature. This company was acquired in 2017 by the Chan Zuckerberg Initiative (CZI), a foundation established by Zuckerberg and his wife, Priscilla Chan, and became one of their projects. Following the rebranding announcement, CZI announced that it had already decided to deprioritize the earlier Meta project, thus it would be transferring its rights to the name to Meta Platforms, and the previous project would end in 2022. Soon after the rebranding, in early February 2022, Meta reported a greater-than-expected decline in profits in the fourth quarter of 2021. It reported no growth in monthly users, and indicated it expected revenue growth to stall. It also expected measures taken by Apple Inc. to protect user privacy to cost it some $10 billion in advertisement revenue, an amount equal to roughly 8% of its revenue for 2021. In meeting with Meta staff the day after earnings were reported, Zuckerberg blamed competition for user attention, particularly from video-based apps such as TikTok. The 27% reduction in the company's share price which occurred in reaction to the news eliminated some $230 billion of value from Meta's market capitalization. Bloomberg described the decline as "an epic rout that, in its sheer scale, is unlike anything Wall Street or Silicon Valley has ever seen". Zuckerberg's net worth fell by as much as $31 billion. Zuckerberg owns 13% of Meta, and the holding makes up the bulk of his wealth. According to published reports by Bloomberg on March 30, 2022, Meta turned over data such as phone numbers, physical addresses, and IP addresses to hackers posing as law enforcement officials using forged documents. The law enforcement requests sometimes included forged signatures of real or fictional officials. When asked about the allegations, a Meta representative said, "We review every data request for legal sufficiency and use advanced systems and processes to validate law enforcement requests and detect abuse." In June 2022, Sheryl Sandberg, the chief operating officer of 14 years, announced she would step down that year. Zuckerberg said that Javier Olivan would replace Sandberg, though in a “more traditional” role. In March 2022, Meta (except Meta-owned WhatsApp) and Instagram were banned in Russia and added to the Russian list of terrorist and extremist organizations for alleged Russophobia and hate speech (up to genocidal calls) amid the ongoing Russian invasion of Ukraine. Meta appealed against the ban, but it was upheld by a Moscow court in June of the same year. Also in March 2022, Meta and Italian eyewear giant Luxottica released Ray-Ban Stories, a series of smartglasses which could play music and take pictures. Meta and Luxottica parent company EssilorLuxottica declined to disclose sales on the line of products as of September 2022, though Meta has expressed satisfaction with its customer feedback. In July 2022, Meta saw its first year-on-year revenue decline when its total revenue slipped by 1% to $28.8bn. Analysts and journalists accredited the loss to its advertising business, which has been limited by Apple's app tracking transparency feature and the number of people who have opted not to be tracked by Meta apps. Zuckerberg also accredited the decline to increasing competition from TikTok. On October 27, 2022, Meta's market value dropped to $268 billion, a loss of around $700 billion compared to 2021, and its shares fell by 24%. It lost its spot among the top 20 US companies by market cap, despite reaching the top 5 in the previous year. In November 2022, Meta laid off 11,000 employees, 13% of its workforce. Zuckerberg said the decision to aggressively increase Meta's investments had been a mistake, as he had wrongly predicted that the surge in e-commerce would last beyond the COVID-19 pandemic. He also attributed the decline to increased competition, a global economic downturn and "ads signal loss". Plans to lay off a further 10,000 employees began in April 2023. The layoffs were part of a general downturn in the technology industry, alongside layoffs by companies including Google, Amazon, Tesla, Snap, Twitter and Lyft. Starting from 2022, Meta scrambled to catch up to other tech companies in adopting specialized artificial intelligence hardware and software. It had been using less expensive CPUs instead of GPUs for AI work, but that approach turned out to be less efficient. The company gifted the Inter-university Consortium for Political and Social Research $1.3 million to finance the Social Media Archive's aim to make their data available to social science research. In 2023, Ireland's Data Protection Commissioner imposed a record EUR 1.2 billion fine on Meta for transferring data from Europe to the United States without adequate protections for EU citizens.: 250 In March 2023, Meta announced a new round of layoffs that would cut 10,000 employees and close 5,000 open positions to make the company more efficient. Meta revenue surpassed analyst expectations for the first quarter of 2023 after announcing that it was increasing its focus on AI. On July 6, Meta launched a new app, Threads, a competitor to Twitter. Meta announced its artificial intelligence model Llama 2 in July 2023, available for commercial use via partnerships with major cloud providers like Microsoft. It was the first project to be unveiled out of Meta's generative AI group after it was set up in February. It would not charge access or usage but instead operate with an open-source model to allow Meta to ascertain what improvements need to be made. Prior to this announcement, Meta said it had no plans to release Llama 2 for commercial use. An earlier version of Llama was released to academics. In August 2023, Meta announced its permanent removal of news content from Facebook and Instagram in Canada due to the Online News Act, which requires Canadian news outlets to be compensated for content shared on its platform. The Online News Act was in effect by year-end, but Meta will not participate in the regulatory process. In October 2023, Zuckerberg said that AI would be Meta's biggest investment area in 2024. Meta finished 2023 as one of the best-performing technology stocks of the year, with its share price up 150 percent. Its stock reached an all-time high in January 2024, bringing Meta within 2% of achieving $1 trillion market capitalization. In November 2023 Meta Platforms launched an ad-free service in Europe, allowing subscribers to opt-out of personal data being collected for targeted advertising. A group of 28 European organizations, including Max Schrems' advocacy group NOYB, the Irish Council for Civil Liberties, Wikimedia Europe, and the Electronic Privacy Information Center, signed a 2024 letter to the European Data Protection Board (EDPB) expressing concern that this subscriber model would undermine privacy protections, specifically GDPR data protection standards. Meta removed the Facebook and Instagram accounts of Iran's Supreme Leader Ali Khamenei in February 2024, citing repeated violations of its Dangerous Organizations & Individuals policy. As of March, Meta was under investigation by the FDA for alleged use of their social media platforms to sell illegal drugs. On 16 May 2024, the European Commission began an investigation into Meta over concerns related to child safety. In May 2023, Iraqi social media influencer Esaa Ahmed-Adnan encountered a troubling issue when Instagram removed his posts, citing false copyright violations despite his content being original and free from copyrighted material. He discovered that extortionists were behind these takedowns, offering to restore his content for $3,000 or provide ongoing protection for $1,000 per month. This scam, exploiting Meta’s rights management tools, became widespread in the Middle East, revealing a gap in Meta’s enforcement in developing regions. An Iraqi nonprofit Tech4Peace’s founder, Aws al-Saadi helped Ahmed-Adnan and others, but the restoration process was slow, leading to significant financial losses for many victims, including prominent figures like Ammar al-Hakim. This situation highlighted Meta’s challenges in balancing global growth with effective content moderation and protection. On 16 September 2024, Meta announced it had banned Russian state media outlets from its platforms worldwide due to concerns about "foreign interference activity." This decision followed allegations that RT and its employees funneled $10 million through shell companies to secretly fund influence campaigns on various social media channels. Meta's actions were part of a broader effort to counter Russian covert influence operations, which had intensified since the invasion. At its 2024 Connect conference, Meta presented Orion, its first pair of augmented reality glasses. Though Orion was originally intended to be sold to consumers, the manufacturing process turned out to be too complex and expensive. Instead, the company pivoted to producing a small number of the glasses to be used internally. On 4 October 2024, Meta announced about its new AI model called Movie Gen, capable of generating realistic video and audio clips based on user prompts. Meta stated it would not release Movie Gen for open development, preferring to collaborate directly with content creators and integrate it into its products by the following year. The model was built using a combination of licensed and publicly available datasets. On October 31, 2024, ProPublica published an investigation into deceptive political advertisement scams that sometimes use hundreds of hijacked profiles and facebook pages run by organized networks of scammers. The authors cited spotty enforcement by Meta as a major reason for the extent of the issue. In November 2024, TechCrunch reported that Meta were considering building a $10bn global underwater cable spanning 25,000 miles. In the same month, Meta closed down 2 million accounts on Facebook and Instagram that were linked to scam centers in Myanmar, Laos, Cambodia, the Philippines, and the United Arab Emirates doing pig butchering scams. In December 2024, Meta announced that, beginning February 2025, they would require advertisers to run ads about financial services in Australia to verify information about who are the beneficiary and the payer in a bid to regulate scams. On December 4, 2024, Meta announced it will invest US$10 billion for its largest AI data center in northeast Louisiana, powered by natural gas facilities. On the 11th of that month, Meta experienced a global outage, impacting accounts on all of their social media and messaging applications. Outage reports from DownDetector reached 70,000+ and 100,000+ within minutes for Instagram and Facebook, respectively. In January 2025, Meta announced plans to roll back its diversity, equity, and inclusion (DEI) initiatives, citing shifts in the "legal and policy landscape" in the United States following the 2024 presidential election. The decision followed reports that CEO Mark Zuckerberg sought to align the company more closely with the incoming Trump administration, including changes to content moderation policies and executive leadership. The new content moderation policies continued to bar insults about a person's intellect or mental illness, but made an exception to allow calling LGBTQ people mentally ill because they are gay or transgender. Later that month, Meta agreed to pay $25 million to settle a 2021 lawsuit brought by Donald Trump for suspending his social media accounts after the January 6 riots. Changes to Meta's moderation policies were controversial among its oversight board, with a significant divide in opinion between the board's US conservatives and its global members. In June 2025, Meta Platforms Inc. has decided to make a multibillion-dollar investment into artificial intelligence startup Scale AI. The financing could exceed $10 billion in value which would make it one of the largest private company funding events of all time. In October 2025, it was announced that Meta would be laying off 600 employees in the artificial intelligence unit to perform better and simpler. They referred to their AI unit as "bloated" and are seeking to trim down the department. This mass layoff is going to impact Meta’s AI infrastructure units, Fundamental Artificial Intelligence Research unit (FAIR) and other product-related positions. Mergers and acquisitions Meta has acquired multiple companies (often identified as talent acquisitions). One of its first major acquisitions was in April 2012, when it acquired Instagram for approximately US$1 billion in cash and stock. In October 2013, Facebook, Inc. acquired Onavo, an Israeli mobile web analytics company. In February 2014, Facebook, Inc. announced it would buy mobile messaging company WhatsApp for US$19 billion in cash and stock. The acquisition was completed on October 6. Later that year, Facebook bought Oculus VR for $2.3 billion in cash and stock, which released its first consumer virtual reality headset in 2016. In late November 2019, Facebook, Inc. announced the acquisition of the game developer Beat Games, responsible for developing one of that year's most popular VR games, Beat Saber. In Late 2022, after Facebook Inc rebranded to Meta Platforms Inc, Oculus was rebranded to Meta Quest. In May 2020, Facebook, Inc. announced it had acquired Giphy for a reported cash price of $400 million. It will be integrated with the Instagram team. However, in August 2021, UK's Competition and Markets Authority (CMA) stated that Facebook, Inc. might have to sell Giphy, after an investigation found that the deal between the two companies would harm competition in display advertising market. Facebook, Inc. was fined $70 million by CMA for deliberately failing to report all information regarding the acquisition and the ongoing antitrust investigation. In October 2022, the CMA ruled for a second time that Meta be required to divest Giphy, stating that Meta already controls half of the advertising in the UK. Meta agreed to the sale, though it stated that it disagrees with the decision itself. In May 2023, Giphy was divested to Shutterstock for $53 million. In November 2020, Facebook, Inc. announced that it planned to purchase the customer-service platform and chatbot specialist startup Kustomer to promote companies to use their platform for business. It has been reported that Kustomer valued at slightly over $1 billion. The deal was closed in February 2022 after regulatory approval. In September 2022, Meta acquired Lofelt, a Berlin-based haptic tech startup. In December 2025, it was announced Meta had acquired the AI-wearables startup, Limitless. In the same month, they also acquired another AI startup, Manus AI, for $2 billion. Manus announced in December that its platform had achieved $100mm in recurring revenue just 8 months after its launch and Meta said it will scale the platform to many other businesses. In January 2026, it was announced Meta proposed acquisition of Manus was undergoing preliminary scrutiny by Chinese regulators. The examination concerns the cross-border transfer of artificial intelligence technology developed in China. Lobbying In 2020, Facebook, Inc. spent $19.7 million on lobbying, hiring 79 lobbyists. In 2019, it had spent $16.7 million on lobbying and had a team of 71 lobbyists, up from $12.6 million and 51 lobbyists in 2018. Facebook was the largest spender of lobbying money among the Big Tech companies in 2020. The lobbying team includes top congressional aide John Branscome, who was hired in September 2021, to help the company fend off threats from Democratic lawmakers and the Biden administration. In December 2024, Meta donated $1 million to the inauguration fund for then-President-elect Donald Trump. In 2025, Meta was listed among the donors funding the construction of the White House State Ballroom. Partnerships February 2026, Meta announced a long-term partnership with Nvidia. Censorship In August 2024, Mark Zuckerberg sent a letter to Jim Jordan indicating that during the COVID-19 pandemic the Biden administration repeatedly asked Meta to limit certain COVID-19 content, including humor and satire, on Facebook and Instagram. In 2016 Meta hired Jordana Cutler, formerly an employee at the Israeli Embassy to the United States, as its policy chief for Israel and the Jewish Diaspora. In this role, Cutler pushed for the censorship of accounts belonging to Students for Justice in Palestine chapters in the United States. Critics have said that Cutler's position gives the Israeli government an undue influence over Meta policy, and that few countries have such high levels of contact with Meta policymakers. Following the election of Donald Trump in 2025, various sources noted possible censorship related to the Democratic Party on Instagram and other Meta platforms. In February 2025, a Meta rep flagged journalist Gil Duran's article and other "critiques of tech industry figures" as spam or sensitive content, limiting their reach. In March 2025, Meta attempted to block former employee Sarah Wynn-Williams from promoting or further distributing her memoir, Careless People, that includes allegations of unaddressed sexual harassment in the workplace by senior executives. The New York Times reports that the arbitration is among Meta's most forcible attempts to repudiate a former employee's account of workplace dynamics. Publisher Macmillan reacted to the ruling by the Emergency International Arbitral Tribunal by stating that it will ignore its provisions. As of 15 March 2025[update], hardback and digital versions of Careless People were being offered for sale by major online retailers. From October 2025, Meta began removing and restricting access for accounts related to LGBTQ, reproductive health and abortion information pages on its platforms. Martha Dimitratou, executive director of Repro Uncensored, called Meta's shadow-banning of these issues "One of the biggest waves of censorship we are seeing". Disinformation concerns Since its inception, Meta has been accused of being a host for fake news and misinformation. In the wake of the 2016 United States presidential election, Zuckerberg began to take steps to eliminate the prevalence of fake news, as the platform had been criticized for its potential influence on the outcome of the election. The company initially partnered with ABC News, the Associated Press, FactCheck.org, Snopes and PolitiFact for its fact-checking initiative; as of 2018, it had over 40 fact-checking partners across the world, including The Weekly Standard. A May 2017 review by The Guardian found that the platform's fact-checking initiatives of partnering with third-party fact-checkers and publicly flagging fake news were regularly ineffective, and appeared to be having minimal impact in some cases. In 2018, journalists working as fact-checkers for the company criticized the partnership, stating that it had produced minimal results and that the company had ignored their concerns. In 2024 Meta's decision to continue to disseminate a falsified video of US president Joe Biden, even after it had been proven to be fake, attracted criticism and concern. In January 2025, Meta ended its use of third-party fact-checkers in favor of a user-run community notes system similar to the one used on X. While Zuckerberg supported these changes, saying that the amount of censorship on the platform was excessive, the decision received criticism by fact-checking institutions, stating that the changes would make it more difficult for users to identify misinformation. Meta also faced criticism for weakening its policies on hate speech that were designed to protect minorities and LGBTQ+ individuals from bullying and discrimination. While moving its content review teams from California to Texas, Meta changed their hateful conduct policy to eliminate restrictions on anti-LGBT and anti-immigrant hate speech, as well as explicitly allowing users to accuse LGBT people of being mentally ill or abnormal based on their sexual orientation or gender identity. In January 2025, Meta faced significant criticism for its role in removing LGBTQ+ content from its platforms, amid its broader efforts to address anti-LGBTQ+ hate speech. The removal of LGBTQ+ themes was noted as part of the wider crackdown on content deemed to violate its community guidelines. Meta's content moderation policies, which were designed to combat harmful speech and protect users from discrimination, inadvertently led to the removal or restriction of LGBTQ+ content, particularly posts highlighting LGBTQ+ identities, support, or political issues. According to reports, LGBTQ+ posts, including those that simply celebrated pride or advocated for LGBTQ+ rights, were flagged and removed for reasons that some critics argue were vague or inconsistently applied. Many LGBTQ+ activists and users on Meta's platforms expressed concern that such actions stifled visibility and expression, potentially isolating LGBTQ+ individuals and communities, especially in spaces that were historically important for outreach and support. Lawsuits Numerous lawsuits have been filed against the company, both when it was known as Facebook, Inc., and as Meta Platforms. In March 2020, the Office of the Australian Information Commissioner (OAIC) sued Facebook, for significant and persistent infringements of the rule on privacy involving the Cambridge Analytica fiasco. Every violation of the Privacy Act is subject to a theoretical cumulative liability of $1.7 million. The OAIC estimated that a total of 311,127 Australians had been exposed. On December 8, 2020, the U.S. Federal Trade Commission and 46 states (excluding Alabama, Georgia, South Carolina, and South Dakota), the District of Columbia and the territory of Guam, launched Federal Trade Commission v. Facebook as an antitrust lawsuit against Facebook. The lawsuit concerns Facebook's acquisition of two competitors—Instagram and WhatsApp—and the ensuing monopolistic situation. FTC alleges that Facebook holds monopolistic power in the U.S. social networking market and seeks to force the company to divest from Instagram and WhatsApp to break up the conglomerate. William Kovacic, a former chairman of the Federal Trade Commission, argued the case will be difficult to win as it would require the government to create a counterfactual argument of an internet where the Facebook-WhatsApp-Instagram entity did not exist, and prove that harmed competition or consumers. In November 2025, it was ruled that Meta did not violate antitrust laws and holds no monopoly in the market. On December 24, 2021, a court in Russia fined Meta for $27 million after the company declined to remove unspecified banned content. The fine was reportedly tied to the company's annual revenue in the country. In May 2022, a lawsuit was filed in Kenya against Meta and its local outsourcing company Sama. Allegedly, Meta has poor working conditions in Kenya for workers moderating Facebook posts. According to the lawsuit, 260 screeners were declared redundant with confusing reasoning. The lawsuit seeks financial compensation and an order that outsourced moderators be given the same health benefits and pay scale as Meta employees. In June 2022, 8 lawsuits were filed across the U.S. over the allege that excessive exposure to platforms including Facebook and Instagram has led to attempted or actual suicides, eating disorders and sleeplessness, among other issues. The litigation follows a former Facebook employee's testimony in Congress that the company refused to take responsibility. The company noted that tools have been developed for parents to keep track of their children's activity on Instagram and set time limits, in addition to Meta's "Take a break" reminders. In addition, the company is providing resources specific to eating disorders as well as developing AI to prevent children under the age of 13 signing up for Facebook or Instagram. In June 2022, Meta settled a lawsuit with the US Department of Justice. The lawsuit, which was filed in 2019, alleged that the company enabled housing discrimination through targeted advertising, as it allowed homeowners and landlords to run housing ads excluding people based on sex, race, religion, and other characteristics. The U.S. Department of Justice stated that this was in violation of the Fair Housing Act. Meta was handed a penalty of $115,054 and given until December 31, 2022, to shadow the algorithm tool. In January 2023, Meta was fined €390 million for violations of the European Union General Data Protection Regulation. In May 2023, the European Data Protection Board fined Meta a record €1.2 billion for breaching European Union data privacy laws by transferring personal data of Facebook users to servers in the U.S. In July 2024, Meta agreed to pay the state of Texas US$1.4 billion to settle a lawsuit brought by Texas Attorney General Ken Paxton accusing the company of collecting users' biometric data without consent, setting a record for the largest privacy-related settlement ever obtained by a state attorney general. In October 2024, Meta Platforms faced lawsuits in Japan from 30 plaintiffs who claimed they were defrauded by fake investment ads on Facebook and Instagram, featuring false celebrity endorsements. The plaintiffs are seeking approximately $2.8 million in damages. In April 2025, the Kenyan High Court ruled that a US$2.4 billion lawsuit in which three plaintiffs claim that Facebook inflamed civil violence in Ethiopia in 2021 could proceed. In April 2025, Meta was fined €200 million ($230 million) for breaking the Digital Markets Act, by imposing a “consent or pay” system that forces users to either allow their personal data to be used to target advertisements, or pay a subscription fee for advertising-free versions of Facebook and Instagram. In late April 2025, a case was filed against Meta in Ghana over the alleged psychological distress experienced by content moderators employed to take down disturbing social media content including depictions of murders, extreme violence and child sexual abuse. Meta moved the moderation service to the Ghanaian capital of Accra after legal issues in the previous location Kenya. The new moderation company is Teleperformance, a multinational corporation with a history of worker's rights violation. Reports suggests the conditions are worse here than in the previous Kenyan location, with many workers afraid of speaking out due to fear of returning to conflict zones. Workers reported developing mental illnesses, attempted suicides, and low pay. In 26 January 2026, a New Mexico state court case was filed, suggesting that Mark Zuckerberg approved allowing minors to access artificial intelligence chatbot companions that safety staffers warned were capable of sexual interactions. In 2020, the company UReputation, which had been involved in several cases concerning the management of digital armies[clarification needed], filed a lawsuit against Facebook, accusing it of unlawfully transmitting personal data to third parties. Legal actions were initiated in Tunisia, France, and the United States. In 2025, the United States District court for the Northern District of Georgia approved a discovery procedure, allowing UReputation to access documents and evidence held by Meta. Structure Meta's key management consists of: As of October 2022[update], Meta had 83,553 employees worldwide. As of June 2024[update], Meta's board consisted of the following directors; Meta Platforms is mainly owned by institutional investors, who hold around 80% of all shares. Insiders control the majority of voting shares. The three largest individual investors in 2024 were Mark Zuckerberg, Sheryl Sandberg and Christopher K. Cox. The largest shareholders in late 2024/early 2025 were: Roger McNamee, an early Facebook investor and Zuckerberg's former mentor, said Facebook had "the most centralized decision-making structure I have ever encountered in a large company". Facebook co-founder Chris Hughes has stated that chief executive officer Mark Zuckerberg has too much power, that the company is now a monopoly, and that, as a result, it should be split into multiple smaller companies. In an op-ed in The New York Times, Hughes said he was concerned that Zuckerberg had surrounded himself with a team that did not challenge him, and that it is the U.S. government's job to hold him accountable and curb his "unchecked power". He also said that "Mark's power is unprecedented and un-American." Several U.S. politicians agreed with Hughes. European Union Commissioner for Competition Margrethe Vestager stated that splitting Facebook should be done only as "a remedy of the very last resort", and that it would not solve Facebook's underlying problems. Revenue Facebook ranked No. 34 in the 2020 Fortune 500 list of the largest United States corporations by revenue, with almost $86 billion in revenue most of it coming from advertising. One analysis of 2017 data determined that the company earned US$20.21 per user from advertising. According to New York, since its rebranding, Meta has reportedly lost $500 billion as a result of new privacy measures put in place by companies such as Apple and Google which prevents Meta from gathering users' data. In February 2015, Facebook announced it had reached two million active advertisers, with most of the gain coming from small businesses. An active advertiser was defined as an entity that had advertised on the Facebook platform in the last 28 days. In March 2016, Facebook announced it had reached three million active advertisers with more than 70% from outside the United States. Prices for advertising follow a variable pricing model based on auctioning ad placements, and potential engagement levels of the advertisement itself. Similar to other online advertising platforms like Google and Twitter, targeting of advertisements is one of the chief merits of digital advertising compared to traditional media. Marketing on Meta is employed through two methods based on the viewing habits, likes and shares, and purchasing data of the audience, namely targeted audiences and "look alike" audiences. The U.S. IRS challenged the valuation Facebook used when it transferred IP from the U.S. to Facebook Ireland (now Meta Platforms Ireland) in 2010 (which Facebook Ireland then revalued higher before charging out), as it was building its double Irish tax structure. The case is ongoing and Meta faces a potential fine of $3–5bn. The U.S. Tax Cuts and Jobs Act of 2017 changed Facebook's global tax calculations. Meta Platforms Ireland is subject to the U.S. GILTI tax of 10.5% on global intangible profits (i.e. Irish profits). On the basis that Meta Platforms Ireland Limited is paying some tax, the effective minimum US tax for Facebook Ireland will be circa 11%. In contrast, Meta Platforms Inc. would incur a special IP tax rate of 13.125% (the FDII rate) if its Irish business relocated to the U.S. Tax relief in the U.S. (21% vs. Irish at the GILTI rate) and accelerated capital expensing, would make this effective U.S. rate around 12%. The insignificance of the U.S./Irish tax difference was demonstrated when Facebook moved 1.5bn non-EU accounts to the U.S. to limit exposure to GDPR. Facilities Users outside of the U.S. and Canada contract with Meta's Irish subsidiary, Meta Platforms Ireland Limited (formerly Facebook Ireland Limited), allowing Meta to avoid US taxes for all users in Europe, Asia, Australia, Africa and South America. Meta is making use of the Double Irish arrangement which allows it to pay 2–3% corporation tax on all international revenue. In 2010, Facebook opened its fourth office, in Hyderabad, India, which houses online advertising and developer support teams and provides support to users and advertisers. In India, Meta is registered as Facebook India Online Services Pvt Ltd. It also has offices or planned sites in Chittagong, Bangladesh; Dublin, Ireland; and Austin, Texas, among other cities. Facebook opened its London headquarters in 2017 in Fitzrovia in central London. Facebook opened an office in Cambridge, Massachusetts in 2018. The offices were initially home to the "Connectivity Lab", a group focused on bringing Internet access to those who do not have access to the Internet. In April 2019, Facebook opened its Taiwan headquarters in Taipei. In March 2022, Meta opened new regional headquarters in Dubai. In September 2023, it was reported that Meta had paid £149m to British Land to break the lease on Triton Square London office. Meta reportedly had another 18 years left on its lease on the site. As of 2023, Facebook operated 21 data centers. It committed to purchase 100% renewable energy and to reduce its greenhouse gas emissions 75% by 2020. Its data center technologies include Fabric Aggregator, a distributed network system that accommodates larger regions and varied traffic patterns. Reception US Representative Alexandria Ocasio-Cortez responded in a tweet to Zuckerberg's announcement about Meta, saying: "Meta as in 'we are a cancer to democracy metastasizing into a global surveillance and propaganda machine for boosting authoritarian regimes and destroying civil society ... for profit!'" Ex-Facebook employee Frances Haugen and whistleblower behind the Facebook Papers responded to the rebranding efforts by expressing doubts about the company's ability to improve while led by Mark Zuckerberg, and urged the chief executive officer to resign. In November 2021, a video published by Inspired by Iceland went viral, in which a Zuckerberg look-alike promoted the Icelandverse, a place of "enhanced actual reality without silly looking headsets". In a December 2021 interview, SpaceX and Tesla chief executive officer Elon Musk said he could not see a compelling use-case for the VR-driven metaverse, adding: "I don't see someone strapping a frigging screen to their face all day." In January 2022, Louise Eccles of The Sunday Times logged into the metaverse with the intention of making a video guide. She wrote: Initially, my experience with the Oculus went well. I attended work meetings as an avatar and tried an exercise class set in the streets of Paris. The headset enabled me to feel the thrill of carving down mountains on a snowboard and the adrenaline rush of climbing a mountain without ropes. Yet switching to the social apps, where you mingle with strangers also using VR headsets, it was at times predatory and vile. Eccles described being sexually harassed by another user, as well as "accents from all over the world, American, Indian, English, Australian, using racist, sexist, homophobic and transphobic language". She also encountered users as young as 7 years old on the platform, despite Oculus headsets being intended for users over 13. See also References External links 37°29′06″N 122°08′54″W / 37.48500°N 122.14833°W / 37.48500; -122.14833
========================================
[SOURCE: https://en.wikipedia.org/wiki/File:Transistor-die-KSY34.jpg] | [TOKENS: 97]
File:Transistor-die-KSY34.jpg Summary Licensing File history Click on a date/time to view the file as it appeared at that time. File usage The following 2 pages use this file: Global file usage The following other wikis use this file: Metadata This file contains additional information, probably added from the digital camera or scanner used to create or digitize it. If the file has been modified from its original state, some details may not fully reflect the modified file.
========================================
[SOURCE: https://en.wikipedia.org/wiki/Wikipedia:Verifiability#Burden_of_evidence] | [TOKENS: 2637]
Contents Wikipedia:Verifiability In the English Wikipedia, verifiability means that people can check that facts or claims correspond to reliable sources. Wikipedia's content is determined by published information rather than editors' beliefs, experiences, or previously unpublished ideas or information. Even if you are sure something is true, it must have been previously published in a reliable source before you can add it.[a] If reliable sources disagree with each other, then maintain a neutral point of view and present what the various sources say, giving each side its due weight. Each fact or claim in an article must be verifiable. Additionally, four types of information must be accompanied by an inline citation to a reliable source that directly supports[b] the material: When material that needs an inline citation appears in two or more articles, an inline citation is needed in each. Any material that needs an inline citation but does not have one may be removed. Please immediately remove contentious material about living people (or existing groups) that is unsourced or poorly sourced. For how to write citations, see citing sources. Verifiability, no original research, and neutral point of view are Wikipedia's core content policies. They work together to determine content, so editors should understand the key points of all three. Articles must also comply with the copyright policy. Responsibility for providing citations All content must be verifiable. A fact or claim is "verifiable" if a reliable source that supports it could be cited, even if there is no citation for it in the article at the moment. The burden to demonstrate verifiability lies with the editor who adds or restores material, and it is satisfied by providing one inline citation to a reliable source that directly supports[b] the contribution.[c] The cited source must clearly support the material as presented in the article. Cite the source clearly, ideally giving page number(s)—though sometimes a section, chapter, or other division is appropriate instead; see Wikipedia:Citing sources for details of how to do this. Facts or claims without an inline citation to a reliable source that directly supports[b] them may be removed. They should not be restored without an inline citation to a reliable source. Whether or how quickly material should be removed for lacking an inline citation to a reliable source depends on the material and the overall state of the article. Consider adding a citation needed tag as an interim step to removing unsourced material, to allow references to be added.[d] When tagging or removing material for lacking an inline citation, state your concern that it may not be possible to find a published reliable source, and the material therefore may not be verifiable.[e] If you think the material is verifiable, you are encouraged to provide an inline citation yourself before removing or tagging it. Do not leave unsourced or poorly sourced material in an article if it might damage the reputation of living people or existing groups, and do not move it to the talk page, per Wikipedia:Biographies of living persons policy (common shortcut: WP:BLP). The latter policy also applies to some groups. Reliable sources A cited source on Wikipedia is often a specific portion of text (such as a short article or a page in a book). But when editors discuss sources (for example, to debate their appropriateness or reliability) they are usually talking about one or more related characteristics: All four can affect reliability. Base articles on reliable, independent, published sources with a reputation for fact-checking and accuracy. Source material must be published, on Wikipedia meaning made available to the public in some form.[f] Unpublished material is not considered reliable. Use sources that directly support the material presented in an article and are appropriate to the claims made. The appropriateness of any source depends on the context. Be especially careful when sourcing content related to living people or medicine. If available, academic and peer-reviewed publications are usually the most reliable sources on topics such as history, medicine, and science. Editors may also use material from reliable non-academic sources, particularly if it appears in respected mainstream publications. Other reliable sources include: Editors may also use electronic media, subject to the same criteria (see details in Wikipedia:Identifying reliable sources). The best sources have a professional structure for checking or analyzing facts, legal issues, evidence, and arguments. The greater the degree of scrutiny given to these issues, the more reliable the source. Some newspapers, magazines, and other news organizations host online pages or columns they call blogs. These may be acceptable sources if the writers are professionals, but use them with caution because blogs may not be subject to the news organization's normal fact-checking process.[g] If a news organization publishes an opinion piece in a blog, attribute the statement to the writer, e.g. "Jane Smith wrote ..." Never use the blog comments that are left by the readers as sources. For personal or group blogs that are not reliable sources, see § Self-published sources below. To discuss the reliability of a specific source for a particular statement, consult Wikipedia:Reliable sources/Noticeboard, which seeks to apply this policy to particular cases. For a guideline discussing the reliability of particular types of sources, see Wikipedia:Reliable sources. In case of conflict between this policy and the Wikipedia:Reliable sources guideline, or any other guideline related to sourcing, this policy has priority. Sources that are usually not reliable Questionable sources are those that have a poor reputation for checking the facts, lack meaningful editorial oversight, or have an apparent conflict of interest. Such sources include websites and publications expressing views widely considered by other sources to be promotional, extremist, or that rely heavily on unsubstantiated gossip, rumor, or personal opinion. Questionable sources should be used only as sources for material on themselves, such as in articles about themselves; see below. They are not suitable sources for contentious claims. Predatory open access journals are questionable due to the absence of quality control in the peer-review process. Anyone can create a personal web page, self-publish a book, or claim to be an expert. Self-published material, such as books, patents, newsletters, personal websites, open wikis, personal or group blogs (as distinguished from newsblogs, above), content farms, podcasts, Internet forum postings, and social media postings, are largely not acceptable as sources. Self-published sources may be considered reliable if published by an established subject-matter expert, whose work in the relevant field has previously been published by reliable, independent publications.[g] Be careful when using such sources: if the information in question is suitable for inclusion, someone else will likely have published it in independent, reliable sources. Never use self-published sources as third-party sources about living people, even if the author is an expert, well-known professional researcher, or writer. Self-published and questionable sources may be used as sources of information about themselves, usually in articles about themselves or their activities, without the self-published source requirement that they are established experts in the field, so long as: This policy also applies to material made public by the source on social networking websites such as Twitter, Tumblr, LinkedIn, Reddit, Instagram and Facebook. Do not use articles from Wikipedia (whether English Wikipedia or Wikipedias in other languages) as sources, since Wikipedia is a user-generated source. Do not use websites mirroring Wikipedia content or publications relying on material from Wikipedia as sources. Content from a Wikipedia article is not considered reliable unless it is backed up by citing reliable sources. Confirm that these sources support the content, then use them directly. An exception is allowed when Wikipedia itself is being discussed in the article. These may cite content from Wikipedia or a sister project to support a statement about Wikipedia. Wikipedia or the sister project is a primary source in this case and may be used following the policy for primary sources. Any such use should avoid original research, undue emphasis on Wikipedia's role or views, and inappropriate self-reference. The article text should say that the material is sourced from Wikipedia. Accessibility Do not reject reliable sources just because they are difficult or costly to access. Some reliable sources are not easily accessible. For example, an online source may require payment, and a print-only source may be available only through libraries. Rare historical sources may even be available only in special museum collections and archives. If you have trouble accessing a source, others might do so for you (see WikiProject Resource Exchange). Citations to non-English reliable sources are allowed on the English Wikipedia. However, because this project is in English, English-language sources are preferred over non-English ones when they are available and of equal quality and relevance. As with sources in English, if a dispute arises involving a citation to a non-English source, editors may request a quotation of relevant portions of the original source be provided, either in text, in a footnote, or on the article talk page.[h] (See Template:Request quotation.) If you quote a non-English reliable source (whether in the main text or in a footnote), a translation into English should accompany the quote. Translations published by reliable sources are preferred over translations by Wikipedians, but translations by Wikipedians are preferred over machine translations. When using a machine translation of source material, editors should be reasonably certain that the translation is accurate and the source is appropriate. Editors should not rely upon machine translations of non-English sources in contentious articles or biographies of living people. If needed, ask an editor who can translate it for you. The original text is usually included with the translated text in articles when translated by Wikipedians, and the translating editor is usually not cited. When quoting any material, whether in English or in some other language, be careful not to violate copyright; see the fair-use guideline. Other issues While information must be verifiable for inclusion in an article, not all verifiable information must be included. Consensus may determine that inclusion of a verifiable fact or claim does not improve an article, and other policies may indicate that the material is inappropriate. Such information should be omitted or presented instead in a different article. The responsibility for achieving consensus for inclusion is on those seeking to include disputed content. If you want to request an inline citation for an unsourced statement, you can tag a sentence with the {{citation needed}} template. You can also leave a note on the talk page asking for a source, or move the material to the talk page and ask for a source there. To request verification that a reference supports the text, tag it with {{verification needed}}. Material that fails verification may be tagged with {{failed verification}} or removed. It helps other editors to explain your rationale for using templates to tag material in the template, edit summary, or on the talk page. Take special care with contentious material about living and recently deceased people. Unsourced or poorly sourced material that is contentious, especially text that is negative, derogatory, or potentially damaging, should be removed immediately rather than tagged or moved to the talk page. Any exceptional claim requires multiple high-quality sources. Warnings (red flags) that should prompt extra caution include: Verifiability and other principles Do not plagiarize or breach copyright when using sources. Summarize source material in your own words as much as possible; when quoting or closely paraphrasing a source, use an inline citation, and in-text attribution where appropriate. Do not link to any source that violates the copyrights of others per contributors' rights and obligations. You can link to websites that display copyrighted works as long as the website has licensed the work or uses the work in a way compliant with fair use. Knowingly directing others to material that violates copyright may be considered contributory copyright infringement. If there is reason to think a source violates copyright, do not cite it. This is particularly relevant when linking to sites such as Scribd or YouTube, where due care should be taken to avoid linking to material violating copyright. Even when information is cited to reliable sources, you must present it with a neutral point of view (NPOV). Articles should be based on thorough research of sources. All articles must adhere to NPOV, fairly representing all majority and significant-minority viewpoints published by reliable sources, in rough proportion to the prominence of each view. Tiny-minority views need not be included, except in articles devoted to them. If there is a disagreement between sources, use in-text attribution: "John Smith argues X, while Paul Jones maintains Y," followed by an inline citation. Sources themselves do not need to maintain a neutral point of view. Indeed, many reliable sources are not neutral. Our job as editors is simply to summarize what reliable sources say. If no reliable, independent sources can be found on a topic, Wikipedia should not have an article on it (i.e., the topic is not notable). However, notability is based on the existence of suitable sources, not on the state of sourcing in an article. The no original research policy (NOR) is closely related to the Verifiability policy. Among its requirements are: See also Notes References Further reading
========================================
[SOURCE: https://en.wikipedia.org/wiki/Birthday#cite_ref-20] | [TOKENS: 4101]
Contents Birthday A birthday is the anniversary of the birth of a person or the figurative birth of an institution. Birthdays of people are celebrated in numerous cultures, often with birthday gifts, birthday cards, a birthday party, or a rite of passage. Many religions celebrate the birth of their founders or religious figures with special holidays (e.g. Christmas, Mawlid, Buddha's Birthday, Krishna Janmashtami, and Gurpurb). There is a distinction between birthday and birthdate (also known as date of birth): the former, except for February 29, occurs each year (e.g. January 15), while the latter is the complete date when a person was born (e.g. January 15, 2001). Coming of age In most legal systems, one becomes a legal adult on a particular birthday when they reach the age of majority (usually between 12 and 21), and reaching age-specific milestones confers particular rights and responsibilities. At certain ages, one may become eligible to leave full-time education, become subject to military conscription or to enlist in the military, to consent to sexual intercourse, to marry with parental consent, to marry without parental consent, to vote, to run for elected office, to legally purchase (or consume) alcohol and tobacco products, to purchase lottery tickets, or to obtain a driver's licence. The age of majority is when minors cease to legally be considered children and assume control over their persons, actions, and decisions, thereby terminating the legal control and responsibilities of their parents or guardians over and for them. Most countries set the age of majority at 18, though it varies by jurisdiction. Many cultures celebrate a coming of age birthday when a person reaches a particular year of life. Some cultures celebrate landmark birthdays in early life or old age. In many cultures and jurisdictions, if a person's real birthday is unknown (for example, if they are an orphan), their birthday may be adopted or assigned to a specific day of the year, such as January 1. Racehorses are reckoned to become one year old in the year following their birth on January 1 in the Northern Hemisphere and August 1 in the Southern Hemisphere.[relevant?] Birthday parties In certain parts of the world, an individual's birthday is celebrated by a party featuring a specially made cake. Presents are bestowed on the individual by the guests appropriate to their age. Other birthday activities may include entertainment (sometimes by a hired professional, i.e., a clown, magician, or musician) and a special toast or speech by the birthday celebrant. The last stanza of Patty Hill's and Mildred Hill's famous song, "Good Morning to You" (unofficially titled "Happy Birthday to You") is typically sung by the guests at some point in the proceedings. In some countries, a piñata takes the place of a cake. The birthday cake may be decorated with lettering and the person's age, or studded with the same number of lit candles as the age of the individual. The celebrated individual may make a silent wish and attempt to blow out the candles in one breath; if successful, superstition holds that the wish will be granted. In many cultures, the wish must be kept secret or it will not "come true". Birthdays as holidays Historically significant people's birthdays, such as national heroes or founders, are often commemorated by an official holiday marking the anniversary of their birth. Some notables, particularly monarchs, have an official birthday on a fixed day of the year, which may not necessarily match the day of their birth, but on which celebrations are held. In Mahayana Buddhism, many monasteries celebrate the anniversary of Buddha's birth, usually in a highly formal, ritualized manner. They treat Buddha's statue as if it was Buddha himself as if he were alive; bathing, and "feeding" him. Jesus Christ's traditional birthday is celebrated as Christmas Eve or Christmas Day around the world, on December 24 or 25, respectively. As some Eastern churches use the Julian calendar, December 25 will fall on January 7 in the Gregorian calendar. These dates are traditional and have no connection with Jesus's actual birthday, which is not recorded in the Gospels. Similarly, the birthdays of the Virgin Mary and John the Baptist are liturgically celebrated on September 8 and June 24, especially in the Roman Catholic and Eastern Orthodox traditions (although for those Eastern Orthodox churches using the Julian calendar the corresponding Gregorian dates are September 21 and July 7 respectively). As with Christmas, the dates of these celebrations are traditional and probably have no connection with the actual birthdays of these individuals. Catholic saints are remembered by a liturgical feast on the anniversary of their "birth" into heaven a.k.a. their day of death. In Hinduism, Ganesh Chaturthi is a festival celebrating the birth of the elephant-headed deity Ganesha in extensive community celebrations and at home. Figurines of Ganesha are made for the holiday and are widely sold. Sikhs celebrate the anniversary of the birth of Guru Nanak and other Sikh gurus, which is known as Gurpurb. Mawlid is the anniversary of the birth of Muhammad and is celebrated on the 12th or 17th day of Rabi' al-awwal by adherents of Sunni and Shia Islam respectively. These are the two most commonly accepted dates of birth of Muhammad. However, there is much controversy regarding the permissibility of celebrating Mawlid, as some Muslims judge the custom as an unacceptable practice according to Islamic tradition. In Iran, Mother's Day is celebrated on the birthday of Fatima al-Zahra, the daughter of Muhammad. Banners reading Ya Fatima ("O Fatima") are displayed on government buildings, private buildings, public streets and car windows. Religious views In Judaism, rabbis are divided about celebrating this custom, although the majority of the faithful accept it. In the Torah, the only mention of a birthday is the celebration of Pharaoh's birthday in Egypt (Genesis 40:20). Although the birthday of Jesus of Nazareth is celebrated as a Christian holiday on December 25, historically the celebrating of an individual person's birthday has been subject to theological debate. Early Christians, notes The World Book Encyclopedia, "considered the celebration of anyone's birth to be a pagan custom." Origen, in his commentary "On Levites," wrote that Christians should not only refrain from celebrating their birthdays but should look at them with disgust as a pagan custom. A saint's day was typically celebrated on the anniversary of their martyrdom or death, considered the occasion of or preparation for their entrance into Heaven or the New Jerusalem. Ordinary folk in the Middle Ages celebrated their saint's day (the saint they were named after), but nobility celebrated the anniversary of their birth.[citation needed] The "Squire's Tale", one of Chaucer's Canterbury Tales, opens as King Cambuskan proclaims a feast to celebrate his birthday. In the Modern era, the Catholic Church, the Eastern Orthodox Church and Protestantism, i.e. the three main branches of Christianity, as well as almost all Christian religious denominations, consider celebrating birthdays acceptable or at most a choice of the individual. An exception is Jehovah's Witnesses, who do not celebrate them for various reasons: in their interpretation this feast has pagan origins, was not celebrated by early Christians, is negatively expounded in the Holy Scriptures and has customs linked to superstition and magic. In some historically Roman Catholic and Eastern Orthodox countries,[a] it is common to have a 'name day', otherwise known as a 'Saint's day'. It is celebrated in much the same way as a birthday, but it is held on the official day of a saint with the same Christian name as the birthday person; the difference being that one may look up a person's name day in a calendar, or easily remember common name days (for example, John or Mary); however in pious traditions, the two were often made to concur by giving a newborn the name of a saint celebrated on its day of confirmation, more seldom one's birthday. Some are given the name of the religious feast of their christening's day or birthday, for example, Noel or Pascal (French for Christmas and "of Easter"); as another example, Togliatti was given Palmiro as his first name because he was born on Palm Sunday. The birthday does not reflect Islamic tradition, and because of this, the majority of Muslims refrain from celebrating it. Others do not object, as long as it is not accompanied by behavior contrary to Islamic tradition. A good portion of Muslims (and Arab Christians) who have emigrated to the United States and Europe celebrate birthdays as customary, especially for children, while others abstain. Hindus celebrate the birth anniversary day every year when the day that corresponds to the lunar month or solar month (Sun Signs Nirayana System – Sourava Mana Masa) of birth and has the same asterism (Star/Nakshatra) as that of the date of birth. That age is reckoned whenever Janma Nakshatra of the same month passes. Hindus regard death to be more auspicious than birth, since the person is liberated from the bondages of material society. Also, traditionally, rituals and prayers for the departed are observed on the 5th and 11th days, with many relatives gathering. Historical and cultural perspectives According to Herodotus (5th century BC), of all the days in the year, the one which the Persians celebrate most is their birthday. It was customary to have the board furnished on that day with an ampler supply than common: the richer people eat wholly baked cow, horse, camel, or donkey (Greek: ὄνον), while the poorer classes use instead the smaller kinds of cattle. On his birthday, the king anointed his head and presented gifts to the Persians. According to the law of the Royal Supper, on that day "no one should be refused a request". The rule for drinking was "No restrictions". In ancient Rome, a birthday (dies natalis) was originally an act of religious cultivation (cultus). A dies natalis was celebrated annually for a temple on the day of its founding, and the term is still used sometimes for the anniversary of an institution such as a university. The temple founding day might become the "birthday" of the deity housed there. March 1, for example, was celebrated as the birthday of the god Mars. Each human likewise had a natal divinity, the guardian spirit called the Genius, or sometimes the Juno for a woman, who was owed religious devotion on the day of birth, usually in the household shrine (lararium). The decoration of a lararium often shows the Genius in the role of the person carrying out the rites. A person marked their birthday with ritual acts that might include lighting an altar, saying prayers, making vows (vota), anointing and wreathing a statue of the Genius, or sacrificing to a patron deity. Incense, cakes, and wine were common offerings. Celebrating someone else's birthday was a way to show affection, friendship, or respect. In exile, the poet Ovid, though alone, celebrated not only his own birthday rite but that of his far distant wife. Birthday parties affirmed social as well as sacred ties. One of the Vindolanda tablets is an invitation to a birthday party from the wife of one Roman officer to the wife of another. Books were a popular birthday gift, sometimes handcrafted as a luxury edition or composed especially for the person honored. Birthday poems are a minor but distinctive genre of Latin literature. The banquets, libations, and offerings or gifts that were a regular part of most Roman religious observances thus became part of birthday celebrations for individuals. A highly esteemed person would continue to be celebrated on their birthday after death, in addition to the several holidays on the Roman calendar for commemorating the dead collectively. Birthday commemoration was considered so important that money was often bequeathed to a social organization to fund an annual banquet in the deceased's honor. The observance of a patron's birthday or the honoring of a political figure's Genius was one of the religious foundations for imperial cult or so-called "emperor worship." The Chinese word for "year(s) old" (t 歲, s 岁, suì) is entirely different from the usual word for "year(s)" (年, nián), reflecting the former importance of Chinese astrology and the belief that one's fate was bound to the stars imagined to be in opposition to the planet Jupiter at the time of one's birth. The importance of this duodecennial orbital cycle only survives in popular culture as the 12 animals of the Chinese zodiac, which change each Chinese New Year and may be used as a theme for some gifts or decorations. Because of the importance attached to the influence of these stars in ancient China and throughout the Sinosphere, East Asian age reckoning previously began with one at birth and then added years at each Chinese New Year, so that it formed a record of the suì one had lived through rather than of the exact amount of time from one's birth. This method—which can differ by as much as two years of age from other systems—is increasingly uncommon and is not used for official purposes in the PRC or on Taiwan, although the word suì is still used for describing age. Traditionally, Chinese birthdays—when celebrated—were reckoned using the lunisolar calendar, which varies from the Gregorian calendar by as much as a month forward or backward depending on the year. Celebrating the lunisolar birthday remains common on Taiwan while growing increasingly uncommon on the mainland. Birthday traditions reflected the culture's deep-seated focus on longevity and wordplay. From the homophony in some dialects between 酒 ("rice wine") and 久 (meaning "long" in the sense of time passing), osmanthus and other rice wines are traditional gifts for birthdays in China. Longevity noodles are another traditional food consumed on the day, although western-style birthday cakes are increasingly common among urban Chinese. Hongbaos—red envelopes stuffed with money, now especially the red 100 RMB notes—are the usual gift from relatives and close family friends for most children. Gifts for adults on their birthdays are much less common, although the birthday for each decade is a larger occasion that might prompt a large dinner and celebration. The Japanese reckoned their birthdays by the Chinese system until the Meiji Reforms. Celebrations remained uncommon or muted until after the American occupation that followed World War II.[citation needed] Children's birthday parties are the most important, typically celebrated with a cake, candles, and singing. Adults often just celebrate with their partner. In North Korea, the Day of the Sun, Kim Il Sung's birthday, is the most important public holiday of the country, and Kim Jong Il's birthday is celebrated as the Day of the Shining Star. North Koreans are not permitted to celebrate birthdays on July 8 and December 17 because these were the dates of the deaths of Kim Il Sung and Kim Jong Il, respectively. More than 100,000 North Koreans celebrate displaced birthdays on July 9 and December 18 instead to avoid these dates. A person born on July 8 before 1994 may change their birthday, with official recognition. South Korea was one of the last countries to use a form of East Asian age reckoning for many official purposes. Prior to June 2023, three systems were used together—"Korean ages" that start with 1 at birth and increase every January 1st with the Gregorian New Year, "year ages" that start with 0 at birth and otherwise increase the same way, and "actual ages" that start with 0 at birth and increase each birthday. First birthday celebrations was heavily celebrated, despite usually having little to do with the child's age. In June 2023, all Korean ages were set back at least one year, and official ages henceforth are reckoned only by birthdays. In Ghana, children wake up on their birthday to a special treat called oto, which is a patty made from mashed sweet potato and eggs fried in palm oil. Later they have a birthday party where they usually eat stew and rice and a dish known as kelewele, which is fried plantain chunks. Distribution through the year Birthdays are fairly evenly distributed throughout the year, with some seasonal effects. In the United States, there tend to be more births in September and October. This may be because there is a holiday season nine months before (the human gestation period is about nine months), or because the longest nights of the year also occur in the Northern Hemisphere nine months before. However, the holidays affect birth rates more than the winter: New Zealand, a Southern Hemisphere country, has the same September and October peak with no corresponding peak in March and April. The least common birthdays tend to fall around public holidays, such as Christmas, New Year's Day and fixed-date holidays such as Independence Day in the US, which falls on July 4. Between 1973 and 1999, September 16 was the most common birthday in the United States, and December 25 was the least common birthday (other than February 29 because of leap years). In 2011, October 5 and 6 were reported as the most frequently occurring birthdays. New Zealand's most common birthday is September 29, and the least common birthday is December 25. The ten most common birthdays all fall within a thirteen-day period, between September 22 and October 4. The ten least common birthdays (other than February 29) are December 24–27, January 1–2, February 6, March 22, April 1, and April 25. This is based on all live births registered in New Zealand between 1980 and 2017. Positive and negative associations with culturally significant dates may influence birth rates. The study shows a 5.3% decrease in spontaneous births and a 16.9% decrease in Caesarean births on Halloween, compared to dates occurring within one week before and one week after the October holiday. In contrast, on Valentine's Day, there is a 3.6% increase in spontaneous births and a 12.1% increase in Caesarean births. In Sweden, 9.3% of the population is born in March and 7.3% in November, when a uniform distribution would give 8.3%. In the Gregorian calendar (a common solar calendar), February in a leap year has 29 days instead of the usual 28, so the year lasts 366 days instead of the usual 365. A person born on February 29 may be called a "leapling" or a "leaper". In common years, they usually celebrate their birthdays on February 28. In some situations, March 1 is used as the birthday in a non-leap year since it is the day following February 28. Technically, a leapling will have fewer birthday anniversaries than their age in years. This phenomenon is exploited when a person claims to be only a quarter of their actual age, by counting their leap-year birthday anniversaries only. In Gilbert and Sullivan's 1879 comic opera The Pirates of Penzance, Frederic the pirate apprentice discovers that he is bound to serve the pirates until his 21st birthday rather than until his 21st year. For legal purposes, legal birthdays depend on how local laws count time intervals. An individual's Beddian birthday, named in tribute to firefighter Bobby Beddia, occurs during the year that their age matches the last two digits of the year they were born. Some studies show people are more likely to die on their birthdays, with explanations including excessive drinking, suicide, cardiovascular events due to high stress or happiness, efforts to postpone death for major social events, and death certificate paperwork errors. See also References Notes External links
========================================
[SOURCE: https://en.wikipedia.org/wiki/Languages_used_on_the_Internet] | [TOKENS: 722]
Contents Languages used on the Internet Slightly over half of the homepages of the most visited websites on the World Wide Web are in English, with varying amounts of information available in many other languages. Other top languages are Chinese, Spanish, Russian, Portuguese, French, German and Japanese. Of the more than 7,000 existing languages, only a few hundred are recognized as being in use for Web pages on the World Wide Web. Languages used There is debate over the most-used languages on the Internet. A 2009 UNESCO report monitoring the languages of websites for 12 years, from 1996 to 2008, found a steady year-on-year decline in the percentage of webpages in English, from 75 percent in 1998 to 45 percent in 2005. The authors found that English remained at 45 percent of content for 2005 to the end of the study but believe this was due to the bias of search engines indexing more English-language content rather than a true stabilization of the percentage of content in English on the World Wide Web. The number of non-English web pages is rapidly expanding. The use of English online increased by around 281 percent from 2001 to 2011, a lower rate of growth than that of Spanish (743 percent), Chinese (1,277 percent), Russian (1,826 percent) or Arabic (2,501 percent) over the same period. According to a 2000 study, the international auxiliary language Esperanto ranked 40 out of all languages in search engine queries, also ranking 27 out of all languages that rely on the Latin script. Usage statistics of content languages for websites As of 3 December 2025, the following table lists the forty most commonly used content languages among the top 10 million websites on the World Wide Web, according to estimates by W3Techs. All other languages are used in less than 0.1% of websites. Even including all languages, percentages may not sum to 100% because some websites contain multiple content languages. The figures from the W3Techs study are based on the one million most visited websites (i.e., approximately 0.27 percent of all websites according to December 2011 figures) as ranked by Alexa.com, and language is identified using only the home page of the sites in most cases (e.g., all of Wikipedia is based on the language detection of http://www.wikipedia.org). As a consequence, the figures show a significantly higher percentage for many languages (especially for English) as compared to the figures for all websites. For all websites, estimates are between 20 and 50% for English. Most used scripts on the Internet Content languages on YouTube Of the top 250 YouTube channels, 66% of the content is in English, 15% in Spanish, 7% in Portuguese, 5% in Hindi, and 2% in Korean, while other languages make up 5%, although other sources point to different percentages.[better source needed] YouTube is available in over 80 languages with more than a hundred different local versions. Of those popular YouTube channels that posted a video in the first week of 2019, just over half contained some content in a language other than English. Internet users by language InternetWorldStats estimates of the number of Internet users by language as of March 31, 2020: Wikipedia page views by language The Wikimedia Analytics API provides the most recent data on page views and page edits, among other statistics, for all language editions of Wikipedia. See also References External links
========================================
[SOURCE: https://en.wikipedia.org/wiki/Environmental_radioactivity] | [TOKENS: 3353]
Contents Environmental radioactivity Environmental radioactivity is part of the overall background radiation and is produced by radioactive materials in the human environment. While some radioisotopes, such as strontium-90 (90Sr) and technetium-99 (99Tc), are only found on Earth as a result of human activity, and some, like potassium-40 (40K), are only present due to natural processes, a few isotopes, such as tritium (3H), result from both natural processes and human activities. The concentration and location of some natural isotopes, particularly uranium-238 (238U), can be affected by human activity, such as nuclear weapons testing. Background level in soils Radioactivity is present everywhere, and has been since the formation of the Earth. Natural radioactivity detected in soil is predominantly due to the following four natural radioisotopes: 40K, 226Ra, 238U, and 232Th. In one kilogram of soil, the potassium-40 amounts to an average 370 Bq of radiation, with a typical range of 100–700 Bq; the others each contribute some 25 Bq, with typical ranges of 10–50 Bq (7–50 Bq for the 232Th). Some soils may vary greatly from these norms. A recent report on the Sava river in Serbia suggests that many of the river silts contain about 100 Bq kg−1 of natural radioisotopes (226Ra, 232Th, and 238U). According to the United Nations the normal concentration of uranium in soil ranges between 300 μg kg−1 and 11.7 mg kg−1. It is well known that some plants, called hyperaccumulators, are able to absorb and concentrate metals within their tissues; iodine was first isolated from seaweed in France, which suggests that seaweed is an iodine hyperaccumulator. Synthetic radioisotopes also can be detected in silt. Busby[citation needed] quotes a report on the plutonium activity in Welsh intertidal sediments by Garland et al. (1989), which suggests that the closer a site is to Sellafield, the higher is the concentration of plutonium in the silt. Some relationship between distance and activity can be seen in their data, when fitted to an exponential curve, but the scatter of the points is large (R2 = 0.3683). Man-made The additional radioactivity in the biosphere caused by human activity due to the releases of man-made radioactivity and of Naturally Occurring Radioactive Materials (NORM) can be divided into several classes. Just because a radioisotope lands on the surface of the soil, does not mean it will enter the human food chain. After release into the environment, radioactive materials can reach humans in a range of different routes, and the chemistry of the element usually dictates the most likely route. Jiří Hála claims in his textbook "Radioactivity, Ionizing Radiation and Nuclear Energy" that cattle only pass a minority of the strontium, caesium, plutonium and americium they ingest to the humans who consume milk and meat. Using milk as an example, if the cow has a daily intake of 1000 Bq of the preceding isotopes then the milk will have the following activities. Jiří Hála's textbook states that soils vary greatly in their ability to bind radioisotopes, the clay particles and humic acids can alter the distribution of the isotopes between the soil water and the soil. The distribution coefficient Kd is the ratio of the soil's radioactivity (Bq g−1) to that of the soil water (Bq ml−1). If the radioactivity is tightly bonded to by the minerals in the soil then less radioactivity can be absorbed by crops and grass growing in the soil. One dramatic source of man-made radioactivity is a nuclear weapons test. The glassy trinitite created by the first atom bomb contains radioisotopes formed by neutron activation and nuclear fission. In addition some natural radioisotopes are present. A recent paper reports the levels of long-lived radioisotopes in the trinitite. The trinitite was formed from feldspar and quartz which were melted by the heat. Two samples of trinitite were used, the first (left-hand-side bars in the graph) was taken from between 40 and 65 meters of ground zero while the other sample was taken from further away from the ground zero point. The 152Eu (half life 13.54 year) and 154Eu (half life 8.59 year) were mainly formed by the neutron activation of the europium in the soil, it is clear that the level of radioactivity for these isotopes is highest where the neutron dose to the soil was larger. Some of the 60Co (half life 5.27 year) is generated by activation of the cobalt in the soil, but some was also generated by the activation of the cobalt in the steel (100 foot) tower. This 60Co from the tower would have been scattered over the site reducing the difference in the soil levels. The 133Ba (half life 10.5 year) and 241Am (half life 432.6 year) are due to the neutron activation of barium and plutonium inside the bomb. The barium was present in the form of the nitrate in the chemical explosives used while the plutonium was the fissile fuel used. The 137Cs level is higher in the sample that was further away from the ground zero point – this is thought to be because the precursors to the 137Cs (137I and 137Xe) and, to a lesser degree, the caesium itself are volatile. The natural radioisotopes in the glass are about the same in both locations. The action of neutrons on stable isotopes can form radioisotopes, for instance the neutron bombardment (neutron activation) of nitrogen-14 forms carbon-14. This radioisotope can be released from the nuclear fuel cycle; this is the radioisotope responsible for the majority of the dose experienced by the population as a result of the activities of the nuclear power industry.[citation needed] Nuclear bomb tests have increased the specific activity of carbon, whereas the use of fossil fuels has decreased it. See the article on radiocarbon dating for further details. Discharges from nuclear plants within the nuclear fuel cycle introduce fission products to the environment. The releases from nuclear reprocessing plants tend to be medium to long-lived radioisotopes; this is because the nuclear fuel is allowed to cool for several years before being dissolved in the nitric acid. The releases from nuclear reactor accidents and bomb detonations will contain a greater amount of the short-lived radioisotopes (when the amounts are expressed in activity Bq)). An example of a short-lived fission product is iodine-131, this can also be formed as an activation product by the neutron activation of tellurium. In both bomb fallout and a release from a power reactor accident, the short-lived isotopes cause the dose rate on day one to be much higher than that which will be experienced at the same site many days later. This holds true even if no attempts at decontamination are made. In the graphs below, the total gamma dose rate and the share of the dose due to each main isotope released by the Chernobyl accident are shown. An example of a medium lived is 137Cs, which has a half-life of 30 years. Caesium is released in bomb fallout and from the nuclear fuel cycle. A paper has been written on the radioactivity in oysters found in the Irish Sea, these were found by gamma spectroscopy to contain 141Ce, 144Ce, 103Ru, 106Ru, 137Cs, 95Zr and 95Nb.[citation needed] In addition, a zinc activation product (65Zn) was found, this is thought to be due to the corrosion of magnox fuel cladding in cooling ponds. The concentration of all these isotopes in the Irish Sea attributable to nuclear facilities such as Sellafield has significantly decreased in recent decades. An important part of the Chernobyl release was the caesium-137, this isotope is responsible for much of the long term (at least one year after the fire) external exposure which has occurred at the site. The caesium isotopes in the fallout have had an effect on farming. A large amount of caesium was released during the Goiânia accident where a radioactive source (made for medical use) was stolen and then smashed open during an attempt to convert it into scrap metal. The accident could have been stopped at several stages; first, the last legal owners of the source failed to make arrangements for the source to be stored in a safe and secure place; and second, the scrap metal workers who took it did not recognise the markings which indicated that it was a radioactive object. Soudek et al. reported in 2006 details of the uptake of 90Sr and 137Cs into sunflowers grown under hydroponic conditions. The caesium was found in the leaf veins, in the stem and in the apical leaves. It was found that 12% of the caesium entered the plant, and 20% of the strontium. This paper also reports details of the effect of potassium, ammonium and calcium ions on the uptake of the radioisotopes. Caesium binds tightly to clay minerals such as illite and montmorillonite; hence it remains in the upper layers of soil where it can be accessed by plants with shallow roots (such as grass). Hence grass and mushrooms can carry a considerable amount of 137Cs which can be transferred to humans through the food chain. One of the best countermeasures in dairy farming against 137Cs is to mix up the soil by deeply ploughing the soil. This has the effect of putting the 137Cs out of reach of the shallow roots of the grass, hence the level of radioactivity in the grass will be lowered. Also, after a nuclear war or serious accident, the removal of top few cm of soil and its burial in a shallow trench will reduce the long term gamma dose to humans due to 137Cs as the gamma photons will be attenuated by their passage through the soil. The more remote the trench is from humans and the deeper the trench is the better the degree of protection which will be afforded to the human population. In livestock farming, an important countermeasure against 137Cs is to feed to animals a little prussian blue. This iron potassium cyanide compound acts as an ion-exchanger. The cyanide is so tightly bonded to the iron that it is safe for a human to eat several grams of prussian blue per day. The prussian blue reduces the biological half-life (not to be confused with the nuclear half-life) of the caesium). The physical or nuclear half-life of 137Cs is about 30 years, which is a constant and can not be changed; however, the biological half-life will change according to the nature and habits of the organism for which it is expressed. Caesium in humans normally has a biological half-life of between one and four months. An added advantage of the prussian blue is that the caesium which is stripped from the animal in the droppings is in a form which is not available to plants. Hence, it prevents the caesium from being recycled. The form of prussian blue required for the treatment of humans or animals is a special grade. Attempts to use the pigment grade used in paints have not been successful. Examples of long-lived isotopes include iodine-129 and Tc-99, which have nuclear half-lives of 15 million and 200,000 years, respectively. In popular culture, plutonium is credited with being the ultimate threat to life and limb which is wrong; while ingesting plutonium is not likely to be good for one's health, other radioisotopes such as radium are more toxic to humans. Regardless, the introduction of the transuranium elements such as plutonium into the environment should be avoided wherever possible. Currently, the activities of the nuclear reprocessing industry have been subject to great debate as one of the fears of those opposed to the industry is that large amounts of plutonium will be either mismanaged or released into the environment. In the past, one of the largest releases of plutonium into the environment has been nuclear bomb testing. Natural Cosmogenic isotopes (or cosmogenic nuclides) are rare isotopes created when a high-energy cosmic ray interacts with the nucleus of an in situ atom. These isotopes are produced within earth materials such as rocks or soil, in Earth's atmosphere, and in extraterrestrial items such as meteorites. By measuring cosmogenic isotopes, scientists are able to gain insight into a range of geological and astronomical processes. There are both radioactive and stable cosmogenic isotopes. Some of these radioisotopes are tritium, carbon-14 and phosphorus-32. Here is a list of radioisotopes formed by the action of cosmic rays on the atmosphere; the list also contains the production mode of the isotope. These data were obtained from the SCOPE50 report, see table 1.9 of chapter 1. The level of beryllium-7 in the air is related to the Sun spot cycle, as radiation from the Sun forms this radioisotope in the atmosphere. The rate at which it is transferred from the air to the ground is controlled in part by the weather. Because cosmogenic isotopes have long half-lives (anywhere from thousands to millions of years), scientists find them useful for geologic dating. Cosmogenic isotopes are produced at or near the surface of the Earth, and thus are commonly applied to problems of measuring ages and rates of geomorphic and sedimentary events and processes. Specific applications of cosmogenic isotopes include: To measure cosmogenic isotopes produced within solid earth materials, such as rock, samples are generally first put through a process of mechanical separation. The sample is crushed and desirable material, such as a particular mineral (quartz in the case of Be-10), is separated from non-desirable material by using a density separation in a heavy liquid medium such as lithium sodium tungstate (LST). The sample is then dissolved, a common isotope carrier added (Be-9 carrier in the case of Be-10), and the aqueous solution is purified down to an oxide or other pure solid. Finally, the ratio of the rare cosmogenic isotope to the common isotope is measured using accelerator mass spectrometry. The original concentration of cosmogenic isotope in the sample is then calculated using the measured isotopic ratio, the mass of the sample, and the mass of carrier added to the sample. Radium and radon are in the environment because they are decay products of uranium and thorium. The radon (222Rn) released into the air decays to 210Pb and other radioisotopes, and the levels of 210Pb can be measured. The rate of deposition of this radioisotope is dependent on the weather. Below is a graph of the deposition rate observed in Japan. Uranium–lead dating is usually performed on the mineral zircon (ZrSiO4), though other materials can be used. Zircon incorporates uranium atoms into its crystalline structure as substitutes for zirconium, but strongly rejects lead. It has a high blocking temperature, is resistant to mechanical weathering and is chemically inert. Zircon also forms multiple crystal layers during metamorphic events, which each may record an isotopic age of the event. These can be dated by a SHRIMP ion microprobe. One of the advantages of this method is that any sample provides two clocks, one based on uranium-235's decay to lead-207 with a half-life of about 703 million years, and one based on uranium-238's decay to lead-206 with a half-life of about 4.5 billion years, providing a built-in crosscheck that allows accurate determination of the age of the sample even if some of the lead has been lost. See also References Further reading External links
========================================
[SOURCE: https://en.wikipedia.org/wiki/PlayStation_(console)#cite_note-FOOTNOTEEdge_staff1995b73-103] | [TOKENS: 10728]
Contents PlayStation (console) The PlayStation[a] (codenamed PSX, abbreviated as PS, and retroactively PS1 or PS one) is a home video game console developed and marketed by Sony Computer Entertainment. It was released in Japan on 3 December 1994, followed by North America on 9 September 1995, Europe on 29 September 1995, and other regions following thereafter. As a fifth-generation console, the PlayStation primarily competed with the Nintendo 64 and the Sega Saturn. Sony began developing the PlayStation after a failed venture with Nintendo to create a CD-ROM peripheral for the Super Nintendo Entertainment System in the early 1990s. The console was primarily designed by Ken Kutaragi and Sony Computer Entertainment in Japan, while additional development was outsourced in the United Kingdom. An emphasis on 3D polygon graphics was placed at the forefront of the console's design. PlayStation game production was designed to be streamlined and inclusive, enticing the support of many third party developers. The console proved popular for its extensive game library, popular franchises, low retail price, and aggressive youth marketing which advertised it as the preferable console for adolescents and adults. Critically acclaimed games that defined the console include Gran Turismo, Crash Bandicoot, Spyro the Dragon, Tomb Raider, Resident Evil, Metal Gear Solid, Tekken 3, and Final Fantasy VII. Sony ceased production of the PlayStation on 23 March 2006—over eleven years after it had been released, and in the same year the PlayStation 3 debuted. More than 4,000 PlayStation games were released, with cumulative sales of 962 million units. The PlayStation signaled Sony's rise to power in the video game industry. It received acclaim and sold strongly; in less than a decade, it became the first computer entertainment platform to ship over 100 million units. Its use of compact discs heralded the game industry's transition from cartridges. The PlayStation's success led to a line of successors, beginning with the PlayStation 2 in 2000. In the same year, Sony released a smaller and cheaper model, the PS one. History The PlayStation was conceived by Ken Kutaragi, a Sony executive who managed a hardware engineering division and was later dubbed "the Father of the PlayStation". Kutaragi's interest in working with video games stemmed from seeing his daughter play games on Nintendo's Famicom. Kutaragi convinced Nintendo to use his SPC-700 sound processor in the Super Nintendo Entertainment System (SNES) through a demonstration of the processor's capabilities. His willingness to work with Nintendo was derived from both his admiration of the Famicom and conviction in video game consoles becoming the main home-use entertainment systems. Although Kutaragi was nearly fired because he worked with Nintendo without Sony's knowledge, president Norio Ohga recognised the potential in Kutaragi's chip and decided to keep him as a protégé. The inception of the PlayStation dates back to a 1988 joint venture between Nintendo and Sony. Nintendo had produced floppy disk technology to complement cartridges in the form of the Family Computer Disk System, and wanted to continue this complementary storage strategy for the SNES. Since Sony was already contracted to produce the SPC-700 sound processor for the SNES, Nintendo contracted Sony to develop a CD-ROM add-on, tentatively titled the "Play Station" or "SNES-CD". The PlayStation name had already been trademarked by Yamaha, but Nobuyuki Idei liked it so much that he agreed to acquire it for an undisclosed sum rather than search for an alternative. Sony was keen to obtain a foothold in the rapidly expanding video game market. Having been the primary manufacturer of the MSX home computer format, Sony had wanted to use their experience in consumer electronics to produce their own video game hardware. Although the initial agreement between Nintendo and Sony was about producing a CD-ROM drive add-on, Sony had also planned to develop a SNES-compatible Sony-branded console. This iteration was intended to be more of a home entertainment system, playing both SNES cartridges and a new CD format named the "Super Disc", which Sony would design. Under the agreement, Sony would retain sole international rights to every Super Disc game, giving them a large degree of control despite Nintendo's leading position in the video game market. Furthermore, Sony would also be the sole benefactor of licensing related to music and film software that it had been aggressively pursuing as a secondary application. The Play Station was to be announced at the 1991 Consumer Electronics Show (CES) in Las Vegas. However, Nintendo president Hiroshi Yamauchi was wary of Sony's increasing leverage at this point and deemed the original 1988 contract unacceptable upon realising it essentially handed Sony control over all games written on the SNES CD-ROM format. Although Nintendo was dominant in the video game market, Sony possessed a superior research and development department. Wanting to protect Nintendo's existing licensing structure, Yamauchi cancelled all plans for the joint Nintendo–Sony SNES CD attachment without telling Sony. He sent Nintendo of America president Minoru Arakawa (his son-in-law) and chairman Howard Lincoln to Amsterdam to form a more favourable contract with Dutch conglomerate Philips, Sony's rival. This contract would give Nintendo total control over their licences on all Philips-produced machines. Kutaragi and Nobuyuki Idei, Sony's director of public relations at the time, learned of Nintendo's actions two days before the CES was due to begin. Kutaragi telephoned numerous contacts, including Philips, to no avail. On the first day of the CES, Sony announced their partnership with Nintendo and their new console, the Play Station. At 9 am on the next day, in what has been called "the greatest ever betrayal" in the industry, Howard Lincoln stepped onto the stage and revealed that Nintendo was now allied with Philips and would abandon their work with Sony. Incensed by Nintendo's renouncement, Ohga and Kutaragi decided that Sony would develop their own console. Nintendo's contract-breaking was met with consternation in the Japanese business community, as they had broken an "unwritten law" of native companies not turning against each other in favour of foreign ones. Sony's American branch considered allying with Sega to produce a CD-ROM-based machine called the Sega Multimedia Entertainment System, but the Sega board of directors in Tokyo vetoed the idea when Sega of America CEO Tom Kalinske presented them the proposal. Kalinske recalled them saying: "That's a stupid idea, Sony doesn't know how to make hardware. They don't know how to make software either. Why would we want to do this?" Sony halted their research, but decided to develop what it had developed with Nintendo and Sega into a console based on the SNES. Despite the tumultuous events at the 1991 CES, negotiations between Nintendo and Sony were still ongoing. A deal was proposed: the Play Station would still have a port for SNES games, on the condition that it would still use Kutaragi's audio chip and that Nintendo would own the rights and receive the bulk of the profits. Roughly two hundred prototype machines were created, and some software entered development. Many within Sony were still opposed to their involvement in the video game industry, with some resenting Kutaragi for jeopardising the company. Kutaragi remained adamant that Sony not retreat from the growing industry and that a deal with Nintendo would never work. Knowing that they had to take decisive action, Sony severed all ties with Nintendo on 4 May 1992. To determine the fate of the PlayStation project, Ohga chaired a meeting in June 1992, consisting of Kutaragi and several senior Sony board members. Kutaragi unveiled a proprietary CD-ROM-based system he had been secretly working on which played games with immersive 3D graphics. Kutaragi was confident that his LSI chip could accommodate one million logic gates, which exceeded the capabilities of Sony's semiconductor division at the time. Despite gaining Ohga's enthusiasm, there remained opposition from a majority present at the meeting. Older Sony executives also opposed it, who saw Nintendo and Sega as "toy" manufacturers. The opposers felt the game industry was too culturally offbeat and asserted that Sony should remain a central player in the audiovisual industry, where companies were familiar with one another and could conduct "civili[s]ed" business negotiations. After Kutaragi reminded him of the humiliation he suffered from Nintendo, Ohga retained the project and became one of Kutaragi's most staunch supporters. Ohga shifted Kutaragi and nine of his team from Sony's main headquarters to Sony Music Entertainment Japan (SMEJ), a subsidiary of the main Sony group, so as to retain the project and maintain relationships with Philips for the MMCD development project. The involvement of SMEJ proved crucial to the PlayStation's early development as the process of manufacturing games on CD-ROM format was similar to that used for audio CDs, with which Sony's music division had considerable experience. While at SMEJ, Kutaragi worked with Epic/Sony Records founder Shigeo Maruyama and Akira Sato; both later became vice-presidents of the division that ran the PlayStation business. Sony Computer Entertainment (SCE) was jointly established by Sony and SMEJ to handle the company's ventures into the video game industry. On 27 October 1993, Sony publicly announced that it was entering the game console market with the PlayStation. According to Maruyama, there was uncertainty over whether the console should primarily focus on 2D, sprite-based graphics or 3D polygon graphics. After Sony witnessed the success of Sega's Virtua Fighter (1993) in Japanese arcades, the direction of the PlayStation became "instantly clear" and 3D polygon graphics became the console's primary focus. SCE president Teruhisa Tokunaka expressed gratitude for Sega's timely release of Virtua Fighter as it proved "just at the right time" that making games with 3D imagery was possible. Maruyama claimed that Sony further wanted to emphasise the new console's ability to utilise redbook audio from the CD-ROM format in its games alongside high quality visuals and gameplay. Wishing to distance the project from the failed enterprise with Nintendo, Sony initially branded the PlayStation the "PlayStation X" (PSX). Sony formed their European division and North American division, known as Sony Computer Entertainment Europe (SCEE) and Sony Computer Entertainment America (SCEA), in January and May 1995. The divisions planned to market the new console under the alternative branding "PSX" following the negative feedback regarding "PlayStation" in focus group studies. Early advertising prior to the console's launch in North America referenced PSX, but the term was scrapped before launch. The console was not marketed with Sony's name in contrast to Nintendo's consoles. According to Phil Harrison, much of Sony's upper management feared that the Sony brand would be tarnished if associated with the console, which they considered a "toy". Since Sony had no experience in game development, it had to rely on the support of third-party game developers. This was in contrast to Sega and Nintendo, which had versatile and well-equipped in-house software divisions for their arcade games and could easily port successful games to their home consoles. Recent consoles like the Atari Jaguar and 3DO suffered low sales due to a lack of developer support, prompting Sony to redouble their efforts in gaining the endorsement of arcade-savvy developers. A team from Epic Sony visited more than a hundred companies throughout Japan in May 1993 in hopes of attracting game creators with the PlayStation's technological appeal. Sony found that many disliked Nintendo's practices, such as favouring their own games over others. Through a series of negotiations, Sony acquired initial support from Namco, Konami, and Williams Entertainment, as well as 250 other development teams in Japan alone. Namco in particular was interested in developing for PlayStation since Namco rivalled Sega in the arcade market. Attaining these companies secured influential games such as Ridge Racer (1993) and Mortal Kombat 3 (1995), Ridge Racer being one of the most popular arcade games at the time, and it was already confirmed behind closed doors that it would be the PlayStation's first game by December 1993, despite Namco being a longstanding Nintendo developer. Namco's research managing director Shegeichi Nakamura met with Kutaragi in 1993 to discuss the preliminary PlayStation specifications, with Namco subsequently basing the Namco System 11 arcade board on PlayStation hardware and developing Tekken to compete with Virtua Fighter. The System 11 launched in arcades several months before the PlayStation's release, with the arcade release of Tekken in September 1994. Despite securing the support of various Japanese studios, Sony had no developers of their own by the time the PlayStation was in development. This changed in 1993 when Sony acquired the Liverpudlian company Psygnosis (later renamed SCE Liverpool) for US$48 million, securing their first in-house development team. The acquisition meant that Sony could have more launch games ready for the PlayStation's release in Europe and North America. Ian Hetherington, Psygnosis' co-founder, was disappointed after receiving early builds of the PlayStation and recalled that the console "was not fit for purpose" until his team got involved with it. Hetherington frequently clashed with Sony executives over broader ideas; at one point it was suggested that a television with a built-in PlayStation be produced. In the months leading up to the PlayStation's launch, Psygnosis had around 500 full-time staff working on games and assisting with software development. The purchase of Psygnosis marked another turning point for the PlayStation as it played a vital role in creating the console's development kits. While Sony had provided MIPS R4000-based Sony NEWS workstations for PlayStation development, Psygnosis employees disliked the thought of developing on these expensive workstations and asked Bristol-based SN Systems to create an alternative PC-based development system. Andy Beveridge and Martin Day, owners of SN Systems, had previously supplied development hardware for other consoles such as the Mega Drive, Atari ST, and the SNES. When Psygnosis arranged an audience for SN Systems with Sony's Japanese executives at the January 1994 CES in Las Vegas, Beveridge and Day presented their prototype of the condensed development kit, which could run on an ordinary personal computer with two extension boards. Impressed, Sony decided to abandon their plans for a workstation-based development system in favour of SN Systems's, thus securing a cheaper and more efficient method for designing software. An order of over 600 systems followed, and SN Systems supplied Sony with additional software such as an assembler, linker, and a debugger. SN Systems produced development kits for future PlayStation systems, including the PlayStation 2 and was bought out by Sony in 2005. Sony strived to make game production as streamlined and inclusive as possible, in contrast to the relatively isolated approach of Sega and Nintendo. Phil Harrison, representative director of SCEE, believed that Sony's emphasis on developer assistance reduced most time-consuming aspects of development. As well as providing programming libraries, SCE headquarters in London, California, and Tokyo housed technical support teams that could work closely with third-party developers if needed. Sony did not favour their own over non-Sony products, unlike Nintendo; Peter Molyneux of Bullfrog Productions admired Sony's open-handed approach to software developers and lauded their decision to use PCs as a development platform, remarking that "[it was] like being released from jail in terms of the freedom you have". Another strategy that helped attract software developers was the PlayStation's use of the CD-ROM format instead of traditional cartridges. Nintendo cartridges were expensive to manufacture, and the company controlled all production, prioritising their own games, while inexpensive compact disc manufacturing occurred at dozens of locations around the world. The PlayStation's architecture and interconnectability with PCs was beneficial to many software developers. The use of the programming language C proved useful, as it safeguarded future compatibility of the machine should developers decide to make further hardware revisions. Despite the inherent flexibility, some developers found themselves restricted due to the console's lack of RAM. While working on beta builds of the PlayStation, Molyneux observed that its MIPS processor was not "quite as bullish" compared to that of a fast PC and said that it took his team two weeks to port their PC code to the PlayStation development kits and another fortnight to achieve a four-fold speed increase. An engineer from Ocean Software, one of Europe's largest game developers at the time, thought that allocating RAM was a challenging aspect given the 3.5 megabyte restriction. Kutaragi said that while it would have been easy to double the amount of RAM for the PlayStation, the development team refrained from doing so to keep the retail cost down. Kutaragi saw the biggest challenge in developing the system to be balancing the conflicting goals of high performance, low cost, and being easy to program for, and felt he and his team were successful in this regard. Its technical specifications were finalised in 1993 and its design during 1994. The PlayStation name and its final design were confirmed during a press conference on May 10, 1994, although the price and release dates had not been disclosed yet. Sony released the PlayStation in Japan on 3 December 1994, a week after the release of the Sega Saturn, at a price of ¥39,800. Sales in Japan began with a "stunning" success with long queues in shops. Ohga later recalled that he realised how important PlayStation had become for Sony when friends and relatives begged for consoles for their children. PlayStation sold 100,000 units on the first day and two million units within six months, although the Saturn outsold the PlayStation in the first few weeks due to the success of Virtua Fighter. By the end of 1994, 300,000 PlayStation units were sold in Japan compared to 500,000 Saturn units. A grey market emerged for PlayStations shipped from Japan to North America and Europe, with buyers of such consoles paying up to £700. "When September 1995 arrived and Sony's Playstation roared out of the gate, things immediately felt different than [sic] they did with the Saturn launch earlier that year. Sega dropped the Saturn $100 to match the Playstation's $299 debut price, but sales weren't even close—Playstations flew out the door as fast as we could get them in stock. Before the release in North America, Sega and Sony presented their consoles at the first Electronic Entertainment Expo (E3) in Los Angeles on 11 May 1995. At their keynote presentation, Sega of America CEO Tom Kalinske revealed that their Saturn console would be released immediately to select retailers at a price of $399. Next came Sony's turn: Olaf Olafsson, the head of SCEA, summoned Steve Race, the head of development, to the conference stage, who said "$299" and left the audience with a round of applause. The attention to the Sony conference was further bolstered by the surprise appearance of Michael Jackson and the showcase of highly anticipated games, including Wipeout (1995), Ridge Racer and Tekken (1994). In addition, Sony announced that no games would be bundled with the console. Although the Saturn had released early in the United States to gain an advantage over the PlayStation, the surprise launch upset many retailers who were not informed in time, harming sales. Some retailers such as KB Toys responded by dropping the Saturn entirely. The PlayStation went on sale in North America on 9 September 1995. It sold more units within two days than the Saturn had in five months, with almost all of the initial shipment of 100,000 units sold in advance and shops across the country running out of consoles and accessories. The well-received Ridge Racer contributed to the PlayStation's early success, — with some critics considering it superior to Sega's arcade counterpart Daytona USA (1994) — as did Battle Arena Toshinden (1995). There were over 100,000 pre-orders placed and 17 games available on the market by the time of the PlayStation's American launch, in comparison to the Saturn's six launch games. The PlayStation released in Europe on 29 September 1995 and in Australia on 15 November 1995. By November it had already outsold the Saturn by three to one in the United Kingdom, where Sony had allocated a £20 million marketing budget during the Christmas season compared to Sega's £4 million. Sony found early success in the United Kingdom by securing listings with independent shop owners as well as prominent High Street chains such as Comet and Argos. Within its first year, the PlayStation secured over 20% of the entire American video game market. From September to the end of 1995, sales in the United States amounted to 800,000 units, giving the PlayStation a commanding lead over the other fifth-generation consoles,[b] though the SNES and Mega Drive from the fourth generation still outsold it. Sony reported that the attach rate of sold games and consoles was four to one. To meet increasing demand, Sony chartered jumbo jets and ramped up production in Europe and North America. By early 1996, the PlayStation had grossed $2 billion (equivalent to $4.106 billion 2025) from worldwide hardware and software sales. By late 1996, sales in Europe totalled 2.2 million units, including 700,000 in the UK. Approximately 400 PlayStation games were in development, compared to around 200 games being developed for the Saturn and 60 for the Nintendo 64. In India, the PlayStation was launched in test market during 1999–2000 across Sony showrooms, selling 100 units. Sony finally launched the console (PS One model) countrywide on 24 January 2002 with the price of Rs 7,990 and 26 games available from start. PlayStation was also doing well in markets where it was never officially released. For example, in Brazil, due to the registration of the trademark by a third company, the console could not be released, which was why the market was taken over by the officially distributed Sega Saturn during the first period, but as the Sega console withdraws, PlayStation imports and large piracy increased. In another market, China, the most popular 32-bit console was Sega Saturn, but after leaving the market, PlayStation grown with a base of 300,000 users until January 2000, although Sony China did not have plans to release it. The PlayStation was backed by a successful marketing campaign, allowing Sony to gain an early foothold in Europe and North America. Initially, PlayStation demographics were skewed towards adults, but the audience broadened after the first price drop. While the Saturn was positioned towards 18- to 34-year-olds, the PlayStation was initially marketed exclusively towards teenagers. Executives from both Sony and Sega reasoned that because younger players typically looked up to older, more experienced players, advertising targeted at teens and adults would draw them in too. Additionally, Sony found that adults reacted best to advertising aimed at teenagers; Lee Clow surmised that people who started to grow into adulthood regressed and became "17 again" when they played video games. The console was marketed with advertising slogans stylised as "LIVE IN YUR WRLD. PLY IN URS" (Live in Your World. Play in Ours.) and "U R NOT E" (red E). The four geometric shapes were derived from the symbols for the four buttons on the controller. Clow thought that by invoking such provocative statements, gamers would respond to the contrary and say "'Bullshit. Let me show you how ready I am.'" As the console's appeal enlarged, Sony's marketing efforts broadened from their earlier focus on mature players to specifically target younger children as well. Shortly after the PlayStation's release in Europe, Sony tasked marketing manager Geoff Glendenning with assessing the desires of a new target audience. Sceptical over Nintendo and Sega's reliance on television campaigns, Glendenning theorised that young adults transitioning from fourth-generation consoles would feel neglected by marketing directed at children and teenagers. Recognising the influence early 1990s underground clubbing and rave culture had on young people, especially in the United Kingdom, Glendenning felt that the culture had become mainstream enough to help cultivate PlayStation's emerging identity. Sony partnered with prominent nightclub owners such as Ministry of Sound and festival promoters to organise dedicated PlayStation areas where demonstrations of select games could be tested. Sheffield-based graphic design studio The Designers Republic was contracted by Sony to produce promotional materials aimed at a fashionable, club-going audience. Psygnosis' Wipeout in particular became associated with nightclub culture as it was widely featured in venues. By 1997, there were 52 nightclubs in the United Kingdom with dedicated PlayStation rooms. Glendenning recalled that he had discreetly used at least £100,000 a year in slush fund money to invest in impromptu marketing. In 1996, Sony expanded their CD production facilities in the United States due to the high demand for PlayStation games, increasing their monthly output from 4 million discs to 6.5 million discs. This was necessary because PlayStation sales were running at twice the rate of Saturn sales, and its lead dramatically increased when both consoles dropped in price to $199 that year. The PlayStation also outsold the Saturn at a similar ratio in Europe during 1996, with 2.2 million consoles sold in the region by the end of the year. Sales figures for PlayStation hardware and software only increased following the launch of the Nintendo 64. Tokunaka speculated that the Nintendo 64 launch had actually helped PlayStation sales by raising public awareness of the gaming market through Nintendo's added marketing efforts. Despite this, the PlayStation took longer to achieve dominance in Japan. Tokunaka said that, even after the PlayStation and Saturn had been on the market for nearly two years, the competition between them was still "very close", and neither console had led in sales for any meaningful length of time. By 1998, Sega, encouraged by their declining market share and significant financial losses, launched the Dreamcast as a last-ditch attempt to stay in the industry. Although its launch was successful, the technically superior 128-bit console was unable to subdue Sony's dominance in the industry. Sony still held 60% of the overall video game market share in North America at the end of 1999. Sega's initial confidence in their new console was undermined when Japanese sales were lower than expected, with disgruntled Japanese consumers reportedly returning their Dreamcasts in exchange for PlayStation software. On 2 March 1999, Sony officially revealed details of the PlayStation 2, which Kutaragi announced would feature a graphics processor designed to push more raw polygons than any console in history, effectively rivalling most supercomputers. The PlayStation continued to sell strongly at the turn of the new millennium: in June 2000, Sony released the PSOne, a smaller, redesigned variant which went on to outsell all other consoles in that year, including the PlayStation 2. In 2005, PlayStation became the first console to ship 100 million units with the PlayStation 2 later achieving this faster than its predecessor. The combined successes of both PlayStation consoles led to Sega retiring the Dreamcast in 2001, and abandoning the console business entirely. The PlayStation was eventually discontinued on 23 March 2006—over eleven years after its release, and less than a year before the debut of the PlayStation 3. Hardware The main microprocessor is a R3000 CPU made by LSI Logic operating at a clock rate of 33.8688 MHz and 30 MIPS. This 32-bit CPU relies heavily on the "cop2" 3D and matrix math coprocessor on the same die to provide the necessary speed to render complex 3D graphics. The role of the separate GPU chip is to draw 2D polygons and apply shading and textures to them: the rasterisation stage of the graphics pipeline. Sony's custom 16-bit sound chip supports ADPCM sources with up to 24 sound channels and offers a sampling rate of up to 44.1 kHz and music sequencing. It features 2 MB of main RAM, with an additional 1 MB of video RAM. The PlayStation has a maximum colour depth of 16.7 million true colours with 32 levels of transparency and unlimited colour look-up tables. The PlayStation can output composite, S-Video or RGB video signals through its AV Multi connector (with older models also having RCA connectors for composite), displaying resolutions from 256×224 to 640×480 pixels. Different games can use different resolutions. Earlier models also had proprietary parallel and serial ports that could be used to connect accessories or multiple consoles together; these were later removed due to a lack of usage. The PlayStation uses a proprietary video compression unit, MDEC, which is integrated into the CPU and allows for the presentation of full motion video at a higher quality than other consoles of its generation. Unusual for the time, the PlayStation lacks a dedicated 2D graphics processor; 2D elements are instead calculated as polygons by the Geometry Transfer Engine (GTE) so that they can be processed and displayed on screen by the GPU. While running, the GPU can also generate a total of 4,000 sprites and 180,000 polygons per second, in addition to 360,000 per second flat-shaded. The PlayStation went through a number of variants during its production run. Externally, the most notable change was the gradual reduction in the number of external connectors from the rear of the unit. This started with the original Japanese launch units; the SCPH-1000, released on 3 December 1994, was the only model that had an S-Video port, as it was removed from the next model. Subsequent models saw a reduction in number of parallel ports, with the final version only retaining one serial port. Sony marketed a development kit for amateur developers known as the Net Yaroze (meaning "Let's do it together" in Japanese). It was launched in June 1996 in Japan, and following public interest, was released the next year in other countries. The Net Yaroze allowed hobbyists to create their own games and upload them via an online forum run by Sony. The console was only available to buy through an ordering service and with the necessary documentation and software to program PlayStation games and applications through C programming compilers. On 7 July 2000, Sony released the PS One (stylised as "PS one" or "PSone"), a smaller, redesigned version of the original PlayStation. It was the highest-selling console through the end of the year, outselling all other consoles—including the PlayStation 2. In 2002, Sony released a 5-inch (130 mm) LCD screen add-on for the PS One, referred to as the "Combo pack". It also included a car cigarette lighter adaptor adding an extra layer of portability. Production of the LCD "Combo Pack" ceased in 2004, when the popularity of the PlayStation began to wane in markets outside Japan. A total of 28.15 million PS One units had been sold by the time it was discontinued in March 2006. Three iterations of the PlayStation's controller were released over the console's lifespan. The first controller, the PlayStation controller, was released alongside the PlayStation in December 1994. It features four individual directional buttons (as opposed to a conventional D-pad), a pair of shoulder buttons on both sides, Start and Select buttons in the centre, and four face buttons consisting of simple geometric shapes: a green triangle, red circle, blue cross, and a pink square (, , , ). Rather than depicting traditionally used letters or numbers onto its buttons, the PlayStation controller established a trademark which would be incorporated heavily into the PlayStation brand. Teiyu Goto, the designer of the original PlayStation controller, said that the circle and cross represent "yes" and "no", respectively (though this layout is reversed in Western versions); the triangle symbolises a point of view and the square is equated to a sheet of paper to be used to access menus. The European and North American models of the original PlayStation controllers are roughly 10% larger than its Japanese variant, to account for the fact the average person in those regions has larger hands than the average Japanese person. Sony's first analogue gamepad, the PlayStation Analog Joystick (often erroneously referred to as the "Sony Flightstick"), was first released in Japan in April 1996. Featuring two parallel joysticks, it uses potentiometer technology previously used on consoles such as the Vectrex; instead of relying on binary eight-way switches, the controller detects minute angular changes through the entire range of motion. The stick also features a thumb-operated digital hat switch on the right joystick, corresponding to the traditional D-pad, and used for instances when simple digital movements were necessary. The Analog Joystick sold poorly in Japan due to its high cost and cumbersome size. The increasing popularity of 3D games prompted Sony to add analogue sticks to its controller design to give users more freedom over their movements in virtual 3D environments. The first official analogue controller, the Dual Analog Controller, was revealed to the public in a small glass booth at the 1996 PlayStation Expo in Japan, and released in April 1997 to coincide with the Japanese releases of analogue-capable games Tobal 2 and Bushido Blade. In addition to the two analogue sticks (which also introduced two new buttons mapped to clicking in the analogue sticks), the Dual Analog controller features an "Analog" button and LED beneath the "Start" and "Select" buttons which toggles analogue functionality on or off. The controller also features rumble support, though Sony decided that haptic feedback would be removed from all overseas iterations before the United States release. A Sony spokesman stated that the feature was removed for "manufacturing reasons", although rumours circulated that Nintendo had attempted to legally block the release of the controller outside Japan due to similarities with the Nintendo 64 controller's Rumble Pak. However, a Nintendo spokesman denied that Nintendo took legal action. Next Generation's Chris Charla theorised that Sony dropped vibration feedback to keep the price of the controller down. In November 1997, Sony introduced the DualShock controller. Its name derives from its use of two (dual) vibration motors (shock). Unlike its predecessor, its analogue sticks feature textured rubber grips, longer handles, slightly different shoulder buttons and has rumble feedback included as standard on all versions. The DualShock later replaced its predecessors as the default controller. Sony released a series of peripherals to add extra layers of functionality to the PlayStation. Such peripherals include memory cards, the PlayStation Mouse, the PlayStation Link Cable, the Multiplayer Adapter (a four-player multitap), the Memory Drive (a disk drive for 3.5-inch floppy disks), the GunCon (a light gun), and the Glasstron (a monoscopic head-mounted display). Released exclusively in Japan, the PocketStation is a memory card peripheral which acts as a miniature personal digital assistant. The device features a monochrome liquid crystal display (LCD), infrared communication capability, a real-time clock, built-in flash memory, and sound capability. Sharing similarities with the Dreamcast's VMU peripheral, the PocketStation was typically distributed with certain PlayStation games, enhancing them with added features. The PocketStation proved popular in Japan, selling over five million units. Sony planned to release the peripheral outside Japan but the release was cancelled, despite receiving promotion in Europe and North America. In addition to playing games, most PlayStation models are equipped to play CD-Audio. The Asian model SCPH-5903 can also play Video CDs. Like most CD players, the PlayStation can play songs in a programmed order, shuffle the playback order of the disc and repeat one song or the entire disc. Later PlayStation models use a music visualisation function called SoundScope. This function, as well as a memory card manager, is accessed by starting the console without either inserting a game or closing the CD tray, thereby accessing a graphical user interface (GUI) for the PlayStation BIOS. The GUI for the PS One and PlayStation differ depending on the firmware version: the original PlayStation GUI had a dark blue background with rainbow graffiti used as buttons, while the early PAL PlayStation and PS One GUI had a grey blocked background with two icons in the middle. PlayStation emulation is versatile and can be run on numerous modern devices. Bleem! was a commercial emulator which was released for IBM-compatible PCs and the Dreamcast in 1999. It was notable for being aggressively marketed during the PlayStation's lifetime, and was the centre of multiple controversial lawsuits filed by Sony. Bleem! was programmed in assembly language, which allowed it to emulate PlayStation games with improved visual fidelity, enhanced resolutions, and filtered textures that was not possible on original hardware. Sony sued Bleem! two days after its release, citing copyright infringement and accusing the company of engaging in unfair competition and patent infringement by allowing use of PlayStation BIOSs on a Sega console. Bleem! were subsequently forced to shut down in November 2001. Sony was aware that using CDs for game distribution could have left games vulnerable to piracy, due to the growing popularity of CD-R and optical disc drives with burning capability. To preclude illegal copying, a proprietary process for PlayStation disc manufacturing was developed that, in conjunction with an augmented optical drive in Tiger H/E assembly, prevented burned copies of games from booting on an unmodified console. Specifically, all genuine PlayStation discs were printed with a small section of deliberate irregular data, which the PlayStation's optical pick-up was capable of detecting and decoding. Consoles would not boot game discs without a specific wobble frequency contained in the data of the disc pregap sector (the same system was also used to encode discs' regional lockouts). This signal was within Red Book CD tolerances, so PlayStation discs' actual content could still be read by a conventional disc drive; however, the disc drive could not detect the wobble frequency (therefore duplicating the discs omitting it), since the laser pick-up system of any optical disc drive would interpret this wobble as an oscillation of the disc surface and compensate for it in the reading process. Early PlayStations, particularly early 1000 models, experience skipping full-motion video or physical "ticking" noises from the unit. The problems stem from poorly placed vents leading to overheating in some environments, causing the plastic mouldings inside the console to warp slightly and create knock-on effects with the laser assembly. The solution is to sit the console on a surface which dissipates heat efficiently in a well vented area or raise the unit up slightly from its resting surface. Sony representatives also recommended unplugging the PlayStation when it is not in use, as the system draws in a small amount of power (and therefore heat) even when turned off. The first batch of PlayStations use a KSM-440AAM laser unit, whose case and movable parts are all built out of plastic. Over time, the plastic lens sled rail wears out—usually unevenly—due to friction. The placement of the laser unit close to the power supply accelerates wear, due to the additional heat, which makes the plastic more vulnerable to friction. Eventually, one side of the lens sled will become so worn that the laser can tilt, no longer pointing directly at the CD; after this, games will no longer load due to data read errors. Sony fixed the problem by making the sled out of die-cast metal and placing the laser unit further away from the power supply on later PlayStation models. Due to an engineering oversight, the PlayStation does not produce a proper signal on several older models of televisions, causing the display to flicker or bounce around the screen. Sony decided not to change the console design, since only a small percentage of PlayStation owners used such televisions, and instead gave consumers the option of sending their PlayStation unit to a Sony service centre to have an official modchip installed, allowing play on older televisions. Game library The PlayStation featured a diverse game library which grew to appeal to all types of players. Critically acclaimed PlayStation games included Final Fantasy VII (1997), Crash Bandicoot (1996), Spyro the Dragon (1998), Metal Gear Solid (1998), all of which became established franchises. Final Fantasy VII is credited with allowing role-playing games to gain mass-market appeal outside Japan, and is considered one of the most influential and greatest video games ever made. The PlayStation's bestselling game is Gran Turismo (1997), which sold 10.85 million units. After the PlayStation's discontinuation in 2006, the cumulative software shipment was 962 million units. Following its 1994 launch in Japan, early games included Ridge Racer, Crime Crackers, King's Field, Motor Toon Grand Prix, Toh Shin Den (i.e. Battle Arena Toshinden), and Kileak: The Blood. The first two games available at its later North American launch were Jumping Flash! (1995) and Ridge Racer, with Jumping Flash! heralded as an ancestor for 3D graphics in console gaming. Wipeout, Air Combat, Twisted Metal, Warhawk and Destruction Derby were among the popular first-year games, and the first to be reissued as part of Sony's Greatest Hits or Platinum range. At the time of the PlayStation's first Christmas season, Psygnosis had produced around 70% of its launch catalogue; their breakthrough racing game Wipeout was acclaimed for its techno soundtrack and helped raise awareness of Britain's underground music community. Eidos Interactive's action-adventure game Tomb Raider contributed substantially to the success of the console in 1996, with its main protagonist Lara Croft becoming an early gaming icon and garnering unprecedented media promotion. Licensed tie-in video games of popular films were also prevalent; Argonaut Games' 2001 adaptation of Harry Potter and the Philosopher's Stone went on to sell over eight million copies late in the console's lifespan. Third-party developers committed largely to the console's wide-ranging game catalogue even after the launch of the PlayStation 2; some of the notable exclusives in this era include Harry Potter and the Philosopher's Stone, Fear Effect 2: Retro Helix, Syphon Filter 3, C-12: Final Resistance, Dance Dance Revolution Konamix and Digimon World 3.[c] Sony assisted with game reprints as late as 2008 with Metal Gear Solid: The Essential Collection, this being the last PlayStation game officially released and licensed by Sony. Initially, in the United States, PlayStation games were packaged in long cardboard boxes, similar to non-Japanese 3DO and Saturn games. Sony later switched to the jewel case format typically used for audio CDs and Japanese video games, as this format took up less retailer shelf space (which was at a premium due to the large number of PlayStation games being released), and focus testing showed that most consumers preferred this format. Reception The PlayStation was mostly well received upon release. Critics in the west generally welcomed the new console; the staff of Next Generation reviewed the PlayStation a few weeks after its North American launch, where they commented that, while the CPU is "fairly average", the supplementary custom hardware, such as the GPU and sound processor, is stunningly powerful. They praised the PlayStation's focus on 3D, and complemented the comfort of its controller and the convenience of its memory cards. Giving the system 41⁄2 out of 5 stars, they concluded, "To succeed in this extremely cut-throat market, you need a combination of great hardware, great games, and great marketing. Whether by skill, luck, or just deep pockets, Sony has scored three out of three in the first salvo of this war." Albert Kim from Entertainment Weekly praised the PlayStation as a technological marvel, rivalling that of Sega and Nintendo. Famicom Tsūshin scored the console a 19 out of 40, lower than the Saturn's 24 out of 40, in May 1995. In a 1997 year-end review, a team of five Electronic Gaming Monthly editors gave the PlayStation scores of 9.5, 8.5, 9.0, 9.0, and 9.5—for all five editors, the highest score they gave to any of the five consoles reviewed in the issue. They lauded the breadth and quality of the games library, saying it had vastly improved over previous years due to developers mastering the system's capabilities in addition to Sony revising their stance on 2D and role playing games. They also complimented the low price point of the games compared to the Nintendo 64's, and noted that it was the only console on the market that could be relied upon to deliver a solid stream of games for the coming year, primarily due to third party developers almost unanimously favouring it over its competitors. Legacy SCE was an upstart in the video game industry in late 1994, as the video game market in the early 1990s was dominated by Nintendo and Sega. Nintendo had been the clear leader in the industry since the introduction of the Nintendo Entertainment System in 1985 and the Nintendo 64 was initially expected to maintain this position. The PlayStation's target audience included the generation which was the first to grow up with mainstream video games, along with 18- to 29-year-olds who were not the primary focus of Nintendo. By the late 1990s, Sony became a highly regarded console brand due to the PlayStation, with a significant lead over second-place Nintendo, while Sega was relegated to a distant third. The PlayStation became the first "computer entertainment platform" to ship over 100 million units worldwide, with many critics attributing the console's success to third-party developers. It remains the sixth best-selling console of all time as of 2025[update], with a total of 102.49 million units sold. Around 7,900 individual games were published for the console during its 11-year life span, the second-most games ever produced for a console. Its success resulted in a significant financial boon for Sony as profits from their video game division contributed to 23%. Sony's next-generation PlayStation 2, which is backward compatible with the PlayStation's DualShock controller and games, was announced in 1999 and launched in 2000. The PlayStation's lead in installed base and developer support paved the way for the success of its successor, which overcame the earlier launch of the Sega's Dreamcast and then fended off competition from Microsoft's newcomer Xbox and Nintendo's GameCube. The PlayStation 2's immense success and failure of the Dreamcast were among the main factors which led to Sega abandoning the console market. To date, five PlayStation home consoles have been released, which have continued the same numbering scheme, as well as two portable systems. The PlayStation 3 also maintained backward compatibility with original PlayStation discs. Hundreds of PlayStation games have been digitally re-released on the PlayStation Portable, PlayStation 3, PlayStation Vita, PlayStation 4, and PlayStation 5. The PlayStation has often ranked among the best video game consoles. In 2018, Retro Gamer named it the third best console, crediting its sophisticated 3D capabilities as one of its key factors in gaining mass success, and lauding it as a "game-changer in every sense possible". In 2009, IGN ranked the PlayStation the seventh best console in their list, noting its appeal towards older audiences to be a crucial factor in propelling the video game industry, as well as its assistance in transitioning game industry to use the CD-ROM format. Keith Stuart from The Guardian likewise named it as the seventh best console in 2020, declaring that its success was so profound it "ruled the 1990s". In January 2025, Lorentio Brodesco announced the nsOne project, attempting to reverse engineer PlayStation's motherboard. Brodesco stated that "detailed documentation on the original motherboard was either incomplete or entirely unavailable". The project was successfully crowdfunded via Kickstarter. In June, Brodesco manufactured the first working motherboard, promising to bring a fully rooted version with multilayer routing as well as documentation and design files in the near future. The success of the PlayStation contributed to the demise of cartridge-based home consoles. While not the first system to use an optical disc format, it was the first highly successful one, and ended up going head-to-head with the proprietary cartridge-relying Nintendo 64,[d] which the industry had expected to use CDs like PlayStation. After the demise of the Sega Saturn, Nintendo was left as Sony's main competitor in Western markets. Nintendo chose not to use CDs for the Nintendo 64; they were likely concerned with the proprietary cartridge format's ability to help enforce copy protection, given their substantial reliance on licensing and exclusive games for their revenue. Besides their larger capacity, CD-ROMs could be produced in bulk quantities at a much faster rate than ROM cartridges, a week compared to two to three months. Further, the cost of production per unit was far cheaper, allowing Sony to offer games about 40% lower cost to the user compared to ROM cartridges while still making the same amount of net revenue. In Japan, Sony published fewer copies of a wide variety of games for the PlayStation as a risk-limiting step, a model that had been used by Sony Music for CD audio discs. The production flexibility of CD-ROMs meant that Sony could produce larger volumes of popular games to get onto the market quickly, something that could not be done with cartridges due to their manufacturing lead time. The lower production costs of CD-ROMs also allowed publishers an additional source of profit: budget-priced reissues of games which had already recouped their development costs. Tokunaka remarked in 1996: Choosing CD-ROM is one of the most important decisions that we made. As I'm sure you understand, PlayStation could just as easily have worked with masked ROM [cartridges]. The 3D engine and everything—the whole PlayStation format—is independent of the media. But for various reasons (including the economies for the consumer, the ease of the manufacturing, inventory control for the trade, and also the software publishers) we deduced that CD-ROM would be the best media for PlayStation. The increasing complexity of developing games pushed cartridges to their storage limits and gradually discouraged some third-party developers. Part of the CD format's appeal to publishers was that they could be produced at a significantly lower cost and offered more production flexibility to meet demand. As a result, some third-party developers switched to the PlayStation, including Square and Enix, whose Final Fantasy VII and Dragon Quest VII respectively had been planned for the Nintendo 64 (both companies later merged to form Square Enix). Other developers released fewer games for the Nintendo 64 (Konami, releasing only thirteen N64 games but over fifty on the PlayStation). Nintendo 64 game releases were less frequent than the PlayStation's, with many being developed by either Nintendo themselves or second-parties such as Rare. The PlayStation Classic is a dedicated video game console made by Sony Interactive Entertainment that emulates PlayStation games. It was announced in September 2018 at the Tokyo Game Show, and released on 3 December 2018, the 24th anniversary of the release of the original console. As a dedicated console, the PlayStation Classic features 20 pre-installed games; the games run off the open source emulator PCSX. The console is bundled with two replica wired PlayStation controllers (those without analogue sticks), an HDMI cable, and a USB-Type A cable. Internally, the console uses a MediaTek MT8167a Quad A35 system on a chip with four central processing cores clocked at @ 1.5 GHz and a Power VR GE8300 graphics processing unit. It includes 16 GB of eMMC flash storage and 1 Gigabyte of DDR3 SDRAM. The PlayStation Classic is 45% smaller than the original console. The PlayStation Classic received negative reviews from critics and was compared unfavorably to Nintendo's rival Nintendo Entertainment System Classic Edition and Super Nintendo Entertainment System Classic Edition. Criticism was directed at its meagre game library, user interface, emulation quality, use of PAL versions for certain games, use of the original controller, and high retail price, though the console's design received praise. The console sold poorly. See also Notes References
========================================
[SOURCE: https://en.wikipedia.org/wiki/Tesla_Roadster_(first_generation)] | [TOKENS: 9592]
Contents Tesla Roadster (first generation) The first generation Tesla Roadster is a battery electric sports car based on the Lotus Elise chassis, and was produced by Tesla Motors (now Tesla, Inc.) from 2008 to 2012. The Roadster was the first highway legal, serial production, all-electric car to use lithium-ion battery cells, and the first production all-electric car with a range of more than 200 miles (320 km) per charge. Tesla sold about 2,450 Roadsters in over 30 countries, and most of the last Roadsters were sold in Europe and Asia during the fourth quarter of 2012. Tesla produced right-hand-drive Roadsters from early 2010. The Roadster qualified for government incentives in several nations. According to the U.S. EPA, the Roadster can travel 244 miles (393 km) on a single charge of its lithium-ion battery pack. The vehicle can accelerate from 0 to 60 mph (0 to 97 km/h) in 3.7 or 3.9 seconds depending on the model. It has a top speed of 125 mph (201 km/h). The Roadster's efficiency, as of September 2008[update], was reported as 120 miles per gallon gasoline equivalent (28 kW⋅h/100 mi) (2.0 L/100 km). It uses 21.7 kWh/100 mi (135 Wh/km) battery-to-wheel, and has an efficiency of 88% on average. History Prototypes of the car were officially revealed to the public on July 19, 2006, in Santa Monica, California, at a 350-person invitation-only event held in Barker Hangar at Santa Monica Airport. It was featured in Time in December 2006 as the recipient of the magazine's "Best Inventions 2006—Transportation Invention" award. The first Tesla Roadster was delivered in February 2008 to Tesla early investor, chairman and product architect Elon Musk. The company produced 500 similar vehicles through June 2009. In July 2009, Tesla began production of its 2010 model-year Roadster—the first major product upgrade. Simultaneously, Tesla began producing the Roadster Sport, the first derivative of Tesla's proprietary, patented powertrain. The car accelerates from 0 to 60 mph (0 to 97 km/h) in 3.7 seconds, compared to 3.9 seconds for the standard Roadster. Changes for the 2010 model-year cars included: All of these features, except for the motor, were available either as standard or as add-on option for the non-sport model. Beginning mid-March 2010, Tesla, in an effort to show off the practicality of its electric cars, sent one of its Roadsters around the world. Starting at the Geneva International Motor Show, the Roadster completed its journey upon its arrival in Paris on September 28, 2010. In July 2010, Tesla introduced the "Roadster 2.5" update including: Tesla produced the Roadster until January 2012, when its supply of gliders ran out, as its contract with Lotus Cars for 2,500 gliders expired at the end of 2011. Tesla stopped taking orders for the Roadster in the U.S. market in August 2011. Featuring new options and enhanced components, the 2012 Tesla Roadster was sold in limited numbers only in Europe, Asia, and Australia. Tesla's U.S. exemption for not having advanced (two-stage) passenger airbags expired for cars made after the end of 2011 so the last Roadsters could not be sold in the American market. Fifteen Final Edition Roadsters were produced to close the manufacturing cycle of Tesla's first electric car. For a time, Tesla offered an optional upgrade to existing Roadsters, the Roadster 3.0. It offered a new battery pack with cells from LG Chem increasing capacity by 50% to 70 kWh, a new aero kit designed to reduce drag, and new tires with lower rolling resistance. The upgrade was offered between September 2015 and late 2016 at a cost of US$29,000 (equivalent to $38,904 in 2025). In November 2023, Tesla open-sourced some of the Roadster's design and engineering documents, as well as diagnostic software. Development After Martin Eberhard sold NuvoMedia to TV Guide, he wanted a sports car, but could not find one to his liking. His battery experience with the Rocket eBook inspired him to develop an electric car. During his search, Eberhard test drove the tzero, a concept car from the small automaker AC Propulsion. Eberhard and Marc Tarpenning, who had also driven the tzero, tried to convince the company to put the car into production, but when they declined, they decided to establish Tesla Motors in Delaware on July 1, 2003, to pursue the idea commercially. South African-born entrepreneur Elon Musk would also test drive a tzero and encouraged AC Propulsion to put the car into production, instead the company connected Musk with Eberhard and Tarpenning. Musk took an active role within the company starting in 2004, including investing US$7.5 million (~$11.9 million in 2024), overseeing Roadster product design from the beginning, and greatly expanding Tesla's long-term strategic sales goals by using the sports car to fund the development of mainstream vehicles. Musk became Tesla's chairman of the board in April 2004 and helped recruit J. B. Straubel as chief technology officer in May 2004. Musk received the Global Green 2006 product design award for the design of the Tesla Roadster, presented by Mikhail Gorbachev, and he received the 2007 Index Design award for the design of the Tesla Roadster. Before Tesla had developed the Roadster's proprietary powertrain, they borrowed a tzero for use as a development mule and converted the vehicle from lead–acid AGM batteries to lithium-ion cells, which substantially increased the range, reduced weight, and boosted 0 to 60 mph performance. Tesla licensed AC Propulsion's EV power system design and reductive charging patent, which covers integration of the charging electronics with the inverter, thus reducing mass, complexity, and cost. Tesla, however, was dissatisfied with how the motor and transmission worked in the chassis. Tesla then designed and built its own power electronics, motor, and other drivetrain components that incorporated this licensed technology from AC Propulsion. Given the extensive redevelopment of the vehicle, Tesla Motors no longer licenses any proprietary technology from AC Propulsion. The Roadster's powertrain is unique. On July 11, 2005, Tesla and British sports car maker Lotus entered an agreement about products and services based on the Lotus Elise, where Lotus provided advice on designing and developing a vehicle as well as producing partly assembled vehicles, and amended in 2009, helped with basic chassis development. The Roadster has a parts overlap of roughly 6% with the Lotus Elise, a 2-inch-longer wheelbase, and a slightly stiffer chassis according to Eberhard. Tesla's designers chose to construct the body panels using resin transfer molded carbon fiber composite to minimize weight; this choice makes the Roadster one of the least expensive cars with an entirely carbon fiber skin. Several prototypes of the Tesla Roadster were produced from 2004 through 2007. Initial studies were done in two development mule vehicles based on Lotus Elises equipped with all-electric drive systems. Tesla then built and tested ten engineering prototypes (EP1 through EP10) in late 2006 and early 2007, which led to many minor changes. Next, Tesla produced at least 26 validation prototypes, which were delivered beginning in March 2007. These final revisions were endurance and crash tested in preparation for series production.[citation needed] In August 2007, Martin Eberhard was replaced by an interim CEO, Michael Marks. Marks accepted the temporary position while a recruitment was undertaken. In December 2007, Ze'ev Drori became the CEO and president of Tesla. In October 2008, Musk succeeded Drori as CEO. Drori left the company in December. In January 2008, the U.S. National Highway Traffic Safety Administration (NHTSA) announced that it would grant Tesla a waiver of the advanced (two-stage) air bag rule noting that the Roadster includes standard air bags. Similar waivers were granted to other small volume manufacturers, including Lotus, Ferrari, and Bugatti. Tesla delivered its first production car in February 2008 to Musk. Tesla announced in early August 2009 that Roadster sales had resulted in overall corporate profitability for the month of July 2009, earning US$1 million on revenue of US$20 million. Tesla, which signed a production contract with Lotus in 2007 to produce "gliders" (complete cars minus electric powertrain) for the Roadster, announced in early 2010 that Roadster production would continue until early 2012. Starting one year prior to the end of the contract, no changes to the order was allowed to give time for tooling changes at Lotus's assembly plant in the UK. Several years later in 2018, Musk would go on to say that using the Lotus Elise as a base for the Roadster was a poor strategy because the Elise was incompatible with the intended AC Propulsion technology and was modified so extensively only 7% of the Elise remained in common with the final production Roadster. Production Tesla's cumulative production of the Roadster reached 1,000 cars in January 2010. The Roadster is considered an American car though many carry a Vehicle Identification Number beginning with the letter "S", which is the designation for the United Kingdom. Some, however, carry a number starting with "5" appropriate to the US. Parts were sourced from around the world. The body panels came from French supplier Sotira. These were sent from France to Hethel, U.K., where Tesla contracted with Lotus to build the Roadster's unique chassis. The Roadster shares roughly 7% of its components with the Lotus Elise including the windshield, airbags, some dashboard parts, and suspension components. The Roadster's single-speed gearbox was made in Detroit by BorgWarner. Brakes and airbags were made by Siemens in Germany, and some crash testing was conducted at Siemens as well. 30–40% of components were sourced from Taiwan. For Roadsters bound for customers in North America, the glider was sent to Tesla's facility in Menlo Park, California for final assembly, and for Roadsters bound for customers in Europe or elsewhere outside of North America, the glider was sent to a facility at Wymondham near Hethel for final assembly. At these locations, Tesla employees installed the powertrain, which consisted of the battery pack, power electronics module, gearbox and motor. Tesla ordered 2,500 gliders from Lotus, which ceased production in December 2011 when their contract expired. Tesla ended production of the Roadster in January 2012. Subsequent to completion of the first production car, the company announced problems with transmission reliability. The development transmission, with first gear enabled to accelerate 0 to 60 mph (0 to 97 km/h) in 4 seconds, was reported to have a life expectancy of as low as only a few thousand miles. Tesla's first two transmission suppliers were unable to produce transmissions, in quantity, that could withstand the gear-shift requirements of the high torque, high rpm electric motor. In December 2007, Tesla announced plans to ship the initial Roadsters with the transmissions locked into second gear, providing 0 to 60 mph (0 to 97 km/h) acceleration in 5.7 seconds and allowing customers to swap out transmissions under warranty when the finalized transmission, power electronics module (PEM), and cooling system became available. The EPA range of the car was also restated downward from 245 to 221 mi (394 to 356 km). The downward revision was attributed to an error in equipment calibration at the laboratory that conducted the original test. Tesla produced a special edition of 15 Final Edition Roadsters to close the production cycle of the electric car. The 15 special-edition cars were sold in each of the three sales regions, North America, Europe and Asia, and five units were allocated to each. The Final Edition Roadster did not have any performance modifications, but featured sporting atomic red paint, a duo of dark silver stripes on its hood and rear clamshell, and exclusive anthracite aluminum wheels. Specifications The Roadster is powered by a 3-phase, 4-pole, induction electric motor with a maximum output power of 248 hp (185 kW). Its maximum torque of 200 lb⋅ft (270 N⋅m) is immediately available and remains constant from 0 to 6,000 rpm; nearly instantaneous torque is a characteristic of electric motors and offers one of the biggest performance differences from internal combustion engines. The motor is air-cooled and does not need a liquid cooling system. The Sport model introduced during the Jan 2009 Detroit Auto Show includes a motor with a higher density, hand-wound stator that produces a maximum of 288 hp (215 kW). Both motors are designed for rotational speeds of up to 14,000 rpm, and the regular motor delivers a typical efficiency of 88% or 90%; 80% at peak power. It weighs less than 70 pounds (32 kg). Starting in September 2008 Tesla selected BorgWarner to manufacture gearboxes and began equipping all Roadsters with a single speed, fixed gear gearbox (8.2752:1 ratio) with an electrically actuated parking pawl mechanism and a mechanical lubrication pump. The company previously worked with several companies, including XTrac and Magna International, to find the right automatic transmission, but a two-gear solution proved to be too challenging. This led to substantial delays in production. At the "Town Hall Meeting" with owners in December 2007, Tesla announced plans to ship the initial 2008 Roadsters with their interim Magna two-speed direct shift manual transmissions locked into second gear, limiting the performance of the car to less than what was originally stated (0 to 60 mph (0 to 97 km/h) in 5.7 seconds instead of the announced 4.0 seconds). Tesla also announced it would upgrade those transmissions under warranty when the final transmission became available. At the "Town Hall Meeting" with owners on January 30, 2008, Tesla Motors described the planned transmission upgrade as a single-speed gearbox with a drive ratio of 8.27:1 combined with improved electronics and motor cooling that retain the acceleration from 0 to 60 mph (0 to 97 km/h) in under 4 seconds and an improved motor limit of 14,000 rpm to retain the 125 mph (201 km/h) top speed. In the interior the gear selector is similar to a push-button automatic with buttons labeled P, R, N and D. Some earlier models have a gear lever similar to that in cars with manual transmission. The Roadster's 0 to 60 mph (0 to 97 km/h) acceleration time is 3.9 seconds for the Standard model and 3.7 seconds for the 2010 V2.5 Sport, which Motor Trend confirmed in the first independent, instrumented testing of the V2.5 Sport model. The magazine also recorded a 0-to-1⁄4-mile (0.00 to 0.40 km) time of 12.6 seconds at 165.1 km/h (102.6 mph). Tesla said the top speed is electronically limited to 125 mph (201 km/h). Tesla claims it has a weight of 2,877 lb (1,305 kg), a drag coefficient of Cd=0.35–0.36 and a rolling resistance of Crr=0.011. Tesla began delivering the higher performance version of the Roadster in July 2009. The Roadster Sport has adjustable dampers and a new hand-wound motor, capable of 0 to 60 mph (0 to 97 km/h) in 3.7 seconds. Scotty Pollacheck, a high-performance driver for Killacycle, drove a 2010 Tesla Roadster Sport at the Wayland Invitational Drag Race in Portland, Oregon, in July 2009. He did a quarter-mile (~400 m) in dry conditions in 12.643 seconds, setting a new record in the National Electric Drag Racing Association among the SP/A3 class of vehicles. The EPA combined range (specifying distance traveled between charges) measured in February 2008 for early production Roadsters was 231 mi (372 km) city, 224 mi (360 km) highway, and 227 mi (365 km) combined (city/highway). In August 2008, additional testing with the newer Powertrain 1.5 resulted in an EPA combined range of 393 km (244 mi). The vehicle set a new distance record when it completed the 241-mile (388 km) Rallye Monte Carlo d'Energies Alternatives with 36 mi (58 km) left on the charge. A Roadster drove around the world (although flying as cargo over oceans) in 2012, and repeated it in 80 days with other electric cars in 2016. Simon Hackett and Emilis Prelgauskas broke the distance record for an electric vehicle, driving 501 km (311 mi) from Alice Springs to Marla, South Australia, in Simon's Tesla Roadster. The car had about 4.8 km (3.0 mi) of range left when the drive was completed. Tesla refers to the Roadster's battery pack as the Energy Storage System or ESS. The 992-pound (450 kg) ESS contains 6,831 lithium ion cells arranged into 11 "sheets" connected in series; each sheet contains 9 "bricks" connected in series; each "brick" contains 69 cells connected in parallel (11S 9S 69P). The cells are of the 18650 form factor commonly found in laptop batteries. Sources disagree on the exact type of Li-Ion cells—GreenCar says lithium cobalt oxide (LiCo), while researchers at DTU/INESC Porto state lithium manganese oxide (LMO). LiCo has higher reaction energy during thermal runaway than LMO. The pack is designed to prevent catastrophic cell failures from propagating to adjacent cells (thermal runaway), even when the cooling system is off. Coolant is pumped continuously through the ESS both when the car is running and when the car is turned off if the pack retains more than a 90% charge. The coolant pump draws 146 watts. The cooling and battery management system keeps the temperatures and voltages within specific limits. A full recharge to 53 kWh requires about 3+1⁄2 hours using the "High Power Wall Connector", which supplies 70-amp, 240-volt electricity. Tesla said in February 2009 that the ESS had expected life span of seven years/100,000 mi (160,000 km), and began selling pre-purchase battery replacements for about one third of the battery's price today, with the replacement to be delivered after seven years. Tesla says the ESS retains 70% capacity after five years and 50,000 mi (80,000 km) of driving, assuming 10,000 mi (16,000 km) driven each year. A July 2013 study found that after 100,000 mi (160,000 km), Roadster batteries still had 80%–85% capacity and the only significant factor is mileage (not temperature). Tesla announced plans to sell the battery system to TH!NK and possibly others through its Tesla Energy Group division. The TH!NK plans were put on hold by interim CEO Michael Marks in September 2007. TH!NK now obtains its lithium-ion batteries from Enerdel.[citation needed] The Roadster uses a proprietary AC charging connector, although Tesla sells a mobile adapter that enables recharging with an SAE J1772 connector. The vehicle was not provided with any DC fast-charging ability and was not retrofitted later on when the Tesla Supercharger network was established. It can be recharged with AC using: Charging times vary depending on the ESS's state-of-charge, the available voltage, and the available circuit breaker amp rating (current). In a best-case scenario using a 240 V charger on a 90 A circuit breaker, Tesla documents a recharging rate of 56 mi (90 km) of range for each hour charging; a complete recharge from empty would require just under four hours. The slowest charging rate using a 120 V outlet on a 15 A circuit breaker would add 5 mi (8.0 km) of range for each hour charging; a complete recharge from empty would require 48 hours. Energy efficiency In June 2006, Tesla reported the Roadster's battery-to-wheel efficiency as 110 Wh/km (17.7 kWh/100 mi) on an unspecified driving cycle—either a constant 60 mph (97 km/h)) or SAE J1634 test—and stated a charging efficiency of 86% for an overall plug-to-wheel efficiency of 128 Wh/km (20.5 kWh/100 mi). In March 2007, Tesla reported the Roadster's efficiency on the EPA highway cycle as "135 mpg [U.S.] equivalent, per the conversion rate used by the EPA" or 133 Wh/km (21.5 kWh/100 mi) battery-to-wheel and 155 Wh/km (24.9 kWh/100 mi) plug-to-wheel. The official U.S. window sticker of the 2009 Tesla Roadster showed an EPA rated energy consumption of 32 kWh/100 mi in city and 33 kWh/100 mi on the highway, equivalent to 105 mpg city and 102 mpg highway. The EPA rating for on board energy efficiency for electric vehicles before 2010 was expressed as kilowatt-hour per 100 miles (kWh/100 mi). Since November 2010, with the introduction of the Nissan Leaf and the Chevrolet Volt, EPA began using a new metric, miles per gallon gasoline equivalent (MPGe). The Roadster was never officially rated by the EPA in MPGe. In August 2007, Tesla dynamometer testing of a validation prototype on the EPA combined cycle yielded a range of 221 mi (356 km) using 23.9 kWh/100 mi (149 Wh/km) battery-to-wheel and 33.6 kWh/100 mi (209 Wh/km) plug-to-wheel. In February 2008, Tesla reported improved plug-to-wheel efficiency after testing a validation prototype car at an EPA-certified location. Those tests yielded a range of 220 mi (350 km) and a plug-to-wheel efficiency of 32.1 kWh/100 mi (199 Wh/km). In August 2008, Tesla reported on testing with the new, single-speed gearbox and upgraded electronics of powertrain 1.5, which yielded an EPA range of 244 mi (393 km) and an EPA combined cycle, plug-to-wheel efficiency of 28 kWh/100 mi (174 Wh/km). In 2007, the Roadster's battery-to-wheel motor efficiency was reported as 88% to 90% on average and 80% at peak power. For comparison, internal combustion engines have a tank-to-wheel efficiency of about 15%. Taking a more complete picture including the cost of energy drawn from its source, Tesla reports that their technology, assuming electricity generated from natural gas-burning power plants, has a high well-to-wheel efficiency of 1.14 km per megajoule, compared to 0.202 km/MJ for gasoline-powered sports cars, 0.478 km/MJ for gasoline-powered commuter cars, 0.556 km/MJ for hybrid cars, and 0.348 km/MJ for hydrogen fuel cell vehicles. As the Roadster does not use gasoline, petroleum efficiency (MPG, L/100 km) cannot be measured directly but instead is calculated using one of several equivalent methods: A number comparable to the typical Monroney sticker's "pump-to-wheel" fuel efficiency is calculated based on regulations from the DOE using its energy content for a U.S. gallon of gasoline of 33,705 Wh⁄gal (also called the Lower Heating Value (LHV) of gasoline): For CAFE regulatory purposes, the DOE's full petroleum-equivalency equation combines the primary energy efficiencies of the US electric grid and the well-to-pump path with a "fuel content factor" that quantifies the value of conservation, scarcity of fuels, and energy security in the US. This regulation uses a factor of 82,049 Wh⁄gal in the above equation and results in a regulatory fuel efficiency of 293 mpggeCAFE. Recharging with electricity from the average US grid, the conversion factor is 12,307 Wh⁄galUS to remove the "fuel content factor" = 1⁄0.15 and the full-cycle energy-equivalency is 44.0 mpgge full-cycle. For full-cycle comparison, the window sticker or "pump-to-wheel" value from a gasoline-fueled vehicle must be multiplied by the fuel's "well-to-pump" efficiency. The DOE regulation specifies a "well-to-pump" efficiency of 83% for gasoline. The 2006 Prius's sticker 46 mpg‑US (5.1 L/100 km; 55 mpg‑imp), for example, converts to a full-cycle energy-equivalent of 38.2 mpgfull-cycle. Recharging with electricity generated by newer, 58% efficiency CCGT power plants, results in a conversion factor of 21,763 Wh⁄gal and yields a fuel efficiency of 77.7 mpgge. Recharging with non-fossil fuel electricity sources such as hydroelectric, solar power, wind or nuclear, the petroleum equivalent efficiency can be even higher as fossil fuel is not directly used in refueling. Service Whereas vehicles with internal combustion engines require more frequent service for oil changes and routine maintenance on engine components and other related systems, Tesla's website recommends the owner bring the vehicle in for service "once a year or every 12,000 miles". For other concerns with vehicles, Tesla created a "mobile service unit" that dispatches company-trained technicians to customers' homes or offices in case the owner is experiencing problems. Tesla charges the customer according to the distance the service unit needs to travel: one US dollar per mile roundtrip with a 100-dollar minimum. Technicians drive company vans equipped with numerous tools and testing equipment to do "in the field" repairs, enhancements and software upgrades. Tesla debuted this "house call" approach in the spring of 2009, when the company announced a recall due to a manufacturing problem in the Lotus assembly plant, which also affected the Lotus Elise and other models from the British sports car maker. The first Tesla service center, in Los Angeles, California, was opened on Santa Monica Boulevard on May 1, 2008. Tesla publicly opened their second showroom and service area in Menlo Park, California on July 22, 2008. The Menlo Park location is also the final assembly area for Tesla Roadsters. Tesla also operates service centers in New York City, Miami, Chicago, and Seattle.[citation needed] In 2007, Tesla announced plans to build additional service centers over the following few years to support sales of its next vehicle, the Model S sports sedan. This included an additional 15 service centers in United States major metropolitan locations. Possible locations for sales and service locations in Europe were announced in a letter to customers in May 2008. As of May 2017[update], Tesla has issued two product safety recalls for the Roadster. In May 2009, Tesla issued a recall for 345 Roadsters manufactured before April 22, 2009. Tesla sent technicians to customers' homes to tighten the rear, inner hub flange bolts. Using wording from the National Highway Traffic and Safety Administration, Tesla told customers that without this adjustment, the driver could lose control of the car. The problem originated at the Lotus assembly line, where the Roadster glider was built. Lotus also recalled some Elise and Exige vehicles for the same reason. On October 1, 2010, Tesla issued a second product safety recall in the US affecting 439 Roadsters. The recall involved the 12 V low-voltage auxiliary cable from a redundant back-up system. The recall followed an incident where the low voltage auxiliary cable in a vehicle chafed against the edge of a carbon fiber panel, causing a short, smoke and a possible fire behind the right front headlamp. This issue was limited to the 12 V low-voltage auxiliary cable and did not involve the main battery pack or main power system. Reviews Tesla Roadster reviews can be grouped in two main categories: older reviews of "validation prototypes" (2006–2008), before Tesla began serial production and customer deliveries, and reviews on cars in serial production (2008–2010).[citation needed] The global online auto review site Autoguide.com tested Tesla's fourth-generation car in October 2010. Autoguide editor Derek Kreindler said "The Tesla Roadster 2.5 S is a massively impressive vehicle, more spacecraft than sports car. Theories like global warming, peak oil and rising oil prices should no longer bring heart palpitations to car fans. The Tesla shows just how good zero-emissions 'green' technology can be. Quite frankly, getting into a normal car at the end of the test drive was a major letdown. The whirr of the engine, the shove in the backside and the little roadster that seems to pivot around you is replaced by a grunting, belching, feedback-free driving experience". He continues on that "but for a $100,000 car, it could use some work" complaining of purposefully cheap work. In the March 2010 print edition of British enthusiast magazine EVO (p. 120), editor Richard Meaden was the first to review the all-new right-hand-drive version of the Roadster. He said the car had "serious, instantaneous muscle". "With so much torque from literally no revs the acceleration punch is wholly alien. Away from traffic lights you'd murder anything, be it a 911 Turbo, GT-R or 599, simply because while they have to mess about with balancing revs and clutch, or fiddle with launch controls and invalid warranties, all you have to do is floor the throttle and wave goodbye". In December 2009, The Wall Street Journal editor Joseph White conducted an extended test-drive and determined that "you can have enormous fun within the legal speed limit as you whoosh around unsuspecting Camry drivers, zapping from 40 to 60 miles per hour in two seconds while the startled victims eat your electric dust". White praised the car's environmental efficiency but said consumer demand reflected not the environmental attributes of the car but its performance. "The Tesla turns the frugal environmentalist aesthetic on its head. Sure, it doesn't burn petroleum, and if plugged into a wind turbine or a nuclear plant, it would be a very low-carbon machine. But anyone who buys one will get the most satisfaction from smoking someone's doors off. The Tesla's message is that 'green' technology can appeal to the id, not just the superego". In December 2009, Motor Trend was the first to independently confirm the Roadster Sport's reported 0 to 60 mph (0 to 97 km/h) time of 3.7 seconds. (Motor Trend recorded 0 to 60 mph (0 to 97 km/h) of 3.70 seconds; it recorded a quarter-mile test at 12.6 sec at 102.6 mph (165.1 km/h).) Engineering editor Kim Reynolds called the acceleration "breathtaking" and said the car confirms "Tesla as an actual car company. ...Tesla is the first maker to crack the EV legitimacy barrier in a century". In November 2009, Automobile Magazine West Coast editor Jason Cammisa spent a week driving a production Tesla Roadster. Cammisa was immediately impressed with the acceleration, saying the car "explodes off the line, pulling like a small jet plane. ... It's like driving a Lamborghini with a big V-12 revved over 6000 rpm at all times, waiting to pounce—without the noise, vibration, or misdemeanor arrest for disturbing the peace". He also took the car to Infineon Raceway in Sonoma, California, and praised the car for its robustness, saying the Roadster: wins the Coolest Car I've Ever Driven award. Why? Despite the flat-out sprints, the drag racing, the donuts, the top-speed runs, and dicing through traffic like there's a jet pack strapped to the trunk, Pacific Gas and Electric—which generated power for the Tesla—released into the atmosphere the same amount of carbon dioxide as would a gasoline-powered car getting 99 mpg. And the Roadster didn't break. It didn't smoke, lock up, freeze, or experience flux-capacitor failure. Over the past ten decades, no company has been able to reinvent the car—not General Motors with the EV1, not Toyota with the Prius. And now, a bunch of dudes from Silicon Valley have created an electric car that really works—as both an environmental fix and a speed fix In 2009 the Tesla Roadster was one of the Scandinavian Sports Car of the Year participants. In a comparison made by Nordic car magazines Tekniikan Maailma (Finland), Teknikens Värld (Sweden) and Bil Magasinet (Denmark), critics praised the torque of the car and a track car structure, but also highlighted more negative aspects such as a short battery life; they were unable to drive a full track lap in dry track conditions. In May 2009, Car and Driver technical editor Aaron Robinson wrote a review based on the first extended test-drive of a production Tesla Roadster. Robinson had the car for nearly a week at his home. He complained of "design anomalies, daily annoyances, absurd ergonomics, and ridiculous economics" and stated he never got to see if the car could go 240 miles on a single charge because the torturous seating forced him to stop driving the car. He also complained of Tesla increasing the car prices on those who had already made deposits and charging extra for previously free necessary components. In February 2009, automotive critic Dan Neil of the Los Angeles Times called the production Tesla Roadster "a superb piece of machinery: stiff, well sorted, highly focused, dead-sexy and eerily quick". Neil said he had the car for 24 hours but "caned it like the Taliban caned Gillette salesmen and it never even blinked". In February 2009, Road & Track tested another production vehicle and conducted the first independently verified metered testing of the Roadster. Engineering editor Dennis Simanaitis said the testing confirmed what he called "extravagant claims", that the Roadster had a 4.0 s 0 to 60 mph (0 to 97 km/h) acceleration and a 200-mile (320 km) range. They said the Roadster felt like "an over-ballasted Lotus Elise", but the weight was well-distributed, so the car remained responsive. "Fit and finish of our Tesla were exemplary", which Road & Track thought fit the target market. Overall, they considered it a "delight" to drive. Testing a pre-production car in early 2008, Road & Track said "The Tesla feels composed and competent at speed with great turn-in and transitioning response", though they recommended against it as a "primary grocery-getter". In January 2009, automotive critic Warren Brown of The Washington Post called the production Roadster "a head-turner, jaw-dropper. It is sexy as all get-out". He described the feeling behind the wheel as, "Wheeeeeee! Drive a Tesla, even if you have to fly to Tesla's Menlo Park, Calif., headquarters, to get your hands on one for a day. ... If this is the future of the automobile, I want it". In a review of a Roadster prototype in March 2008, before the cars were in serial production, Motor Trend gave a generally favorable review, stating that it was "undeniably, unbelievably efficient" and would be "profoundly humbling to just about any rumbling Ferrari or Porsche that makes the mistake of pulling up next to a silent, 105 mpg‑US (2.2 L/100 km; 126 mpg‑imp) Tesla Roadster at a stoplight"; they nonetheless detected a "nasty drive-train buck" during the test drive of an early Roadster with the older, two-speed transmission. In a July 8, 2007, review of a prototype Roadster, Jay Leno wrote, "If you like sports cars and you want to be green, this is the only way to go. The Tesla is a car that you can live with, drive and enjoy as a sports car. I had a brief drive in the car and it was quite impressive. This is an electric car that is fun to drive". In a November 27, 2006, review of a prototype Roadster in Slate, Paul Boutin wrote, "A week ago, I went for a spin in the fastest, most fun car I've ever ridden in—and that includes the Aston Martin I tried to buy once. I was so excited, in fact, that I decided to take a few days to calm down before writing about it. Well, my waiting period is over, I'm thinking rationally, and I'm still unbelievably stoked about the Tesla". In the third quarter of 2008, Top Gear's Jeremy Clarkson reviewed two production Roadsters with the v1.5 transmission and described the driving experience with the exclamations "God Almighty! Wave goodbye to dial-up, and say hello to the world of broadband motoring!" and "This car is biblically quick!" when comparing the acceleration versus the car the Roadster was based on, a Lotus Elise. Clarkson also noted, however, that the handling of the car was not as sharp as that of the Elise: "through the corners things are less rosy". The segment also claimed that the car's batteries would run flat after 55 mi (89 km) of heavy use on a track and showed the car being pushed off the track. Tesla spokesperson responded with statements in blogs and to mainstream news organizations that the cars provided to Top Gear never had less than 20% charge and never experienced brake failure. In addition, neither car provided to Top Gear needed to be pushed off the track at any point. Clarkson also showed a wind turbine with stationary rotor blades and complained that it would take countless hours to refuel the car using such a source of electricity, although the car can be charged from a 240 V 70 A outlet in as little as 3.5 hours. After numerous blogs and several large news organizations began following the controversy, the BBC issued a statement saying "the tested Tesla was filmed being pushed into the shed in order to show what would happen if the Roadster had run out of charge. Top Gear stands by the findings in this film and is content that it offers a fair representation of the Tesla's performance on the day it was tested", without addressing the other alleged misrepresentations that Tesla highlighted to the media. After several weeks of increasing pressure and inquiries from the BBC, Clarkson wrote a blog entry for The Times, acknowledging that "Inevitably, the film we had shot was a bit of a mess. There was a handful of shots of a silver car. Some of a grey car". "But as a device for moving you and your things around, it is about as much use as a bag of muddy spinach". In the months that followed Clarkson's acknowledgment, the original episode—including the misstatements—reran on BBC America and elsewhere without any editing.[citation needed] On March 29, 2011, Tesla sued the programme over libel and malicious falsehood, while simultaneously launching the website TeslaVsTopGear.com. The current position of Tesla is found on its web page. In a blogpost, producer Andy Wilman has referred to Tesla's allegations as a "crusade" and contested the truth value of Tesla's statements. On October 19, 2011, the High Court in London rejected Tesla's libel claim. Tesla appealed High Court's decision to the Court of Appeal, where a three-judge panel of Lords Justice upheld the lower court's decision, and ordered Tesla to pay the BBC's legal costs of £100,000. Sales Tesla delivered approximately 2,450 Roadsters worldwide between February 2008 and December 2012. Featuring new options and enhanced features, the 2012 Tesla Roadster was sold in limited numbers only in mainland Europe, Asia and Australia, and as of July 2012, less than 140 units were available for sale in Europe and Asia before the remaining inventory would be sold out. Tesla's US exemption for not having special two-stage passenger airbags expired for cars made after the end of 2011 so the last Roadsters were not sold in the American market for regulatory reasons. The U.S. was the leading market with about 1,800 Roadsters sold. There were fewer than 50 right-hand-drive models of the Tesla Roadster produced and hand built in the UK.[citation needed] The Roadster had a three-year, 36,000-mile (58,000 km) warranty. Tesla also offered an extended powertrain warranty and a battery replacement warranty.[citation needed] In July 2009, Tesla announced that US consumers could finance the Roadster through Bank of America. Financing was available for up to 75% of the total vehicle purchase price.[better source needed] Tesla sold Roadsters directly to customers. It sold them online, in 13 showrooms and over the phone in North America and Europe. Tesla does not operate through franchise dealerships but operates company-owned stores. The company said that it took its retail cues from Apple, Starbucks and other non-automotive retailers. The company has been shipping cars to European customers since mid-2009. Tesla sold out of its EU special-edition vehicle, which had a 2010 model-year production run of 250 cars. A total of 575 units have been sold in Europe through October 2012. Tesla first overseas showroom opened in London in 2009, with right-hand-drive models promised for early 2010. Showrooms in Munich and Monaco were also added in 2009, followed by Zurich and Copenhagen in 2010 and Milan in 2011. Reservations for the 2010 Roadster were available for a €3,000 refundable reservation fee. From 2009 to 2014, Hansjoerg von Gemmingen of Karlsruhe, Germany drove his Tesla Roadster 400,000 km (250,000 mi), this being the mileage world record for an all-electric vehicle and reached 520,000 km (320,000 mi) in 2017. He also drove another 200,000 km (120,000 mi) in a Tesla Model S and voiced his plan to become the first man to travel a million kilometres in an electric vehicle. Kevin Yu, the director of Tesla Motors Asia Pacific, said Roadsters in Japan had additional yearly taxes for exceeding the width limit of normal sized cars.[citation needed] In 2009, Roadster reservation holders who had already placed deposits up to US$50,000 (~$70,845 in 2024) to lock in their orders were informed that their orders had been unlocked and that they had to re-option their ordered vehicles on the threat of losing their spot on the orders list. Tesla then raised the prices of several options, and a new Tesla Roadster with the same set of features that had previously been standard became US$6,700 more expensive than before. For example, the high performance charger that was previously claimed to be standard on all vehicles was changed to be an optional feature costing US$3,000, and the previously claimed standard forged alloy wheels became a US$2,300 upgrade. One person who pre-ordered a Tesla Roadster complained: I am [pre-ordered owner] number 395. I am not a rich person dabbling in a plaything. I thought I was actually doing some good by supporting a company that was moving us to a more sustainable future. I put $50,000 of my own money down on this car in May of 2007. I withstood the delays. I held in there when it almost seemed the company was going bankrupt. Now, after locking in my options, they pull this on me. Awards The world distance record of 501 km (311 mi) for a production electric car on a single charge was set by a Roadster on October 27, 2009, during the Global Green Challenge in outback Australia, in which it averaged a speed of 25 mph (40 km/h). In March 2010, a Tesla Roadster became the first electric vehicle to win the Monte Carlo Alternative Energy Rally and the first to win any Federation Internationale de l'Automobile-sanctioned championship when a Roadster driven by former Formula One driver Érik Comas beat 96 competitors for range, efficiency and performance in the three-day, nearly 1,000-kilometre (620 mi) challenge. Space launch In December 2017, Elon Musk announced that his personal Tesla Roadster, sn:686, would be launched into space, serving as dummy payload on the maiden flight of the SpaceX Falcon Heavy rocket. The launch on February 6, 2018, was successful; the vehicle was placed into a heliocentric orbit that took it beyond Mars's orbital path around the Sun. The Roadster was the first production car to be launched into space. See also References External links
========================================
[SOURCE: https://en.wikipedia.org/wiki/Animal#cite_note-Mora2011-71] | [TOKENS: 6011]
Contents Animal Animals are multicellular, eukaryotic organisms belonging to the biological kingdom Animalia (/ˌænɪˈmeɪliə/). With few exceptions, animals consume organic material, breathe oxygen, have myocytes and are able to move, can reproduce sexually, and grow from a hollow sphere of cells, the blastula, during embryonic development. Animals form a clade, meaning that they arose from a single common ancestor. Over 1.5 million living animal species have been described, of which around 1.05 million are insects, over 85,000 are molluscs, and around 65,000 are vertebrates. It has been estimated there are as many as 7.77 million animal species on Earth. Animal body lengths range from 8.5 μm (0.00033 in) to 33.6 m (110 ft). They have complex ecologies and interactions with each other and their environments, forming intricate food webs. The scientific study of animals is known as zoology, and the study of animal behaviour is known as ethology. The animal kingdom is divided into five major clades, namely Porifera, Ctenophora, Placozoa, Cnidaria and Bilateria. Most living animal species belong to the clade Bilateria, a highly proliferative clade whose members have a bilaterally symmetric and significantly cephalised body plan, and the vast majority of bilaterians belong to two large clades: the protostomes, which includes organisms such as arthropods, molluscs, flatworms, annelids and nematodes; and the deuterostomes, which include echinoderms, hemichordates and chordates, the latter of which contains the vertebrates. The much smaller basal phylum Xenacoelomorpha have an uncertain position within Bilateria. Animals first appeared in the fossil record in the late Cryogenian period and diversified in the subsequent Ediacaran period in what is known as the Avalon explosion. Nearly all modern animal phyla first appeared in the fossil record as marine species during the Cambrian explosion, which began around 539 million years ago (Mya), and most classes during the Ordovician radiation 485.4 Mya. Common to all living animals, 6,331 groups of genes have been identified that may have arisen from a single common ancestor that lived about 650 Mya during the Cryogenian period. Historically, Aristotle divided animals into those with blood and those without. Carl Linnaeus created the first hierarchical biological classification for animals in 1758 with his Systema Naturae, which Jean-Baptiste Lamarck expanded into 14 phyla by 1809. In 1874, Ernst Haeckel divided the animal kingdom into the multicellular Metazoa (now synonymous with Animalia) and the Protozoa, single-celled organisms no longer considered animals. In modern times, the biological classification of animals relies on advanced techniques, such as molecular phylogenetics, which are effective at demonstrating the evolutionary relationships between taxa. Humans make use of many other animal species for food (including meat, eggs, and dairy products), for materials (such as leather, fur, and wool), as pets and as working animals for transportation, and services. Dogs, the first domesticated animal, have been used in hunting, in security and in warfare, as have horses, pigeons and birds of prey; while other terrestrial and aquatic animals are hunted for sports, trophies or profits. Non-human animals are also an important cultural element of human evolution, having appeared in cave arts and totems since the earliest times, and are frequently featured in mythology, religion, arts, literature, heraldry, politics, and sports. Etymology The word animal comes from the Latin noun animal of the same meaning, which is itself derived from Latin animalis 'having breath or soul'. The biological definition includes all members of the kingdom Animalia. In colloquial usage, the term animal is often used to refer only to nonhuman animals. The term metazoa is derived from Ancient Greek μετα meta 'after' (in biology, the prefix meta- stands for 'later') and ζῷᾰ zōia 'animals', plural of ζῷον zōion 'animal'. A metazoan is any member of the group Metazoa. Characteristics Animals have several characteristics that they share with other living things. Animals are eukaryotic, multicellular, and aerobic, as are plants and fungi. Unlike plants and algae, which produce their own food, animals cannot produce their own food, a feature they share with fungi. Animals ingest organic material and digest it internally. Animals have structural characteristics that set them apart from all other living things: Typically, there is an internal digestive chamber with either one opening (in Ctenophora, Cnidaria, and flatworms) or two openings (in most bilaterians). Animal development is controlled by Hox genes, which signal the times and places to develop structures such as body segments and limbs. During development, the animal extracellular matrix forms a relatively flexible framework upon which cells can move about and be reorganised into specialised tissues and organs, making the formation of complex structures possible, and allowing cells to be differentiated. The extracellular matrix may be calcified, forming structures such as shells, bones, and spicules. In contrast, the cells of other multicellular organisms (primarily algae, plants, and fungi) are held in place by cell walls, and so develop by progressive growth. Nearly all animals make use of some form of sexual reproduction. They produce haploid gametes by meiosis; the smaller, motile gametes are spermatozoa and the larger, non-motile gametes are ova. These fuse to form zygotes, which develop via mitosis into a hollow sphere, called a blastula. In sponges, blastula larvae swim to a new location, attach to the seabed, and develop into a new sponge. In most other groups, the blastula undergoes more complicated rearrangement. It first invaginates to form a gastrula with a digestive chamber and two separate germ layers, an external ectoderm and an internal endoderm. In most cases, a third germ layer, the mesoderm, also develops between them. These germ layers then differentiate to form tissues and organs. Repeated instances of mating with a close relative during sexual reproduction generally leads to inbreeding depression within a population due to the increased prevalence of harmful recessive traits. Animals have evolved numerous mechanisms for avoiding close inbreeding. Some animals are capable of asexual reproduction, which often results in a genetic clone of the parent. This may take place through fragmentation; budding, such as in Hydra and other cnidarians; or parthenogenesis, where fertile eggs are produced without mating, such as in aphids. Ecology Animals are categorised into ecological groups depending on their trophic levels and how they consume organic material. Such groupings include carnivores (further divided into subcategories such as piscivores, insectivores, ovivores, etc.), herbivores (subcategorised into folivores, graminivores, frugivores, granivores, nectarivores, algivores, etc.), omnivores, fungivores, scavengers/detritivores, and parasites. Interactions between animals of each biome form complex food webs within that ecosystem. In carnivorous or omnivorous species, predation is a consumer–resource interaction where the predator feeds on another organism, its prey, who often evolves anti-predator adaptations to avoid being fed upon. Selective pressures imposed on one another lead to an evolutionary arms race between predator and prey, resulting in various antagonistic/competitive coevolutions. Almost all multicellular predators are animals. Some consumers use multiple methods; for example, in parasitoid wasps, the larvae feed on the hosts' living tissues, killing them in the process, but the adults primarily consume nectar from flowers. Other animals may have very specific feeding behaviours, such as hawksbill sea turtles which mainly eat sponges. Most animals rely on biomass and bioenergy produced by plants and phytoplanktons (collectively called producers) through photosynthesis. Herbivores, as primary consumers, eat the plant material directly to digest and absorb the nutrients, while carnivores and other animals on higher trophic levels indirectly acquire the nutrients by eating the herbivores or other animals that have eaten the herbivores. Animals oxidise carbohydrates, lipids, proteins and other biomolecules in cellular respiration, which allows the animal to grow and to sustain basal metabolism and fuel other biological processes such as locomotion. Some benthic animals living close to hydrothermal vents and cold seeps on the dark sea floor consume organic matter produced through chemosynthesis (via oxidising inorganic compounds such as hydrogen sulfide) by archaea and bacteria. Animals originated in the ocean; all extant animal phyla, except for Micrognathozoa and Onychophora, feature at least some marine species. However, several lineages of arthropods begun to colonise land around the same time as land plants, probably between 510 and 471 million years ago, during the Late Cambrian or Early Ordovician. Vertebrates such as the lobe-finned fish Tiktaalik started to move on to land in the late Devonian, about 375 million years ago. Other notable animal groups that colonized land environments are Mollusca, Platyhelmintha, Annelida, Tardigrada, Onychophora, Rotifera, Nematoda. Animals occupy virtually all of earth's habitats and microhabitats, with faunas adapted to salt water, hydrothermal vents, fresh water, hot springs, swamps, forests, pastures, deserts, air, and the interiors of other organisms. Animals are however not particularly heat tolerant; very few of them can survive at constant temperatures above 50 °C (122 °F) or in the most extreme cold deserts of continental Antarctica. The collective global geomorphic influence of animals on the processes shaping the Earth's surface remains largely understudied, with most studies limited to individual species and well-known exemplars. Diversity The blue whale (Balaenoptera musculus) is the largest animal that has ever lived, weighing up to 190 tonnes and measuring up to 33.6 metres (110 ft) long. The largest extant terrestrial animal is the African bush elephant (Loxodonta africana), weighing up to 12.25 tonnes and measuring up to 10.67 metres (35.0 ft) long. The largest terrestrial animals that ever lived were titanosaur sauropod dinosaurs such as Argentinosaurus, which may have weighed as much as 73 tonnes, and Supersaurus which may have reached 39 metres. Several animals are microscopic; some Myxozoa (obligate parasites within the Cnidaria) never grow larger than 20 μm, and one of the smallest species (Myxobolus shekel) is no more than 8.5 μm when fully grown. The following table lists estimated numbers of described extant species for the major animal phyla, along with their principal habitats (terrestrial, fresh water, and marine), and free-living or parasitic ways of life. Species estimates shown here are based on numbers described scientifically; much larger estimates have been calculated based on various means of prediction, and these can vary wildly. For instance, around 25,000–27,000 species of nematodes have been described, while published estimates of the total number of nematode species include 10,000–20,000; 500,000; 10 million; and 100 million. Using patterns within the taxonomic hierarchy, the total number of animal species—including those not yet described—was calculated to be about 7.77 million in 2011.[a] 3,000–6,500 4,000–25,000 Evolutionary origin Evidence of animals is found as long ago as the Cryogenian period. 24-Isopropylcholestane (24-ipc) has been found in rocks from roughly 650 million years ago; it is only produced by sponges and pelagophyte algae. Its likely origin is from sponges based on molecular clock estimates for the origin of 24-ipc production in both groups. Analyses of pelagophyte algae consistently recover a Phanerozoic origin, while analyses of sponges recover a Neoproterozoic origin, consistent with the appearance of 24-ipc in the fossil record. The first body fossils of animals appear in the Ediacaran, represented by forms such as Charnia and Spriggina. It had long been doubted whether these fossils truly represented animals, but the discovery of the animal lipid cholesterol in fossils of Dickinsonia establishes their nature. Animals are thought to have originated under low-oxygen conditions, suggesting that they were capable of living entirely by anaerobic respiration, but as they became specialised for aerobic metabolism they became fully dependent on oxygen in their environments. Many animal phyla first appear in the fossil record during the Cambrian explosion, starting about 539 million years ago, in beds such as the Burgess Shale. Extant phyla in these rocks include molluscs, brachiopods, onychophorans, tardigrades, arthropods, echinoderms and hemichordates, along with numerous now-extinct forms such as the predatory Anomalocaris. The apparent suddenness of the event may however be an artefact of the fossil record, rather than showing that all these animals appeared simultaneously. That view is supported by the discovery of Auroralumina attenboroughii, the earliest known Ediacaran crown-group cnidarian (557–562 mya, some 20 million years before the Cambrian explosion) from Charnwood Forest, England. It is thought to be one of the earliest predators, catching small prey with its nematocysts as modern cnidarians do. Some palaeontologists have suggested that animals appeared much earlier than the Cambrian explosion, possibly as early as 1 billion years ago. Early fossils that might represent animals appear for example in the 665-million-year-old rocks of the Trezona Formation of South Australia. These fossils are interpreted as most probably being early sponges. Trace fossils such as tracks and burrows found in the Tonian period (from 1 gya) may indicate the presence of triploblastic worm-like animals, roughly as large (about 5 mm wide) and complex as earthworms. However, similar tracks are produced by the giant single-celled protist Gromia sphaerica, so the Tonian trace fossils may not indicate early animal evolution. Around the same time, the layered mats of microorganisms called stromatolites decreased in diversity, perhaps due to grazing by newly evolved animals. Objects such as sediment-filled tubes that resemble trace fossils of the burrows of wormlike animals have been found in 1.2 gya rocks in North America, in 1.5 gya rocks in Australia and North America, and in 1.7 gya rocks in Australia. Their interpretation as having an animal origin is disputed, as they might be water-escape or other structures. Phylogeny Animals are monophyletic, meaning they are derived from a common ancestor. Animals are the sister group to the choanoflagellates, with which they form the Choanozoa. Ros-Rocher and colleagues (2021) trace the origins of animals to unicellular ancestors, providing the external phylogeny shown in the cladogram. Uncertainty of relationships is indicated with dashed lines. The animal clade had certainly originated by 650 mya, and may have come into being as much as 800 mya, based on molecular clock evidence for different phyla. Holomycota (inc. fungi) Ichthyosporea Pluriformea Filasterea The relationships at the base of the animal tree have been debated. Other than Ctenophora, the Bilateria and Cnidaria are the only groups with symmetry, and other evidence shows they are closely related. In addition to sponges, Placozoa has no symmetry and was often considered a "missing link" between protists and multicellular animals. The presence of hox genes in Placozoa shows that they were once more complex. The Porifera (sponges) have long been assumed to be sister to the rest of the animals, but there is evidence that the Ctenophora may be in that position. Molecular phylogenetics has supported both the sponge-sister and ctenophore-sister hypotheses. In 2017, Roberto Feuda and colleagues, using amino acid differences, presented both, with the following cladogram for the sponge-sister view that they supported (their ctenophore-sister tree simply interchanging the places of ctenophores and sponges): Porifera Ctenophora Placozoa Cnidaria Bilateria Conversely, a 2023 study by Darrin Schultz and colleagues uses ancient gene linkages to construct the following ctenophore-sister phylogeny: Ctenophora Porifera Placozoa Cnidaria Bilateria Sponges are physically very distinct from other animals, and were long thought to have diverged first, representing the oldest animal phylum and forming a sister clade to all other animals. Despite their morphological dissimilarity with all other animals, genetic evidence suggests sponges may be more closely related to other animals than the comb jellies are. Sponges lack the complex organisation found in most other animal phyla; their cells are differentiated, but in most cases not organised into distinct tissues, unlike all other animals. They typically feed by drawing in water through pores, filtering out small particles of food. The Ctenophora and Cnidaria are radially symmetric and have digestive chambers with a single opening, which serves as both mouth and anus. Animals in both phyla have distinct tissues, but these are not organised into discrete organs. They are diploblastic, having only two main germ layers, ectoderm and endoderm. The tiny placozoans have no permanent digestive chamber and no symmetry; they superficially resemble amoebae. Their phylogeny is poorly defined, and under active research. The remaining animals, the great majority—comprising some 29 phyla and over a million species—form the Bilateria clade, which have a bilaterally symmetric body plan. The Bilateria are triploblastic, with three well-developed germ layers, and their tissues form distinct organs. The digestive chamber has two openings, a mouth and an anus, and in the Nephrozoa there is an internal body cavity, a coelom or pseudocoelom. These animals have a head end (anterior) and a tail end (posterior), a back (dorsal) surface and a belly (ventral) surface, and a left and a right side. A modern consensus phylogenetic tree for the Bilateria is shown below. Xenacoelomorpha Ambulacraria Chordata Ecdysozoa Spiralia Having a front end means that this part of the body encounters stimuli, such as food, favouring cephalisation, the development of a head with sense organs and a mouth. Many bilaterians have a combination of circular muscles that constrict the body, making it longer, and an opposing set of longitudinal muscles, that shorten the body; these enable soft-bodied animals with a hydrostatic skeleton to move by peristalsis. They also have a gut that extends through the basically cylindrical body from mouth to anus. Many bilaterian phyla have primary larvae which swim with cilia and have an apical organ containing sensory cells. However, over evolutionary time, descendant spaces have evolved which have lost one or more of each of these characteristics. For example, adult echinoderms are radially symmetric (unlike their larvae), while some parasitic worms have extremely simplified body structures. Genetic studies have considerably changed zoologists' understanding of the relationships within the Bilateria. Most appear to belong to two major lineages, the protostomes and the deuterostomes. It is often suggested that the basalmost bilaterians are the Xenacoelomorpha, with all other bilaterians belonging to the subclade Nephrozoa. However, this suggestion has been contested, with other studies finding that xenacoelomorphs are more closely related to Ambulacraria than to other bilaterians. Protostomes and deuterostomes differ in several ways. Early in development, deuterostome embryos undergo radial cleavage during cell division, while many protostomes (the Spiralia) undergo spiral cleavage. Animals from both groups possess a complete digestive tract, but in protostomes the first opening of the embryonic gut develops into the mouth, and the anus forms secondarily. In deuterostomes, the anus forms first while the mouth develops secondarily. Most protostomes have schizocoelous development, where cells simply fill in the interior of the gastrula to form the mesoderm. In deuterostomes, the mesoderm forms by enterocoelic pouching, through invagination of the endoderm. The main deuterostome taxa are the Ambulacraria and the Chordata. Ambulacraria are exclusively marine and include acorn worms, starfish, sea urchins, and sea cucumbers. The chordates are dominated by the vertebrates (animals with backbones), which consist of fishes, amphibians, reptiles, birds, and mammals. The protostomes include the Ecdysozoa, named after their shared trait of ecdysis, growth by moulting, Among the largest ecdysozoan phyla are the arthropods and the nematodes. The rest of the protostomes are in the Spiralia, named for their pattern of developing by spiral cleavage in the early embryo. Major spiralian phyla include the annelids and molluscs. History of classification In the classical era, Aristotle divided animals,[d] based on his own observations, into those with blood (roughly, the vertebrates) and those without. The animals were then arranged on a scale from man (with blood, two legs, rational soul) down through the live-bearing tetrapods (with blood, four legs, sensitive soul) and other groups such as crustaceans (no blood, many legs, sensitive soul) down to spontaneously generating creatures like sponges (no blood, no legs, vegetable soul). Aristotle was uncertain whether sponges were animals, which in his system ought to have sensation, appetite, and locomotion, or plants, which did not: he knew that sponges could sense touch and would contract if about to be pulled off their rocks, but that they were rooted like plants and never moved about. In 1758, Carl Linnaeus created the first hierarchical classification in his Systema Naturae. In his original scheme, the animals were one of three kingdoms, divided into the classes of Vermes, Insecta, Pisces, Amphibia, Aves, and Mammalia. Since then, the last four have all been subsumed into a single phylum, the Chordata, while his Insecta (which included the crustaceans and arachnids) and Vermes have been renamed or broken up. The process was begun in 1793 by Jean-Baptiste de Lamarck, who called the Vermes une espèce de chaos ('a chaotic mess')[e] and split the group into three new phyla: worms, echinoderms, and polyps (which contained corals and jellyfish). By 1809, in his Philosophie Zoologique, Lamarck had created nine phyla apart from vertebrates (where he still had four phyla: mammals, birds, reptiles, and fish) and molluscs, namely cirripedes, annelids, crustaceans, arachnids, insects, worms, radiates, polyps, and infusorians. In his 1817 Le Règne Animal, Georges Cuvier used comparative anatomy to group the animals into four embranchements ('branches' with different body plans, roughly corresponding to phyla), namely vertebrates, molluscs, articulated animals (arthropods and annelids), and zoophytes (radiata) (echinoderms, cnidaria and other forms). This division into four was followed by the embryologist Karl Ernst von Baer in 1828, the zoologist Louis Agassiz in 1857, and the comparative anatomist Richard Owen in 1860. In 1874, Ernst Haeckel divided the animal kingdom into two subkingdoms: Metazoa (multicellular animals, with five phyla: coelenterates, echinoderms, articulates, molluscs, and vertebrates) and Protozoa (single-celled animals), including a sixth animal phylum, sponges. The protozoa were later moved to the former kingdom Protista, leaving only the Metazoa as a synonym of Animalia. In human culture The human population exploits a large number of other animal species for food, both of domesticated livestock species in animal husbandry and, mainly at sea, by hunting wild species. Marine fish of many species are caught commercially for food. A smaller number of species are farmed commercially. Humans and their livestock make up more than 90% of the biomass of all terrestrial vertebrates, and almost as much as all insects combined. Invertebrates including cephalopods, crustaceans, insects—principally bees and silkworms—and bivalve or gastropod molluscs are hunted or farmed for food, fibres. Chickens, cattle, sheep, pigs, and other animals are raised as livestock for meat across the world. Animal fibres such as wool and silk are used to make textiles, while animal sinews have been used as lashings and bindings, and leather is widely used to make shoes and other items. Animals have been hunted and farmed for their fur to make items such as coats and hats. Dyestuffs including carmine (cochineal), shellac, and kermes have been made from the bodies of insects. Working animals including cattle and horses have been used for work and transport from the first days of agriculture. Animals such as the fruit fly Drosophila melanogaster serve a major role in science as experimental models. Animals have been used to create vaccines since their discovery in the 18th century. Some medicines such as the cancer drug trabectedin are based on toxins or other molecules of animal origin. People have used hunting dogs to help chase down and retrieve animals, and birds of prey to catch birds and mammals, while tethered cormorants have been used to catch fish. Poison dart frogs have been used to poison the tips of blowpipe darts. A wide variety of animals are kept as pets, from invertebrates such as tarantulas, octopuses, and praying mantises, reptiles such as snakes and chameleons, and birds including canaries, parakeets, and parrots all finding a place. However, the most kept pet species are mammals, namely dogs, cats, and rabbits. There is a tension between the role of animals as companions to humans, and their existence as individuals with rights of their own. A wide variety of terrestrial and aquatic animals are hunted for sport. The signs of the Western and Chinese zodiacs are based on animals. In China and Japan, the butterfly has been seen as the personification of a person's soul, and in classical representation the butterfly is also the symbol of the soul. Animals have been the subjects of art from the earliest times, both historical, as in ancient Egypt, and prehistoric, as in the cave paintings at Lascaux. Major animal paintings include Albrecht Dürer's 1515 The Rhinoceros, and George Stubbs's c. 1762 horse portrait Whistlejacket. Insects, birds and mammals play roles in literature and film, such as in giant bug movies. Animals including insects and mammals feature in mythology and religion. The scarab beetle was sacred in ancient Egypt, and the cow is sacred in Hinduism. Among other mammals, deer, horses, lions, bats, bears, and wolves are the subjects of myths and worship. See also Notes References External links
========================================
[SOURCE: https://en.wikipedia.org/wiki/TikTok] | [TOKENS: 18870]
Contents TikTok TikTok, known in mainland China, Macau, and Hong Kong as Douyin (Chinese: 抖音; pinyin: Dǒuyīn; lit. 'Shaking Sound'), is a social media and short-form online video platform. It hosts user-submitted videos, which range in duration from three seconds to 60 minutes. It can be accessed through a mobile app or through its website. Since its launch, TikTok has become one of the world's most popular social media platforms, using recommendation algorithms to connect content creators and influencers with new audiences. In April 2020, TikTok surpassed two billion mobile downloads worldwide. Cloudflare ranked TikTok the most popular website of 2021, surpassing Google. The popularity of TikTok has allowed viral trends in food, fashion, and music to take off and increase the platform's cultural impact worldwide. TikTok has come under scrutiny due to data privacy violations, mental health concerns, misinformation, offensive content, addictive algorithm, its role during the Gaza war, and, following its 2026 divestiture in the U.S., alleged censorship of criticism of Donald Trump and discussions of Jeffrey Epstein. While TikTok remains accessible to users in most countries, a minority of countries (including India and Afghanistan) have implemented full or partial bans. Many other countries limit TikTok's use on government-issued devices for security or privacy reasons. Corporate structure TikTok Ltd was incorporated in the Cayman Islands in the Caribbean and is based in both Singapore and Los Angeles. It owns entities which are based respectively in Australia (which also runs the New Zealand business), United Kingdom (also owns subsidiaries in the European Union), and Singapore (owns operations in Southeast Asia and India). A spin-off company, TikTok USDS Joint Venture LLC was formed on 22 January 2026 to handle TikTok and other ByteDance properties in the United States, Oracle Corporation, MGX Fund Management Limited, Silver Lake each holding a 15% stake, ByteDance holds a 19.9% stake and the remaining 35.1% is shared between Dell Technologies founder Michael Dell and Vastmere Strategic Investments. Its parent company, Beijing-based ByteDance, is owned by founders and Chinese investors, other global investors, and employees. One of ByteDance's main domestic subsidiaries is owned by Chinese state funds and entities through a 1% golden share. Employees have reported that multiple overlaps exist between TikTok and ByteDance in terms of personnel management and product development. TikTok says that since 2020, its US-based CEO is responsible for making important decisions, and has downplayed its China connection. History Douyin was launched on September 20, 2016, by ByteDance, originally under the name A.me, before changing its name to Douyin (抖音) in December 2016. Douyin was developed in nearly 7 months and within a year had 100 million users, with more than one billion videos viewed every day. While TikTok and Douyin share a similar user interface, the platforms operate separately. Douyin includes an in-video search feature that can search by people's faces for more videos of them, along with other features such as buying, booking hotels, and making geo-tagged reviews. ByteDance planned on Douyin expanding overseas. The founder of ByteDance, Zhang Yiming, stated that "China is home to only one-fifth of Internet users globally. If we don't expand on a global scale, we are bound to lose to peers eyeing the four-fifths. So, going global is a must." ByteDance created TikTok as an overseas version of Douyin. TikTok was launched in the international market in September 2017. On 9 November 2017, ByteDance spent nearly $1 billion to purchase Musical.ly, a startup headquartered in Shanghai with an overseas office in Santa Monica, California. Musical.ly was a social media video platform that allowed users to create short lip-sync and comedy videos, initially released in August 2014. TikTok merged with Musical.ly on 2 August 2018 with existing accounts and data consolidated into one app, keeping the title TikTok. On 23 January 2018, the TikTok app ranked first among free application downloads on app stores in Thailand and other countries. TikTok has been downloaded more than 130 million times in the United States and has reached 2 billion downloads worldwide, according to data from mobile research firm Sensor Tower (those numbers exclude Android users in China). In the United States, Jimmy Fallon, Tony Hawk, and other celebrities began using the app in 2018. Other celebrities like Jennifer Lopez, Jessica Alba, Will Smith, and Justin Bieber joined TikTok. In January 2019, TikTok allowed creators to embed merchandise sale links into their videos. On 3 September 2019, TikTok and the US National Football League (NFL) announced a multi-year partnership. The agreement came just two days before the NFL's 100th season kick-off at Soldier Field in Chicago where TikTok hosted activities for fans in honor of the deal. The partnership entails the launch of an official NFL TikTok account, which is to bring about new marketing opportunities such as sponsored videos and hashtag challenges. In July 2020, TikTok, excluding Douyin, reported close to 800 million monthly active users worldwide after less than four years of existence. In May 2021, TikTok appointed Shou Zi Chew as their new CEO who assumed the position from interim CEO Vanessa Pappas, following the resignation of Kevin A. Mayer on 27 August 2020. In September 2021, TikTok reported that it had reached 1 billion users. In 2021, TikTok earned $4 billion in advertising revenue. In October 2022, TikTok was reported to be planning an expansion into the e-commerce market in the US, following the launch of TikTok Shop in the United Kingdom. The company posted job listings for staff for a series of order fulfillment centers in the US and is reportedly planning to start the new live shopping business before the end of the year. The Financial Times reported that TikTok will launch a video gaming channel, but the report was denied in a statement to Digiday, with TikTok instead aiming to be a social hub for the gaming community. According to data from app analytics group Sensor Tower, advertising on TikTok in the US grew by 11% in March 2023, with companies including Pepsi, DoorDash, Amazon, and Apple among the top spenders. According to estimates from research group Insider Intelligence, TikTok is projected to generate $14.15 billion in revenue in 2023, up from $9.89 billion in 2022. In March 2024, The Wall Street Journal reported that TikTok's growth in the US had stagnated. Since at least 2020, following calls to ban TikTok in the country, the Committee on Foreign Investment in the United States (CFIUS) has been investigating the company's 2017 merger with Musical.ly but has not finalized any of its negotiations with TikTok, such as the Project Texas proposal, waiting instead for Congress to act. In January 2025, Chinese officials began preliminary talks about potentially selling TikTok's US operations to Elon Musk if the app faced an impending ban due to national security concerns. While Beijing preferred TikTok remain under ByteDance's control, the sale could happen through a competitive process or with US government involvement. One possibility involved Musk's platform, X, taking over TikTok's US business. The move came ahead of a Supreme Court case that upheld the constitutionality of a law that would force a sale or ban of TikTok in the US by 19 January 2025, due to national security concerns regarding its ties to China. Other potential buyers included Project Liberty's "The People's Bid For TikTok" consortium of Frank McCourt with Kevin O'Leary, Steven Mnuchin, MrBeast and Bobby Kotick, the seriousness of these potential buyers was unclear. The day before the impending ban, California-based conversational search engine company Perplexity AI submitted a bid for a merger with TikTok US. On September 14, 2025, the Wall Street Journal reported the US and China have reached the "framework of a deal" for the US operations of TikTok to be sold to a consortium of investors in the US including close Trump ally Larry Ellison of Oracle. The deal was completed by January 22, 2026, with investors including Oracle, Silver Lake, MGX, and others including the personal investment entity for Michael Dell owning more than 80% of the new venture, with ByteDance retaining 19.9% ownership. Under the deal, the app will remain the same, and the algorithm will be adjusted over time to favor American topics for those users. TikTok was downloaded over 104 million times on Apple's App Store during the first half of 2018, according to data provided to CNBC by Sensor Tower. After merging with musical.ly in August, downloads increased and TikTok became the most downloaded app in the US in October 2018, which musical.ly had done once before. In February 2019, TikTok, together with Douyin, hit one billion downloads globally, excluding Android installs in China. In 2019, media outlets cited TikTok as the 7th-most-downloaded mobile app of the decade, from 2010 to 2019. It was also the most-downloaded app on Apple's App Store in 2018 and 2019, surpassing Facebook, YouTube and Instagram. In September 2020, a deal was confirmed between ByteDance and Oracle in which the latter will serve as a partner to provide cloud hosting. In November 2020, TikTok signed a licensing deal with Sony Music. In December 2020, Warner Music Group signed a licensing deal with TikTok. The advertising revenue of short video clips is lower than other social media: while users spend more time, American audience is monetized at a rate of $0.31 per hour, a third the rate of Facebook and a fifth the rate of Instagram, $67 per year while Instagram will make more than $200. In July 2023, Iranian Mehr News Agency reported "experts from Douyin" will meet Iranian business in Tehran to enable Iranian exports to China. In 2023, several high-level executives transferred from ByteDance to TikTok to focus on moneymaking operations. Some moved from Beijing to the US. According to sources for The Wall Street Journal, the personnel move led to concerns from some TikTok employees and was reported to the office of US senator Ted Cruz for further investigation. In December 2023, TikTok invested $1.5 billion in GoTo's Indonesian e-commerce business, Tokopedia. In March 2024, The Information reported that it is an open secret among investors that TikTok loses billions of dollars annually. Although the size of its user base falls short of that of Facebook, Instagram, or YouTube, TikTok reached 1 billion active monthly users faster than any of them. Competition from TikTok prompted Instagram, which is owned by Facebook, to spend $120 million as of 2022 to entice more content creators to its Reels service, although engagement level remained low. Snapchat had likewise paid out $250 million in 2021 to its creators. Many platforms and services, including YouTube Shorts, began to imitate TikTok's format and recommendation page. Those changes caused a backlash from users of Instagram, Spotify, and Twitter. In March 2022, The Washington Post reported that Facebook's owner Meta Platforms paid Targeted Victory—a consulting firm backed by supporters of the US Republican Party—to coordinate lobbying and media campaigns against TikTok and portray it as "a danger to American children and society". Its efforts included asking local reporters to serve as "back channels" of anti-TikTok messages, writing opinion articles and letters to the editor, including one in the name of a concerned parent, amplifying stories about TikTok trends, such as "devious licks" and "Slap a Teacher", that actually originated on Facebook, and promoting Facebook's own corporate initiatives. Ties to Meta were not disclosed to the other parties involved. Targeted Victory said that it is "proud of the work". A Meta spokesperson said that all platforms, including TikTok, should face scrutiny. The Wall Street Journal reported that Silicon Valley executives met with US lawmakers to build an "anti-China alliance" before TikTok CEO's congressional hearing in March 2023. In April 2024, TikTok users started receiving notifications that their current and future picture posts would be shown on a new app called TikTok Notes. The app is not released yet; however, TikTok confirmed it is being worked on. TikTok Notes will be a direct competitor to Instagram for photo sharing. Jasmine Enberg, a principal social media analyst at eMarketer, observes that launching Notes as a separate app, instead of as a new feature in TikTok, may be done in response to regulatory and consumer scrutiny. On 18 April 2024, Notes first released to users in Canada and Australia for limited testing. On 1 April 2025, it was announced that Notes would be shut down on 8 May. It went defunct on May 8, 2025. Features The mobile app allows users to create short videos, which often feature music in the background and can be sped up, slowed down, or edited with a filter. They can also add their own sound on top of the background music. To create a music video with the app, users can choose background music from a wide variety of music genres, edit with a filter and record a 15-second video with speed adjustments before uploading it to share with others on TikTok or other social platforms. The "For You" page on TikTok is a endless feed of videos that are recommended to users based on their activity on the app. Content is curated by TikTok's artificial intelligence depending on the content a user liked, interacted with, or searched. This helps users find new content and creators reach new audiences, in contrast to other social networks that base recommendations on the interactions and relationships between users. TikTok's algorithm, recognized by The New York Times in 2020 as one of the most advanced for shaping user experiences and social interactions, stands out from traditional social media. While typical platforms focus on active user actions like likes, clicks, or follows, TikTok monitors a wider array of behaviors during video viewing. This comprehensive observation is then used to refine its algorithms, as noted by Wired in 2020. Furthermore, The Wall Street Journal in 2021 highlighted its superiority over other social media platforms in understanding users' preferences and emotions. TikTok's algorithm leverages this insight to present similar content, creating an environment that users often find hard to disengage from. The app's "react" feature allows users to film their reaction to a specific video, over which it is placed in a small window that is movable around the screen. Its "duet" feature allows users to film a video aside from another video. The "duet" feature was another trademark of Musical.ly. The duet feature is also only able to be used if both parties adjust the privacy settings. Videos that users do not want to post yet can be stored in their "drafts". The user is allowed to see their "drafts" and post when they find it fitting. The app allows users to set their accounts as "private". When first downloading the app, the user's account is public by default. The user can change to private in their settings. Private content remains visible to TikTok but is blocked from TikTok users who the account holder has not authorized to view their content. Users can choose whether any other user, or only their "friends", may interact with them through the app via comments, messages, or "react" or "duet" videos. Users also can set specific videos to either "public", "friends only", or "private" regardless if the account is private or not. Users may send their friends videos, emojis, and messages with direct messaging. TikTok has also included a feature to create a video based on the user's comments. Influencers often use the "live" feature. This feature is only available for those who have at least 1,000 followers and are over 16 years old. If over 18, the user's followers can send virtual "gifts" that can be later exchanged for money. TikTok announced a "family safety mode" in February 2020 for parents to be able to control their children's presence on the app. There is a screen time management option, restricted mode, and the option to put a limit on direct messages. The app expanded its parental controls feature called "Family Pairing" in September 2020 to provide parents and guardians with educational resources to understand what children on TikTok are exposed to. Content for the feature was created in partnership with online safety nonprofit, Internet Matters. In October 2021, TikTok launched a test feature that allows users to directly tip certain creators. Accounts of users that are of age, have at least 100,000 followers and agree to the terms can activate a "Tip" button on their profile, which allows followers to tip any amount, starting from $1. In December 2021, TikTok started beta-testing Live Studio, a streaming software that would let users broadcast applications open on their computers, including games. The software also launched with support for mobile and PC streaming. However, a few days later, users on Twitter discovered that the software uses code from the open-source OBS Studio. OBS made a statement saying that, under the GNU GPL version 2, TikTok has to make the code of Live Studio publicly available if it wants to use any code from OBS. In May 2022, TikTok announced TikTok Pulse, an ad revenue-sharing program. It covers the "top 4% of all videos on TikTok" and is only available to creators with more than 100,000 followers. If an eligible creator's video reaches the top 4%, they will receive a 50% share of the revenue from ads displayed with the video. In July 2023, TikTok launched a new streaming service called TikTok Music, currently available only in Brazil and Indonesia. This service allows users to listen to, download and share songs. It is reported that TikTok Music features songs from major record companies like Universal Music Group, Sony Music and Warner Music Group. On 19 July 2023, TikTok Music was expanded for select users in Australia, Mexico and Singapore. After a dispute with TikTok regarding payouts for artists and regulation of AI-generated music content on the platform, Universal Music Group decided not to renew its licensing agreement with TikTok, causing its catalogue of 3 million recordings to become unavailable for usage after 31 January 2024. This marked the company's first instance of withdrawing its music from a major platform, in contrast to Warner Music, which had recently renewed its own licensing deal with TikTok. In March 2024, Universal Music Publishing Group removed its catalogue of 4 million compositions from TikTok. In April 2024, Taylor Swift's music returned to the platform. In December 2025, TikTok launched digital gift cards. In February 2026, TikTok launched Local Feed, a feature intended to help users connect with their community and discover local events and updates. Content The app has spawned numerous viral trends, Internet celebrities, and music trends around the world. Duets, a feature that allows users to add their own video to an existing video with the original content's audio, have sparked many of these trends. Many stars got their start on Musical.ly, which merged with TikTok on 2 August 2018. These include Loren Gray, Baby Ariel, Zach King, Lisa and Lena, Jacob Sartorius, and many others. Loren Gray remained the most-followed individual on TikTok until Charli D'Amelio surpassed her on 25 March 2020. Gray's was the first TikTok account to reach 40 million followers on the platform. She was surpassed with 41.3 million followers. D'Amelio was the first to ever reach 50, 60, and 70 million followers. Charli D'Amelio remained the most-followed individual on the platform until she was surpassed by Khaby Lame on 23 June 2022. Other creators rose to fame after the platform merged with musical.ly on 2 August 2018. TikTok also played a major part in making "Old Town Road" by Lil Nas X one of the biggest songs of 2019 and the longest-running number-one song in the history of the US Billboard Hot 100. TikTok has allowed many music artists to gain a wider audience, often including foreign fans. For example, despite never having toured in Asia, the band Fitz and the Tantrums developed a large following in South Korea following the widespread popularity of their 2016 song "HandClap" on the platform. "Any Song" by R&B and rap artist Zico became number one on the Korean music charts due to the popularity of the #anysongchallenge, where users dance to the choreography of the song. The platform has also launched many songs that failed to garner initial commercial success into sleeper hits, particularly since the outbreak of the COVID-19 pandemic. However, it has received criticism for not paying royalties to artists whose music is used on the platform. Classic stars are able to connect with younger audiences born decades after a musician's first debut and across traditional genres. In 2020, Fleetwood Mac's "Dreams" was used in a skating video and a recreation by Mick Fleetwood. The song re-entered Billboard Hot 100 after 43 years and topped Apple Music. In 2022, Kate Bush's "Running Up That Hill" went viral among fans of Stranger Things, topping the UK singles chart 37 years after its original release. In 2023 Kylie Minogue's "Padam Padam" entered the Radio 1 playlist after being shared by Gen Z, even though many youth radio stations had refused to play it. Other older artists with strong engagement on TikTok include Elton John and Rod Stewart. In Japan, artists from the 1970s to 1990s, such as Kohmi Hirose, Yōko Oginome, Akina Nakamori, Seiko Matsuda, Momoe Yamaguchi and Saki Kubota, have become popular on TikTok during the Showa (and early Heisei) retro boom. In June 2020, TikTok users and K-pop fans "claimed to have registered potentially hundreds of thousands of tickets" for Donald Trump's campaign rally in Tulsa, Oklahoma through communication on TikTok, contributing to "rows of empty seats" at the event. Later, in October 2020, an organization called TikTok for Biden was created to support then-presidential candidate Joe Biden. After the election, the organization was renamed to Gen-Z for Change. On 10 August 2020, Emily Jacobssen wrote and sang "Ode to Remy", a song praising the protagonist from Pixar's 2007 computer-animated film Ratatouille. The song rose to popularity when musician Daniel Mertzlufft composed a backing track to the song. In response, began creating a "crowdsourced" project called Ratatouille the Musical. Since Mertzlufft's video, many new elements including costume design, additional songs, and a playbill have been created. On 1 January 2021, a full one-hour virtual presentation of Ratatouille the Musical premiered on TodayTix. It starred Titus Burgess as Remy, Wayne Brady as Django, Adam Lambert as Emile, Kevin Chamberlin as Gusteau, Andrew Barth Feldman as Linguini, Ashley Park as Colette, Priscilla Lopez as Mabel, Mary Testa as Skinner, and André De Shields as Ego. A viral TikTok trend known as "devious licks" involves students vandalizing or stealing school property and posting videos of the action on the platform. The trend has led to increasing school vandalism and subsequent measures taken by some schools to prevent damage. Some students have been arrested for participating in the trend. TikTok has taken measures to remove and prevent access to content displaying the trend. Another TikTok trend known as the Kia Challenge involves users stealing certain models of Kia and Hyundai cars manufactured without immobilizers, which was a standard feature at the time, between 2010 and 2021. As of February 2023, it had resulted in at least 14 crashes and eight deaths according to the National Highway Traffic Safety Administration. In May, Kia and Hyundai settled a $200-million class-action lawsuit by agreeing to provide software updates to affected vehicles and over 26,000 steering wheel locks. In 2023, a trend emerged where streamers acted as if they were video-game characters following prompts from their viewers. On Douyin, the Chinese version of TikTok, some celebrities who had garnered large followings as of August 2019 include Dilraba Dilmurat, Angelababy, Luo Zhixiang, Ouyang Nana, and Pan Changjiang. In the 2022 FIFA World Cup, a Qatari teenage royal became an Internet celebrity after his angry expressions were recorded in Qatar's opening match loss to Ecuador; he amassed more than 15 million followers in less than a week after creating a Douyin account. TikTok food trends refer to popular recipes and food-related fads on the social media platform TikTok. These trends amassed popularity in 2020 during the COVID-19 pandemic, as many people spent more time cooking at home while engaging with social media for entertainment. Food-related content on TikTok is often categorized under the hashtags #TikTokFood and #FoodTok. These hashtags have amassed 4.6 million and 4.5 million posts, respectively, according to the platform. Some TikTok users share personal recipes and dietary habits, while others use step-by-step cooking videos to grow their online presence. The widespread popularity of these trends has influenced various aspects of society, including interest in cooking among younger generations, discussions about body image, the marketing of food products on social media, and temporary food shortages. Several TikTok content creators, such as Eitan Bernath and Emily Mariko have gained recognition through their recipes and content. Some of the most notable TikTok food trends include the leftover salmon bowl, baked feta cheese pasta, and pesto eggs. "Midsize" fashion gained greater exposure on TikTok after many creators opened up about not able to find clothing sizes that fit them well. Women's apparel can roughly be divided into petite, straight, and plus sizes, leaving gaps in between. Realistic videos about how differently pieces of garment fit on a model compared to how they fit on a typical consumer resonated with many who had believed that they were alone in their struggle. Content promoting cosmetic surgery is popular on TikTok and has spawned several viral trends on the platform. In December 2021, Plastic and Reconstructive Surgery, the journal of the American Society of Plastic Surgeons, published an article about the popularity of some plastic surgeons on TikTok. In the article, it was noted that plastic surgeons were some of the earliest adopters of social media in the medical field and many had been recognized as influencers on the platform. The article published stats about the most popular plastic surgeons on TikTok up to February 2021 and at the time, five different plastic surgeons had surpassed 1 million followers on the platform. In 2021, it was reported that a trend known as the #NoseJobCheck trend was going viral on TikTok. TikTok content creators used a specific audio on their videos while showing how their noses looked before and after having their rhinoplasty surgeries. By January 2021, the hashtag #nosejob had accumulated 1.6 billion views, #nosejobcheck had accumulated 1 billion views, and the audio used in the #NoseJobCheck trend had been used in 120,000 videos. In 2020, Charli D'Amelio, the most-followed person on TikTok at the time, also made a #NoseJobCheck video to show the results of her surgery to repair a broken nose. In April 2022, NBC News reported that surgeons were giving influencers on the platform discounted or free cosmetic surgeries in order to advertise the procedures to their audiences. They also reported that facilities that offered these surgeries were also posting about them on TikTok. TikTok has banned the advertising of cosmetic surgeries on the platform but cosmetic surgeons are still able to reach large audiences using unpaid photo and video posts. NBC reported that videos using the hashtags '#plasticsurgery' and '#lipfiller' had amassed a combined 26 billion views on the platform. In December 2022, it was reported that a cosmetic surgery procedure known as buccal fat removal was going viral on the platform. The procedure involves surgically removing fat from the cheeks in order to give the face a slimmer and more chiseled appearance. Videos using hashtags related to buccal fat removal had collectively amassed over 180 million views. Some TikTok users criticized the trend for promoting an unobtainable beauty standard. A significant number of users on TikTok, such as Juju Green, create content surrounding Film criticism and Easter eggs. However, as reported by The New York Times, these people often do not see themselves necessarily as film critics. These creators would often attend red carpet premieres of movies and interview the celebrities in attendance, which was the subject of significant debate as some considered the questions the creators asked to be disrespectful. By 2022, TikTok released a Showbiz List, highlighting individuals who were having a larger impact on the film industry. During the 2023 SAG-AFTRA strike, such influencers were told that they would be denied future entry into the union if they partnered with struck studios. This led many creators to stop creating new content which they were not already contractually obligated to create. Creators who posted saying that they would not be changing their content, such as Green, were met with significant criticism. In March 2023, TikTok introduced a dedicated feed for science, technology, engineering, and mathematics (STEM) content. It works with Common Sense Networks to check for safety and age appropriateness and with the Poynter Institute for reliability of information. In January 2023, Forbes reported that a "heating" tool allows TikTok to manually promote certain videos, comprising 1–2% of daily views. The practice began as a way to grow and diversify content and influencers that were not automatically picked up by the recommendation algorithm. It was also used to promote brands, artists, and NGOs, such as the FIFA World Cup and Taylor Swift. However, some employees have abused it to promote their own accounts or those of their spouses, while others have felt that their guidelines leave too much room for discretion. TikTok said only a few individuals can approve heating in the US and the promoted videos take up less than 0.002% of user feeds. To address concerns of Chinese influence, the company is negotiating with the Committee on Foreign Investment in the United States (CFIUS) such that future heating could only be performed by vetted security personnel in the US and the process would be audited by third-parties such as Oracle. TikTok's and Douyin's censorship policies have been criticized as non-transparent. Internal guidelines against the promotion of violence, separatism, and "demonization of countries" could be used to prohibit content related to the 1989 Tiananmen Square protests and massacre, Falun Gong, human rights in Tibet, Taiwan, Chechnya, Northern Ireland, the Cambodian genocide, the 1998 Indonesian riots, Kurdish nationalism, ethnic conflicts between blacks and whites, or between different Islamic sects. A more specific list banned criticisms against world leaders, including past and present ones from Russia, the United States, Japan, North and South Korea, India, Indonesia, and Turkey. In 2019, The Guardian reported that TikTok had censored videos of topics not favored by the Chinese government. That year, TikTok took down a video about human rights abuses in the Xinjiang internment camps against Uyghurs but restored it after 50 minutes as well as the creator's account, saying that the action was a mistake and triggered by a brief "satirical" image of Osama bin Laden in another post. Other human rights activists have also said that their TikTok videos discussing human rights violations of the Uyghurs have been taken down. TikTok moderators were instructed to suppress posts from "For You" recommendations if the users shown were deemed "too ugly, poor, or disabled". The consumption of alcohol, full or partial nudity, LGBT, and intersex contents were restricted even in places where they are legal. TikTok has since apologized and instituted a ban against anti-LGBTQ ideology, but censorship continues on Douyin due to regulations in China. Douyin guidelines also forbid live broadcasting by unregistered foreigners, "feudal superstition", "money worship", smoking and drinking, competitive eating by the "already obese", "toxic" slime, "pornographic" ASMR such as ear-licking, and female anchors wearing revealing clothes. ByteDance said its early guidelines were global and aimed at reducing online harassment and divisiveness when its platforms were still growing. They have been replaced by versions customized by local teams for users in different regions. A March 2021 study by the Citizen Lab found that TikTok did not censor searches politically but was inconclusive about whether posts are. A 2023 paper by the Internet Governance Project at Georgia Institute of Technology concluded that TikTok is "not exporting censorship, either directly by blocking material, or indirectly via its recommendation algorithm." After increased scrutiny, TikTok said it is granting some outside experts access to the platform's anonymized data sets and protocols, including filters, keywords, criteria for heating, and source code. A December 2023 study by the Network Contagion Research Institute (NCRI) found a "strong possibility that content on TikTok is either amplified or suppressed based on its alignment with the interests of the Chinese government." According to its director, the NCRI is an independent non-profit research organization funded by Rutgers University, the British government, and private donors. The New York Times commented that "[a]lready, there is evidence that China uses TikTok as a propaganda tool. Posts related to subjects that the Chinese government wants to suppress — like Hong Kong protests and Tibet — are strangely missing from the platform." TikTok subsequently restricted the number of hashtags that can be searched under its Creative Center, saying it was "misused to draw inaccurate conclusions". A historian from the Cato Institute said that there were "basic errors" in the Rutgers University study and criticized the uncritical news coverage that followed. The study compares data from before TikTok even existed to show the app has fewer hashtags about historically sensitive topics, distorting the findings. In August 2024, the NCRI released a subsequent report based on user journey data from 24 accounts that they created across TikTok, Instagram, and YouTube. By searching for four keywords—Uyghur, Xinjiang, Tibet, and Tiananmen, the researchers found that TikTok returned a higher percentage of positive, neutral, or irrelevant content related to human rights in China. For example, more than 25% of results for "Tiananmen" on TikTok were considered pro-China by the researchers, compared to 16% on Instagram and 8% on YouTube. In other cases, however, Instagram and YouTube showed higher rates of pro-China content than TikTok. For example, 50% of searches about "Uyghur" and "Xinjiang" on YouTube were considered positive, compared to less than 25% on TikTok. The researchers said this is because some YouTube accounts are linked to state actors. According to their survey, people who use TikTok more than three hours daily are significantly more positive about China's human rights record compared to those who do not use the app. TikTok pushed back against the NCRI, saying that making "accounts that interact with the app in a prescribed manner" is not the same as the experience of real users and some of the events being compared happened before TikTok existed. Concerns have been voiced regarding content relating to, and the promotion and spreading of, hate speech and far-right extremism, such as antisemitism, islamophobia, racism, and xenophobia. Some videos were shown to expressly promote Holocaust denial and told viewers to take up arms and fight in the name of white supremacy and the swastika. As TikTok has gained popularity among young children, and the popularity of extremist and hateful content is growing, calls for tighter restrictions on their flexible boundaries have been made. TikTok has since released tougher parental controls to filter out inappropriate content and to ensure they can provide sufficient protection and security. In October 2019, TikTok removed about two dozen accounts that were responsible for posting ISIL propaganda and execution videos on the app. In Malaysia, TikTok is used by some users to engage in hate speech against race and religion especially mentioning the 13 May incident after the 2022 election. TikTok responded by taking down videos with content that violated their community guidelines. In March 2023, The Jewish Chronicle reported that TikTok still hosted videos that promoted the neo-Nazi propaganda film Europa: The Last Battle, despite having been alerted to the issue four months prior. TikTok said it removed and would continue to remove the content and associated accounts and has blocked the search term as well. In July 2024, the Institute for Strategic Dialogue reported that an organized neo-Nazi TikTok network promoting neo-Nazi propaganda, including Europa: The Last Battle, was receiving millions of views and was having its content promoted by TikTok's algorithm. In September 2024, Sky News reported that clips of Adolf Hitler's speeches with added music were attracting high levels of engagement on TikTok. Although they were removed by TikTok after the report, mixing audio remains an effective way to evade content moderation on many platforms. In July 2025, Media Matters reported that Google's Veo 3 text-to-video model for AI-generated content is being used to generate large numbers of dehumanizing and violent videos with racist and antisemitic tropes which are being shared on TikTok. In June 2021, TikTok made an apology after a shock video, showing a girl dancing which then cuts to a graphic scene of a man being beheaded by a saw, went viral. The video has been put on TikTok's blacklist, which detects it before being uploaded. TikTok has previously worked to remove graphic content from its platform, including a suicide video that circulated in September 2020, which had appeared among the recommended clips of TikTok's For You section. TikTok has banned Holocaust denial, but other conspiracy theories have become popular on the platform, such as Pizzagate and QAnon (two conspiracy theories popular among the US alt-right) whose hashtags reached almost 80 million views and 50 million views respectively by June 2020. The platform has also been used to spread misinformation about the COVID-19 pandemic, such as clips from Plandemic. TikTok removed some of these videos and has generally added links to accurate COVID-19 information on videos with tags related to the pandemic. In January 2020, left-leaning media watchdog Media Matters for America said that TikTok hosted misinformation related to the COVID-19 pandemic despite a recent policy against misinformation. In April 2020, the government of India asked TikTok to remove users posting misinformation related to the COVID-19 pandemic. There were also multiple conspiracy theories that the government is involved with the spread of the pandemic. It reported that in the second half of 2020, over 340,000 videos in the US about election misinformation and 50,000 videos of COVID-19 misinformation were removed. To combat misinformation in the 2022 midterm election in the US, TikTok announced a midterms Elections Center available in-app to users in 40 different languages. TikTok partnered with the National Association of Secretaries of State to give accurate local information to users. In September 2022, NewsGuard Technologies reported that among the TikTok searches it had conducted and analyzed from the US, 19.4% surfaced misinformation such as questionable or harmful content about COVID-19 vaccines, homemade remedies, the 2020 US elections, the Russian invasion of Ukraine, the Robb Elementary School shooting, and abortion. NewsGuard suggested that in contrast, results from Google were of higher quality. Mashable's own test from Australia found innocuous results after searching for "getting my COVID vaccine" but suggestions such as "climate change is a myth" after typing in "climate change". In November 2023, Singaporean Law and Home Affairs Minister K. Shanmugam applied for court orders requiring TikTok to provide information on the identity of three users that he accuses of spreading false and defamatory information about him. The users had shared on TikTok an article published by celebscritic.com alleging that K. Shanmugam was involved in an extra-marital affair. According to an investigation published in September 2025 by Moldovan newspaper Ziarul de Gardă, before the 2025 Moldovan parliamentary election, hundreds of accounts with false identities were created to spread Russian propaganda in TikTok and Facebook. That month, Bloomberg reported, citing European officials and documents of undisclosed origin, that Russia had prepared a plan to interfere in the election, which would have included a disinformation campaign in both Romanian and Russian on Facebook, Telegram and TikTok. In the run-up to the election, TikTok removed over 134,000 fake accounts, almost 2 million fake followers, 1,173 accounts impersonating Moldovan officials and over 9,300 videos that violated rules on civic integrity, disinformation and AI content generation; prevented 2.9 million fake likes and 1.8 million fake follow requests, also blocking the creation of over 268,000 spam accounts; and dismantled five coordinated networks with at least 7,593 accounts that "promoted pro-Russian politicians and attempted to discredit the current government". Furthermore, TikTok developed an in-app Electoral Center in partnership with the Central Electoral Commission of Moldova and worked with Reuters and the STOP FALS! Moldovan organization to fight, identify and educate about disinformation. As of 2022, TikTok is the 10th most popular app in Russia. After a new set of Russian war censorship laws was installed in March 2022, the company announced a series of restrictions on Russian and non-Russian posts and livestreams. Tracking Exposed, a user data rights group, learned of what was likely a technical glitch that became exploited by pro-Russia posters. It stated that although this and other loopholes were patched by TikTok before the end of March, the initial failure to correctly implement the restrictions, in addition to the effects from Kremlin's "fake news" laws, contributed to the formation of a "splInternet ... dominated by pro-war content" in Russia. TikTok said that it had removed 204 accounts for swaying public opinion about the war while obscuring their origins and that its fact checkers had removed 41,191 videos for violating its misinformation policies. In December 2023, BBC News reported that it had discovered nearly 800 fake TikTok accounts promoting Russian propaganda and disinformation. TikTok's own investigation found more than 12,000 fake accounts, including ones using additional languages such as English and Italian. In September 2024, TikTok removed the accounts of Russian state media outlets RT and Sputnik. The growth of popularity and access to TikTok has contributed to a growth in popularity of digital feminist movements and discourse originating from the platform. Digital spaces like TikTok enable marginalized communities and activists, such as feminists, to feel safer and have an easier place to engage in discussion and dialogue or build an identity which might otherwise be impossible due to circumstances. The momentum of digital feminist movements through platforms like TikTok have additionally encouraged many social media agents and marketing campaigns around the world to adopt some degree of feminism as a part of their online image or personal brand. TikTok's unique platform organization, of spontaneous peer-peer information sharing, has enabled its utilization for community-engaged, digital knowledge mobilization and exchange between social justice communities. However inversely enabled by the platform's organic potential, both feminist challenges and anti-feminist reinforcement of dominant social, hierarchical, and gender values are widespread and instigated through TikTok, and content labeled as anti-feminist is itself popularized on TikTok. Usage TikTok tends to appeal to younger users, as 41% of its users are between the ages of 16 and 24. As of 2021[update], these individuals are considered Generation Z. Among these TikTok users, 90% said they used the app daily. TikTok's geographical use in 2019 has shown that 43% of new users were from India before the social platform was banned in the country. But adults have also seen growth on TikTok. The share of US adults who regularly get news from TikTok hit 5% in 2025. By July 2023, TikTok has become the primary news source for British teenagers on social media, with 28% of 12 to 15-year-olds relying on the platform, while traditional sources like BBC One/Two are more trusted at 82%, according to a report by UK regulator Ofcom. As of the first quarter of 2022, there were over 100 million monthly active users in the United States and 23 million in the UK. The average user, daily, was spending 1 hour and 25 minutes on the app and opening TikTok 17 times. Out of TikTok's top 100 male creators, a 2022 analysis reported 67% were white, with 54% having near-perfect facial symmetry. China heavily regulates how Douyin is used by minors in the country, especially after 2018. Under government pressure, ByteDance introduced parental controls and a "teenage mode" that shows only whitelisted content, such as knowledge sharing, and bans pranks, superstition, dance clubs, and pro-LGBT content.[a] A mandatory screen time limit was put in place for users under the age of 14 and a requirement to link accounts to a real identity to prevent minors from lying about their age or using an adult's account. The differences between Douyin and TikTok have led some US politicians and commentators to accuse the company or the Chinese government of malicious intent. In March 2023, TikTok announced default screen time limits for users under the age of 18. Those under the age of 13 would need a passcode from their parents to extend their time. As with other platforms popular with children, underage users may inadvertently reveal their daily routine and whereabouts, raising concerns of potential misuse by sexual predators. At the time of reporting (2018), TikTok had only two privacy settings, either private or completely public, without any middle ground. Comment sections of "sexy" videos, such as young girls dancing in revealing clothes, were found to contain requests for nude pictures. Despite safety concerns, TikTok began rolling out photo, video, and voice messaging around 2025. As of January 2026, the features are only available in certain regions, including the United States and Canada. In recent years, the US has charged and sentenced sexual predators for illegal activities on TikTok against underage girls. On 22 January 2021, the Italian Data Protection Authority demanded that TikTok temporarily suspend Italian users whose age could not be established. The order came after the death of a 10-year-old Sicilian girl involved in an Internet challenge. TikTok asked its users in Italy to confirm again that they were over 13 years old. By May, over 500,000 accounts had been removed for failing the age check. In July 2021, the Dutch Data Protection Authority fined TikTok €750,000 for offering privacy statements only in English but not in Dutch. It noted that TikTok had implemented positive measures, such as forbidding direct messaging for users younger than 16 and allowing their parents to manage privacy settings directly through a paired family account, but the risk of children pretending to be older when creating their account remains. TikTok raised the minimum age for livestreaming from 16 to 18 after a BBC News investigation found hundreds of accounts going live from Syrian refugee camps. Thirty of them showed children begging for digital donation. TikTok reportedly made as much as a 70% commission on some of them, a figure that the company disputed. In March 2024, the Italian Competition Authority fined TikTok €10 million for not protecting underage users adequately from harmful content such as the "French scar" challenge, which left heavy pinch marks on a person's cheeks. On December 30, 2024, Venezuela's Supreme Court fined TikTok $10 million over viral challenges that authorities say led to the deaths of three children. The court cited TikTok's negligence in failing to implement "necessary and adequate measures" to prevent the viral video challenges. TikTok has provided a platform for users to create content not only for fun but also for money. As the platform has grown significantly over the past few years, it has allowed companies to advertise and rapidly reach their intended demographic through influencer marketing. The platform's algorithm also contributes to the influencer marketing potential, as it picks out content according to the user's preference. Sponsored content is not as prevalent on the platform as it is on other social media apps, but brands and influencers still can make as much as they would if not more in comparison to other platforms. Influencers on the platform who earn money through engagement, such as likes and comments, are referred to as "meme machines". In 2021, The New York Times reported that viral TikTok videos by young people relating the emotional impact of books on them, tagged with the label "BookTok", significantly drove sales of literature. Publishers were increasingly using the platform as a venue for influencer marketing. In December 2022, NBC News reported in a television segment that some TikTok and YouTube influencers were being given free and discounted cosmetic surgeries in order for them to advertise the surgeries to users of the platforms. In 2022, it was reported that a trend called "de-influencing" had become popular on the platform as a backlash to influencer marketing. TikTok creators participating in this trend made videos criticizing products promoted by influencers and asked their audiences not to buy products they did not need. However, some creators participating in the trend started promoting alternative products to their audiences and earning commission from sales made through their affiliate links in the same manner as the influencers they were originally criticizing. In June 2022, NBC News reported that some of the influencers paid by FeetFinder, a website that sells foot fetish content, did not disclose their videos were ads. FeetFinder said that it has suggested to influencers to be upfront about who was funding them. Existing sellers on FeetFinder said that the videos often misrepresented how "easy" it is to make money from posting feet pictures. Other TikTok creators have spoken out against accepting sponsorship deals indiscriminately and criticized those who posted undisclosed FeetFinder ads. In October 2020, the e-commerce platform Shopify added TikTok to its portfolio of social media platforms, allowing online merchants to sell their products directly to consumers on TikTok. Some small businesses have used TikTok to advertise and to reach an audience wider than the geographical region they would normally serve. The viral response to many small business TikTok videos has been attributed to TikTok's algorithm, which shows content that viewers at large are drawn to, but which they are unlikely to actively search for (such as videos on unconventional types of businesses, like beekeeping and logging). In 2020, digital media companies such as Group Nine Media and Global used TikTok increasingly, focusing on tactics such as brokering partnerships with TikTok influencers and developing branded content campaigns. Notable collaborations between larger brands and top TikTok influencers have included Chipotle's partnership with David Dobrik in May 2019 and Dunkin' Donuts' partnership with Charli D'Amelio in September 2020. TikTok is regularly used by sex workers to promote pornographic content sold on platforms such as OnlyFans. One porn actor posted a viral song referring to himself as an "accountant", starting a trend. In 2020, TikTok updated their terms of service to ban content that promotes "premium sexual content", impacting a large number of adult content creators. In response, they began substituting words in their captions and videos and using filters to censor explicit images. Some adult content creators have found a way to game TikTok's recommendation algorithm by posting riddles, attracting a large number of viewers that struggled to solve them. This increased potential Web traffic linked to the riddle posters' accounts on OnlyFans. The Israeli Defense Force (IDF) actively recruits influencers on TikTok and other social media platforms, often with what commentators have dubbed "Thirst traps". In a separate case in 2021, the IDF awarded a TikToker who was one of its military police officers for promoting the image of Israel. She had more followers than IDF spokesperson Ran Kochav or Prime Minister Benjamin Netanyahu. Since 2021, TikTok has created "election centres" on its platform leading up to European Parliament elections. About 30% of EP lawmakers use TikTok to get their messages across and to dispel misinformation. In February 2024, the re-election campaign for then-US president Joe Biden announced that it had opened a TikTok account while taking "advanced safety precautions". Biden posted his first video during Super Bowl LVIII. The move was criticized by a number of lawmakers over security concerns. Since 2022, the Biden administration had briefed TikTokers on news items such as the Russo-Ukrainian war and student debt relief in the United States. Many charities use TikTok for fundraising and education, especially with younger audiences, charities using TikTok include; Oregon Zoo, Shelter, Battersea Dogs and Cats Home, British Red Cross, American Heart Association, United Way, Catskill Animal Sanctuary and the Black Country Living Museum. Some TikTok influencers run their own events to raise money for charitable causes e.g. Mercury Stardust runs the TikTok-A-Thon for Trans Healthcare. 'TikTok For Good' was created by TikTok to support fundraising on the platform. In 2019, TikTok announced the #EduTok Mentorship program, a live workshop series in the Indian states of Bihar, Andhra Pradesh, Telangana, Rajasthan, Jharkhand and Jammu inspired by the hashtag #EduTok, in which video creators present factually for purposes of education. In 2022, TikTok banned fundraising for political accounts. Privacy and security concerns Privacy concerns have been brought up regarding the app. TikTok's privacy policy lists that the app collects usage information, IP addresses, a user's mobile carrier, unique device identifiers, keystroke patterns, and location data, among other data. Other information collected includes users inferred interests based on the content they view as well as content created by users. TikTok is also able to track information about web users even if they are not users of the TikTok app. It collects information such as IP address, online browsing habits and web search history. TikTok can share data with its corporate group, including ByteDance. The company says that it employs access control and approval process overseen by a US-based team. In June 2021, TikTok updated its privacy policy to include potential collection of biometric data, including "faceprints and voiceprints", for special effects and other purposes. The terms said that user authorization would be requested if local law demands such. Experts considered them to be "vague" and their implications "problematic" for the United States due to the country's general lack of robust data privacy laws. In a November 2022 update to its European privacy policy, TikTok stated that its global corporate group employees from China and other countries could gain remote access to the user information of accounts from Europe based on "demonstrated need". A March 2021 study by the Citizen Lab found that TikTok did not collect data beyond the industry norms, what its policy stated, or without additional user permission. In May 2023, The Wall Street Journal reported that former employees complained about TikTok tracking users who had viewed LGBT-related content. The company said its algorithm tracks interests not identity, and non-LGBT users also view such content. Concerns have been raised about the potential control and influence of the Chinese government over TikTok's owner, ByteDance, in particular the extraterritorial implications of China's 2017 National Intelligence Law. An article in the law insists that all organizations and citizens shall "support, assist and cooperate with national intelligence efforts." Analysts differ in their assessments of the data collection risks. Jim Lewis of the Center for Strategic and International Studies said TikTok would have no right to appeal requests for data made by the Chinese government. Some cybersecurity experts say individual users are not at risk. The United States has not offered any evidence of TikTok sharing such information with Chinese authorities. Keeping user data within the United States became the motivation behind TikTok's Project Texas. In October 2021, following the 2021 Facebook leak and controversies about social media ethics, a bipartisan group of United States lawmakers also pressed TikTok, YouTube, and Snapchat on questions of data privacy and moderation for age-appropriate content. Lawmakers also "hammered" TikTok about whether consumer data could be turned over to the Chinese government through ByteDance, its parent company in China. TikTok said it does not give information to China's government and "US user data" is stored within the country with backups in Singapore. In June 2022, BuzzFeed News reported that leaked audio recordings of internal TikTok meetings reveal employees in China had access to overseas data, including a "master admin" who could see "everything". Some of the recordings were made during consultations with Booz Allen Hamilton, a US government contractor. A spokesperson of the contractor said some of the report's information was inaccurate but would neither confirm nor deny whether TikTok was one of its clients. As a consequence, the Senate Intelligence Committee including US lawmakers Mark Warner and Marco Rubio called for the Federal Communications Commission (FCC) to investigate ByteDance and whether TikTok had misled them. Following the reports, TikTok confirmed that employees in China could have access to US data. It also announced that US user traffic would now be routed through Oracle Cloud and that backup copies would be deleted from other servers. In June 2022, FCC Commissioner Brendan Carr called for Google and Apple to remove TikTok from their app stores, saying sensitive data were being accessed from Beijing and ByteDance would be "required by law to comply with [Chinese government] surveillance demands." In November 2022, Christopher A. Wray, director of the Federal Bureau of Investigation (FBI), said the Chinese government could use TikTok for influence operations on its users. In May 2023, a former ByteDance employee filed a wrongful termination lawsuit alleging that Hong Kong users' device information and communications, particularly those of demonstrators in the 2019–2020 Hong Kong protests, were accessed by Chinese Communist Party members in 2018. ByteDance denied the claims, saying the employee worked on a defunct project and that TikTok was pulled out of Hong Kong in 2020. The whistleblower claimed in a sworn court statement that his father in mainland China had been detained by the authorities in retaliation for his speaking to the media about alleged censorship by TikTok. In June 2023, TikTok confirmed that some financial information, such as tax forms and Social Security numbers, of American content creators are stored in China. This applies to those signing contracts with and receiving payment transactions from ByteDance. Whether similar information will remain exempt from being treated as "protected user data" is being negotiated with Committee on Foreign Investment in the United States (CFIUS). A 2024 unclassified threat assessment by the Director of National Intelligence said "TikTok accounts run by a [Chinese] propaganda arm reportedly targeted candidates" during the 2022 United States elections. In April 2024, it was discovered that former employee Zen Goziker—allegedly the source of various leaks about TikTok to The Washington Post, Forbes, and BuzzFeed News—had made improbable claims. He has also spoken with law enforcement agencies and lawmakers hostile to TikTok. He has accused not only his former employer but also the Attorney General, the Director of National Intelligence, and the Department of Homeland Security for getting him fired. In response to security concerns of the United States government, TikTok has been working to silo privileged user data within the United States under oversight from the US government or a third party such as Oracle. Named Project Texas, the initiative focuses on unauthorized access, state influence, and software security. A new subsidiary, TikTok US Data Security Inc. (USDS), was created to manage user data, software code, back-end systems, and content moderation. It would report to the Committee on Foreign Investment in the United States (CFIUS), not ByteDance or TikTok, even for hiring practices. Oracle would review and spot check the data flows through USDS. It would also digitally sign software code, approve updates, and oversee content moderation and recommendation. Physical locations would be established so that Oracle and the US government could conduct their own reviews. The company has been engaged in confidential negotiations over the project with CFIUS since 2021 and submitted its proposal but received little response from the panel afterward. In March 2023, a former employee of the company said Project Texas did not go far enough and that a complete "re-engineering" would be needed. TikTok responded by saying that Project Texas already is a re-engineering of the app and that the former employee left in 2022 before the project specifications were finalized. Other former employees had their own takes on the situation. A data scientist said US user data were emailed to ByteDance workers in China to identify viewer interests. A manager recounted that there was a lot more separation on the technical side between TikTok and ByteDance by the time he left. Another said TikTok had to employ better data collection practices than Meta or Google due to the scrutiny it received. On 22 January 2026, US President Donald Trump and TikTok announced that the company had joined a joint venture, which will run as an independent entity while operating under defined safeguards that protect national security for US users. Adam Presser who was previously TikTok's head of operations and trust and safety will serve as CEO, while TikTok's current CEO Shou Chew will be a director of the new venture. The new investors are loyal to Trump. TikTok has subsequently been accused of censoring content critical of Trump and his administration. The censorship includes censoring words such as "Epstein" and videos that criticize Trump or Immigration and Customs Enforcement (ICE). The resulting backlash saw TikTok competitor UpScrolled receive record downloads and rise to second in the Apple's App Store rankings. TikTok has faced criticism for transferring European user data to servers in the United States. It is holding discussions with UK's National Cyber Security Centre about a "Project Clover" for storing European information locally. The company plans to build two data centers in Ireland and one more in Norway. A third party will oversee the cybersecurity policies, data flows, and personnel access independently of TikTok. In October 2022, Forbes reported that a team at ByteDance planned to surveil certain US citizens for undisclosed reasons. TikTok said that the tracking method suggested by the report would not be feasible because precise GPS information is not collected by the platform. In December 2022, ByteDance confirmed after internal investigation that the data of two journalists and their close contacts had been accessed by its employees from China and the United States. It was intended to uncover sources of leaks who might have met with the journalists from Forbes and the Financial Times. The data included IP addresses, which can be used to approximate a user's location. ByteDance stated that it fired four employees in response. The incident is being investigated by the US Department of Justice and FBI. The US Attorney for the Eastern District of Virginia reportedly subpoenaed information from ByteDance regarding its surveillance of journalists on TikTok. In December 2023, the United States House Select Committee on Strategic Competition between the United States and the Chinese Communist Party inquired the FBI about the status of the case. In January 2020, Check Point Research discovered a vulnerability through which a hacker could spoof TikTok's official SMS messages and replace them with malicious links to gain access to user accounts. It was later patched by TikTok. In August 2020, The Wall Street Journal reported that TikTok tracked Android user data, including MAC addresses and IMEIs, with a tactic in violation of Google's policies. In August 2022, software engineer and security researcher Felix Krause found that in-app browsers from TikTok and other platforms contained codes for keylogger functionality but did not have the means to further investigate whether any data was tracked or recorded. TikTok said that the code is disabled. On 27 February 2019, the United States Federal Trade Commission (FTC) reached a consent decree with ByteDance, fining it US$5.7 million for collecting information from minors under the age of 13 in violation of the Children's Online Privacy Protection Act (COPPA). ByteDance responded by adding a kids-only mode to TikTok which blocks the upload of videos, the building of user profiles, direct messaging, and commenting on others' videos, while still allowing the viewing and recording of content. In May 2020, an advocacy group filed a complaint with the FTC saying that TikTok had violated the terms of the February 2019 consent decree with the FTC, which sparked subsequent congressional calls for a renewed FTC investigation. In March 2022, following a class action lawsuit for violations of COPPA, TikTok settled for US$1.1 million. In March 2024, it was reported that the FTC continues to investigate TikTok. In August 2024, the FTC and US Department of Justice filed a joint lawsuit alleging violations of the 2019 consent decree. In September 2021, the Ireland Data Protection Commission (DPC) launched investigations into TikTok concerning the protection of minors' data and transfers of personal data to China. The Irish DPC became the lead agency to handle such matters after TikTok established an office in the country, taking over investigations started by Dutch and Italian authorities. In September 2023, the DPC fined TikTok €345 million for violations of the General Data Protection Regulation (GDPR) vis-à-vis the mishandling of children data. In April 2025, the Ireland DPC fined TikTok over €500 million for illegally sending European user data to China. In July 2025, the DPC opened a new investigation into TikTok for unauthorized transfers of user data to China. In February 2019, the United Kingdom's Information Commissioner's Office (ICO) launched an investigation of TikTok following the fine ByteDance received from the United States Federal Trade Commission (FTC). Speaking to a parliamentary committee, Information Commissioner Elizabeth Denham said that the investigation focuses on the issues of private data collection, the kind of videos collected and shared by children online, as well as the platform's open messaging system which allows any adult to message any child. She noted that the company was potentially violating the GDPR which requires the company to provide different services and different protections for children. In April 2023, the ICO imposed a £12.7 million fine on TikTok for misusing children's data. In March 2025, the ICO opened another investigation into TikTok concerning its use of children's personal information to recommend content to them. In February 2022, Texas Attorney General Ken Paxton initiated an investigation into TikTok for alleged violations of children's privacy and facilitation of human trafficking. Paxton claimed that the Texas Department of Public Safety gathered several pieces of content showing the attempted recruitment of teenagers to smuggle people or goods across the Mexico–United States border. He claimed the evidence may prove the company's involvement in "human smuggling, sex trafficking and drug trafficking". The company claimed that no illegal activity of any kind is supported on the platform. In 2022, Turkey's Financial Crimes Investigation Board (MASAK) initiated a probe into TikTok in relation to millions of dollars in fund transfers involving TikTok accounts that were suspected of money laundering or terrorism financing. In February 2023, the Privacy Commissioner of Canada, along with its counterparts in Alberta, British Columbia, and Quebec, launched an investigation into TikTok's data collection practices. In 2025, a Canadian investigation found that TikTok had collected sensitive personal data from children in violation of federal privacy laws. In February 2024, the European Commission launched an investigation into TikTok for potential violations of the Digital Services Act (DSA), involving content aiming at children and advertising transparency. In April 2024, the European Commission opened a second investigation into TikTok to assess whether it broke EU law. In October 2024, the European Commission requested additional information from TikTok relating to its algorithm and risks around elections, mental health, and protection of minors. In December 2024, the European Commission announced an investigation into TikTok over accusations of Russian interference in the 2024 Romanian presidential election. In May 2025, the European Commission found TikTok had violated digital advertising rules under the DSA. In December 2023, the Office of the Australian Information Commissioner announced an inquiry into TikTok's data harvesting of Australian citizens amid allegations that it contravened Australian privacy law. Controversies Vox noted in 2018 that bullies and trolls were relatively rare on TikTok compared to other platforms. Nonetheless, several users have reported cyberbullying via features such as Duet or React, which is used to interact with followers. A trend making fun of autism eventually created a huge backlash, even on the platform itself, and the company ended up removing the hashtag altogether. Parents filming how their children reacted to people with disability, often in terror, led to criticisms of ableism.[citation needed] In December 2019, following a report by German digital rights group netzpolitik.org, TikTok admitted that it had suppressed videos by disabled users as well as LGBTQ+ users in a purported temporary effort to limit cyberbullying. TikTok: Murder Gone Viral, a documentary series produced by ITV, highlights how cyberbullying and TikTok's role as a social media platform has led to the murder of certain individuals. There are concerns that some users may find it hard to stop using TikTok. Internal TikTok research has documented the addiction potential of the app. In April 2018, an addiction-reduction feature was added to Douyin. This encouraged users to take a break every 90 minutes. Later in 2018, the feature was rolled out to the TikTok app. TikTok uses popular influencers to encourage viewers to stop using the app and take a break. Many were also concerned with the app affecting users' attention spans due to the short-form nature of the content. This is a concern as many of TikTok's audience are younger children, whose brains are still developing. TikTok executives and representatives have noted and made aware to advertisers on the platform that users have poor attention spans. The company's survey reported that nearly 50% of social media users find it stressful to watch a video longer than a minute and a third of users watch videos at double speed. Their short attention spans posed a challenge for TikTok to pivot towards longer content formats. TikTok has also received criticism for enabling children to purchase coins which they can send to other users. Daily hours of entertainment screen media (Social Medias) may displace healthy behaviors such as socializing face to face, chores, hobbies, homework, family meals time, exercise, and sufficient sleep. Insomnia is considered a strong mediator between screen media time and mental health symptoms which implies that engaging in screen time pushes out adequate sleep and leads to decreased mental health. In February 2022, The Wall Street Journal reported that "Mental-health professionals around the country are growing increasingly concerned about the effects on teen girls of posting sexualized TikTok videos." In March 2022, a coalition of US state attorneys general launched an investigation into TikTok's effect on children's mental health. In June 2022, TikTok introduced the ability to set a maximum uninterrupted screen time allowance, after which the app blocks off the ability to navigate the feed. The block only lifts after the app is exited and left unused for a set period of time. Additionally, the app features a dashboard with statistics on how often the app is opened, how much time is spent browsing it and when the browsing occurs. Since 2021, it has been reported that accounts engaging with contents related to suicide, self-harm, or eating disorders were shown more similar videos. Some users were able to circumvent TikTok filters by writing in code or using unconventional spelling. The company has faced multiple lawsuits pertaining to wrongful deaths. TikTok said it is working to break up these "rabbit holes" of similar recommendations. US searches for eating disorder receive a prompt that offers mental health resources. In 2021, the platform revealed that it will be introducing a feature that will prevent teenagers from receiving notifications past their bedtime. The company will no longer send push notifications after 9 pm to users aged between 13 and 15. For 16 to 17 year olds, notifications will not be sent after 10 pm. In March 2023, TikTok announced default screen time limits for users under the age of 18. The Wall Street Journal has reported that doctors experienced a surge in reported cases of tics, tied to an increasing number of TikTok videos from content creators with Tourette syndrome. Doctors suggested that the cause may be a social one as users who consumed content showcasing various tics would sometimes develop tics of their own, akin to mass psychogenic illness. In May 2024, Nebraska Attorney General Mike Hilgers filed a lawsuit against TikTok for allegedly harming minors' mental health through an algorithm designed to be cultivate compulsive behavior. In October 2024, US senators Richard Blumenthal and Marsha Blackburn requested that TikTok turn over "all documents and information" related to child safety disclosures that were uncovered by NPR and Kentucky Public Radio. As of March 2025[update], strong scientific understanding of TikTok's effects on user's mental health "remains elusive". A 2025 meta-analysis found that use of TikTok was correlated with symptoms of anxiety and depression, with stronger links in females and users under 24 years old. In November 2025, French prosecutors opened a probe into TikTok, citing concerns that its algorithms could push vulnerable young people to suicide. The Ministry of Education (Taiwan) has warned against the harmful impact of TikTok and related app RedNote on body image and ensuing self harm related to weight loss especially in teenagers. Civil society groups warn that these apps push users down "rabbit holes" and then surround them in an "information cocoon". In November 2022, Australia's medical regulatory agency, the Therapeutic Goods Administration (TGA) reported that there was a global shortage of the diabetes medication Ozempic. According to the TGA, the rise in demand was caused by an increase in off-label prescription of the drug for weight loss purposes. In December 2022, with the United States experiencing a shortage as well, it was reported that the huge increase in demand for the medicine was caused by a weight loss trend on TikTok, where videos about the drug exceeded 360 million views. Wegovy, a drug that has been specifically approved for treating obesity, also became popular on the platform after Elon Musk credited it for helping him lose weight. Several former employees of the company have claimed of poor workplace conditions, including the start of the workweek on Sunday to cooperate with Chinese time zones and excessive workload. Employees claimed they averaged 85 hours of meetings per week and would frequently stay up all night in order to complete tasks. Some employees claimed the workplace's schedule operated similarly to the 996 schedule. The company has a stated policy of working from 10 AM to 7 PM five days per week (63 hours per week), but employees noted that it was encouraged for employees to work after hours. One female worker complained that the company did not allow her adequate time to change her feminine hygiene product because of back-to-back meetings. Another employee noted that working at the company caused her to seek marriage therapy and lose an unhealthy amount of weight. In response to the allegations, the company noted that they were committed to allowing employees "support and flexibility". In September 2023, two former ByteDance employees filed a formal complaint with the US Equal Employment Opportunity Commission (EEOC) asking the EEOC to investigate TikTok's practice of retaliation against workers who complain about discrimination. With reports that Palestinians resorted to TikTok for promoting their cause after platforms like Facebook and Twitter blocked their content, Israeli analyst Yoni Ben-Menachem called the app a "tool of dangerous influence" inciting violence against Israelis. According to Ynet, the Palestinian militant group Lion's Den gained much of their popularity through TikTok. In February 2023, Otzma Yehudit politician Almog Cohen advocated blocking TikTok for all of East Jerusalem. US lawmakers wanting to ban TikTok accused the platform of pushing pro-Hamas and pro-Palestine content. According to The Times of Israel, antisemitism at the company was "rampant" after the October 7 attacks, allowing anti-Jewish and anti-Israel content to increase on the platform. Prominent Jewish individuals such as Sacha Baron Cohen, Debra Messing, Amy Schumer, and TikTok creator Miriam Ezagui raised the issue with Adam Presser, TikTok's head of operations, and Seth Melnick, its global head of user operations, both also Jewish. TikTok said that a significant proportion of its userbase comes from non-US regions such as the Middle East and Southeast Asia and that hashtags should not be cherry-picked due to differences in the number of views per post and the age of a post or tag. The popularity of pro-Palestine content has also been explained by the app's younger user base, which has shifted its sympathy away from Israel towards the Palestinians. The Jewish Federations of North America expressed support for TikTok to be banned, while Israel's critics denounced the "criminalisation of pro-Palestinian voices", including on TikTok, which has been used to condemn "Israel's atrocities", according to The New Arab. TikTok was also accused by Malaysia's minister of communications, Fahmi Fadzil of suppressing pro-Palestinian content. The company stated it banned praising Hamas and removed more than 775,000 videos and 14,000 livestreams. There has also been reporting on the trend of Israeli soldiers using TikTok to mock Palestinians and boast about their actions in the Gaza Strip. According to Huda Abudagga, a legal advisor for the British nonprofit Law for Palestine, many of the videos show soldiers partaking in actions, such as the destruction of Palestinian homes as well as the theft of private and public property, that are considered war crimes under international law. Similarly, four legal experts consulted by the New York Times said that the videos "could be used to show unlawful destruction, a violation of the Geneva Conventions." One such video, which showed Israeli soldiers dancing and singing “There are no uninvolved civilians,” was included as evidence of genocidal intent in South Africa's case against Israel before the International Court of Justice. In November 2023, Osama bin Laden's 2002 "Letter to the American people" went viral on TikTok and other social media. In the letter, he denounced the US and its support for Israel, and supported al-Qaeda's war against the US as a defensive struggle. Numerous social media users, including Americans, expressed their opposition to US foreign policy by sharing the resurfaced copies of the letter and its contents. The Guardian website removed the letter after displaying it for more than 20 years, and TikTok began issuing takedowns of videos featuring the letter. Reporting in The Washington Post suggested that the virality of the letter had been limited prior to media coverage, having never trended on TikTok. Many of the TikTok videos covering the letter were critical of bin Laden, and media coverage had exaggerated its significance while elevating the virality of the letter. In July 2025, TikTok hired Erica Mindel, a former Israel Defense Forces instructor and contractor for the US State Department's Special Envoy to Monitor and Combat Antisemitism, as its Public Policy Manager of Hate Speech. Starting in 2026, TikTok began cracking down on the accounts of Palestinian activists and journalists, including Bisan Owda, who had reported on the human rights abuses taking place in Gaza. Media outlets such as Al Jazeera verified that TikTok's new algorithm hid content depending on the user's region, with the app hiding Palestinian content from users in the Middle East. An investigation in 2025 found that TikTok was profiting from sexual livestreams in Kenya involving minors, with teenagers as young as 15 using the platform to solicit explicit content. Women in Kenya reported earning money through TikTok Lives, where coded sexual slang and emoji gifts facilitated transactions, with explicit content often delivered via other platforms. TikTok takes a 70% cut of these livestream earnings and has been aware of child exploitation since at least 2022. Many moderators say the company's content policies are ineffective, and digital pimps exploit underage users on the large. Kenya lacks adequate moderation, and TikTok denies any sort of wrongdoing. In response to the exposé, Kenya's Communications Authority (CA) launched a formal inquiry, directing TikTok to remove all sexual content involving minors and submit a detailed plan to strengthen its moderation and child protection systems. According to The Washington Post, Meta hired the Republican consulting firm Targeted Victory to run a campaign aimed at turning public opinion against TikTok. Internal emails revealed that the firm sought to portray TikTok as "the real threat" and encouraged headlines such as "From dances to danger: how TikTok has become the most harmful social media space for kids." Operatives promoted stories to local media that tied TikTok to allegedly dangerous trends among teenagers, including the 2021 "devious lick" vandalism challenge, which evidence shows originated on Facebook. The campaign also sought to deflect attention from criticisms of Meta's own privacy and antitrust issues. A Meta spokesperson defended the effort, saying that all platforms, including TikTok, should face equal scrutiny. An analysis estimated that Meta, the parent company of Facebook and Instagram, could gain between $2.46 billion and $3.38 billion in advertising revenue if TikTok were banned. In 2026 TikTok was criticized for allegedly censoring criticism of Donald Trump and references to Jeffrey Epstein, with whom Trump had a long-standing relationship. California governor Gavin Newsom said the Government of California was launching an investigation of TikTok for violating California law by censoring Trump-critical content. TikTok claimed it was not deliberately censoring content and blamed alleged power outages. Restrictions and bans On 21 December 2024, Albanian Prime Minister Edi Rama announced that the Albanian government will shut down TikTok in 2025 for at least a year, following a deadly incident in November 2024 in which a teenager fatally stabbed another teen after a dispute that began on the platform. On 7 March 2025, the shutdown was officially enacted after the Albanian Cabinet cited concerns over the app's role in promoting violence and bullying among children. On 6 November 2024, Canada ordered TikTok to shut down its offices and subsidiary company (TikTok Technology Canada, Inc.) in the country due to national security concerns, but access to the app was not banned. Users will still be able to access the video app and upload content to it. In February 2023, the European Parliament, the European Commission, and the Council of the European Union, have all banned TikTok on staff devices, citing cybersecurity concerns. In 2020, TikTok was banned indefinitely in India after the country had a border clash with China. In January 2020, the United States Army and Navy banned TikTok on government devices after the Defense Department labeled it a security risk. Recruiters had been using the app to help fill quotas, and some continue to maintain a level of engagement through their personal accounts. According to a 2020 article in The New York Times, Central Intelligence Agency analysts determined that while it is possible the Chinese government could obtain user information from the app, there was no evidence it had done so. On 6 August 2020, US President Donald Trump signed an order which would ban TikTok transactions in 45 days if it was not sold by ByteDance. On 14 August 2020, Trump issued another order giving ByteDance 90 days to sell or spin off its US TikTok business. In the order, Trump said that there is "credible evidence" that leads him to believe that ByteDance "might take action that threatens to impair the national security of the United States". In June 2021, US President Joe Biden signed an executive order revoking the Trump administration ban on TikTok, and instead ordered the Secretary of Commerce to investigate the app to determine if it poses a threat to US national security. On 27 December 2022, the Chief Administrative Officer of the United States House of Representatives banned TikTok from all devices managed by the House of Representatives. On 30 December 2022, President Joe Biden signed the No TikTok on Government Devices Act, prohibiting the use of the app on devices owned by the federal government, with some exceptions. On 13 March 2024, the United States House of Representatives passed H.R. 7521, which would ban TikTok entirely unless it was divested from its Chinese parent company, ByteDance. In April, the House of Representatives included a revised version of the bill in a foreign aid package, which was passed by the Senate on 23 April 2024, and signed into law by President Joe Biden the following day. The law was challenged in TikTok, Inc. v. Garland but was upheld as constitutional. In response to the potential ban, many users signed up for the Chinese app Xiaohongshu (Chinese: 小红书), known internationally as REDnote. On 18 January 2025, hours before the bill went into effect, TikTok became unavailable across the country. The next day, TikTok restored access to their service after re-elected US President Donald Trump assured TikTok he would not enforce the law. President Trump signed an executive order on 20 January 2025, delaying the enforcement of the TikTok ban by 75 days. In April 2025, President Trump signed another executive order further delaying the enforcement of the TikTok ban by 75 days, and did so again in June 2025 with a 90-day extension. As of February 2023, at least 32 (of 50) states have announced or enacted bans on state government agencies, employees, and contractors using TikTok on government-issued devices. State bans only affect government employees and do not prohibit civilians from having or using the app on their personal devices. Critics say the United States itself surveils individuals abroad via tech companies under FISA laws. Data collected by TikTok and other social networks can already be purchased through other means. Some theorize that, if passed, H.R. 7521 could "embolden authoritarian censorship" of American Internet companies and affect US interests, reputation, and online speech. They have also labeled a potential ban on the app an assault on freedom of speech, including Republican congressmen Rand Paul and Thomas Massie. Observers have argued that the national security concerns raised are largely hypothetical. There is insufficient public evidence to show that American user data has been accessed by or shared with the PRC government, with some claims reportedly exaggerated. Biden himself was on TikTok as the president, while Trump has reversed his previous position. According to computer security specialist Bruce Schneier, which company owns TikTok may not matter, as Russia had interfered in the 2016 US elections using Facebook without owning it. Partnerships In April 2021, the Abu Dhabi Department of Culture and Tourism partnered with TikTok to promote tourism for the city. It came following the January 2021 winter campaign with the United Arab Emirates Government Media Office. In June 2023, The New Zealand Herald reported that TikTok, working in cooperation with both New Zealand and Australian police, deleted 340 accounts and 2,000 videos associated with criminal gangs including the Mongrel Mob, Black Power, Killer Beez, the Comancheros, Mongols, and Rebels. TikTok had earlier drawn criticism for hosting content by organized crime groups promoting the gang lifestyle and fights. A TikTok spokesperson reiterated the platform's efforts to countering "violent" and "hateful" organizations' content and cooperating with police. New Zealand Police Commissioner Andrew Coster praised the platform for taking a "socially-responsible stance" against gangs. TikTok has partnered with the Hispanic Heritage Foundation to support small Latino businesses, setting aside $5000 each for 40 grant recipients based on entrepreneurship. After digital advertising rules for the Olympics were relaxed, TikTok and Team GB signed a sponsorship deal to help UK athletes connect with new audiences for the 2024 Summer Olympics. Starting in 2021, TikTok became the primary sponsor/partner of the English Football League club Wrexham A.F.C. located in Wrexham, Wales. A large version of the TikTok logo was emblazoned on the front of the player's red coloured home and away uniforms below the Wrexham A.F.C. crest as well as on shirts sold by Wrexham's brick and mortar and virtual fan stores, the partnership ended in 2023 when Wrexham was promoted to the EFL League Two after which US airline United took over the partnership/sponsorship. On 8 January 2026, TikTok announced a new partnership with the FIFA World Cup. In early 2026, the United States and China approved a deal transferring operational control of TikTok's US business to a joint venture led by Oracle and Silver Lake, following a 2024 US law requiring ByteDance to divest the platform over national security concerns. Shou Chew announced in a memo addressed to employees that TikTok will henceforth operate in the United States under a joint investment entity in which the majority ownership will be American. Under the agreement reached, it is guaranteed that Americans will be protected on the social network TikTok through “data protection,” “content oversight,” and “algorithm security.” See also Notes References Further reading External links
========================================
[SOURCE: https://en.wikipedia.org/wiki/Birthday#cite_ref-21] | [TOKENS: 4101]
Contents Birthday A birthday is the anniversary of the birth of a person or the figurative birth of an institution. Birthdays of people are celebrated in numerous cultures, often with birthday gifts, birthday cards, a birthday party, or a rite of passage. Many religions celebrate the birth of their founders or religious figures with special holidays (e.g. Christmas, Mawlid, Buddha's Birthday, Krishna Janmashtami, and Gurpurb). There is a distinction between birthday and birthdate (also known as date of birth): the former, except for February 29, occurs each year (e.g. January 15), while the latter is the complete date when a person was born (e.g. January 15, 2001). Coming of age In most legal systems, one becomes a legal adult on a particular birthday when they reach the age of majority (usually between 12 and 21), and reaching age-specific milestones confers particular rights and responsibilities. At certain ages, one may become eligible to leave full-time education, become subject to military conscription or to enlist in the military, to consent to sexual intercourse, to marry with parental consent, to marry without parental consent, to vote, to run for elected office, to legally purchase (or consume) alcohol and tobacco products, to purchase lottery tickets, or to obtain a driver's licence. The age of majority is when minors cease to legally be considered children and assume control over their persons, actions, and decisions, thereby terminating the legal control and responsibilities of their parents or guardians over and for them. Most countries set the age of majority at 18, though it varies by jurisdiction. Many cultures celebrate a coming of age birthday when a person reaches a particular year of life. Some cultures celebrate landmark birthdays in early life or old age. In many cultures and jurisdictions, if a person's real birthday is unknown (for example, if they are an orphan), their birthday may be adopted or assigned to a specific day of the year, such as January 1. Racehorses are reckoned to become one year old in the year following their birth on January 1 in the Northern Hemisphere and August 1 in the Southern Hemisphere.[relevant?] Birthday parties In certain parts of the world, an individual's birthday is celebrated by a party featuring a specially made cake. Presents are bestowed on the individual by the guests appropriate to their age. Other birthday activities may include entertainment (sometimes by a hired professional, i.e., a clown, magician, or musician) and a special toast or speech by the birthday celebrant. The last stanza of Patty Hill's and Mildred Hill's famous song, "Good Morning to You" (unofficially titled "Happy Birthday to You") is typically sung by the guests at some point in the proceedings. In some countries, a piñata takes the place of a cake. The birthday cake may be decorated with lettering and the person's age, or studded with the same number of lit candles as the age of the individual. The celebrated individual may make a silent wish and attempt to blow out the candles in one breath; if successful, superstition holds that the wish will be granted. In many cultures, the wish must be kept secret or it will not "come true". Birthdays as holidays Historically significant people's birthdays, such as national heroes or founders, are often commemorated by an official holiday marking the anniversary of their birth. Some notables, particularly monarchs, have an official birthday on a fixed day of the year, which may not necessarily match the day of their birth, but on which celebrations are held. In Mahayana Buddhism, many monasteries celebrate the anniversary of Buddha's birth, usually in a highly formal, ritualized manner. They treat Buddha's statue as if it was Buddha himself as if he were alive; bathing, and "feeding" him. Jesus Christ's traditional birthday is celebrated as Christmas Eve or Christmas Day around the world, on December 24 or 25, respectively. As some Eastern churches use the Julian calendar, December 25 will fall on January 7 in the Gregorian calendar. These dates are traditional and have no connection with Jesus's actual birthday, which is not recorded in the Gospels. Similarly, the birthdays of the Virgin Mary and John the Baptist are liturgically celebrated on September 8 and June 24, especially in the Roman Catholic and Eastern Orthodox traditions (although for those Eastern Orthodox churches using the Julian calendar the corresponding Gregorian dates are September 21 and July 7 respectively). As with Christmas, the dates of these celebrations are traditional and probably have no connection with the actual birthdays of these individuals. Catholic saints are remembered by a liturgical feast on the anniversary of their "birth" into heaven a.k.a. their day of death. In Hinduism, Ganesh Chaturthi is a festival celebrating the birth of the elephant-headed deity Ganesha in extensive community celebrations and at home. Figurines of Ganesha are made for the holiday and are widely sold. Sikhs celebrate the anniversary of the birth of Guru Nanak and other Sikh gurus, which is known as Gurpurb. Mawlid is the anniversary of the birth of Muhammad and is celebrated on the 12th or 17th day of Rabi' al-awwal by adherents of Sunni and Shia Islam respectively. These are the two most commonly accepted dates of birth of Muhammad. However, there is much controversy regarding the permissibility of celebrating Mawlid, as some Muslims judge the custom as an unacceptable practice according to Islamic tradition. In Iran, Mother's Day is celebrated on the birthday of Fatima al-Zahra, the daughter of Muhammad. Banners reading Ya Fatima ("O Fatima") are displayed on government buildings, private buildings, public streets and car windows. Religious views In Judaism, rabbis are divided about celebrating this custom, although the majority of the faithful accept it. In the Torah, the only mention of a birthday is the celebration of Pharaoh's birthday in Egypt (Genesis 40:20). Although the birthday of Jesus of Nazareth is celebrated as a Christian holiday on December 25, historically the celebrating of an individual person's birthday has been subject to theological debate. Early Christians, notes The World Book Encyclopedia, "considered the celebration of anyone's birth to be a pagan custom." Origen, in his commentary "On Levites," wrote that Christians should not only refrain from celebrating their birthdays but should look at them with disgust as a pagan custom. A saint's day was typically celebrated on the anniversary of their martyrdom or death, considered the occasion of or preparation for their entrance into Heaven or the New Jerusalem. Ordinary folk in the Middle Ages celebrated their saint's day (the saint they were named after), but nobility celebrated the anniversary of their birth.[citation needed] The "Squire's Tale", one of Chaucer's Canterbury Tales, opens as King Cambuskan proclaims a feast to celebrate his birthday. In the Modern era, the Catholic Church, the Eastern Orthodox Church and Protestantism, i.e. the three main branches of Christianity, as well as almost all Christian religious denominations, consider celebrating birthdays acceptable or at most a choice of the individual. An exception is Jehovah's Witnesses, who do not celebrate them for various reasons: in their interpretation this feast has pagan origins, was not celebrated by early Christians, is negatively expounded in the Holy Scriptures and has customs linked to superstition and magic. In some historically Roman Catholic and Eastern Orthodox countries,[a] it is common to have a 'name day', otherwise known as a 'Saint's day'. It is celebrated in much the same way as a birthday, but it is held on the official day of a saint with the same Christian name as the birthday person; the difference being that one may look up a person's name day in a calendar, or easily remember common name days (for example, John or Mary); however in pious traditions, the two were often made to concur by giving a newborn the name of a saint celebrated on its day of confirmation, more seldom one's birthday. Some are given the name of the religious feast of their christening's day or birthday, for example, Noel or Pascal (French for Christmas and "of Easter"); as another example, Togliatti was given Palmiro as his first name because he was born on Palm Sunday. The birthday does not reflect Islamic tradition, and because of this, the majority of Muslims refrain from celebrating it. Others do not object, as long as it is not accompanied by behavior contrary to Islamic tradition. A good portion of Muslims (and Arab Christians) who have emigrated to the United States and Europe celebrate birthdays as customary, especially for children, while others abstain. Hindus celebrate the birth anniversary day every year when the day that corresponds to the lunar month or solar month (Sun Signs Nirayana System – Sourava Mana Masa) of birth and has the same asterism (Star/Nakshatra) as that of the date of birth. That age is reckoned whenever Janma Nakshatra of the same month passes. Hindus regard death to be more auspicious than birth, since the person is liberated from the bondages of material society. Also, traditionally, rituals and prayers for the departed are observed on the 5th and 11th days, with many relatives gathering. Historical and cultural perspectives According to Herodotus (5th century BC), of all the days in the year, the one which the Persians celebrate most is their birthday. It was customary to have the board furnished on that day with an ampler supply than common: the richer people eat wholly baked cow, horse, camel, or donkey (Greek: ὄνον), while the poorer classes use instead the smaller kinds of cattle. On his birthday, the king anointed his head and presented gifts to the Persians. According to the law of the Royal Supper, on that day "no one should be refused a request". The rule for drinking was "No restrictions". In ancient Rome, a birthday (dies natalis) was originally an act of religious cultivation (cultus). A dies natalis was celebrated annually for a temple on the day of its founding, and the term is still used sometimes for the anniversary of an institution such as a university. The temple founding day might become the "birthday" of the deity housed there. March 1, for example, was celebrated as the birthday of the god Mars. Each human likewise had a natal divinity, the guardian spirit called the Genius, or sometimes the Juno for a woman, who was owed religious devotion on the day of birth, usually in the household shrine (lararium). The decoration of a lararium often shows the Genius in the role of the person carrying out the rites. A person marked their birthday with ritual acts that might include lighting an altar, saying prayers, making vows (vota), anointing and wreathing a statue of the Genius, or sacrificing to a patron deity. Incense, cakes, and wine were common offerings. Celebrating someone else's birthday was a way to show affection, friendship, or respect. In exile, the poet Ovid, though alone, celebrated not only his own birthday rite but that of his far distant wife. Birthday parties affirmed social as well as sacred ties. One of the Vindolanda tablets is an invitation to a birthday party from the wife of one Roman officer to the wife of another. Books were a popular birthday gift, sometimes handcrafted as a luxury edition or composed especially for the person honored. Birthday poems are a minor but distinctive genre of Latin literature. The banquets, libations, and offerings or gifts that were a regular part of most Roman religious observances thus became part of birthday celebrations for individuals. A highly esteemed person would continue to be celebrated on their birthday after death, in addition to the several holidays on the Roman calendar for commemorating the dead collectively. Birthday commemoration was considered so important that money was often bequeathed to a social organization to fund an annual banquet in the deceased's honor. The observance of a patron's birthday or the honoring of a political figure's Genius was one of the religious foundations for imperial cult or so-called "emperor worship." The Chinese word for "year(s) old" (t 歲, s 岁, suì) is entirely different from the usual word for "year(s)" (年, nián), reflecting the former importance of Chinese astrology and the belief that one's fate was bound to the stars imagined to be in opposition to the planet Jupiter at the time of one's birth. The importance of this duodecennial orbital cycle only survives in popular culture as the 12 animals of the Chinese zodiac, which change each Chinese New Year and may be used as a theme for some gifts or decorations. Because of the importance attached to the influence of these stars in ancient China and throughout the Sinosphere, East Asian age reckoning previously began with one at birth and then added years at each Chinese New Year, so that it formed a record of the suì one had lived through rather than of the exact amount of time from one's birth. This method—which can differ by as much as two years of age from other systems—is increasingly uncommon and is not used for official purposes in the PRC or on Taiwan, although the word suì is still used for describing age. Traditionally, Chinese birthdays—when celebrated—were reckoned using the lunisolar calendar, which varies from the Gregorian calendar by as much as a month forward or backward depending on the year. Celebrating the lunisolar birthday remains common on Taiwan while growing increasingly uncommon on the mainland. Birthday traditions reflected the culture's deep-seated focus on longevity and wordplay. From the homophony in some dialects between 酒 ("rice wine") and 久 (meaning "long" in the sense of time passing), osmanthus and other rice wines are traditional gifts for birthdays in China. Longevity noodles are another traditional food consumed on the day, although western-style birthday cakes are increasingly common among urban Chinese. Hongbaos—red envelopes stuffed with money, now especially the red 100 RMB notes—are the usual gift from relatives and close family friends for most children. Gifts for adults on their birthdays are much less common, although the birthday for each decade is a larger occasion that might prompt a large dinner and celebration. The Japanese reckoned their birthdays by the Chinese system until the Meiji Reforms. Celebrations remained uncommon or muted until after the American occupation that followed World War II.[citation needed] Children's birthday parties are the most important, typically celebrated with a cake, candles, and singing. Adults often just celebrate with their partner. In North Korea, the Day of the Sun, Kim Il Sung's birthday, is the most important public holiday of the country, and Kim Jong Il's birthday is celebrated as the Day of the Shining Star. North Koreans are not permitted to celebrate birthdays on July 8 and December 17 because these were the dates of the deaths of Kim Il Sung and Kim Jong Il, respectively. More than 100,000 North Koreans celebrate displaced birthdays on July 9 and December 18 instead to avoid these dates. A person born on July 8 before 1994 may change their birthday, with official recognition. South Korea was one of the last countries to use a form of East Asian age reckoning for many official purposes. Prior to June 2023, three systems were used together—"Korean ages" that start with 1 at birth and increase every January 1st with the Gregorian New Year, "year ages" that start with 0 at birth and otherwise increase the same way, and "actual ages" that start with 0 at birth and increase each birthday. First birthday celebrations was heavily celebrated, despite usually having little to do with the child's age. In June 2023, all Korean ages were set back at least one year, and official ages henceforth are reckoned only by birthdays. In Ghana, children wake up on their birthday to a special treat called oto, which is a patty made from mashed sweet potato and eggs fried in palm oil. Later they have a birthday party where they usually eat stew and rice and a dish known as kelewele, which is fried plantain chunks. Distribution through the year Birthdays are fairly evenly distributed throughout the year, with some seasonal effects. In the United States, there tend to be more births in September and October. This may be because there is a holiday season nine months before (the human gestation period is about nine months), or because the longest nights of the year also occur in the Northern Hemisphere nine months before. However, the holidays affect birth rates more than the winter: New Zealand, a Southern Hemisphere country, has the same September and October peak with no corresponding peak in March and April. The least common birthdays tend to fall around public holidays, such as Christmas, New Year's Day and fixed-date holidays such as Independence Day in the US, which falls on July 4. Between 1973 and 1999, September 16 was the most common birthday in the United States, and December 25 was the least common birthday (other than February 29 because of leap years). In 2011, October 5 and 6 were reported as the most frequently occurring birthdays. New Zealand's most common birthday is September 29, and the least common birthday is December 25. The ten most common birthdays all fall within a thirteen-day period, between September 22 and October 4. The ten least common birthdays (other than February 29) are December 24–27, January 1–2, February 6, March 22, April 1, and April 25. This is based on all live births registered in New Zealand between 1980 and 2017. Positive and negative associations with culturally significant dates may influence birth rates. The study shows a 5.3% decrease in spontaneous births and a 16.9% decrease in Caesarean births on Halloween, compared to dates occurring within one week before and one week after the October holiday. In contrast, on Valentine's Day, there is a 3.6% increase in spontaneous births and a 12.1% increase in Caesarean births. In Sweden, 9.3% of the population is born in March and 7.3% in November, when a uniform distribution would give 8.3%. In the Gregorian calendar (a common solar calendar), February in a leap year has 29 days instead of the usual 28, so the year lasts 366 days instead of the usual 365. A person born on February 29 may be called a "leapling" or a "leaper". In common years, they usually celebrate their birthdays on February 28. In some situations, March 1 is used as the birthday in a non-leap year since it is the day following February 28. Technically, a leapling will have fewer birthday anniversaries than their age in years. This phenomenon is exploited when a person claims to be only a quarter of their actual age, by counting their leap-year birthday anniversaries only. In Gilbert and Sullivan's 1879 comic opera The Pirates of Penzance, Frederic the pirate apprentice discovers that he is bound to serve the pirates until his 21st birthday rather than until his 21st year. For legal purposes, legal birthdays depend on how local laws count time intervals. An individual's Beddian birthday, named in tribute to firefighter Bobby Beddia, occurs during the year that their age matches the last two digits of the year they were born. Some studies show people are more likely to die on their birthdays, with explanations including excessive drinking, suicide, cardiovascular events due to high stress or happiness, efforts to postpone death for major social events, and death certificate paperwork errors. See also References Notes External links
========================================
[SOURCE: https://en.wikipedia.org/wiki/Julius_Edgar_Lilienfeld] | [TOKENS: 865]
Contents Julius Edgar Lilienfeld Julius Edgar Lilienfeld (April 18, 1882 – August 28, 1963) was an Austro-Hungarian, and later American (where he moved in 1921) electrical engineer and physicist who has been credited with the first patent on the field-effect transistor in 1925. He was never able to build a working practical semiconductor device based on his concept. Additionally, because he didn't publish articles in learned journals and since high-purity semiconductor materials were not available to him, his FET patent never achieved fame, causing confusion for later inventors. Early life Lilienfeld was born to a Jewish family in Lemberg (present-day Lviv) in the Austrian part of the Austro-Hungarian Empire. Lilienfeld's father was the lawyer Sigmund Lilienfeld, his mother Sarah Jampoler Lilienfeld. Education After graduating high school in 1899, between 1900 and 1904, Lilienfeld studied at the Friedrich-Wilhelms-Universität (renamed Humboldt University in 1949), in Berlin, where he received his Ph.D. on February 18, 1905. In 1905, he started work at the physics institute at Leipzig University as an untenured professor. Career Lilienfeld's early career, at the University of Leipzig, saw him conduct important early work on electrical discharges in "vacuum", between metal electrodes, from about 1910 onwards. His early passion was to clarify how this phenomenon changed as vacuum preparation techniques improved. More than any other scientist, he was responsible for the identification of field electron emission as a separate physical effect. (He called it "auto-electronic emission", and was interested in it as a possible electron source for miniaturised X-ray tubes, in medical applications.) Lilienfeld was responsible for the first reliable account in English of the experimental phenomenology of field electron emission, in 1922. The effect was explained by Fowler and Nordheim in 1928. Lilienfeld moved to the United States in 1921 to pursue his patent claims, resigning his professorship at Leipzig to stay permanently in 1926. In 1928, he began working at Amrad in Malden, Massachusetts, later called Ergon Research Laboratories owned by Magnavox, which closed in 1935. In the United States, Lilienfeld did research on anodic aluminum oxide films, patenting the electrolytic capacitor in 1931, the method continuing to be used throughout the century. He also invented a "FET-like" transistor, filing several patents describing the construction and operation of transistors, as well as many features of modern transistors. (US patent #1,745,175 for a FET-like transistor was granted January 28, 1930.) When Brattain, Bardeen, and their colleague chemist Robert Gibney tried to get patents on their earliest devices, most of their claims were rejected due to the Lilienfeld patents. The optical radiation emitted when electrons strike a metal surface is named "Lilienfeld radiation" after he first discovered it close to X-ray tube anodes. Its origin is attributed to the excitation of plasmons in the metal surface. The American Physical Society has named one of its major prizes after Lilienfeld. Personal life Lilienfeld was a German-speaking Ashkenazi Jew who was a citizen of Austria-Hungary and later had dual citizenship in the United States and in Poland. He married an American, Beatrice Ginsburg, in New York City on May 2, 1926. They lived in Winchester, Massachusetts, where Lilienfeld was director of the Ergon Research Laboratories in Malden, becoming a United States citizen in 1934. After it closed in 1935, he and his wife built a house on St. Thomas in the U.S. Virgin Islands in hope of escaping an allergy associated with wheat fields, from which Lilienfeld had suffered for most of his life. Lilienfeld frequently traveled between St. Thomas and various mainland locations and continued to test new ideas and patent the resulting products. Patents See also References External links
========================================
[SOURCE: https://en.wikipedia.org/wiki/Rad_(radiation_unit)] | [TOKENS: 1161]
Contents Rad (radiation unit) The rad is a unit of absorbed radiation dose, defined as 1 rad = 0.01 Gy = 0.01 J/kg. It was originally defined in CGS units in 1953 as the dose causing 100 ergs of energy to be absorbed by one gram of matter. The material absorbing the radiation can be human tissue, air, water, or any other substance. It has been replaced by the gray (symbol Gy) in SI derived units. The rad is still used in the United States, although this is "strongly discouraged" in Chapter 5.2 of the Guide to the SI, which was written and published by the U.S. National Institute of Standards and Technology. However, the numerically equivalent SI unit submultiple, the centigray (symbol cGy), is widely used to report absorbed doses within radiotherapy. The roentgen, used to quantify the radiation exposure, may be related to the corresponding absorbed dose by use of the F-factor. Health effects A dose of under 100 rad will typically produce no immediate symptoms other than blood changes. A dose of 100 to 200 rad delivered to the entire body in less than a day may cause acute radiation syndrome (ARS), but is usually not fatal. Doses of 200 to 1,000 rad delivered in a few hours will cause serious illness, with poor prognosis at the upper end of the range. Whole body doses of more than 1,000 rad are almost invariably fatal. Therapeutic doses of radiation therapy are often given and tolerated well even at higher doses to treat discrete, well-defined anatomical structures. The same dose given over a longer period of time is less likely to cause ARS. Dose thresholds are about 50% higher for dose rates of 20 rad/h, and even higher for lower dose rates. The International Commission on Radiological Protection maintains a model of health risks as a function of absorbed dose and other factors. That model calculates an effective radiation dose, measured in units of rem, which is more representative of the stochastic risk than the absorbed dose in rad. In most power plant scenarios, where the radiation environment is dominated by X- or gamma rays applied uniformly to the whole body, 1 rad of absorbed dose gives 1 rem of effective dose. In other situations, the effective dose in rem might be thirty times higher or thousands of times lower than the absorbed dose in rad. History In the 1930s the roentgen was the most commonly used unit of radiation exposure. This unit is obsolete and no longer clearly defined. One roentgen deposits 0.877 rad in dry air, 0.96 rad in soft tissue, or anywhere from 1 to more than 4 rad in bone depending on the beam energy. These conversions to absorbed energy all depend on the ionizing energy of a standard medium, which is ambiguous in the latest NIST definition. Even where the standard medium is fully defined, the ionizing energy is often not precisely known. In 1940, British physicist Louis Harold Gray, who had been studying the effect of neutron damage on human tissue, together with William Valentine Mayneord and John Read published a paper in which a unit of measure, dubbed the "gram roentgen" (symbol: gr) defined as "that amount of neutron radiation which produces an increment in energy in unit volume of tissue equal to the increment of energy produced in unit volume of water by one roentgen of radiation" was proposed. This unit was found to be equivalent to 88 ergs in air. It marked a shift towards measurements based on energy rather than charge. The Röntgen equivalent physical (rep), introduced by Herbert Parker in 1945, was the absorbed energetic dose to tissue before factoring in relative biological effectiveness. The rep has variously been defined as 83 or 93 ergs per gram of tissue (8.3/9.3 mGy) or per cc of tissue. In 1953 the ICRU recommended the rad, equal to 100 erg/g as a new unit of absorbed radiation, but then promoted a switch to the gray in the 1970s. The International Committee for Weights and Measures (CIPM) has not accepted the use of the rad. From 1977 to 1998, the US NIST's translations of the SI brochure stated that the CIPM had temporarily accepted the use of the rad (and other radiology units) with SI units since 1969. However, the only related CIPM decisions shown in the appendix are with regards to the curie in 1964 and the radian (symbol: rad) in 1960. The NIST brochures redefined the rad as 0.01 Gy. The CIPM's current SI brochure excludes the rad from the tables of non-SI units accepted for use with the SI. The US NIST clarified in 1998 that it was providing its own interpretations of the SI system, whereby it accepted the rad for use in the US with the SI, while recognizing that the CIPM did not. NIST recommends defining the rad in relation to SI units in every document where this unit is used. Nevertheless, use of the rad remains widespread in the US, where it is still an industry standard. Although the United States Nuclear Regulatory Commission still permits the use of the units curie, rad, and rem alongside SI units, the European Union required that its use for "public health ... purposes" be phased out by 31 December 1985. Radiation-related quantities The following table shows radiation quantities in SI and non-SI units: See also References
========================================
[SOURCE: https://en.wikipedia.org/wiki/Assignment_(computer_science)] | [TOKENS: 3093]
Contents Assignment (computer science) In computer programming, an assignment statement sets and/or re-sets the value stored in the storage location(s) denoted by a variable name; in other words, it copies a value into the variable. In most imperative programming languages, the assignment statement (or expression) is a fundamental construct. Today, the most commonly used notation for this operation is x = expr (originally Superplan 1949–51, popularized by Fortran 1957 and C). The second most commonly used notation is x := expr (originally ALGOL 1958, popularised by Pascal). Many other notations are also in use. In some languages, the symbol used is regarded as an operator (meaning that the assignment statement as a whole returns a value). Other languages define assignment as a statement (meaning that it cannot be used in an expression). Assignments typically allow a variable to hold different values at different times during its life-span and scope. However, some languages (primarily strictly functional languages) do not allow that kind of "destructive" reassignment, as it might imply changes of non-local state. The purpose is to enforce referential transparency, i.e. functions that do not depend on the state of some variable(s), but produce the same results for a given set of parametric inputs at any point in time. Modern programs in other languages also often use similar strategies, although less strict, and only in certain parts, in order to reduce complexity, normally in conjunction with complementing methodologies such as data structuring, structured programming and object orientation. Semantics An assignment operation is a process in imperative programming in which different values are associated with a particular variable name as time passes. The program, in such model, operates by changing its state using successive assignment statements. Primitives of imperative programming languages rely on assignment to do iteration. At the lowest level, assignment is implemented using machine operations such as MOVE or STORE. Variables are containers for values. It is possible to put a value into a variable and later replace it with a new one. An assignment operation modifies the current state of the executing program. Consequently, assignment is dependent on the concept of variables. In an assignment: Example: Assuming that a is a numeric variable, the assignment a := 2*a means that the content of the variable a is doubled after the execution of the statement. An example segment of C code: In this sample, the variable x is first declared as an int, and is then assigned the value of 10. Notice that the declaration and assignment occur in the same statement. In the second line, y is declared without an assignment. In the third line, x is reassigned the value of 23. Finally, y is assigned the value of 32.4. For an assignment operation, it is necessary that the value of the expression is well-defined (it is a valid rvalue) and that the variable represents a modifiable entity (it is a valid modifiable (non-const) lvalue). In some languages, typically dynamic ones, it is not necessary to declare a variable prior to assigning it a value. In such languages, a variable is automatically declared the first time it is assigned to, with the scope it is declared in varying by language. Single assignment Any assignment that changes an existing value (e.g. x := x + 1) is disallowed in purely functional languages. In functional programming, assignment is discouraged in favor of single assignment, more commonly known as initialization. Single assignment is an example of name binding and differs from assignment as described in this article in that it can only be done once, usually when the variable is created; no subsequent reassignment is allowed. An evaluation of an expression does not have a side effect if it does not change an observable state of the machine, other than producing the result, and always produces same value for the same input. Imperative assignment can introduce side effects while destroying and making the old value unavailable while substituting it with a new one, and is referred to as destructive assignment for that reason in LISP and functional programming, similar to destructive updating. Single assignment is the only form of assignment available in purely functional languages, such as Haskell, which do not have variables in the sense of imperative programming languages but rather named constant values possibly of compound nature, with their elements progressively defined on-demand, for the lazy languages. Purely functional languages can provide an opportunity for computation to be performed in parallel, avoiding the von Neumann bottleneck of sequential one step at a time execution, since values are independent of each other. Impure functional languages provide both single assignment as well as true assignment (though true assignment is typically used with less frequency than in imperative programming languages). For example, in Scheme, both single assignment (with let) and true assignment (with set!) can be used on all variables, and specialized primitives are provided for destructive update inside lists, vectors, strings, etc. In OCaml, only single assignment is allowed for variables, via the let name = value syntax; however destructive update can be used on elements of arrays and strings with separate <- operator, as well as on fields of records and objects that have been explicitly declared mutable (meaning capable of being changed after their initial declaration) by the programmer. Functional programming languages that use single assignment include Clojure (for data structures, not vars), Erlang (it accepts multiple assignment if the values are equal, in contrast to Haskell), F#, Haskell, JavaScript (for constants), Lava, OCaml, Oz (for dataflow variables, not cells), Racket (for some data structures like lists, not symbols), SASL, Scala (for vals), SISAL, Standard ML. Non-backtracking Prolog code can be considered explicit single-assignment, explicit in a sense that its (named) variables can be in explicitly unassigned state, or be set exactly once. In Haskell, by contrast, there can be no unassigned variables, and every variable can be thought of as being implicitly set, when it is created, to its value (or rather to a computational object that will produce its value on demand). Value of an assignment In some programming languages, an assignment statement returns a value, while in others it does not. In most expression-oriented programming languages (for example, C), the assignment statement returns the assigned value, allowing such idioms as x = y = a, in which the assignment statement y = a returns the value of a, which is then assigned to x. In a statement such as while ((ch = getchar()) != EOF) {…}, the return value of a function is used to control a loop while assigning that same value to a variable. In other programming languages, Scheme for example, the return value of an assignment is undefined and such idioms are invalid. In Haskell, there is no variable assignment; but operations similar to assignment (like assigning to a field of an array or a field of a mutable data structure) usually evaluate to the unit type, which is represented as (). This type has only one possible value, therefore containing no information. It is typically the type of an expression that is evaluated purely for its side effects. Variant forms of assignment Certain use patterns are very common, and thus often have special syntax to support them. These are primarily syntactic sugar to reduce redundancy in the source code, but also assists readers of the code in understanding the programmer's intent, and provides the compiler with a clue to possible optimization. The case where the assigned value depends on a previous one is so common that many imperative languages, most notably C and the majority of its descendants, provide special operators called augmented assignment, like *=, so a = 2*a can instead be written as a *= 2. Beyond syntactic sugar, this assists the task of the compiler by making clear that in-place modification of the variable a is possible. A statement like w = x = y = z is called a chained assignment in which the value of z is assigned to multiple variables w, x, and y. Chained assignments are often used to initialize multiple variables, as in a = b = c = d = f = 0 Not all programming languages support chained assignment. Chained assignments are equivalent to a sequence of assignments, but the evaluation strategy differs between languages. For simple chained assignments, like initializing multiple variables, the evaluation strategy does not matter, but if the targets (l-values) in the assignment are connected in some way, the evaluation strategy affects the result. In some programming languages (C for example), chained assignments are supported because assignments are expressions, and have values. In this case chain assignment can be implemented by having a right-associative assignment, and assignments happen right-to-left. For example, i = arr[i] = f() is equivalent to arr[i] = f(); i = arr[i]. In C++ they are also available for values of class types by declaring the appropriate return type for the assignment operator. In Python, assignment statements are not expressions and thus do not have a value. Instead, chained assignments are a series of statements with multiple targets for a single expression. The assignments are executed left-to-right so that i = arr[i] = f() evaluates the expression f(), then assigns the result to the leftmost target, i, and then assigns the same result to the next target, arr[i], using the new value of i. This is essentially equivalent to tmp = f(); i = tmp; arr[i] = tmp though no actual variable is produced for the temporary value. Some programming languages, such as APL, Common Lisp, Go, JavaScript (since 1.7), Julia, PHP, Maple, Lua, occam 2, Perl, Python, REBOL, Ruby, and PowerShell allow several variables to be assigned in parallel, with syntax like: which simultaneously assigns 0 to a and 1 to b. This is most often known as parallel assignment; it was introduced in CPL in 1963, under the name simultaneous assignment, and is sometimes called multiple assignment, though this is confusing when used with "single assignment", as these are not opposites. If the right-hand side of the assignment is a single variable (e.g. an array or structure), the feature is called unpacking or destructuring assignment: The list will be unpacked so that 0 is assigned to a and 1 to b. Furthermore, swaps the values of a and b. In languages without parallel assignment, this would have to be written to use a temporary variable since a := b; b := a leaves both a and b with the original value of b. Some languages, such as Go, F# and Python, combine parallel assignment, tuples, and automatic tuple unpacking to allow multiple return values from a single function, as in this Python example, while other languages, such as C# and Rust, shown here, require explicit tuple construction and deconstruction with parentheses: This provides an alternative to the use of output parameters for returning multiple values from a function. This dates to CLU (1974), and CLU helped popularize parallel assignment generally. C# additionally allows generalized deconstruction assignment with implementation defined by the expression on the right-hand side, as the compiler searches for an appropriate instance or extension Deconstruct method on the expression, which must have output parameters for the variables being assigned to. For example, one such method that would give the class it appears in the same behavior as the return value of f() above would be In C and C++, the comma operator is similar to parallel assignment in allowing multiple assignments to occur within a single statement, writing a = 1, b = 2 instead of a, b = 1, 2. This is primarily used in for loops, and is replaced by parallel assignment in other languages such as Go. However, the above C++ code does not ensure perfect simultaneity, since the right side of the following code a = b, b = a+1 is evaluated after the left side. In languages such as Python, a, b = b, a+1 will assign the two variables concurrently, using the initial value of a to compute the new b. Assignment versus equality The use of the equals sign = as an assignment operator has been frequently criticized, due to the conflict with equals as comparison for equality. This results both in confusion by novices in writing code, and confusion even by experienced programmers in reading code. The use of equals for assignment dates back to Heinz Rutishauser's language Superplan, designed from 1949 to 1951, and was particularly popularized by Fortran: A notorious example for a bad idea was the choice of the equal sign to denote assignment. It goes back to Fortran in 1957[a] and has blindly been copied by armies of language designers. Why is it a bad idea? Because it overthrows a century old tradition to let “=” denote a comparison for equality, a predicate which is either true or false. But Fortran made it to mean assignment, the enforcing of equality. In this case, the operands are on unequal footing: The left operand (a variable) is to be made equal to the right operand (an expression). x = y does not mean the same thing as y = x. — Niklaus Wirth, Good Ideas, Through the Looking Glass Beginning programmers sometimes confuse assignment with the relational operator for equality, as "=" means equality in mathematics, and is used for assignment in many languages. But assignment alters the value of a variable, while equality testing tests whether two expressions have the same value. In some languages, such as BASIC, a single equals sign ("=") is used for both the assignment operator and the equality relational operator, with context determining which is meant. Other languages use different symbols for the two operators. For example: The similarity in the two symbols can lead to errors if the programmer forgets which form ("=", "==", ":=") is appropriate, or mistypes "=" when "==" was intended. This is a common programming problem with languages such as C (including one famous attempt to backdoor the Linux kernel), where the assignment operator also returns the value assigned (in the same way that a function returns a value), and can be validly nested inside expressions. If the intention was to compare two values in an if statement, for instance, an assignment is quite likely to return a value interpretable as Boolean true, in which case the then clause will be executed, leading the program to behave unexpectedly. Some language processors (such as gcc) can detect such situations, and warn the programmer of the potential error. Notation The two most common representations for the copying assignment are equals sign (=) and colon-equals (:=). Both forms may semantically denote either an assignment statement or an assignment operator (which also has a value), depending on language and/or usage. Other possibilities include a left arrow or a keyword, though there are other, rarer, variants: Mathematical pseudo code assignments are generally depicted with a left-arrow. Some platforms put the expression on the left and the variable on the right: Some expression-oriented languages, such as Lisp and Tcl, uniformly use prefix (or postfix) syntax for all statements, including assignment. See also Notes References
========================================
[SOURCE: https://en.wikipedia.org/wiki/Blocking_of_Twitter_in_Brazil] | [TOKENS: 2381]
Contents Blocking of Twitter in Brazil X Corp. Justice Alexandre de Moraes Elon Musk From 30 August 2024 to 8 October 2024, Brazil's Supreme Federal Court judge Alexandre de Moraes unilaterally imposed a block of X (formerly Twitter) in Brazil. This occurred after the social network's chairman, Elon Musk, refused to appoint a legal representative in the country, which led Moraes to freeze Starlink's financial assets and impose fines for non-compliance. X began to be suspended at the start of the following day. The decision followed an investigation by the Brazilian Supreme Court into Musk due to X reinstating accounts that were suspended under the terms of a court order. The Court reportedly ordered the removal of far-right accounts associated with the 8 January 2023 attacks in Brasília. The investigation began in April 2024, after Musk stated that he would reinstate the accounts. After the platform settled $5.2 million in fines, appointed a legal representative in Brazil and complied with orders to deactivate accounts, the block was lifted on 8 October 2024. Background Musk recognizes that the various countries in which X operates have different laws with respect to freedom of speech. On 26 April 2022, Musk tweeted "By 'free speech,' I simply mean that which matches the law." In June 2023, he said that "X doesn't have a choice but to obey local governments. If we don't obey local government law, then we will get shut down. The best we can do is really to hew close to the law in any given country, but it's impossible for us to do more than that or we will be blocked and our people will be arrested." Brazil enacted an Internet Bill of Rights in 2014. Among other things, the law says that platforms are not legally responsible for user-generated content unless a court orders them to remove the content and the platform refuses. In the lead-up to the 2023 Brazilian Congress attack, election misinformation is said to have circulated on a variety of social media platforms, and people used social media to help plan the attack. It is unclear what entity adjudicates which information has to be regarded as disinformation. After the attack, Moraes ordered several social media platforms, including X, to block specific accounts that had been involved in the planning, stating that the companies would be fined if they failed to comply. The orders were initially sealed before they were disclosed by a committee of the US Congress. In April 2023, Brazil's Ministry of Justice and Public Security requested X to remove five hundred accounts and posts encouraging school violence. The platform did not comply with the request until an executive decree was issued, threatening fines and a potential ban. In April 2024, American journalist Michael Shellenberger publicly criticized judge Alexandre de Moraes in what he called the "X Files Brazil". Shellenberger shared emails from a former X executive criticizing requests from the Brazilian judiciary for data from users of the platform, which would go against the social network's policy. Supreme Federal Court investigation On 6 April 2024, X Corp. stated that it had received a court order by Alexandre de Moraes to suspend several accounts. Elon Musk wrote that he would defy the ruling several hours later and also suggested that users could get around the block by using a virtual private network. In response, Moraes said that he was opening an investigation into Musk, which the Associated Press described as focused on "the dissemination of defamatory fake news and ... obstruction, incitement and criminal organization." Musk suffered a vampetaço from Brazilians on twitter after his allegations. On 17 August 2024, after Moraes threatened to arrest its legal representative, X Corp. announced that it was closing its office in the country and recalled its staff. On 28 August, Moraes gave X a 24-hour deadline to appoint a new legal representative or face suspension. The summons was issued via a post on X itself. The deadline passed without a new representative being named. Brazil's Supreme Court suspended the social network in the country after Musk did not meet legal obligations during a crackdown on disinformation. The tension grew as Musk refused to block accounts tied to former president Jair Bolsonaro. Musk responded by accusing Moraes of undermining democracy. On 21 September 2024, Moraes stated in a decision that X had still not fully complied with requirements to lift its suspension and gave it five days to submit additional documents. Blocking On 30 August 2024, Alexandre de Moraes ordered internet service providers to block access to X, instaured a daily fine of fifty thousand reals (US$9,104) for users who bypass the ban through virtual private networks (VPNs), and froze Starlink's finances in Brazil. To enforce the suspension order against X, Moraes directed Brazil's National Telecommunications Agency (ANATEL) to take action. The order will remain in force until the platform complies with the decisions of the Supreme Federal Court, pays fines totaling R$18.3 million (US$3.33 million), and appoints a representative in Brazil, a requirement grounded in the country's law. Moraes had also instructed Apple and Google to remove X and VPN applications from their virtual stores, but mitigated that decision on the same day, instead suspending the withdrawal of VPN apps in virtual stores until there is a manifestation of X in the records of the apps, citing concerns about potential "unnecessary" disruptions. In the order, Moraes described Musk as an "outlaw" who would "allow the massive spread of disinformation, hate speech and attacks on the democratic rule of law, violating the free choice of the electorate, by keeping voters away from real and accurate information." X began to be suspended at approximately 12:10 a.m. (UTC−03:00) on 31 August. On 1 September, Starlink told ANATEL that it would not obey the order to block X until its assets were unfrozen, but it reversed course two days later, saying that it would comply. On 2 September, a five-judge panel of the Supreme Court affirmed the ban. On 13 September, Moraes unfroze Starlink's finances after the Brazilian government took R$18.35 million (US$3.34 million) from X and Starlink to cover the fines owed; however, X remained blocked at the time. On 18 September 2024, X evaded the block by rerouting its traffic through Cloudflare for requests originating from Brazil. X said the restoration of service was an "inadvertent and temporary" side-effect of switching network providers, and that the change was made because the block ordered in Brazil affected its internet infrastructure that allowed it to provide service to the rest of Latin America, according to the company. By 19 September Cloudflare had reportedly agreed to isolate X traffic, enabling Brazilian internet service providers to resume blocking traffic. On 23 September, Cloudflare CEO Matthew Prince stated that Cloudflare neither helped X evade the block in Brazil nor assisted the country's regulators as they sought to restore the block. Moraes fined X five million reals (US$910,400) for violating the suspension. Reactions Reacting to the decision, Musk wrote on his X account: "Free speech is the bedrock of democracy and an unelected pseudo-judge in Brazil is destroying it for political purposes." Shortly after its suspension, X created the "@AlexandreFiles" account, purportedly in order to shed "a light on the abuses of Brazilian law committed by Alexandre de Moraes". The account began posting sealed orders from Moraes on 31 August. The orders have not been redacted, and have revealed private information including full names and Social Security numbers. Musk retweeted calls for protest and for Moraes' impeachment. He also suggested that the U.S. government intervene, confiscating Brazilian assets and halting foreign aid. In an interview on 30 August, Brazilian president Luiz Inácio Lula da Silva emphasized that Elon Musk should respect the decisions of the Supreme Federal Court, criticizing him for allegedly offending local authorities. Lula da Silva stated that Brazilian society does not have an "inferiority complex". He later said that the world could learn from Brazil's example, and it's not "obliged to put up with Musk's far-right free-for-all just because he is rich." According to journalist Andréia Sadi [pt], behind the scenes at the Supreme Federal Court, the judges assessed that the decision was severe but necessary, as Musk was believed to have intentionally escalated the situation. Sadi reported that there was a consensus between judges that, despite this, Moraes would have committed excesses by blocking the Starlink accounts, which could potentially tarnish the image of the Brazilian judiciary. Members of the National Congress of Brazil expressed varied opinions about the decision. Right-wing Nikolas Ferreira, a member of the Chamber of Deputies, stated: "Tirants [sic] want to turn Brazil into another commie dictatorship but we won't back down. I repeat: do not vote on those who don't respect free speech. Orwell was right". Right-wing congresswoman Bia Kicis stated that "the consequences of Alexandre de Moraes' attacks on Elon Musk, X and Starlink will be regrettable for Brazilians". She also urged Rodrigo Pacheco, the President of the Federal Senate, to act. Congressman Marcel van Hattem wrote on X: "I am tweeting this with VPN." On the other hand, left-wing deputy Erika Hilton wrote, "If billionaires want to have companies that make billions in these parts, they need to learn to respect the laws. Long live the rule of law and national sovereignty." The decision to block Starlink's financial assets in the country was criticized by Arthur Lira, the President of the Chamber of Deputies, at an event for investors promoted by XP Inc. In the United States, Wisconsin politician John Macco called for the relocation of the National Football League's game in São Paulo on 6 September (which featured the Green Bay Packers) to the United States, citing the blocking of X as one of his reasons. Beto Simonetti, the president of the Brazilian Bar Association, announced that the association would request the Supreme Court to review the section of the decision imposing a fine for VPN users, calling it a breach of due process. The classical liberal New Party announced that it would challenge the ban in court. The social network Bluesky, which was founded as a competitor to X, gained over one million new Brazilian users between 30 and 31 August. On 7 September, thousands protested against the ban in São Paulo. The decision received widespread international attention. The New York Times described the situation as the most significant test so far for Musk's efforts to turn the social network into a platform where almost anything is possible. The Associated Press highlighted that the measure intensified the ongoing conflict between Musk and Moraes over freedom of speech, far-right accounts, and disinformation. The Washington Post noted that the action followed Musk's refusal to appoint a legal representative in Brazil. Meanwhile, El País stated that Moraes' decision represented a stringent public sanction concerning the limits of free speech and efforts to combat disinformation. As of now, it is unclear who would get to decide what is to be regarded as disinformation. American journalist Glenn Greenwald questioned the legal basis of Moraes' actions, suggesting that the judge was effectively creating new laws without the oversight of Brazil's Congress. See also References External links
========================================
[SOURCE: https://en.wikipedia.org/wiki/Plow] | [TOKENS: 9306]
Contents Plough A plough or (in the United States) plow (both pronounced /plaʊ/) is a farm tool for loosening or turning soil before sowing seed or planting. Ploughs were traditionally drawn by oxen and horses but modern ploughs are drawn by tractors. A plough may have a wooden, iron or steel frame with a blade attached to cut and loosen the soil. It has been fundamental to farming for most of history. The earliest ploughs had no wheels; such a plough was known to the Romans as an aratrum. Celtic peoples first came to use wheeled ploughs in the Roman era. The prime purpose of ploughing is to turn over the uppermost soil, bringing fresh nutrients to the surface while burying weeds and crop remains to decay. Trenches cut by the plough are called furrows. In modern use, a ploughed field is normally left to dry and then harrowed before planting. Ploughing and cultivating soil evens the content of the upper 12 to 25 centimetres (5 to 10 in) layer of soil, where most plant feeder roots grow. Ploughs were initially powered by humans, but the use of farm animals is considerably more efficient. The earliest animals worked were oxen. Later, horses and mules were used in many areas. With the Industrial Revolution came the possibility of steam engines to pull ploughs. These in turn were superseded by internal-combustion-powered tractors in the early 20th century. The Petty Plough was a notable invention for ploughing out orchard strips in Australia in the 1930s. Use of the traditional plough has decreased in some areas threatened by soil damage and erosion. Used instead is shallower ploughing or other less-invasive conservation tillage. The plough appears in one of the oldest surviving pieces of written literature, from the 3rd millennium BC, where it is personified and debating with another tool, the hoe, over which is better: a Sumerian disputation poem known as the Debate between the hoe and the plough. Etymology In older English, as in other Germanic languages, the plough was traditionally known by other names, e.g. Old English sulh (modern dialectal sullow), Old High German medela, geiza, huohilī(n), Old Norse arðr (Swedish årder), and Gothic hōha, all presumably referring to the ard (scratch plough). The modern word comes from the Old Norse plógr, and is therefore Germanic, but it appears relatively late (it is not attested in Gothic) and is thought to be a loan from one of the north Italic languages. The German cognate is "Pflug", the Dutch "ploeg" and the Swedish "plog". In many Slavic languages and in Romanian the word is "plug". Words with the same root appeared with related meanings: in Raetic plaumorati "wheeled heavy plough" (Pliny, Nat. Hist. 18, 172), and in Latin plaustrum "farm cart", plōstrum, plōstellum "cart", and plōxenum, plōximum "cart box". The word must have originally referred to the wheeled heavy plough, common in Roman north-western Europe by the 5th century AD. Many view plough as a derivative of the verb *plehan ~ *plegan 'to take responsibility' (cf. German pflegen 'to look after, nurse'), which would explain, for example, Old High German pfluog with its double meaning of 'plough' and 'livelihood'. Guus Kroonen (2013) proposes a vṛddhi-derivative of *plag/kkōn 'sod' (cf. Dutch plag 'sod', Old Norse plagg 'cloth', Middle High German pflacke 'rag, patch, stain'). Finally, Vladimir Orel (2003) tentatively attaches plough to a PIE stem *blōkó-, which supposedly gave Old Armenian peɫem "to dig" and Welsh bwlch "crack", though the word may not be of Indo-European origin. Parts The basic parts of the modern plough are: Other parts include the frog (or frame), runner, landside, shin, trashboard, and stilts (handles). On modern ploughs and some older ploughs, the mould board is separate from the share and runner, so these parts can be replaced without replacing the mould board. Abrasion eventually wears out all parts of a plough that come into contact with the soil. History When agriculture was first developed, soil was turned using simple hand-held digging sticks and hoes. These were used in highly fertile areas, such as the banks of the Nile, where the annual flood rejuvenates the soil, to create drills (furrows) in which to plant seeds. Digging sticks, hoes and mattocks were not invented in any one place, and hoe cultivation must have been common everywhere agriculture was practised. Hoe-farming is the traditional tillage method in tropical or sub-tropical regions, which are marked by stony soils, steep slope gradients, predominant root crops, and coarse grains grown at wide intervals. While hoe-agriculture is best suited to these regions, it is used in some fashion everywhere. Some ancient hoes, like the Egyptian mr, were pointed and strong enough to clear rocky soil and make seed drills, which is why they are called hand-ards. However, domestication of oxen in Mesopotamia and the Indus Valley Civilisation, perhaps as early as the 6th millennium BC, provided mankind with the draft power needed to develop the larger, animal-drawn true ard (or scratch plough). A ploughed field, from c. 2800 BCE, was also discovered at Kalibangan, India. A terracotta model of the early ards was found at Banawali, India, giving insight into the form of the tool used. The ard remained easy to replace if it became damaged and easy to replicate. The earliest was the bow ard, which consists of a draft-pole (or beam) pierced by a thinner vertical pointed stick called the head (or body), with one end being the stilt (handle) and the other a share (cutting blade) dragged through the topsoil to cut a shallow furrow suitable for most cereal crops. The ard does not clear new land well, so hoes or mattocks had to be used to pull up grass and undergrowth, and a hand-held, coulter-like ristle could be made to cut deeper furrows ahead of the share. Because the ard left a strip of undisturbed earth between furrows, the fields were often cross-ploughed lengthwise and breadth-wise, which tended to form squarish Celtic fields.: 42 The heavy soils of Northern Europe were difficult to work with a scratch plough.:43 In fact, the ard is best suited to loamy or sandy soils that are naturally fertilised by annual flooding, as in the Nile Delta and Fertile Crescent, and to a lesser extent any other cereal-growing region with light or thin soil. To grow crops regularly in less-fertile areas, it was once believed that the soil must be turned to bring nutrients to the surface. A major advance for this type of farming was the turn plough, also known as the mould-board plough (UK), moldboard plow (U.S.), or frame-plough. A coulter (or skeith) could be added to cut vertically into the ground just ahead of the share (in front of the frog), a wedge-shaped cutting edge at the bottom front of the mould board with the landside of the frame supporting the under-share (below-ground component). The heavy iron moldboard plow was invented in China's Han Empire in the 1st and 2nd century, and from there it spread to the Netherlands, which led the Agricultural Revolution.: 20 The mould-board plough introduced in the 18th century was a major advance in technology. Chinese ploughs from Han times on fulfill all these conditions of efficiency nicely, which is presumably why the standard Han plough team consisted of two animals only, and later teams usually of a single animal, rather than the four, six or eight draught animals common in Europe before the introduction of the curved mould-board and other new principles of design in the 18th century. Though the mould-board plough first appeared in Europe in early medieval, if not in late Roman, times, pre-eighteenth century mould-boards were usually wooden and straight (Fig. 59). The enormous labour involved in pulling such a clumsy construction necessitated large plough-teams, and this meant that large areas of land had to be reserved as pasture. In China, where much less animal power was required, it was not necessary to maintain the mixed arable-pasture economy typical of Europe: fallows could be reduced and the arable area expanded, and a considerably larger population could be supported than on the same amount of land in Europe. — Francesca Bray The upper parts of the frame carry (from the front) the coupling for the motive power (horses), the coulter, and the landside frame. Depending on the size of the implement, and the number of furrows it is designed to plough at one time, a fore-carriage with a wheel or wheels (known as a furrow wheel and support wheel) may be added to support the frame (wheeled plough). In the case of a single-furrow plough there is one wheel at the front and handles at the rear for the ploughman to maneuver it. When dragged through a field, the coulter cuts down into the soil and the share cuts horizontally from the previous furrow to the vertical cut. This releases a rectangular strip of sod to be lifted by the share and carried by the mould board up and over, so that the strip of sod (slice of the topsoil) that is being cut lifts and rolls over as the plough moves forward, dropping back upside down into the furrow and onto the turned soil from the previous run down the field. Each gap in the ground where the soil has been lifted and moved across (usually to the right) is called a furrow. The sod lifted from it rests at an angle of about 45 degrees in the adjacent furrow, up the back of the sod from the previous run. A series of ploughings run down a field leaves a row of sods partly in the furrows and partly on the ground lifted earlier. Visually, across the rows, there is the land on the left, a furrow (half the width of the removed strip of soil) and the removed strip almost upside-down lying on about half of the previous strip of inverted soil, and so on across the field. Each layer of soil and the gutter it came from forms a classic furrow. The mould-board plough greatly reduced the time needed to prepare a field and so allowed a farmer to work a larger area of land. In addition, the resulting pattern of low (under the mould board) and high (beside it) ridges in the soil forms water channels, allowing the soil to drain. In areas where snow build-up causes difficulties, this lets farmers plant the soil earlier, as the meltwater run-off drains away more quickly. There are five major parts of a mouldboard plough: The share, landside and mould board are bolted to the frog, which is an irregular piece of cast iron at the base of the plough body, to which the soil-wearing parts are bolted. The share is the edge that makes the horizontal cut to separate the furrow slice from the soil below. Conventional shares are shaped to penetrate soil efficiently: the tip is pointed downward to pull the share into the ground to a regular depth. The clearance, usually referred to as suction or down suction, varies with different makes and types of plough. Share configuration is related to soil type, particularly in the down suction or concavity of its lower surface. Generally three degrees of clearance or down suction are recognised: regular for light soil, deep for ordinary dry soil, and double-deep for clay and gravelly soils. As the share wears away, it becomes blunt and the plough will require more power to pull it through the soil. A plough body with a worn share will not have enough "suck" to ensure it delves the ground to its full working depth. In addition, the share has horizontal suction related to the amount its point is bent out of line with the land side. Down suction causes the plough to penetrate to proper depth when pulled forward, while horizontal suction causes the plough to create the desired width of furrow. The share is a plane part with a trapezoidal shape. It cuts the soil horizontally and lifts it. Common types are regular, winged-plane, bar-point, and share with mounted or welded point. The regular share conserves a good cut but is recommended on stone-free soils. The winged-plane share is used on heavy soil with a moderate amount of stones. The bar-point share can be used in extreme conditions (hard and stony soils). The share with a mounted point is somewhere between the last two types. Makers have designed shares of various shapes (trapesium, diamond, etc.) with bolted point and wings, often separately renewable. Sometimes the share-cutting edge is placed well in advance of the mould board to reduce the pulverizing action of the soil. The mould board is the part of the plough that receives the furrow slice from the share. It is responsible for lifting and turning the furrow slice and sometimes for shattering it, depending on the type of mould board, ploughing depth and soil conditions. The intensity of this depends on the type of mould board. To suit different soil conditions and crop requirements, mould boards have been designed in different shapes, each producing its own furrow profile and surface finish, but essentially they still conform to the original plough body classification. The various types have been traditionally classified as general purpose, digger, and semi-digger, as described below. The land side is the flat plate which presses against and transmits the lateral thrust of the plough bottom to the furrow wall. It helps to resist the side pressure exerted by the furrow slice on the mould board. It also helps to stabilise the plough while in operation. The rear bottom end of the landslide, which rubs against the furrow sole, is known as the heel. A heel iron is bolted to the end of the rear of the land side and helps to support the back of the plough. The land side and share are arranged to give a "lead" towards the unploughed land, so helping to sustain the correct furrow width. The land side is usually made of solid medium-carbon steel and is very short, except at the rear bottom of the plough. The heel or rear end of the rear land side may be subject to excessive wear if the rear wheel is out of adjustment, and so a chilled iron heel piece is frequently used. This is inexpensive and can be easily replaced. The land side is fastened to the frog by plough bolts.[citation needed] The frog (standard) is the central part of the plough bottom to which the other components of the bottom are attached. It is an irregular piece of metal, which may be made of cast iron for cast iron ploughs or welded steel for steel ploughs. The frog is the foundation of the plough bottom. It takes the shock resulting from hitting rocks, and therefore should be tough and strong. The frog is in turn fastened to the plough frame.[citation needed] A runner extending from behind the share to the rear of the plough controls the direction of the plough, because it is held against the bottom land-side corner of the new furrow being formed. The holding force is the weight of the sod, as it is raised and rotated, on the curved surface of the mould board. Because of this runner, the mould board plough is harder to turn around than the scratch plough, and its introduction brought about a change in the shape of fields – from mostly square fields into longer rectangular "strips" (hence the introduction of the furlong).[citation needed] An advance on the basic design was the iron ploughshare, a replaceable horizontal cutting surface mounted on the tip of the share. The earliest ploughs with a detachable and replaceable share date from around 1000 BC in the Ancient Near East, and the earliest iron ploughshares from about 500 BC in China. Early mould boards were wedges that sat inside the cut formed by the coulter, turning over the soil to the side. The ploughshare spread the cut horizontally below the surface, so that when the mould board lifted it, a wider area of soil was turned over. Mould boards are known in Britain from the late 6th century onwards. There are multiple types of ploughs available. When a plough hits a rock or other solid obstruction, serious damage may result unless the plough is equipped with some safety device. The damage may be bent or broken shares, bent standards, beams or braces.[citation needed] The three basic types of safety devices used on mould-board ploughs are a spring release device in the plough drawbar, a trip beam construction on each bottom, and an automatic reset design on each bottom. The spring release was used in the past almost universally on trailing-type ploughs with one to three or four bottoms. It is not practical on larger ploughs. When an obstruction is encountered, the spring release mechanism in the hitch permits the plough to uncouple from the tractor. When a hydraulic lift is used on the plough, the hydraulic hoses will also usually uncouple automatically when the plough uncouples. Most plough makers offer an automatic reset system for tough conditions or rocky soils. The re-set mechanism allows each body to move rearward and upward to pass without damage over obstacles such as rocks hidden below soil surface. A heavy leaf or coil-spring mechanism that holds the body in its working position under normal conditions resets the plough after the obstruction is passed.[citation needed] Another type of auto-reset mechanism uses an oil (hydraulic) and gas accumulator. Shock loads cause the oil to compress the gas. When the gas expands again, the leg returns to its working ploughing position after passing over the obstacle. The simplest mechanism is a breaking (shear) bolt that needs replacement. Shear bolts that break when a plough body hits an obstruction are a cheaper overload protection device.[citation needed] Trip-beam ploughs are constructed with a hinge point in the beam. This is usually located some distance above the top of the plough bottom. The bottom is held in normal ploughing position by a spring-operated latch. When an obstruction is encountered, the entire bottom is released and hinges back and up to pass over the obstruction. It is necessary to back up the tractor and plough to reset the bottom. This construction is used to protect the individual bottoms. The automatic reset design has only recently[when?] been introduced on US ploughs, but has been used extensively on European and Australian ploughs. Here the beam is hinged at a point almost above the point of the share. The bottom is held in the normal position by a set of springs or a hydraulic cylinder on each bottom.[citation needed] When an obstruction is encountered, the plough bottom hinges back and up in such a way as to pass over the obstruction, without stopping the tractor and plough. The bottom automatically returns to normal ploughing position as soon as the obstruction is passed, without any interruption of forward motion. The automatic reset design permits higher field efficiencies since stopping for stones is practically eliminated. It also reduces costs for broken shares, beams and other parts. The fast resetting action helps produce a better job of ploughing, as large areas of unploughed land are not left, as they are when lifting a plough over a stone.[citation needed] Manual loy ploughing was a form used on small farms in Ireland where farmers could not afford more, or on hilly ground that precluded horses. It was used up until the 1960s in poorer land. It suited the moist Irish climate, as the trenches formed by turning in the sods provided drainage. It allowed potatoes to be grown in bogs (peat swamps) and on otherwise unfarmed mountain slopes. In the basic mould-board plough, the depth of cut is adjusted by lifting against the runner in the furrow, which limited the weight of the plough to what a ploughman could easily lift. This limited the construction to a small amount of wood (although metal edges were possible). These ploughs were fairly fragile and unsuitable for the heavier soils of northern Europe. The introduction of wheels to replace the runner allowed the weight of the plough to increase, and in turn the use of a larger mould-board faced in metal. These heavy ploughs led to greater food production and eventually a marked population increase, beginning around AD 1000. Before the Han dynasty (202 BC – AD 220), Chinese ploughs were made almost wholly of wood except for the iron blade of the ploughshare. These were V-shaped iron pieces mounted on wooden blades and handles.: 18 By the Han period the entire ploughshare was made of cast iron. These are the earliest known heavy, mould-board iron ploughs. Several advancements such as the three-shared plow, the plow-and-sow implement, and the harrow were developed subsequently. By the end of the Song dynasty in 1279, Chinese ploughs had reached a state of development that would not be seen in Holland until the 17th century.: 18 The Romans achieved a heavy-wheeled mould-board plough in the late 3rd and 4th century AD, for which archaeological evidence appears, for instance, in Roman Britain. The Greek and Roman mould-boards were usually tied to the bottom of the shaft with bits of rope, which made them more fragile than the Chinese ones, and iron mould-boards did not appear in Europe until the 10th century.: 17 The first indisputable appearance after the Roman period is in a northern Italian document of 643.: 50 Old words connected with the heavy plough and its use appear in Slavic, suggesting possible early use in that region.: 49ff General adoption of the carruca heavy plough in Europe seems to have accompanied adoption of the three-field system in the later 8th and early 9th centuries, leading to improved agricultural productivity per unit of land in northern Europe.: 69–78 This was accompanied by larger fields, known variously as carucates, ploughlands, and plough gates. The basic plough with coulter, ploughshare and mould board remained in use for a millennium. The basic heavy plough worked well in the Northern European soil. It was strong enough to tear up the soil so that cross-ploughing was not needed. Because of this, and the fact that heavy ploughs needed up to eight oxen to pull, the fields became long and narrow. Too many oxen were difficult to turn frequently. Also, the weaker, ard (or scratch) plough necessitated a short and square-shaped field. The way that the plough allowed peasants to work the land meant that the nutrient-dense soil could yield better, more reliable, crops. Major changes in design spread widely in the Age of Enlightenment, when there was rapid progress in design. Joseph Foljambe in Rotherham, England, in 1730, used new shapes based on the Rotherham plough, which covered the mould board with iron. Unlike the heavy plough, the Rotherham, or Rotherham swing plough consisted entirely of the coulter, mould board and handles. It was much lighter than earlier designs and became common in England. It may have been the first plough widely built in factories and commercially successful there. In 1789 Robert Ransome, an iron founder in Ipswich, started casting ploughshares in a disused malting at St Margaret's Ditches. A broken mould in his foundry caused molten metal to come into contact with cold metal, making the metal surface extremely hard. This process, chilled casting, resulted in what Ransome advertised as "self-sharpening" ploughs. He received patents for his discovery. James Small further advanced the design. Using mathematical methods, he eventually arrived at a shape cast from a single piece of iron, an improvement on the Scots plough of James Anderson of Hermiston. A single-piece cast-iron plough was also developed and patented by Charles Newbold in the United States. This was again improved on by Jethro Wood, a blacksmith of Scipio, New York, who made a three-part Scots plough that allowed a broken piece to be replaced. In 1833 John Lane invented a steel plough. Then in 1837 John Deere introduced a steel plough; it was so much stronger than iron designs that it could work soil in US areas previously thought unsuitable for farming. Improvements on this followed developments in metallurgy: steel coulters and shares with softer iron mould boards to prevent breakage, the chilled plough (an early example of surface-hardened steel), and eventually mould boards with faces strong enough to dispense with the coulter. By the time of the early 1900s, the steel plough had many uses, shapes and names. The "two horse breaking plough" had a point and wing used to break the soil's surface and turn the dirt out and over. The "shovel plough" was used to lay off the rows. The "harrow plough" was used to cover the planted seed. The "scratcher" or "geewhiz" was used to deweed or cultivate the crop. The "bulltongue" and "sweeps" were used to plough the middle of the rows. All these metal plough points required being re-sharpened about every ten days, due to their use on rough and rocky ground. The first mould-board ploughs could only turn the soil over in one direction (conventionally to the right), as dictated by the shape of the mould board; therefore, a field had to be ploughed in long strips, or lands. The plough was usually worked clockwise around each land, ploughing the long sides and being dragged across the short sides without ploughing. The length of the strip was limited by the distance oxen (later horses) could comfortably work without rest, and their width by the distance the plough could conveniently be dragged. These distances determined the traditional size of the strips: a furlong, (or "furrow's length", 220 yards (200 m)) by a chain (22 yards (20 m)) – an area of one acre (about 0.4 hectares); this is the origin of the acre. The one-sided action gradually moved soil from the sides to the centre line of the strip. If the strip was in the same place each year, the soil built up into a ridge, creating the ridge and furrow topography still seen in some ancient fields. The turn-wrest plough allows ploughing to be done to either side. The mould board is removable, turning to the right for one furrow, then being moved to the other side of the plough to turn to the left. (The coulter and ploughshare are fixed.) Thus adjacent furrows can be ploughed in opposite directions, allowing ploughing to proceed continuously along the field and so avoid the ridge–furrow topography.[citation needed] The reversible (or roll-over) plough has two mould-board ploughs mounted back to back, one turning right, the other left. While one works the land, the other is borne upside-down in the air. At the end of each row the paired ploughs are turned over so that the other can be used along the next furrow, again working the field in a consistent direction. These ploughs date back to the days of the steam engine and the horse. In almost universal use on farms, they have right and left-handed mould boards, enabling them to work up and down the same furrow. Reversible ploughs may either be mounted or semi-mounted and are heavier and more expensive than right-handed models, but have the great advantage of leaving a level surface that facilitates seedbed preparation and harvesting. Very little marking out is necessary before ploughing can start; idle running on the headland is minimal compared with conventional ploughs. Driving a tractor with furrow-side wheels in the furrow bottom provides the most efficient line of draught between tractor and plough. It is also easier to steer the tractor; driving with the front wheel against the furrow wall will keep the front furrow at the correct width. This is less satisfactory when using a tractor with wide front tyres. Although these make better use of the tractor power, the tyres may compact some of the last furrow slice turned on the previous run. The problem is overcome by using a furrow widener or longer mould board on the rear body. The latter moves the soil further towards the ploughed land, leaving more room for the tractor wheels on the next run. Driving with all four wheels on unploughed land is another solution to the problem of wide tyres. Semi-mounted ploughs can be hitched in a way that allows the tractor to run on unbroken land and pull the plough in correct alignment without any sideways movement (crabbing). Early steel ploughs were walking ploughs, directed by a ploughman holding handles on either side of the plough. Steel ploughs were so much easier to draw through the soil that constant adjustment of the blade to deal with roots or clods was no longer necessary, as the plough could easily cut through them. Not long after that the first riding ploughs appeared, whose wheels kept the plough at an adjustable level above the ground, while the ploughman sat on a seat instead of walking. Direction was now controlled mostly through the draught team, with levers allowing fine adjustments. This led quickly to riding ploughs with multiple mould boards, which dramatically increased ploughing performance.[citation needed] A single draught horse can normally pull a single-furrow plough in clean light soil, but in heavier soils two horses are needed, one walking on the land and one in the furrow. Ploughs with two or more furrows call for more than two horses, and usually one or more have to walk on the ploughed sod, which is hard going for them and means they tread newly ploughed land down. It is usual to rest such horses every half-hour for about ten minutes. John Deere, an Illinois blacksmith, noted that ploughing many sticky, non-sandy soils might benefit from modifications in the design of the mould board and the metals used. A polished needle would enter leather and fabric with greater ease and a polished pitchfork also require less effort. Looking for a polished, slicker surface for a plough, he experimented with portions of saw blades, and by 1837 was making polished, cast steel ploughs. The invention of the mobile steam engine allowed steam power to be applied to ploughing from about 1850. In Europe, soil conditions were often too soft to support the weight of a traction engine. Instead, counterbalanced, wheeled ploughs, known as balance ploughs, were drawn by cables across the fields by pairs of ploughing engines on opposite field edges, or by a single engine drawing directly towards it at one end and drawing away from it via a pulley at the other. The balance plough had two sets of facing ploughs arranged so that when one was in the ground, the other was lifted in the air. When pulled in one direction, the trailing ploughs were lowered onto the ground by the tension on the cable. When the plough reached the edge of the field, the other engine pulled the opposite cable, and the plough tilted (balanced), putting the other set of shares into the ground, and the plough worked back across the field.[citation needed] One set of ploughs was right-handed and the other left-handed, allowing continuous ploughing along the field, as with the turn-wrest and reversible ploughs. The man credited with inventing the ploughing engine and associated balance plough in the mid-19th century was John Fowler, an English agricultural engineer and inventor. However the Fisken brothers demonstrated (and went on to patent) a balance plough about 4 years before Fowler. One notable producer of steam-powered ploughs was J.Kemna of Eastern Prussia, who became the "leading steam plough company on the European continent and penetrated the monopoly of English companies on the world market" at the beginning of the 20th century. The stump-jump plough was invented in 1876 by Australian Richard Bowyer Smith alongside his brother Clarence Herbert Smith. It is designed to break up new farming land that contains tree stumps and rocks expensive to remove. It uses a moveable weight to hold the ploughshare in position. When a tree stump or rock is encountered, the ploughshare is thrown up clear of the obstacle, to avoid breaking its harness or linkage. Ploughing can continue when the weight is returned to the earth. A simpler, later system uses a concave disc (or pair of them) set at a wide angle to the direction of progress, using a concave shape to hold the disc into the soil – unless something hard strikes the circumference of the disc, causing it to roll up and over the obstruction. As this is dragged forward, the sharp edge of the disc cuts the soil, and the concave surface of the rotating disc lifts and throws the soil to the side. It does not work as well as a mould-board plough (but this is not seen as a drawback, because it helps to fight wind erosion), but it does lift and break up the soil (see disc harrow).[citation needed] Modern ploughs are usually multiple reversible, mounted on a tractor with a three-point linkage. These commonly have from two to as many as seven mould boards – and semi-mounted ploughs (whose lifting is assisted by a wheel about halfway along their length) can have as many as 18. The tractor's hydraulics are used to lift and reverse the implement and to adjust furrow width and depth. The plougher still has to set the draughting linkage from the tractor, so that the plough keeps the proper angle in the soil. This angle and depth can be controlled automatically by modern tractors. As a complement to the rear plough a two or three mould-board plough can be mounted on the front of the tractor if it is equipped with front three-point linkage.[citation needed] Specialist ploughs The chisel plough is a common tool for deep tillage (prepared land) with limited soil disruption. Its main function is to loosen and aerate the soils, while leaving crop residue on top. This plough can be used to reduce the effects of soil compaction and to help break up ploughpan and hardpan. Unlike many other ploughs, the chisel will not invert or turn the soil. This feature has made it a useful addition to no-till and low-till farming practices that attempt to maximise the erosion-preventing benefits of keeping organic matter and farming residues present on the soil surface throughout the year. Thus the chisel plough is considered by some[who?] to be more sustainable than other types of plough, such as the mould-board plough. Chisel ploughs are becoming more popular as a primary tillage tool in row-crop farming areas. Basically the chisel plough is a heavy-duty field cultivator intended to operate at depths from 15 cm (5.9 in) to as much as 46 cm (18 in). However some models may run much deeper.[clarification needed] Each individual plough or shank is typically set from 230 mm (9 in) to 360 mm (14 in) apart. Such a plough can meet significant soil drag: a tractor of sufficient power and traction is required. When ploughing with a chisel plough, 10–20 horsepower (7.5–14.9 kW) per shank is required, depending on depth.[citation needed] Pull-type chisel ploughs are made in working widths from about 2.5 metres (8 ft 2 in) up to 13.7 metres (45 ft). They are tractor-mounted, and working depth is hydraulically controlled. Those more than about 4 metres (13 ft) wide may be equipped with folding wings to reduce transport width. Wider machines may have the wings supported by individual wheels and hinge joints to allow flexing of the machine over uneven ground. The wider models usually have a wheel each side to control working depth. Three-point hitch-mounted units are made in widths from about 1.5 to 9 metres (4 ft 11 in to 29 ft 6 in). Cultivators are often similar in form to chisel ploughs, but their goals are different. Cultivator teeth work near the surface, usually for weed control, whereas chisel plough shanks work deep under the surface; therefore, cultivation takes much less power per shank than does chisel ploughing. The country plough is a slanted plough. The most common plough in India, it is recommended for crops like groundnut after the use of a tractor. A ridging plough is used for crops such as potatoes or scallions grown buried in ridges of soil, using a technique called ridging or hilling. A ridging plough has two back-to-back mould boards cutting a deep furrow on each pass with high ridges either side. The same plough may be used to split the ridges to harvest the crop. This variety of ridge plough is notable for having a blade pointing towards the operator. It is used solely by human effort rather than with animal or machine assistance and pulled backwards by the operator, requiring great physical effort. It is particularly used for second breaking of ground and for potato planting. It is found in Shetland, some western crofts, and more rarely Central Scotland, typically on holdings too small or poor to merit the use of animals. Functionally operating as a saw, but pulled as a plough, this device was created in the 19th century and was mainly used in Scandinavia, as part of the ice export industry, creating blocks of ice to ship to Great Britain. The mole plough allows under-drainage to be installed without trenches, or breaks up the deep impermeable soil layers that impede it. It is a deep plough with a torpedo or wedge-shaped tip and a narrow blade connecting it to the body. When dragged over ground, it leaves a channel deep under it that acts as a drain. Modern mole ploughs may also bury a flexible perforated plastic drain pipe as they go, making a more permanent drain – or may be used to lay pipes for water supply or other purposes. Similar machines, so-called pipe-and-cable-laying ploughs, are even used under the sea for laying cables or for preparing the earth for side-scan sonar in a process used in oil exploration. Compacting a tennis ball-sized sample from moling depth by hand, then pushing a pencil through is a simple check to find if the subsoil is in the right condition for mole ploughing. If the hole stays intact without splitting the ball, the soil is in ideal condition for the mole plough. Heavy land requires draining to reduce its water content to a level efficient for plant growth. Heavy soils usually have a system of permanent drains, using perforated plastic or clay pipes that discharge into a ditch. The small tunnels (mole drains) that mole ploughs form lie at a depth of up to 950 mm (37 in) at an angle to the pipe drains. Water from the mole drains seeps into the pipes and runs along them into a ditch. Mole ploughs are usually trailed and pulled by a crawler tractor, but lighter models for use on the three-point linkage of powerful four-wheel drive tractors are also made. A mole plough has a strong frame that slides along the ground when the machine is at work. A heavy leg, similar to a sub-soiler leg, is attached to the frame and a circular section with a larger diameter expander on a flexible link is bolted to the leg. The bullet-shaped share forms a tunnel in the soil about 75 mm (3.0 in) diameter and the expander presses the soil outwards to form a long-lasting drainage channel. The para-plough, or paraplow, loosens compacted soil layers 3 to 4 dm (12 to 16 inches) deep while maintaining high surface residue levels. It is primary tillage implement for deep ploughing without inversion. The spade plough is designed to cut the soil and turn it on its side, minimising damage to earthworms, soil microorganism and fungi. This increases the sustainability and long-term fertility of the soil. Using a bar with square shares mounted perpendicularly and a pivot point to change the bar's angle, the switch plough allows ploughing in either direction. It is best in previously-worked soils, as the ploughshares are designed more to turn the soil over than for deep tillage. At the headland, the operator pivots the bar (and so the ploughshares) to turn the soil to the opposite side of the direction of travel. Switch ploughs are usually lighter than roll-over ploughs, requiring less horsepower to operate. Effects of mould-board ploughing Mould-board ploughing in cold and temperate climates, down to 20 cm (7.9 in), aerates the soil by loosening it. It incorporates crop residues, solid manures, limestone, and commercial fertilisers alongside oxygen, reducing nitrogen losses by denitrification, accelerating mineralisation, and raising short-term nitrogen availability for turning organic matter into humus. It erases wheel tracks and ruts from harvesting equipment. It controls many perennial weeds and delays the growth of others until spring. It accelerates spring soil warming and water evaporation due to lower residues on the soil surface. It facilitates seeding with a lighter seed, controls many crop enemies (slugs, crane flies, seedcorn maggots-bean seed flies, borers), and raises the number of "soil-eating" earthworms (endogic), but deters vertical-dwelling earthworms (anecic).[citation needed] Ploughing leaves little crop residue on the surface that might otherwise reduce both wind and water erosion. Over-ploughing can lead to the formation of hardpan. Typically, farmers break that up with a subsoiler, which acts as a long, sharp knife slicing through the hardened layer of soil deep below the surface. Soil erosion due to improper land and plough utilisation is possible. Contour ploughing mitigates soil erosion by ploughing across a slope, along elevation lines. Alternatives to ploughing, such as a no-till method, have the potential to build soil levels and humus. These may be suitable for smaller, intensively cultivated plots and for farming on poor, shallow or degraded soils that ploughing would further degrade.[citation needed] Use of the mouldboard plough in Europe typically required cooperation between multiple households, since four to eight oxen (later horses) were needed to pull the heavy instrument through the soil, and this was more than a typical peasant household could muster on its own. Thus, the introduction of the mouldboard plough made farming more of a communal effort and contributed to the rise of the manorial system in Northern Europe. It also led to farming being seen as men's work, because of the weight of the instrument. Historian William H. McNeill suggested that the experience of cooperation with non-family members as a consequence of mouldboard tillage was conducive to the prevalence of the corporation in medieval Europe; peasants who engaged in this kind of work, including those who migrated to towns, were used to the idea of entering into agreements with people outside of their family and more readily formed corporations. Depictions See also References Further reading External links
========================================
[SOURCE: https://en.wikipedia.org/wiki/Black_hole#Structure] | [TOKENS: 13839]
Contents Black hole A black hole is an astronomical body so compact that its gravity prevents anything, including light, from escaping. Albert Einstein's theory of general relativity predicts that a sufficiently compact mass will form a black hole. The boundary of no escape is called the event horizon. In general relativity, a black hole's event horizon seals an object's fate but produces no locally detectable change when crossed. General relativity also predicts that every black hole should have a central singularity, where the curvature of spacetime is infinite. In many ways, a black hole acts like an ideal black body, as it reflects no light. Quantum field theory in curved spacetime predicts that event horizons emit Hawking radiation, with the same spectrum as a black body of a temperature inversely proportional to its mass. This temperature is of the order of billionths of a kelvin for stellar black holes, making it essentially impossible to observe directly. Objects whose gravitational fields are too strong for light to escape were first considered in the 18th century by John Michell and Pierre-Simon Laplace. In 1916, Karl Schwarzschild found the first modern solution of general relativity that would characterise a black hole. Due to his influential research, the Schwarzschild metric is named after him. David Finkelstein, in 1958, first interpreted Schwarzschild's model as a region of space from which nothing can escape. Black holes were long considered a mathematical curiosity; it was not until the 1960s that theoretical work showed they were a generic prediction of general relativity. The first black hole known was Cygnus X-1, identified by several researchers independently in 1971. Black holes typically form when massive stars collapse at the end of their life cycle. After a black hole has formed, it can grow by absorbing mass from its surroundings. Supermassive black holes of millions of solar masses may form by absorbing other stars and merging with other black holes, or via direct collapse of gas clouds. There is consensus that supermassive black holes exist in the centres of most galaxies. The presence of a black hole can be inferred through its interaction with other matter and with electromagnetic radiation such as visible light. Matter falling toward a black hole can form an accretion disk of infalling plasma, heated by friction and emitting light. In extreme cases, this creates a quasar, some of the brightest objects in the universe. Merging black holes can also be detected by observation of the gravitational waves they emit. If other stars are orbiting a black hole, their orbits can be used to determine the black hole's mass and location. Such observations can be used to exclude possible alternatives such as neutron stars. In this way, astronomers have identified numerous stellar black hole candidates in binary systems and established that the radio source known as Sagittarius A*, at the core of the Milky Way galaxy, contains a supermassive black hole of about 4.3 million solar masses. History The idea of a body so massive that even light could not escape was first proposed in the late 18th century by English astronomer and clergyman John Michell and independently by French scientist Pierre-Simon Laplace. Both scholars proposed very large stars in contrast to the modern concept of an extremely dense object. Michell's idea, in a short part of a letter published in 1784, calculated that a star with the same density but 500 times the radius of the sun would not let any emitted light escape; the surface escape velocity would exceed the speed of light.: 122 Michell correctly hypothesized that such supermassive but non-radiating bodies might be detectable through their gravitational effects on nearby visible bodies. In 1796, Laplace mentioned that a star could be invisible if it were sufficiently large while speculating on the origin of the Solar System in his book Exposition du Système du Monde. Franz Xaver von Zach asked Laplace for a mathematical analysis, which Laplace provided and published in a journal edited by von Zach. In 1905, Albert Einstein showed that the laws of electromagnetism would be invariant under a Lorentz transformation: they would be identical for observers travelling at different velocities relative to each other. This discovery became known as the principle of special relativity. Although the laws of mechanics had already been shown to be invariant, gravity remained yet to be included.: 19 In 1907, Einstein published a paper proposing his equivalence principle, the hypothesis that inertial mass and gravitational mass have a common cause. Using the principle, Einstein predicted the redshift and half of the lensing effect of gravity on light; the full prediction of gravitational lensing required development of general relativity.: 19 By 1915, Einstein refined these ideas into his general theory of relativity, which explained how matter affects spacetime, which in turn affects the motion of other matter. This formed the basis for black hole physics. Only a few months after Einstein published the field equations describing general relativity, astrophysicist Karl Schwarzschild set out to apply the idea to stars. He assumed spherical symmetry with no spin and found a solution to Einstein's equations.: 124 A few months after Schwarzschild, Johannes Droste, a student of Hendrik Lorentz, independently gave the same solution. At a certain radius from the center of the mass, the Schwarzschild solution became singular, meaning that some of the terms in the Einstein equations became infinite. The nature of this radius, which later became known as the Schwarzschild radius, was not understood at the time. Many physicists of the early 20th century were skeptical of the existence of black holes. In a 1926 popular science book, Arthur Eddington critiqued the idea of a star with mass compressed to its Schwarzschild radius as a flaw in the then-poorly-understood theory of general relativity.: 134 In 1939, Einstein himself used his theory of general relativity in an attempt to prove that black holes were impossible. His work relied on increasing pressure or increasing centrifugal force balancing the force of gravity so that the object would not collapse beyond its Schwarzschild radius. He missed the possibility that implosion would drive the system below this critical value.: 135 By the 1920s, astronomers had classified a number of white dwarf stars as too cool and dense to be explained by the gradual cooling of ordinary stars. In 1926, Ralph Fowler showed that quantum-mechanical degeneracy pressure was larger than thermal pressure at these densities.: 145 In 1931, Subrahmanyan Chandrasekhar calculated that a non-rotating body of electron-degenerate matter below a certain limiting mass is stable, and by 1934 he showed that this explained the catalog of white dwarf stars.: 151 When Chandrasekhar announced his results, Eddington pointed out that stars above this limit would radiate until they were sufficiently dense to prevent light from exiting, a conclusion he considered absurd. Eddington and, later, Lev Landau argued that some yet unknown mechanism would stop the collapse. In the 1930s, Fritz Zwicky and Walter Baade studied stellar novae, focusing on exceptionally bright ones they called supernovae. Zwicky promoted the idea that supernovae produced stars with the density of atomic nuclei—neutron stars—but this idea was largely ignored.: 171 In 1939, based on Chandrasekhar's reasoning, J. Robert Oppenheimer and George Volkoff predicted that neutron stars below a certain mass limit, later called the Tolman–Oppenheimer–Volkoff limit, would be stable due to neutron degeneracy pressure. Above that limit, they reasoned that either their model would not apply or that gravitational contraction would not stop.: 380 John Archibald Wheeler and two of his students resolved questions about the model behind the Tolman–Oppenheimer–Volkoff (TOV) limit. Harrison and Wheeler developed the equations of state relating density to pressure for cold matter all the way through electron degeneracy and neutron degeneracy. Masami Wakano and Wheeler then used the equations to compute the equilibrium curve for stars, relating mass to circumference. They found no additional features that would invalidate the TOV limit. This meant that the only thing that could prevent black holes from forming was a dynamic process ejecting sufficient mass from a star as it cooled.: 205 The modern concept of black holes was formulated by Robert Oppenheimer and his student Hartland Snyder in 1939.: 80 In the paper, Oppenheimer and Snyder solved Einstein's equations of general relativity for an idealized imploding star, in a model later called the Oppenheimer–Snyder model, then described the results from far outside the star. The implosion starts as one might expect: the star material rapidly collapses inward. However, as the density of the star increases, gravitational time dilation increases and the collapse, viewed from afar, seems to slow down further and further until the star reaches its Schwarzschild radius, where it appears frozen in time.: 217 In 1958, David Finkelstein identified the Schwarzschild surface as an event horizon, calling it "a perfect unidirectional membrane: causal influences can cross it in only one direction". In this sense, events that occur inside of the black hole cannot affect events that occur outside of the black hole. Finkelstein created a new reference frame to include the point of view of infalling observers.: 103 Finkelstein's new frame of reference allowed events at the surface of an imploding star to be related to events far away. By 1962 the two points of view were reconciled, convincing many skeptics that implosion into a black hole made physical sense.: 226 The era from the mid-1960s to the mid-1970s was the "golden age of black hole research", when general relativity and black holes became mainstream subjects of research.: 258 In this period, more general black hole solutions were found. In 1963, Roy Kerr found the exact solution for a rotating black hole. Two years later, Ezra Newman found the cylindrically symmetric solution for a black hole that is both rotating and electrically charged. In 1967, Werner Israel found that the Schwarzschild solution was the only possible solution for a nonspinning, uncharged black hole, meaning that a Schwarzschild black hole would be defined by its mass alone. Similar identities were later found for Reissner-Nordstrom and Kerr black holes, defined only by their mass and their charge or spin respectively. Together, these findings became known as the no-hair theorem, which states that a stationary black hole is completely described by the three parameters of the Kerr–Newman metric: mass, angular momentum, and electric charge. At first, it was suspected that the strange mathematical singularities found in each of the black hole solutions only appeared due to the assumption that a black hole would be perfectly spherically symmetric, and therefore the singularities would not appear in generic situations where black holes would not necessarily be symmetric. This view was held in particular by Vladimir Belinski, Isaak Khalatnikov, and Evgeny Lifshitz, who tried to prove that no singularities appear in generic solutions, although they would later reverse their positions. However, in 1965, Roger Penrose proved that general relativity without quantum mechanics requires that singularities appear in all black holes. Astronomical observations also made great strides during this era. In 1967, Antony Hewish and Jocelyn Bell Burnell discovered pulsars and by 1969, these were shown to be rapidly rotating neutron stars. Until that time, neutron stars, like black holes, were regarded as just theoretical curiosities, but the discovery of pulsars showed their physical relevance and spurred a further interest in all types of compact objects that might be formed by gravitational collapse. Based on observations in Greenwich and Toronto in the early 1970s, Cygnus X-1, a galactic X-ray source discovered in 1964, became the first astronomical object commonly accepted to be a black hole. Work by James Bardeen, Jacob Bekenstein, Carter, and Hawking in the early 1970s led to the formulation of black hole thermodynamics. These laws describe the behaviour of a black hole in close analogy to the laws of thermodynamics by relating mass to energy, area to entropy, and surface gravity to temperature. The analogy was completed: 442 when Hawking, in 1974, showed that quantum field theory implies that black holes should radiate like a black body with a temperature proportional to the surface gravity of the black hole, predicting the effect now known as Hawking radiation. While Cygnus X-1, a stellar-mass black hole, was generally accepted by the scientific community as a black hole by the end of 1973, it would be decades before a supermassive black hole would gain the same broad recognition. Although, as early as the 1960s, physicists such as Donald Lynden-Bell and Martin Rees had suggested that powerful quasars in the center of galaxies were powered by accreting supermassive black holes, little observational proof existed at the time. However, the Hubble Space Telescope, launched decades later, found that supermassive black holes were not only present in these active galactic nuclei, but that supermassive black holes in the center of galaxies were ubiquitous: Almost every galaxy had a supermassive black hole at its center, many of which were quiescent. In 1999, David Merritt proposed the M–sigma relation, which related the dispersion of the velocity of matter in the center bulge of a galaxy to the mass of the supermassive black hole at its core. Subsequent studies confirmed this correlation. Around the same time, based on telescope observations of the velocities of stars at the center of the Milky Way galaxy, independent work groups led by Andrea Ghez and Reinhard Genzel concluded that the compact radio source in the center of the galaxy, Sagittarius A*, was likely a supermassive black hole. On 11 February 2016, the LIGO Scientific Collaboration and Virgo Collaboration announced the first direct detection of gravitational waves, named GW150914, representing the first observation of a black hole merger. At the time of the merger, the black holes were approximately 1.4 billion light-years away from Earth and had masses of 30 and 35 solar masses.: 6 In 2017, Rainer Weiss, Kip Thorne, and Barry Barish, who had spearheaded the project, were awarded the Nobel Prize in Physics for their work. Since the initial discovery in 2015, hundreds more gravitational waves have been observed by LIGO and another interferometer, Virgo. On 10 April 2019, the first direct image of a black hole and its vicinity was published, following observations made by the Event Horizon Telescope (EHT) in 2017 of the supermassive black hole in Messier 87's galactic centre. In 2022, the Event Horizon Telescope collaboration released an image of the black hole in the center of the Milky Way galaxy, Sagittarius A*; The data had been collected in 2017. In 2020, the Nobel Prize in Physics was awarded for work on black holes. Andrea Ghez and Reinhard Genzel shared one-half for their discovery that Sagittarius A* is a supermassive black hole. Penrose received the other half for his work showing that the mathematics of general relativity requires the formation of black holes. Cosmologists lamented that Hawking's extensive theoretical work on black holes would not be honored since he died in 2018. In December 1967, a student reportedly suggested the phrase black hole at a lecture by John Wheeler; Wheeler adopted the term for its brevity and "advertising value", and Wheeler's stature in the field ensured it quickly caught on, leading some to credit Wheeler with coining the phrase. However, the term was used by others around that time. Science writer Marcia Bartusiak traces the term black hole to physicist Robert H. Dicke, who in the early 1960s reportedly compared the phenomenon to the Black Hole of Calcutta, notorious as a prison where people entered but never left alive. The term was used in print by Life and Science News magazines in 1963, and by science journalist Ann Ewing in her article "'Black Holes' in Space", dated 18 January 1964, which was a report on a meeting of the American Association for the Advancement of Science held in Cleveland, Ohio. Definition A black hole is generally defined as a region of spacetime from which no information-carrying signals or objects can escape. However, verifying an object as a black hole by this definition would require waiting for an infinite time and at an infinite distance from the black hole to verify that indeed, nothing has escaped, and thus cannot be used to identify a physical black hole. Broadly, physicists do not have a precisely-agreed-upon definition of a black hole. Among astrophysicists, a black hole is a compact object with a mass larger than four solar masses. A black hole may also be defined as a reservoir of information: 142 or a region where space is falling inwards faster than the speed of light. Properties The no-hair theorem postulates that, once it achieves a stable condition after formation, a black hole has only three independent physical properties: mass, electric charge, and angular momentum; the black hole is otherwise featureless. If the conjecture is true, any two black holes that share the same values for these properties, or parameters, are indistinguishable from one another. The degree to which the conjecture is true for real black holes is currently an unsolved problem. The simplest static black holes have mass but neither electric charge nor angular momentum. According to Birkhoff's theorem, these Schwarzschild black holes are the only vacuum solution that is spherically symmetric. Solutions describing more general black holes also exist. Non-rotating charged black holes are described by the Reissner–Nordström metric, while the Kerr metric describes a non-charged rotating black hole. The most general stationary black hole solution known is the Kerr–Newman metric, which describes a black hole with both charge and angular momentum. The simplest static black holes have mass but neither electric charge nor angular momentum. Contrary to the popular notion of a black hole "sucking in everything" in its surroundings, from far away, the external gravitational field of a black hole is identical to that of any other body of the same mass. While a black hole can theoretically have any positive mass, the charge and angular momentum are constrained by the mass. The total electric charge Q and the total angular momentum J are expected to satisfy the inequality Q 2 4 π ϵ 0 + c 2 J 2 G M 2 ≤ G M 2 {\displaystyle {\frac {Q^{2}}{4\pi \epsilon _{0}}}+{\frac {c^{2}J^{2}}{GM^{2}}}\leq GM^{2}} for a black hole of mass M. Black holes with the maximum possible charge or spin satisfying this inequality are called extremal black holes. Solutions of Einstein's equations that violate this inequality exist, but they do not possess an event horizon. These are so-called naked singularities that can be observed from the outside. Because these singularities make the universe inherently unpredictable, many physicists believe they could not exist. The weak cosmic censorship hypothesis, proposed by Sir Roger Penrose, rules out the formation of such singularities, when they are created through the gravitational collapse of realistic matter. However, this theory has not yet been proven, and some physicists believe that naked singularities could exist. It is also unknown whether black holes could even become extremal, forming naked singularities, since natural processes counteract increasing spin and charge when a black hole becomes near-extremal. The total mass of a black hole can be estimated by analyzing the motion of objects near the black hole, such as stars or gas. All black holes spin, often fast—One supermassive black hole, GRS 1915+105 has been estimated to spin at over 1,000 revolutions per second. The Milky Way's central black hole Sagittarius A* rotates at about 90% of the maximum rate. The spin rate can be inferred from measurements of atomic spectral lines in the X-ray range. As gas near the black hole plunges inward, high energy X-ray emission from electron-positron pairs illuminates the gas further out, appearing red-shifted due to relativistic effects. Depending on the spin of the black hole, this plunge happens at different radii from the hole, with different degrees of redshift. Astronomers can use the gap between the x-ray emission of the outer disk and the redshifted emission from plunging material to determine the spin of the black hole. A newer way to estimate spin is based on the temperature of gasses accreting onto the black hole. The method requires an independent measurement of the black hole mass and inclination angle of the accretion disk followed by computer modeling. Gravitational waves from coalescing binary black holes can also provide the spin of both progenitor black holes and the merged hole, but such events are rare. A spinning black hole has angular momentum. The supermassive black hole in the center of the Messier 87 (M87) galaxy appears to have an angular momentum very close to the maximum theoretical value. That uncharged limit is J ≤ G M 2 c , {\displaystyle J\leq {\frac {GM^{2}}{c}},} allowing definition of a dimensionless spin magnitude such that 0 ≤ c J G M 2 ≤ 1. {\displaystyle 0\leq {\frac {cJ}{GM^{2}}}\leq 1.} Most black holes are believed to have an approximately neutral charge. For example, Michal Zajaček, Arman Tursunov, Andreas Eckart, and Silke Britzen found the electric charge of Sagittarius A* to be at least ten orders of magnitude below the theoretical maximum. A charged black hole repels other like charges just like any other charged object. If a black hole were to become charged, particles with an opposite sign of charge would be pulled in by the extra electromagnetic force, while particles with the same sign of charge would be repelled, neutralizing the black hole. This effect may not be as strong if the black hole is also spinning. The presence of charge can reduce the diameter of the black hole by up to 38%. The charge Q for a nonspinning black hole is bounded by Q ≤ G M , {\displaystyle Q\leq {\sqrt {G}}M,} where G is the gravitational constant and M is the black hole's mass. Classification Black holes can have a wide range of masses. The minimum mass of a black hole formed by stellar gravitational collapse is governed by the maximum mass of a neutron star and is believed to be approximately two-to-four solar masses. However, theoretical primordial black holes, believed to have formed soon after the Big Bang, could be far smaller, with masses as little as 10−5 grams at formation. These very small black holes are sometimes called micro black holes. Black holes formed by stellar collapse are called stellar black holes. Estimates of their maximum mass at formation vary, but generally range from 10 to 100 solar masses, with higher estimates for black holes progenated by low-metallicity stars. The mass of a black hole formed via a supernova has a lower bound: If the progenitor star is too small, the collapse may be stopped by the degeneracy pressure of the star's constituents, allowing the condensation of matter into an exotic denser state. Degeneracy pressure occurs from the Pauli exclusion principle—Particles will resist being in the same place as each other. Smaller progenitor stars, with masses less than about 8 M☉, will be held together by the degeneracy pressure of electrons and will become a white dwarf. For more massive progenitor stars, electron degeneracy pressure is no longer strong enough to resist the force of gravity and the star will be held together by neutron degeneracy pressure, which can occur at much higher densities, forming a neutron star. If the star is still too massive, even neutron degeneracy pressure will not be able to resist the force of gravity and the star will collapse into a black hole.: 5.8 Stellar black holes can also gain mass via accretion of nearby matter, often from a companion object such as a star. Black holes that are larger than stellar black holes but smaller than supermassive black holes are called intermediate-mass black holes, with masses of approximately 102 to 105 solar masses. These black holes seem to be rarer than their stellar and supermassive counterparts, with relatively few candidates having been observed. Physicists have speculated that such black holes may form from collisions in globular and star clusters or at the center of low-mass galaxies. They may also form as the result of mergers of smaller black holes, with several LIGO observations finding merged black holes within the 110-350 solar mass range. The black holes with the largest masses are called supermassive black holes, with masses more than 106 times that of the Sun. These black holes are believed to exist at the centers of almost every large galaxy, including the Milky Way. Some scientists have proposed a subcategory of even larger black holes, called ultramassive black holes, with masses greater than 109-1010 solar masses. Theoretical models predict that the accretion disc that feeds black holes will be unstable once a black hole reaches 50-100 billion times the mass of the Sun, setting a rough upper limit to black hole mass. Structure While black holes are conceptually invisible sinks of all matter and light, in astronomical settings, their enormous gravity alters the motion of surrounding objects and pulls nearby gas inwards at near-light speed, making the area around black holes the brightest objects in the universe. Some black holes have relativistic jets—thin streams of plasma travelling away from the black hole at more than one-tenth of the speed of light. A small faction of the matter falling towards the black hole gets accelerated away along the hole rotation axis. These jets can extend as far as millions of parsecs from the black hole itself. Black holes of any mass can have jets. However, they are typically observed around spinning black holes with strongly-magnetized accretion disks. Relativistic jets were more common in the early universe, when galaxies and their corresponding supermassive black holes were rapidly gaining mass. All black holes with jets also have an accretion disk, but the jets are usually brighter than the disk. Quasars, typically found in other galaxies, are believed to be supermassive black holes with jets; microquasars are believed to be stellar-mass objects with jets, typically observed in the Milky Way. The mechanism of formation of jets is not yet known, but several options have been proposed. One method proposed to fuel these jets is the Blandford-Znajek process, which suggests that the dragging of magnetic field lines by a black hole's rotation could launch jets of matter into space. The Penrose process, which involves extraction of a black hole's rotational energy, has also been proposed as a potential mechanism of jet propulsion. Due to conservation of angular momentum, gas falling into the gravitational well created by a massive object will typically form a disk-like structure around the object.: 242 As the disk's angular momentum is transferred outward due to internal processes, its matter falls farther inward, converting its gravitational energy into heat and releasing a large flux of x-rays. The temperature of these disks can range from thousands to millions of Kelvin, and temperatures can differ throughout a single accretion disk. Accretion disks can also emit in other parts of the electromagnetic spectrum, depending on the disk's turbulence and magnetization and the black hole's mass and angular momentum. Accretion disks can be defined as geometrically thin or geometrically thick. Geometrically thin disks are mostly confined to the black hole's equatorial plane and have a well-defined edge at the innermost stable circular orbit (ISCO), while geometrically thick disks are supported by internal pressure and temperature and can extend inside the ISCO. Disks with high rates of electron scattering and absorption, appearing bright and opaque, are called optically thick; optically thin disks are more translucent and produce fainter images when viewed from afar. Accretion disks of black holes accreting beyond the Eddington limit are often referred to as polish donuts due to their thick, toroidal shape that resembles that of a donut. Quasar accretion disks are expected to usually appear blue in color. The disk for a stellar black hole, on the other hand, would likely look orange, yellow, or red, with its inner regions being the brightest. Theoretical research suggests that the hotter a disk is, the bluer it should be, although this is not always supported by observations of real astronomical objects. Accretion disk colors may also be altered by the Doppler effect, with the part of the disk travelling towards an observer appearing bluer and brighter and the part of the disk travelling away from the observer appearing redder and dimmer. In Newtonian gravity, test particles can stably orbit at arbitrary distances from a central object. In general relativity, however, there exists a smallest possible radius for which a massive particle can orbit stably. Any infinitesimal inward perturbations to this orbit will lead to the particle spiraling into the black hole, and any outward perturbations will, depending on the energy, cause the particle to spiral in, move to a stable orbit further from the black hole, or escape to infinity. This orbit is called the innermost stable circular orbit, or ISCO. The location of the ISCO depends on the spin of the black hole and the spin of the particle itself. In the case of a Schwarzschild black hole (spin zero) and a particle without spin, the location of the ISCO is: r I S C O = 3 r s = 6 G M c 2 , {\displaystyle r_{\rm {ISCO}}=3\,r_{\text{s}}={\frac {6\,GM}{c^{2}}},} where r I S C O {\displaystyle r_{\rm {_{ISCO}}}} is the radius of the ISCO, r s {\displaystyle r_{\text{s}}} is the Schwarzschild radius of the black hole, G {\displaystyle G} is the gravitational constant, and c {\displaystyle c} is the speed of light. The radius of this orbit changes slightly based on particle spin. For charged black holes, the ISCO moves inwards. For spinning black holes, the ISCO is moved inwards for particles orbiting in the same direction that the black hole is spinning (prograde) and outwards for particles orbiting in the opposite direction (retrograde). For example, the ISCO for a particle orbiting retrograde can be as far out as about 9 r s {\displaystyle 9r_{\text{s}}} , while the ISCO for a particle orbiting prograde can be as close as at the event horizon itself. The photon sphere is a spherical boundary for which photons moving on tangents to that sphere are bent completely around the black hole, possibly orbiting multiple times. Light rays with impact parameters less than the radius of the photon sphere enter the black hole. For Schwarzschild black holes, the photon sphere has a radius 1.5 times the Schwarzschild radius; the radius for non-Schwarzschild black holes is at least 1.5 times the radius of the event horizon. When viewed from a great distance, the photon sphere creates an observable black hole shadow. Since no light emerges from within the black hole, this shadow is the limit for possible observations.: 152 The shadow of colliding black holes should have characteristic warped shapes, allowing scientists to detect black holes that are about to merge. While light can still escape from the photon sphere, any light that crosses the photon sphere on an inbound trajectory will be captured by the black hole. Therefore, any light that reaches an outside observer from the photon sphere must have been emitted by objects between the photon sphere and the event horizon. Light emitted towards the photon sphere may also curve around the black hole and return to the emitter. For a rotating, uncharged black hole, the radius of the photon sphere depends on the spin parameter and whether the photon is orbiting prograde or retrograde. For a photon orbiting prograde, the photon sphere will be 1-3 Schwarzschild radii from the center of the black hole, while for a photon orbiting retrograde, the photon sphere will be between 3-5 Schwarzschild radii from the center of the black hole. The exact location of the photon sphere depends on the magnitude of the black hole's rotation. For a charged, nonrotating black hole, there will only be one photon sphere, and the radius of the photon sphere will decrease for increasing black hole charge. For non-extremal, charged, rotating black holes, there will always be two photon spheres, with the exact radii depending on the parameters of the black hole. Near a rotating black hole, spacetime rotates similar to a vortex. The rotating spacetime will drag any matter and light into rotation around the spinning black hole. This effect of general relativity, called frame dragging, gets stronger closer to the spinning mass. The region of spacetime in which it is impossible to stay still is called the ergosphere. The ergosphere of a black hole is a volume bounded by the black hole's event horizon and the ergosurface, which coincides with the event horizon at the poles but bulges out from it around the equator. Matter and radiation can escape from the ergosphere. Through the Penrose process, objects can emerge from the ergosphere with more energy than they entered with. The extra energy is taken from the rotational energy of the black hole, slowing down the rotation of the black hole.: 268 A variation of the Penrose process in the presence of strong magnetic fields, the Blandford–Znajek process, is considered a likely mechanism for the enormous luminosity and relativistic jets of quasars and other active galactic nuclei. The observable region of spacetime around a black hole closest to its event horizon is called the plunging region. In this area it is no longer possible for free falling matter to follow circular orbits or stop a final descent into the black hole. Instead, it will rapidly plunge toward the black hole at close to the speed of light, growing increasingly hot and producing a characteristic, detectable thermal emission. However, light and radiation emitted from this region can still escape from the black hole's gravitational pull. For a nonspinning, uncharged black hole, the radius of the event horizon, or Schwarzschild radius, is proportional to the mass, M, through r s = 2 G M c 2 ≈ 2.95 M M ⊙ k m , {\displaystyle r_{\mathrm {s} }={\frac {2GM}{c^{2}}}\approx 2.95\,{\frac {M}{M_{\odot }}}~\mathrm {km,} } where rs is the Schwarzschild radius and M☉ is the mass of the Sun.: 124 For a black hole with nonzero spin or electric charge, the radius is smaller,[Note 1] until an extremal black hole could have an event horizon close to r + = G M c 2 , {\displaystyle r_{\mathrm {+} }={\frac {GM}{c^{2}}},} half the radius of a nonspinning, uncharged black hole of the same mass. Since the volume within the Schwarzschild radius increase with the cube of the radius, average density of a black hole inside its Schwarzschild radius is inversely proportional to the square of its mass: supermassive black holes are much less dense than stellar black holes. The average density of a 108 M☉ black hole is comparable to that of water. The defining feature of a black hole is the existence of an event horizon, a boundary in spacetime through which matter and light can pass only inward towards the center of the black hole. Nothing, not even light, can escape from inside the event horizon. The event horizon is referred to as such because if an event occurs within the boundary, information from that event cannot reach or affect an outside observer, making it impossible to determine whether such an event occurred.: 179 For non-rotating black holes, the geometry of the event horizon is precisely spherical, while for rotating black holes, the event horizon is oblate. To a distant observer, a clock near a black hole would appear to tick more slowly than one further from the black hole.: 217 This effect, known as gravitational time dilation, would also cause an object falling into a black hole to appear to slow as it approached the event horizon, never quite reaching the horizon from the perspective of an outside observer.: 218 All processes on this object would appear to slow down, and any light emitted by the object to appear redder and dimmer, an effect known as gravitational redshift. An object falling from half of a Schwarzschild radius above the event horizon would fade away until it could no longer be seen, disappearing from view within one hundredth of a second. It would also appear to flatten onto the black hole, joining all other material that had ever fallen into the hole. On the other hand, an observer falling into a black hole would not notice any of these effects as they cross the event horizon. Their own clocks appear to them to tick normally, and they cross the event horizon after a finite time without noting any singular behaviour. In general relativity, it is impossible to determine the location of the event horizon from local observations, due to Einstein's equivalence principle.: 222 Black holes that are rotating and/or charged have an inner horizon, often called the Cauchy horizon, inside of the black hole. The inner horizon is divided up into two segments: an ingoing section and an outgoing section. At the ingoing section of the Cauchy horizon, radiation and matter that fall into the black hole would build up at the horizon, causing the curvature of spacetime to go to infinity. This would cause an observer falling in to experience tidal forces. This phenomenon is often called mass inflation, since it is associated with a parameter dictating the black hole's internal mass growing exponentially, and the buildup of tidal forces is called the mass-inflation singularity or Cauchy horizon singularity. Some physicists have argued that in realistic black holes, accretion and Hawking radiation would stop mass inflation from occurring. At the outgoing section of the inner horizon, infalling radiation would backscatter off of the black hole's spacetime curvature and travel outward, building up at the outgoing Cauchy horizon. This would cause an infalling observer to experience a gravitational shock wave and tidal forces as the spacetime curvature at the horizon grew to infinity. This buildup of tidal forces is called the shock singularity. Both of these singularities are weak, meaning that an object crossing them would only be deformed a finite amount by tidal forces, even though the spacetime curvature would still be infinite at the singularity. This is as opposed to a strong singularity, where an object hitting the singularity would be stretched and squeezed by an infinite amount. They are also null singularities, meaning that a photon could travel parallel to the them without ever being intercepted. Ignoring quantum effects, every black hole has a singularity inside, points where the curvature of spacetime becomes infinite, and geodesics terminate within a finite proper time.: 205 For a non-rotating black hole, this region takes the shape of a single point; for a rotating black hole it is smeared out to form a ring singularity that lies in the plane of rotation.: 264 In both cases, the singular region has zero volume. All of the mass of the black hole ends up in the singularity.: 252 Since the singularity has nonzero mass in an infinitely small space, it can be thought of as having infinite density. Observers falling into a Schwarzschild black hole (i.e., non-rotating and not charged) cannot avoid being carried into the singularity once they cross the event horizon. As they fall further into the black hole, they will be torn apart by the growing tidal forces in a process sometimes referred to as spaghettification or the noodle effect. Eventually, they will reach the singularity and be crushed into an infinitely small point.: 182 However any perturbations, such as those caused by matter or radiation falling in, would cause space to oscillate chaotically near the singularity. Any matter falling in would experience intense tidal forces rapidly changing in direction, all while being compressed into an increasingly small volume. Alternative forms of general relativity, including addition of some quatum effects, can lead to regular, or nonsingular, black holes without singularities. For example, the fuzzball model, based on string theory, states that black holes are actually made up of quantum microstates and need not have a singularity or an event horizon. The theory of loop quantum gravity proposes that the curvature and density at the center of a black hole is large, but not infinite. Formation Black holes are formed by gravitational collapse of massive stars, either by direct collapse or during a supernova explosion in a process called fallback. Black holes can result from the merger of two neutron stars or a neutron star and a black hole. Other more speculative mechanisms include primordial black holes created from density fluctuations in the early universe, the collapse of dark stars, a hypothetical object powered by annihilation of dark matter, or from hypothetical self-interacting dark matter. Gravitational collapse occurs when an object's internal pressure is insufficient to resist the object's own gravity. At the end of a star's life, it will run out of hydrogen to fuse, and will start fusing more and more massive elements, until it gets to iron. Since the fusion of elements heavier than iron would require more energy than it would release, nuclear fusion ceases. If the iron core of the star is too massive, the star will no longer be able to support itself and will undergo gravitational collapse. While most of the energy released during gravitational collapse is emitted very quickly, an outside observer does not actually see the end of this process. Even though the collapse takes a finite amount of time from the reference frame of infalling matter, a distant observer would see the infalling material slow and halt just above the event horizon, due to gravitational time dilation. Light from the collapsing material takes longer and longer to reach the observer, with the delay growing to infinity as the emitting material reaches the event horizon. Thus the external observer never sees the formation of the event horizon; instead, the collapsing material seems to become dimmer and increasingly red-shifted, eventually fading away. Observations of quasars at redshift z ∼ 7 {\displaystyle z\sim 7} , less than a billion years after the Big Bang, has led to investigations of other ways to form black holes. The accretion process to build supermassive black holes has a limiting rate of mass accumulation and a billion years is not enough time to reach quasar status. One suggestion is direct collapse of nearly pure hydrogen gas (low metalicity) clouds characteristic of the young universe, forming a supermassive star which collapses into a black hole. It has been suggested that seed black holes with typical masses of ~105 M☉ could have formed in this way which then could grow to ~109 M☉. However, the very large amount of gas required for direct collapse is not typically stable to fragmentation to form multiple stars. Thus another approach suggests massive star formation followed by collisions that seed massive black holes which ultimately merge to create a quasar.: 85 A neutron star in a common envelope with a regular star can accrete sufficient material to collapse to a black hole or two neutron stars can merge. These avenues for the formation of black holes are considered relatively rare. In the current epoch of the universe, conditions needed to form black holes are rare and are mostly only found in stars. However, in the early universe, conditions may have allowed for black hole formations via other means. Fluctuations of spacetime soon after the Big Bang may have formed areas that were denser then their surroundings. Initially, these regions would not have been compact enough to form a black hole, but eventually, the curvature of spacetime in the regions become large enough to cause them to collapse into a black hole. Different models for the early universe vary widely in their predictions of the scale of these fluctuations. Various models predict the creation of primordial black holes ranging from a Planck mass (~2.2×10−8 kg) to hundreds of thousands of solar masses. Primordial black holes with masses less than 1015 g would have evaporated by now due to Hawking radiation. Despite the early universe being extremely dense, it did not re-collapse into a black hole during the Big Bang, since the universe was expanding rapidly and did not have the gravitational differential necessary for black hole formation. Models for the gravitational collapse of objects of relatively constant size, such as stars, do not necessarily apply in the same way to rapidly expanding space such as the Big Bang. In principle, black holes could be formed in high-energy particle collisions that achieve sufficient density, although no such events have been detected. These hypothetical micro black holes, which could form from the collision of cosmic rays and Earth's atmosphere or in particle accelerators like the Large Hadron Collider, would not be able to aggregate additional mass. Instead, they would evaporate in about 10−25 seconds, posing no threat to the Earth. Evolution Black holes can also merge with other objects such as stars or even other black holes. This is thought to have been important, especially in the early growth of supermassive black holes, which could have formed from the aggregation of many smaller objects. The process has also been proposed as the origin of some intermediate-mass black holes. Mergers of supermassive black holes may take a long time: As a binary of supermassive black holes approach each other, most nearby stars are ejected, leaving little for the remaining black holes to gravitationally interact with that would allow them to get closer to each other. This phenomenon has been called the final parsec problem, as the distance at which this happens is usually around one parsec. When a black hole accretes matter, the gas in the inner accretion disk orbits at very high speeds because of its proximity to the black hole. The resulting friction heats the inner disk to temperatures at which it emits vast amounts of electromagnetic radiation (mainly X-rays) detectable by telescopes. By the time the matter of the disk reaches the ISCO, between 5.7% and 42% of its mass will have been converted to energy, depending on the black hole's spin. About 90% of this energy is released within about 20 black hole radii. In many cases, accretion disks are accompanied by relativistic jets that are emitted along the black hole's poles, which carry away much of the energy. The mechanism for the creation of these jets is currently not well understood, in part due to insufficient data. Many of the universe's most energetic phenomena have been attributed to the accretion of matter on black holes. Active galactic nuclei and quasars are believed to be the accretion disks of supermassive black holes. X-ray binaries are generally accepted to be binary systems in which one of the two objects is a compact object accreting matter from its companion. Ultraluminous X-ray sources may be the accretion disks of intermediate-mass black holes. At a certain rate of accretion, the outward radiation pressure will become as strong as the inward gravitational force, and the black hole should unable to accrete any faster. This limit is called the Eddington limit. However, many black holes accrete beyond this rate due to their non-spherical geometry or instabilities in the accretion disk. Accretion beyond the limit is called Super-Eddington accretion and may have been commonplace in the early universe. Stars have been observed to get torn apart by tidal forces in the immediate vicinity of supermassive black holes in galaxy nuclei, in what is known as a tidal disruption event (TDE). Some of the material from the disrupted star forms an accretion disk around the black hole, which emits observable electromagnetic radiation. The correlation between the masses of supermassive black holes in the centres of galaxies with the velocity dispersion and mass of stars in their host bulges suggests that the formation of galaxies and the formation of their central black holes are related. Black hole winds from rapid accretion, particularly when the galaxy itself is still accreting matter, can compress gas nearby, accelerating star formation. However, if the winds become too strong, the black hole may blow nearly all of the gas out of the galaxy, quenching star formation. Black hole jets may also energize nearby cavities of plasma and eject low-entropy gas from out of the galactic core, causing gas in galactic centers to be hotter than expected. If Hawking's theory of black hole radiation is correct, then black holes are expected to shrink and evaporate over time as they lose mass by the emission of photons and other particles. The temperature of this thermal spectrum (Hawking temperature) is proportional to the surface gravity of the black hole, which is inversely proportional to the mass. Hence, large black holes emit less radiation than small black holes.: Ch. 9.6 A stellar black hole of 1 M☉ has a Hawking temperature of 62 nanokelvins. This is far less than the 2.7 K temperature of the cosmic microwave background radiation. Stellar-mass or larger black holes receive more mass from the cosmic microwave background than they emit through Hawking radiation and thus will grow instead of shrinking. To have a Hawking temperature larger than 2.7 K (and be able to evaporate), a black hole would need a mass less than the Moon. Such a black hole would have a diameter of less than a tenth of a millimetre. The Hawking radiation for an astrophysical black hole is predicted to be very weak and would thus be exceedingly difficult to detect from Earth. A possible exception is the burst of gamma rays emitted in the last stage of the evaporation of primordial black holes. Searches for such flashes have proven unsuccessful and provide stringent limits on the possibility of existence of low mass primordial black holes, with modern research predicting that primordial black holes must make up less than a fraction of 10−7 of the universe's total mass. NASA's Fermi Gamma-ray Space Telescope, launched in 2008, has searched for these flashes, but has not yet found any. The properties of a black hole are constrained and interrelated by the theories that predict these properties. When based on general relativity, these relationships are called the laws of black hole mechanics. For a black hole that is not still forming or accreting matter, the zeroth law of black hole mechanics states the black hole's surface gravity is constant across the event horizon. The first law relates changes in the black hole's surface area, angular momentum, and charge to changes in its energy. The second law says the surface area of a black hole never decreases on its own. Finally, the third law says that the surface gravity of a black hole is never zero. These laws are mathematical analogs of the laws of thermodynamics. They are not equivalent, however, because, according to general relativity without quantum mechanics, a black hole can never emit radiation, and thus its temperature must always be zero.: 11 Quantum mechanics predicts that a black hole will continuously emit thermal Hawking radiation, and therefore must always have a nonzero temperature. It also predicts that all black holes have entropy which scales with their surface area. When quantum mechanics is accounted for, the laws of black hole mechanics become equivalent to the classical laws of thermodynamics. However, these conclusions are derived without a complete theory of quantum gravity, although many potential theories do predict black holes having entropy and temperature. Thus, the true quantum nature of black hole thermodynamics continues to be debated.: 29 Observational evidence Millions of black holes with around 30 solar masses derived from stellar collapse are expected to exist in the Milky Way. Even a dwarf galaxy like Draco should have hundreds. Only a few of these have been detected. By nature, black holes do not themselves emit any electromagnetic radiation other than the hypothetical Hawking radiation, so astrophysicists searching for black holes must generally rely on indirect observations. The defining characteristic of a black hole is its event horizon. The horizon itself cannot be imaged, so all other possible explanations for these indirect observations must be considered and eliminated before concluding that a black hole has been observed.: 11 The Event Horizon Telescope (EHT) is a global system of radio telescopes capable of directly observing a black hole shadow. The angular resolution of a telescope is based on its aperture and the wavelengths it is observing. Because the angular diameters of Sagittarius A* and Messier 87* in the sky are very small, a single telescope would need to be about the size of the Earth to clearly distinguish their horizons using radio wavelengths. By combining data from several different radio telescopes around the world, the Event Horizon Telescope creates an effective aperture the diameter size of the Earth. The EHT team used imaging algorithms to compute the most probable image from the data in its observations of Sagittarius A* and M87*. Gravitational-wave interferometry can be used to detect merging black holes and other compact objects. In this method, a laser beam is split down two long arms of a tunnel. The laser beams reflect off of mirrors in the tunnels and converge at the intersection of the arms, cancelling each other out. However, when a gravitational wave passes, it warps spacetime, changing the lengths of the arms themselves. Since each laser beam is now travelling a slightly different distance, they do not cancel out and produce a recognizable signal. Analysis of the signal can give scientists information about what caused the gravitational waves. Since gravitational waves are very weak, gravitational-wave observatories such as LIGO must have arms several kilometers long and carefully control for noise from Earth to be able to detect these gravitational waves. Since the first measurements in 2016, multiple gravitational waves from black holes have been detected and analyzed. The proper motions of stars near the centre of the Milky Way provide strong observational evidence that these stars are orbiting a supermassive black hole. Since 1995, astronomers have tracked the motions of 90 stars orbiting an invisible object coincident with the radio source Sagittarius A*. In 1998, by fitting the motions of the stars to Keplerian orbits, the astronomers were able to infer that Sagittarius A* must be a 2.6×106 M☉ object must be contained within a radius of 0.02 light-years. Since then, one of the stars—called S2—has completed a full orbit. From the orbital data, astronomers were able to refine the calculations of the mass of Sagittarius A* to 4.3×106 M☉, with a radius of less than 0.002 light-years. This upper limit radius is larger than the Schwarzschild radius for the estimated mass, so the combination does not prove Sagittarius A* is a black hole. Nevertheless, these observations strongly suggest that the central object is a supermassive black hole as there are no other plausible scenarios for confining so much invisible mass into such a small volume. Additionally, there is some observational evidence that this object might possess an event horizon, a feature unique to black holes. The Event Horizon Telescope image of Sagittarius A*, released in 2022, provided further confirmation that it is indeed a black hole. X-ray binaries are binary systems that emit a majority of their radiation in the X-ray part of the electromagnetic spectrum. These X-ray emissions result when a compact object accretes matter from an ordinary star. The presence of an ordinary star in such a system provides an opportunity for studying the central object and to determine if it might be a black hole. By measuring the orbital period of the binary, the distance to the binary from Earth, and the mass of the companion star, scientists can estimate the mass of the compact object. The Tolman-Oppenheimer-Volkoff limit (TOV limit) dictates the largest mass a nonrotating neutron star can be, and is estimated to be about two solar masses. While a rotating neutron star can be slightly more massive, if the compact object is much more massive than the TOV limit, it cannot be a neutron star and is generally expected to be a black hole. The first strong candidate for a black hole, Cygnus X-1, was discovered in this way by Charles Thomas Bolton, Louise Webster, and Paul Murdin in 1972. Observations of rotation broadening of the optical star reported in 1986 lead to a compact object mass estimate of 16 solar masses, with 7 solar masses as the lower bound. In 2011, this estimate was updated to 14.1±1.0 M☉ for the black hole and 19.2±1.9 M☉ for the optical stellar companion. X-ray binaries can be categorized as either low-mass or high-mass; This classification is based on the mass of the companion star, not the compact object itself. In a class of X-ray binaries called soft X-ray transients, the companion star is of relatively low mass, allowing for more accurate estimates of the black hole mass. These systems actively emit X-rays for only several months once every 10–50 years. During the period of low X-ray emission, called quiescence, the accretion disk is extremely faint, allowing detailed observation of the companion star. Numerous black hole candidates have been measured by this method. Black holes are also sometimes found in binaries with other compact objects, such as white dwarfs, neutron stars, and other black holes. The centre of nearly every galaxy contains a supermassive black hole. The close observational correlation between the mass of this hole and the velocity dispersion of the host galaxy's bulge, known as the M–sigma relation, strongly suggests a connection between the formation of the black hole and that of the galaxy itself. Astronomers use the term active galaxy to describe galaxies with unusual characteristics, such as unusual spectral line emission and very strong radio emission. Theoretical and observational studies have shown that the high levels of activity in the centers of these galaxies, regions called active galactic nuclei (AGN), may be explained by accretion onto supermassive black holes. These AGN consist of a central black hole that may be millions or billions of times more massive than the Sun, a disk of interstellar gas and dust called an accretion disk, and two jets perpendicular to the accretion disk. Although supermassive black holes are expected to be found in most AGN, only some galaxies' nuclei have been more carefully studied in attempts to both identify and measure the actual masses of the central supermassive black hole candidates. Some of the most notable galaxies with supermassive black hole candidates include the Andromeda Galaxy, Messier 32, Messier 87, the Sombrero Galaxy, and the Milky Way itself. Another way black holes can be detected is through observation of effects caused by their strong gravitational field. One such effect is gravitational lensing: The deformation of spacetime around a massive object causes light rays to be deflected, making objects behind them appear distorted. When the lensing object is a black hole, this effect can be strong enough to create multiple images of a star or other luminous source. However, the distance between the lensed images may be too small for contemporary telescopes to resolve—this phenomenon is called microlensing. Instead of seeing two images of a lensed star, astronomers see the star brighten slightly as the black hole moves towards the line of sight between the star and Earth and then return to its normal luminosity as the black hole moves away. The turn of the millennium saw the first 3 candidate detections of black holes in this way, and in January 2022, astronomers reported the first confirmed detection of a microlensing event from an isolated black hole. This was also the first determination of an isolated black hole mass, 7.1±1.3 M☉. Alternatives While there is a strong case for supermassive black holes, the model for stellar-mass black holes assumes of an upper limit for the mass of a neutron star: objects observed to have more mass are assumed to be black holes. However, the properties of extremely dense matter are poorly understood. New exotic phases of matter could allow other kinds of massive objects. Quark stars would be made up of quark matter and supported by quark degeneracy pressure, a form of degeneracy pressure even stronger than neutron degeneracy pressure. This would halt gravitational collapse at a higher mass than for a neutron star. Even stronger stars called electroweak stars would convert quarks in their cores into leptons, providing additional pressure to stop the star from collapsing. If, as some extensions of the Standard Model posit, quarks and leptons are made up of the even-smaller fundamental particles called preons, a very compact star could be supported by preon degeneracy pressure. While none of these hypothetical models can explain all of the observations of stellar black hole candidates, a Q star is the only alternative which could significantly exceed the mass limit for neutron stars and thus provide an alternative for supermassive black holes.: 12 A few theoretical objects have been conjectured to match observations of astronomical black hole candidates identically or near-identically, but which function via a different mechanism. A dark energy star would convert infalling matter into vacuum energy; This vacuum energy would be much larger than the vacuum energy of outside space, exerting outwards pressure and preventing a singularity from forming. A black star would be gravitationally collapsing slowly enough that quantum effects would keep it just on the cusp of fully collapsing into a black hole. A gravastar would consist of a very thin shell and a dark-energy interior providing outward pressure to stop the collapse into a black hole or formation of a singularity; It could even have another gravastar inside, called a 'nestar'. Open questions According to the no-hair theorem, a black hole is defined by only three parameters: its mass, charge, and angular momentum. This seems to mean that all other information about the matter that went into forming the black hole is lost, as there is no way to determine anything about the black hole from outside other than those three parameters. When black holes were thought to persist forever, this information loss was not problematic, as the information can be thought of as existing inside the black hole. However, black holes slowly evaporate by emitting Hawking radiation. This radiation does not appear to carry any additional information about the matter that formed the black hole, meaning that this information is seemingly gone forever. This is called the black hole information paradox. Theoretical studies analyzing the paradox have led to both further paradoxes and new ideas about the intersection of quantum mechanics and general relativity. While there is no consensus on the resolution of the paradox, work on the problem is expected to be important for a theory of quantum gravity.: 126 Observations of faraway galaxies have found that ultraluminous quasars, powered by supermassive black holes, existed in the early universe as far as redshift z ≥ 7 {\displaystyle z\geq 7} . These black holes have been assumed to be the products of the gravitational collapse of large population III stars. However, these stellar remnants were not massive enough to produce the quasars observed at early times without accreting beyond the Eddington limit, the theoretical maximum rate of black hole accretion. Physicists have suggested a variety of different mechanisms by which these supermassive black holes may have formed. It has been proposed that smaller black holes may have also undergone mergers to produce the observed supermassive black holes. It is also possible that they were seeded by direct-collapse black holes, in which a large cloud of hot gas avoids fragmentation that would lead to multiple stars, due to low angular momentum or heating from a nearby galaxy. Given the right circumstances, a single supermassive star forms and collapses directly into a black hole without undergoing typical stellar evolution. Additionally, these supermassive black holes in the early universe may be high-mass primordial black holes, which could have accreted further matter in the centers of galaxies. Finally, certain mechanisms allow black holes to grow faster than the theoretical Eddington limit, such as dense gas in the accretion disk limiting outward radiation pressure that prevents the black hole from accreting. However, the formation of bipolar jets prevent super-Eddington rates. In fiction Black holes have been portrayed in science fiction in a variety of ways. Even before the advent of the term itself, objects with characteristics of black holes appeared in stories such as the 1928 novel The Skylark of Space with its "black Sun" and the "hole in space" in the 1935 short story Starship Invincible. As black holes grew to public recognition in the 1960s and 1970s, they began to be featured in films as well as novels, such as Disney's The Black Hole. Black holes have also been used in works of the 21st century, such as Christopher Nolan's science fiction epic Interstellar. Authors and screenwriters have exploited the relativistic effects of black holes, particularly gravitational time dilation. For example, Interstellar features a black hole planet with a time dilation factor of over 60,000:1, while the 1977 novel Gateway depicts a spaceship approaching but never crossing the event horizon of a black hole from the perspective of an outside observer due to time dilation effects. Black holes have also been appropriated as wormholes or other methods of faster-than-light travel, such as in the 1974 novel The Forever War, where a network of black holes is used for interstellar travel. Additionally, black holes can feature as hazards to spacefarers and planets: A black hole threatens a deep-space outpost in 1978 short story The Black Hole Passes, and a binary black hole dangerously alters the orbit of a planet in the 2018 Netflix reboot of Lost in Space. Notes References Further reading External links
========================================
[SOURCE: https://en.wikipedia.org/w/index.php?title=Non-player_character&action=edit&section=5] | [TOKENS: 1431]
Editing Non-player character (section) Copy and paste: – — ° ′ ″ ≈ ≠ ≤ ≥ ± − × ÷ ← → · § Cite your sources: <ref></ref> {{}} {{{}}} | [] [[]] [[Category:]] #REDIRECT [[]] &nbsp; <s></s> <sup></sup> <sub></sub> <code></code> <pre></pre> <blockquote></blockquote> <ref></ref> <ref name="" /> {{Reflist}} <references /> <includeonly></includeonly> <noinclude></noinclude> {{DEFAULTSORT:}} <nowiki></nowiki> <!-- --> <span class="plainlinks"></span> Symbols: ~ | ¡ ¿ † ‡ ↔ ↑ ↓ • ¶ # ∞ ‹› «» ¤ ₳ ฿ ₵ ¢ ₡ ₢ $ ₫ ₯ € ₠ ₣ ƒ ₴ ₭ ₤ ℳ ₥ ₦ ₧ ₰ £ ៛ ₨ ₪ ৳ ₮ ₩ ¥ ♠ ♣ ♥ ♦ 𝄫 ♭ ♮ ♯ 𝄪 © ¼ ½ ¾ Latin: A a Á á À à  â Ä ä Ǎ ǎ Ă ă Ā ā à ã Å å Ą ą Æ æ Ǣ ǣ B b C c Ć ć Ċ ċ Ĉ ĉ Č č Ç ç D d Ď ď Đ đ Ḍ ḍ Ð ð E e É é È è Ė ė Ê ê Ë ë Ě ě Ĕ ĕ Ē ē Ẽ ẽ Ę ę Ẹ ẹ Ɛ ɛ Ǝ ǝ Ə ə F f G g Ġ ġ Ĝ ĝ Ğ ğ Ģ ģ H h Ĥ ĥ Ħ ħ Ḥ ḥ I i İ ı Í í Ì ì Î î Ï ï Ǐ ǐ Ĭ ĭ Ī ī Ĩ ĩ Į į Ị ị J j Ĵ ĵ K k Ķ ķ L l Ĺ ĺ Ŀ ŀ Ľ ľ Ļ ļ Ł ł Ḷ ḷ Ḹ ḹ M m Ṃ ṃ N n Ń ń Ň ň Ñ ñ Ņ ņ Ṇ ṇ Ŋ ŋ O o Ó ó Ò ò Ô ô Ö ö Ǒ ǒ Ŏ ŏ Ō ō Õ õ Ǫ ǫ Ọ ọ Ő ő Ø ø Œ œ Ɔ ɔ P p Q q R r Ŕ ŕ Ř ř Ŗ ŗ Ṛ ṛ Ṝ ṝ S s Ś ś Ŝ ŝ Š š Ş ş Ș ș Ṣ ṣ ß T t Ť ť Ţ ţ Ț ț Ṭ ṭ Þ þ U u Ú ú Ù ù Û û Ü ü Ǔ ǔ Ŭ ŭ Ū ū Ũ ũ Ů ů Ų ų Ụ ụ Ű ű Ǘ ǘ Ǜ ǜ Ǚ ǚ Ǖ ǖ V v W w Ŵ ŵ X x Y y Ý ý Ŷ ŷ Ÿ ÿ Ỹ ỹ Ȳ ȳ Z z Ź ź Ż ż Ž ž ß Ð ð Þ þ Ŋ ŋ Ə ə Greek: Ά ά Έ έ Ή ή Ί ί Ό ό Ύ ύ Ώ ώ Α α Β β Γ γ Δ δ Ε ε Ζ ζ Η η Θ θ Ι ι Κ κ Λ λ Μ μ Ν ν Ξ ξ Ο ο Π π Ρ ρ Σ σ ς Τ τ Υ υ Φ φ Χ χ Ψ ψ Ω ω {{Polytonic|}} Cyrillic: А а Б б В в Г г Ґ ґ Ѓ ѓ Д д Ђ ђ Е е Ё ё Є є Ж ж З з Ѕ ѕ И и І і Ї ї Й й Ј ј К к Ќ ќ Л л Љ љ М м Н н Њ њ О о П п Р р С с Т т Ћ ћ У у Ў ў Ф ф Х х Ц ц Ч ч Џ џ Ш ш Щ щ Ъ ъ Ы ы Ь ь Э э Ю ю Я я ́ IPA: t̪ d̪ ʈ ɖ ɟ ɡ ɢ ʡ ʔ ɸ β θ ð ʃ ʒ ɕ ʑ ʂ ʐ ç ʝ ɣ χ ʁ ħ ʕ ʜ ʢ ɦ ɱ ɳ ɲ ŋ ɴ ʋ ɹ ɻ ɰ ʙ ⱱ ʀ ɾ ɽ ɫ ɬ ɮ ɺ ɭ ʎ ʟ ɥ ʍ ɧ ʼ ɓ ɗ ʄ ɠ ʛ ʘ ǀ ǃ ǂ ǁ ɨ ʉ ɯ ɪ ʏ ʊ ø ɘ ɵ ɤ ə ɚ ɛ œ ɜ ɝ ɞ ʌ ɔ æ ɐ ɶ ɑ ɒ ʰ ʱ ʷ ʲ ˠ ˤ ⁿ ˡ ˈ ˌ ː ˑ ̪ {{IPA|}} This page is a member of 11 hidden categories (help):
========================================
[SOURCE: https://en.wikipedia.org/wiki/Black_hole#External_geometry] | [TOKENS: 13839]
Contents Black hole A black hole is an astronomical body so compact that its gravity prevents anything, including light, from escaping. Albert Einstein's theory of general relativity predicts that a sufficiently compact mass will form a black hole. The boundary of no escape is called the event horizon. In general relativity, a black hole's event horizon seals an object's fate but produces no locally detectable change when crossed. General relativity also predicts that every black hole should have a central singularity, where the curvature of spacetime is infinite. In many ways, a black hole acts like an ideal black body, as it reflects no light. Quantum field theory in curved spacetime predicts that event horizons emit Hawking radiation, with the same spectrum as a black body of a temperature inversely proportional to its mass. This temperature is of the order of billionths of a kelvin for stellar black holes, making it essentially impossible to observe directly. Objects whose gravitational fields are too strong for light to escape were first considered in the 18th century by John Michell and Pierre-Simon Laplace. In 1916, Karl Schwarzschild found the first modern solution of general relativity that would characterise a black hole. Due to his influential research, the Schwarzschild metric is named after him. David Finkelstein, in 1958, first interpreted Schwarzschild's model as a region of space from which nothing can escape. Black holes were long considered a mathematical curiosity; it was not until the 1960s that theoretical work showed they were a generic prediction of general relativity. The first black hole known was Cygnus X-1, identified by several researchers independently in 1971. Black holes typically form when massive stars collapse at the end of their life cycle. After a black hole has formed, it can grow by absorbing mass from its surroundings. Supermassive black holes of millions of solar masses may form by absorbing other stars and merging with other black holes, or via direct collapse of gas clouds. There is consensus that supermassive black holes exist in the centres of most galaxies. The presence of a black hole can be inferred through its interaction with other matter and with electromagnetic radiation such as visible light. Matter falling toward a black hole can form an accretion disk of infalling plasma, heated by friction and emitting light. In extreme cases, this creates a quasar, some of the brightest objects in the universe. Merging black holes can also be detected by observation of the gravitational waves they emit. If other stars are orbiting a black hole, their orbits can be used to determine the black hole's mass and location. Such observations can be used to exclude possible alternatives such as neutron stars. In this way, astronomers have identified numerous stellar black hole candidates in binary systems and established that the radio source known as Sagittarius A*, at the core of the Milky Way galaxy, contains a supermassive black hole of about 4.3 million solar masses. History The idea of a body so massive that even light could not escape was first proposed in the late 18th century by English astronomer and clergyman John Michell and independently by French scientist Pierre-Simon Laplace. Both scholars proposed very large stars in contrast to the modern concept of an extremely dense object. Michell's idea, in a short part of a letter published in 1784, calculated that a star with the same density but 500 times the radius of the sun would not let any emitted light escape; the surface escape velocity would exceed the speed of light.: 122 Michell correctly hypothesized that such supermassive but non-radiating bodies might be detectable through their gravitational effects on nearby visible bodies. In 1796, Laplace mentioned that a star could be invisible if it were sufficiently large while speculating on the origin of the Solar System in his book Exposition du Système du Monde. Franz Xaver von Zach asked Laplace for a mathematical analysis, which Laplace provided and published in a journal edited by von Zach. In 1905, Albert Einstein showed that the laws of electromagnetism would be invariant under a Lorentz transformation: they would be identical for observers travelling at different velocities relative to each other. This discovery became known as the principle of special relativity. Although the laws of mechanics had already been shown to be invariant, gravity remained yet to be included.: 19 In 1907, Einstein published a paper proposing his equivalence principle, the hypothesis that inertial mass and gravitational mass have a common cause. Using the principle, Einstein predicted the redshift and half of the lensing effect of gravity on light; the full prediction of gravitational lensing required development of general relativity.: 19 By 1915, Einstein refined these ideas into his general theory of relativity, which explained how matter affects spacetime, which in turn affects the motion of other matter. This formed the basis for black hole physics. Only a few months after Einstein published the field equations describing general relativity, astrophysicist Karl Schwarzschild set out to apply the idea to stars. He assumed spherical symmetry with no spin and found a solution to Einstein's equations.: 124 A few months after Schwarzschild, Johannes Droste, a student of Hendrik Lorentz, independently gave the same solution. At a certain radius from the center of the mass, the Schwarzschild solution became singular, meaning that some of the terms in the Einstein equations became infinite. The nature of this radius, which later became known as the Schwarzschild radius, was not understood at the time. Many physicists of the early 20th century were skeptical of the existence of black holes. In a 1926 popular science book, Arthur Eddington critiqued the idea of a star with mass compressed to its Schwarzschild radius as a flaw in the then-poorly-understood theory of general relativity.: 134 In 1939, Einstein himself used his theory of general relativity in an attempt to prove that black holes were impossible. His work relied on increasing pressure or increasing centrifugal force balancing the force of gravity so that the object would not collapse beyond its Schwarzschild radius. He missed the possibility that implosion would drive the system below this critical value.: 135 By the 1920s, astronomers had classified a number of white dwarf stars as too cool and dense to be explained by the gradual cooling of ordinary stars. In 1926, Ralph Fowler showed that quantum-mechanical degeneracy pressure was larger than thermal pressure at these densities.: 145 In 1931, Subrahmanyan Chandrasekhar calculated that a non-rotating body of electron-degenerate matter below a certain limiting mass is stable, and by 1934 he showed that this explained the catalog of white dwarf stars.: 151 When Chandrasekhar announced his results, Eddington pointed out that stars above this limit would radiate until they were sufficiently dense to prevent light from exiting, a conclusion he considered absurd. Eddington and, later, Lev Landau argued that some yet unknown mechanism would stop the collapse. In the 1930s, Fritz Zwicky and Walter Baade studied stellar novae, focusing on exceptionally bright ones they called supernovae. Zwicky promoted the idea that supernovae produced stars with the density of atomic nuclei—neutron stars—but this idea was largely ignored.: 171 In 1939, based on Chandrasekhar's reasoning, J. Robert Oppenheimer and George Volkoff predicted that neutron stars below a certain mass limit, later called the Tolman–Oppenheimer–Volkoff limit, would be stable due to neutron degeneracy pressure. Above that limit, they reasoned that either their model would not apply or that gravitational contraction would not stop.: 380 John Archibald Wheeler and two of his students resolved questions about the model behind the Tolman–Oppenheimer–Volkoff (TOV) limit. Harrison and Wheeler developed the equations of state relating density to pressure for cold matter all the way through electron degeneracy and neutron degeneracy. Masami Wakano and Wheeler then used the equations to compute the equilibrium curve for stars, relating mass to circumference. They found no additional features that would invalidate the TOV limit. This meant that the only thing that could prevent black holes from forming was a dynamic process ejecting sufficient mass from a star as it cooled.: 205 The modern concept of black holes was formulated by Robert Oppenheimer and his student Hartland Snyder in 1939.: 80 In the paper, Oppenheimer and Snyder solved Einstein's equations of general relativity for an idealized imploding star, in a model later called the Oppenheimer–Snyder model, then described the results from far outside the star. The implosion starts as one might expect: the star material rapidly collapses inward. However, as the density of the star increases, gravitational time dilation increases and the collapse, viewed from afar, seems to slow down further and further until the star reaches its Schwarzschild radius, where it appears frozen in time.: 217 In 1958, David Finkelstein identified the Schwarzschild surface as an event horizon, calling it "a perfect unidirectional membrane: causal influences can cross it in only one direction". In this sense, events that occur inside of the black hole cannot affect events that occur outside of the black hole. Finkelstein created a new reference frame to include the point of view of infalling observers.: 103 Finkelstein's new frame of reference allowed events at the surface of an imploding star to be related to events far away. By 1962 the two points of view were reconciled, convincing many skeptics that implosion into a black hole made physical sense.: 226 The era from the mid-1960s to the mid-1970s was the "golden age of black hole research", when general relativity and black holes became mainstream subjects of research.: 258 In this period, more general black hole solutions were found. In 1963, Roy Kerr found the exact solution for a rotating black hole. Two years later, Ezra Newman found the cylindrically symmetric solution for a black hole that is both rotating and electrically charged. In 1967, Werner Israel found that the Schwarzschild solution was the only possible solution for a nonspinning, uncharged black hole, meaning that a Schwarzschild black hole would be defined by its mass alone. Similar identities were later found for Reissner-Nordstrom and Kerr black holes, defined only by their mass and their charge or spin respectively. Together, these findings became known as the no-hair theorem, which states that a stationary black hole is completely described by the three parameters of the Kerr–Newman metric: mass, angular momentum, and electric charge. At first, it was suspected that the strange mathematical singularities found in each of the black hole solutions only appeared due to the assumption that a black hole would be perfectly spherically symmetric, and therefore the singularities would not appear in generic situations where black holes would not necessarily be symmetric. This view was held in particular by Vladimir Belinski, Isaak Khalatnikov, and Evgeny Lifshitz, who tried to prove that no singularities appear in generic solutions, although they would later reverse their positions. However, in 1965, Roger Penrose proved that general relativity without quantum mechanics requires that singularities appear in all black holes. Astronomical observations also made great strides during this era. In 1967, Antony Hewish and Jocelyn Bell Burnell discovered pulsars and by 1969, these were shown to be rapidly rotating neutron stars. Until that time, neutron stars, like black holes, were regarded as just theoretical curiosities, but the discovery of pulsars showed their physical relevance and spurred a further interest in all types of compact objects that might be formed by gravitational collapse. Based on observations in Greenwich and Toronto in the early 1970s, Cygnus X-1, a galactic X-ray source discovered in 1964, became the first astronomical object commonly accepted to be a black hole. Work by James Bardeen, Jacob Bekenstein, Carter, and Hawking in the early 1970s led to the formulation of black hole thermodynamics. These laws describe the behaviour of a black hole in close analogy to the laws of thermodynamics by relating mass to energy, area to entropy, and surface gravity to temperature. The analogy was completed: 442 when Hawking, in 1974, showed that quantum field theory implies that black holes should radiate like a black body with a temperature proportional to the surface gravity of the black hole, predicting the effect now known as Hawking radiation. While Cygnus X-1, a stellar-mass black hole, was generally accepted by the scientific community as a black hole by the end of 1973, it would be decades before a supermassive black hole would gain the same broad recognition. Although, as early as the 1960s, physicists such as Donald Lynden-Bell and Martin Rees had suggested that powerful quasars in the center of galaxies were powered by accreting supermassive black holes, little observational proof existed at the time. However, the Hubble Space Telescope, launched decades later, found that supermassive black holes were not only present in these active galactic nuclei, but that supermassive black holes in the center of galaxies were ubiquitous: Almost every galaxy had a supermassive black hole at its center, many of which were quiescent. In 1999, David Merritt proposed the M–sigma relation, which related the dispersion of the velocity of matter in the center bulge of a galaxy to the mass of the supermassive black hole at its core. Subsequent studies confirmed this correlation. Around the same time, based on telescope observations of the velocities of stars at the center of the Milky Way galaxy, independent work groups led by Andrea Ghez and Reinhard Genzel concluded that the compact radio source in the center of the galaxy, Sagittarius A*, was likely a supermassive black hole. On 11 February 2016, the LIGO Scientific Collaboration and Virgo Collaboration announced the first direct detection of gravitational waves, named GW150914, representing the first observation of a black hole merger. At the time of the merger, the black holes were approximately 1.4 billion light-years away from Earth and had masses of 30 and 35 solar masses.: 6 In 2017, Rainer Weiss, Kip Thorne, and Barry Barish, who had spearheaded the project, were awarded the Nobel Prize in Physics for their work. Since the initial discovery in 2015, hundreds more gravitational waves have been observed by LIGO and another interferometer, Virgo. On 10 April 2019, the first direct image of a black hole and its vicinity was published, following observations made by the Event Horizon Telescope (EHT) in 2017 of the supermassive black hole in Messier 87's galactic centre. In 2022, the Event Horizon Telescope collaboration released an image of the black hole in the center of the Milky Way galaxy, Sagittarius A*; The data had been collected in 2017. In 2020, the Nobel Prize in Physics was awarded for work on black holes. Andrea Ghez and Reinhard Genzel shared one-half for their discovery that Sagittarius A* is a supermassive black hole. Penrose received the other half for his work showing that the mathematics of general relativity requires the formation of black holes. Cosmologists lamented that Hawking's extensive theoretical work on black holes would not be honored since he died in 2018. In December 1967, a student reportedly suggested the phrase black hole at a lecture by John Wheeler; Wheeler adopted the term for its brevity and "advertising value", and Wheeler's stature in the field ensured it quickly caught on, leading some to credit Wheeler with coining the phrase. However, the term was used by others around that time. Science writer Marcia Bartusiak traces the term black hole to physicist Robert H. Dicke, who in the early 1960s reportedly compared the phenomenon to the Black Hole of Calcutta, notorious as a prison where people entered but never left alive. The term was used in print by Life and Science News magazines in 1963, and by science journalist Ann Ewing in her article "'Black Holes' in Space", dated 18 January 1964, which was a report on a meeting of the American Association for the Advancement of Science held in Cleveland, Ohio. Definition A black hole is generally defined as a region of spacetime from which no information-carrying signals or objects can escape. However, verifying an object as a black hole by this definition would require waiting for an infinite time and at an infinite distance from the black hole to verify that indeed, nothing has escaped, and thus cannot be used to identify a physical black hole. Broadly, physicists do not have a precisely-agreed-upon definition of a black hole. Among astrophysicists, a black hole is a compact object with a mass larger than four solar masses. A black hole may also be defined as a reservoir of information: 142 or a region where space is falling inwards faster than the speed of light. Properties The no-hair theorem postulates that, once it achieves a stable condition after formation, a black hole has only three independent physical properties: mass, electric charge, and angular momentum; the black hole is otherwise featureless. If the conjecture is true, any two black holes that share the same values for these properties, or parameters, are indistinguishable from one another. The degree to which the conjecture is true for real black holes is currently an unsolved problem. The simplest static black holes have mass but neither electric charge nor angular momentum. According to Birkhoff's theorem, these Schwarzschild black holes are the only vacuum solution that is spherically symmetric. Solutions describing more general black holes also exist. Non-rotating charged black holes are described by the Reissner–Nordström metric, while the Kerr metric describes a non-charged rotating black hole. The most general stationary black hole solution known is the Kerr–Newman metric, which describes a black hole with both charge and angular momentum. The simplest static black holes have mass but neither electric charge nor angular momentum. Contrary to the popular notion of a black hole "sucking in everything" in its surroundings, from far away, the external gravitational field of a black hole is identical to that of any other body of the same mass. While a black hole can theoretically have any positive mass, the charge and angular momentum are constrained by the mass. The total electric charge Q and the total angular momentum J are expected to satisfy the inequality Q 2 4 π ϵ 0 + c 2 J 2 G M 2 ≤ G M 2 {\displaystyle {\frac {Q^{2}}{4\pi \epsilon _{0}}}+{\frac {c^{2}J^{2}}{GM^{2}}}\leq GM^{2}} for a black hole of mass M. Black holes with the maximum possible charge or spin satisfying this inequality are called extremal black holes. Solutions of Einstein's equations that violate this inequality exist, but they do not possess an event horizon. These are so-called naked singularities that can be observed from the outside. Because these singularities make the universe inherently unpredictable, many physicists believe they could not exist. The weak cosmic censorship hypothesis, proposed by Sir Roger Penrose, rules out the formation of such singularities, when they are created through the gravitational collapse of realistic matter. However, this theory has not yet been proven, and some physicists believe that naked singularities could exist. It is also unknown whether black holes could even become extremal, forming naked singularities, since natural processes counteract increasing spin and charge when a black hole becomes near-extremal. The total mass of a black hole can be estimated by analyzing the motion of objects near the black hole, such as stars or gas. All black holes spin, often fast—One supermassive black hole, GRS 1915+105 has been estimated to spin at over 1,000 revolutions per second. The Milky Way's central black hole Sagittarius A* rotates at about 90% of the maximum rate. The spin rate can be inferred from measurements of atomic spectral lines in the X-ray range. As gas near the black hole plunges inward, high energy X-ray emission from electron-positron pairs illuminates the gas further out, appearing red-shifted due to relativistic effects. Depending on the spin of the black hole, this plunge happens at different radii from the hole, with different degrees of redshift. Astronomers can use the gap between the x-ray emission of the outer disk and the redshifted emission from plunging material to determine the spin of the black hole. A newer way to estimate spin is based on the temperature of gasses accreting onto the black hole. The method requires an independent measurement of the black hole mass and inclination angle of the accretion disk followed by computer modeling. Gravitational waves from coalescing binary black holes can also provide the spin of both progenitor black holes and the merged hole, but such events are rare. A spinning black hole has angular momentum. The supermassive black hole in the center of the Messier 87 (M87) galaxy appears to have an angular momentum very close to the maximum theoretical value. That uncharged limit is J ≤ G M 2 c , {\displaystyle J\leq {\frac {GM^{2}}{c}},} allowing definition of a dimensionless spin magnitude such that 0 ≤ c J G M 2 ≤ 1. {\displaystyle 0\leq {\frac {cJ}{GM^{2}}}\leq 1.} Most black holes are believed to have an approximately neutral charge. For example, Michal Zajaček, Arman Tursunov, Andreas Eckart, and Silke Britzen found the electric charge of Sagittarius A* to be at least ten orders of magnitude below the theoretical maximum. A charged black hole repels other like charges just like any other charged object. If a black hole were to become charged, particles with an opposite sign of charge would be pulled in by the extra electromagnetic force, while particles with the same sign of charge would be repelled, neutralizing the black hole. This effect may not be as strong if the black hole is also spinning. The presence of charge can reduce the diameter of the black hole by up to 38%. The charge Q for a nonspinning black hole is bounded by Q ≤ G M , {\displaystyle Q\leq {\sqrt {G}}M,} where G is the gravitational constant and M is the black hole's mass. Classification Black holes can have a wide range of masses. The minimum mass of a black hole formed by stellar gravitational collapse is governed by the maximum mass of a neutron star and is believed to be approximately two-to-four solar masses. However, theoretical primordial black holes, believed to have formed soon after the Big Bang, could be far smaller, with masses as little as 10−5 grams at formation. These very small black holes are sometimes called micro black holes. Black holes formed by stellar collapse are called stellar black holes. Estimates of their maximum mass at formation vary, but generally range from 10 to 100 solar masses, with higher estimates for black holes progenated by low-metallicity stars. The mass of a black hole formed via a supernova has a lower bound: If the progenitor star is too small, the collapse may be stopped by the degeneracy pressure of the star's constituents, allowing the condensation of matter into an exotic denser state. Degeneracy pressure occurs from the Pauli exclusion principle—Particles will resist being in the same place as each other. Smaller progenitor stars, with masses less than about 8 M☉, will be held together by the degeneracy pressure of electrons and will become a white dwarf. For more massive progenitor stars, electron degeneracy pressure is no longer strong enough to resist the force of gravity and the star will be held together by neutron degeneracy pressure, which can occur at much higher densities, forming a neutron star. If the star is still too massive, even neutron degeneracy pressure will not be able to resist the force of gravity and the star will collapse into a black hole.: 5.8 Stellar black holes can also gain mass via accretion of nearby matter, often from a companion object such as a star. Black holes that are larger than stellar black holes but smaller than supermassive black holes are called intermediate-mass black holes, with masses of approximately 102 to 105 solar masses. These black holes seem to be rarer than their stellar and supermassive counterparts, with relatively few candidates having been observed. Physicists have speculated that such black holes may form from collisions in globular and star clusters or at the center of low-mass galaxies. They may also form as the result of mergers of smaller black holes, with several LIGO observations finding merged black holes within the 110-350 solar mass range. The black holes with the largest masses are called supermassive black holes, with masses more than 106 times that of the Sun. These black holes are believed to exist at the centers of almost every large galaxy, including the Milky Way. Some scientists have proposed a subcategory of even larger black holes, called ultramassive black holes, with masses greater than 109-1010 solar masses. Theoretical models predict that the accretion disc that feeds black holes will be unstable once a black hole reaches 50-100 billion times the mass of the Sun, setting a rough upper limit to black hole mass. Structure While black holes are conceptually invisible sinks of all matter and light, in astronomical settings, their enormous gravity alters the motion of surrounding objects and pulls nearby gas inwards at near-light speed, making the area around black holes the brightest objects in the universe. Some black holes have relativistic jets—thin streams of plasma travelling away from the black hole at more than one-tenth of the speed of light. A small faction of the matter falling towards the black hole gets accelerated away along the hole rotation axis. These jets can extend as far as millions of parsecs from the black hole itself. Black holes of any mass can have jets. However, they are typically observed around spinning black holes with strongly-magnetized accretion disks. Relativistic jets were more common in the early universe, when galaxies and their corresponding supermassive black holes were rapidly gaining mass. All black holes with jets also have an accretion disk, but the jets are usually brighter than the disk. Quasars, typically found in other galaxies, are believed to be supermassive black holes with jets; microquasars are believed to be stellar-mass objects with jets, typically observed in the Milky Way. The mechanism of formation of jets is not yet known, but several options have been proposed. One method proposed to fuel these jets is the Blandford-Znajek process, which suggests that the dragging of magnetic field lines by a black hole's rotation could launch jets of matter into space. The Penrose process, which involves extraction of a black hole's rotational energy, has also been proposed as a potential mechanism of jet propulsion. Due to conservation of angular momentum, gas falling into the gravitational well created by a massive object will typically form a disk-like structure around the object.: 242 As the disk's angular momentum is transferred outward due to internal processes, its matter falls farther inward, converting its gravitational energy into heat and releasing a large flux of x-rays. The temperature of these disks can range from thousands to millions of Kelvin, and temperatures can differ throughout a single accretion disk. Accretion disks can also emit in other parts of the electromagnetic spectrum, depending on the disk's turbulence and magnetization and the black hole's mass and angular momentum. Accretion disks can be defined as geometrically thin or geometrically thick. Geometrically thin disks are mostly confined to the black hole's equatorial plane and have a well-defined edge at the innermost stable circular orbit (ISCO), while geometrically thick disks are supported by internal pressure and temperature and can extend inside the ISCO. Disks with high rates of electron scattering and absorption, appearing bright and opaque, are called optically thick; optically thin disks are more translucent and produce fainter images when viewed from afar. Accretion disks of black holes accreting beyond the Eddington limit are often referred to as polish donuts due to their thick, toroidal shape that resembles that of a donut. Quasar accretion disks are expected to usually appear blue in color. The disk for a stellar black hole, on the other hand, would likely look orange, yellow, or red, with its inner regions being the brightest. Theoretical research suggests that the hotter a disk is, the bluer it should be, although this is not always supported by observations of real astronomical objects. Accretion disk colors may also be altered by the Doppler effect, with the part of the disk travelling towards an observer appearing bluer and brighter and the part of the disk travelling away from the observer appearing redder and dimmer. In Newtonian gravity, test particles can stably orbit at arbitrary distances from a central object. In general relativity, however, there exists a smallest possible radius for which a massive particle can orbit stably. Any infinitesimal inward perturbations to this orbit will lead to the particle spiraling into the black hole, and any outward perturbations will, depending on the energy, cause the particle to spiral in, move to a stable orbit further from the black hole, or escape to infinity. This orbit is called the innermost stable circular orbit, or ISCO. The location of the ISCO depends on the spin of the black hole and the spin of the particle itself. In the case of a Schwarzschild black hole (spin zero) and a particle without spin, the location of the ISCO is: r I S C O = 3 r s = 6 G M c 2 , {\displaystyle r_{\rm {ISCO}}=3\,r_{\text{s}}={\frac {6\,GM}{c^{2}}},} where r I S C O {\displaystyle r_{\rm {_{ISCO}}}} is the radius of the ISCO, r s {\displaystyle r_{\text{s}}} is the Schwarzschild radius of the black hole, G {\displaystyle G} is the gravitational constant, and c {\displaystyle c} is the speed of light. The radius of this orbit changes slightly based on particle spin. For charged black holes, the ISCO moves inwards. For spinning black holes, the ISCO is moved inwards for particles orbiting in the same direction that the black hole is spinning (prograde) and outwards for particles orbiting in the opposite direction (retrograde). For example, the ISCO for a particle orbiting retrograde can be as far out as about 9 r s {\displaystyle 9r_{\text{s}}} , while the ISCO for a particle orbiting prograde can be as close as at the event horizon itself. The photon sphere is a spherical boundary for which photons moving on tangents to that sphere are bent completely around the black hole, possibly orbiting multiple times. Light rays with impact parameters less than the radius of the photon sphere enter the black hole. For Schwarzschild black holes, the photon sphere has a radius 1.5 times the Schwarzschild radius; the radius for non-Schwarzschild black holes is at least 1.5 times the radius of the event horizon. When viewed from a great distance, the photon sphere creates an observable black hole shadow. Since no light emerges from within the black hole, this shadow is the limit for possible observations.: 152 The shadow of colliding black holes should have characteristic warped shapes, allowing scientists to detect black holes that are about to merge. While light can still escape from the photon sphere, any light that crosses the photon sphere on an inbound trajectory will be captured by the black hole. Therefore, any light that reaches an outside observer from the photon sphere must have been emitted by objects between the photon sphere and the event horizon. Light emitted towards the photon sphere may also curve around the black hole and return to the emitter. For a rotating, uncharged black hole, the radius of the photon sphere depends on the spin parameter and whether the photon is orbiting prograde or retrograde. For a photon orbiting prograde, the photon sphere will be 1-3 Schwarzschild radii from the center of the black hole, while for a photon orbiting retrograde, the photon sphere will be between 3-5 Schwarzschild radii from the center of the black hole. The exact location of the photon sphere depends on the magnitude of the black hole's rotation. For a charged, nonrotating black hole, there will only be one photon sphere, and the radius of the photon sphere will decrease for increasing black hole charge. For non-extremal, charged, rotating black holes, there will always be two photon spheres, with the exact radii depending on the parameters of the black hole. Near a rotating black hole, spacetime rotates similar to a vortex. The rotating spacetime will drag any matter and light into rotation around the spinning black hole. This effect of general relativity, called frame dragging, gets stronger closer to the spinning mass. The region of spacetime in which it is impossible to stay still is called the ergosphere. The ergosphere of a black hole is a volume bounded by the black hole's event horizon and the ergosurface, which coincides with the event horizon at the poles but bulges out from it around the equator. Matter and radiation can escape from the ergosphere. Through the Penrose process, objects can emerge from the ergosphere with more energy than they entered with. The extra energy is taken from the rotational energy of the black hole, slowing down the rotation of the black hole.: 268 A variation of the Penrose process in the presence of strong magnetic fields, the Blandford–Znajek process, is considered a likely mechanism for the enormous luminosity and relativistic jets of quasars and other active galactic nuclei. The observable region of spacetime around a black hole closest to its event horizon is called the plunging region. In this area it is no longer possible for free falling matter to follow circular orbits or stop a final descent into the black hole. Instead, it will rapidly plunge toward the black hole at close to the speed of light, growing increasingly hot and producing a characteristic, detectable thermal emission. However, light and radiation emitted from this region can still escape from the black hole's gravitational pull. For a nonspinning, uncharged black hole, the radius of the event horizon, or Schwarzschild radius, is proportional to the mass, M, through r s = 2 G M c 2 ≈ 2.95 M M ⊙ k m , {\displaystyle r_{\mathrm {s} }={\frac {2GM}{c^{2}}}\approx 2.95\,{\frac {M}{M_{\odot }}}~\mathrm {km,} } where rs is the Schwarzschild radius and M☉ is the mass of the Sun.: 124 For a black hole with nonzero spin or electric charge, the radius is smaller,[Note 1] until an extremal black hole could have an event horizon close to r + = G M c 2 , {\displaystyle r_{\mathrm {+} }={\frac {GM}{c^{2}}},} half the radius of a nonspinning, uncharged black hole of the same mass. Since the volume within the Schwarzschild radius increase with the cube of the radius, average density of a black hole inside its Schwarzschild radius is inversely proportional to the square of its mass: supermassive black holes are much less dense than stellar black holes. The average density of a 108 M☉ black hole is comparable to that of water. The defining feature of a black hole is the existence of an event horizon, a boundary in spacetime through which matter and light can pass only inward towards the center of the black hole. Nothing, not even light, can escape from inside the event horizon. The event horizon is referred to as such because if an event occurs within the boundary, information from that event cannot reach or affect an outside observer, making it impossible to determine whether such an event occurred.: 179 For non-rotating black holes, the geometry of the event horizon is precisely spherical, while for rotating black holes, the event horizon is oblate. To a distant observer, a clock near a black hole would appear to tick more slowly than one further from the black hole.: 217 This effect, known as gravitational time dilation, would also cause an object falling into a black hole to appear to slow as it approached the event horizon, never quite reaching the horizon from the perspective of an outside observer.: 218 All processes on this object would appear to slow down, and any light emitted by the object to appear redder and dimmer, an effect known as gravitational redshift. An object falling from half of a Schwarzschild radius above the event horizon would fade away until it could no longer be seen, disappearing from view within one hundredth of a second. It would also appear to flatten onto the black hole, joining all other material that had ever fallen into the hole. On the other hand, an observer falling into a black hole would not notice any of these effects as they cross the event horizon. Their own clocks appear to them to tick normally, and they cross the event horizon after a finite time without noting any singular behaviour. In general relativity, it is impossible to determine the location of the event horizon from local observations, due to Einstein's equivalence principle.: 222 Black holes that are rotating and/or charged have an inner horizon, often called the Cauchy horizon, inside of the black hole. The inner horizon is divided up into two segments: an ingoing section and an outgoing section. At the ingoing section of the Cauchy horizon, radiation and matter that fall into the black hole would build up at the horizon, causing the curvature of spacetime to go to infinity. This would cause an observer falling in to experience tidal forces. This phenomenon is often called mass inflation, since it is associated with a parameter dictating the black hole's internal mass growing exponentially, and the buildup of tidal forces is called the mass-inflation singularity or Cauchy horizon singularity. Some physicists have argued that in realistic black holes, accretion and Hawking radiation would stop mass inflation from occurring. At the outgoing section of the inner horizon, infalling radiation would backscatter off of the black hole's spacetime curvature and travel outward, building up at the outgoing Cauchy horizon. This would cause an infalling observer to experience a gravitational shock wave and tidal forces as the spacetime curvature at the horizon grew to infinity. This buildup of tidal forces is called the shock singularity. Both of these singularities are weak, meaning that an object crossing them would only be deformed a finite amount by tidal forces, even though the spacetime curvature would still be infinite at the singularity. This is as opposed to a strong singularity, where an object hitting the singularity would be stretched and squeezed by an infinite amount. They are also null singularities, meaning that a photon could travel parallel to the them without ever being intercepted. Ignoring quantum effects, every black hole has a singularity inside, points where the curvature of spacetime becomes infinite, and geodesics terminate within a finite proper time.: 205 For a non-rotating black hole, this region takes the shape of a single point; for a rotating black hole it is smeared out to form a ring singularity that lies in the plane of rotation.: 264 In both cases, the singular region has zero volume. All of the mass of the black hole ends up in the singularity.: 252 Since the singularity has nonzero mass in an infinitely small space, it can be thought of as having infinite density. Observers falling into a Schwarzschild black hole (i.e., non-rotating and not charged) cannot avoid being carried into the singularity once they cross the event horizon. As they fall further into the black hole, they will be torn apart by the growing tidal forces in a process sometimes referred to as spaghettification or the noodle effect. Eventually, they will reach the singularity and be crushed into an infinitely small point.: 182 However any perturbations, such as those caused by matter or radiation falling in, would cause space to oscillate chaotically near the singularity. Any matter falling in would experience intense tidal forces rapidly changing in direction, all while being compressed into an increasingly small volume. Alternative forms of general relativity, including addition of some quatum effects, can lead to regular, or nonsingular, black holes without singularities. For example, the fuzzball model, based on string theory, states that black holes are actually made up of quantum microstates and need not have a singularity or an event horizon. The theory of loop quantum gravity proposes that the curvature and density at the center of a black hole is large, but not infinite. Formation Black holes are formed by gravitational collapse of massive stars, either by direct collapse or during a supernova explosion in a process called fallback. Black holes can result from the merger of two neutron stars or a neutron star and a black hole. Other more speculative mechanisms include primordial black holes created from density fluctuations in the early universe, the collapse of dark stars, a hypothetical object powered by annihilation of dark matter, or from hypothetical self-interacting dark matter. Gravitational collapse occurs when an object's internal pressure is insufficient to resist the object's own gravity. At the end of a star's life, it will run out of hydrogen to fuse, and will start fusing more and more massive elements, until it gets to iron. Since the fusion of elements heavier than iron would require more energy than it would release, nuclear fusion ceases. If the iron core of the star is too massive, the star will no longer be able to support itself and will undergo gravitational collapse. While most of the energy released during gravitational collapse is emitted very quickly, an outside observer does not actually see the end of this process. Even though the collapse takes a finite amount of time from the reference frame of infalling matter, a distant observer would see the infalling material slow and halt just above the event horizon, due to gravitational time dilation. Light from the collapsing material takes longer and longer to reach the observer, with the delay growing to infinity as the emitting material reaches the event horizon. Thus the external observer never sees the formation of the event horizon; instead, the collapsing material seems to become dimmer and increasingly red-shifted, eventually fading away. Observations of quasars at redshift z ∼ 7 {\displaystyle z\sim 7} , less than a billion years after the Big Bang, has led to investigations of other ways to form black holes. The accretion process to build supermassive black holes has a limiting rate of mass accumulation and a billion years is not enough time to reach quasar status. One suggestion is direct collapse of nearly pure hydrogen gas (low metalicity) clouds characteristic of the young universe, forming a supermassive star which collapses into a black hole. It has been suggested that seed black holes with typical masses of ~105 M☉ could have formed in this way which then could grow to ~109 M☉. However, the very large amount of gas required for direct collapse is not typically stable to fragmentation to form multiple stars. Thus another approach suggests massive star formation followed by collisions that seed massive black holes which ultimately merge to create a quasar.: 85 A neutron star in a common envelope with a regular star can accrete sufficient material to collapse to a black hole or two neutron stars can merge. These avenues for the formation of black holes are considered relatively rare. In the current epoch of the universe, conditions needed to form black holes are rare and are mostly only found in stars. However, in the early universe, conditions may have allowed for black hole formations via other means. Fluctuations of spacetime soon after the Big Bang may have formed areas that were denser then their surroundings. Initially, these regions would not have been compact enough to form a black hole, but eventually, the curvature of spacetime in the regions become large enough to cause them to collapse into a black hole. Different models for the early universe vary widely in their predictions of the scale of these fluctuations. Various models predict the creation of primordial black holes ranging from a Planck mass (~2.2×10−8 kg) to hundreds of thousands of solar masses. Primordial black holes with masses less than 1015 g would have evaporated by now due to Hawking radiation. Despite the early universe being extremely dense, it did not re-collapse into a black hole during the Big Bang, since the universe was expanding rapidly and did not have the gravitational differential necessary for black hole formation. Models for the gravitational collapse of objects of relatively constant size, such as stars, do not necessarily apply in the same way to rapidly expanding space such as the Big Bang. In principle, black holes could be formed in high-energy particle collisions that achieve sufficient density, although no such events have been detected. These hypothetical micro black holes, which could form from the collision of cosmic rays and Earth's atmosphere or in particle accelerators like the Large Hadron Collider, would not be able to aggregate additional mass. Instead, they would evaporate in about 10−25 seconds, posing no threat to the Earth. Evolution Black holes can also merge with other objects such as stars or even other black holes. This is thought to have been important, especially in the early growth of supermassive black holes, which could have formed from the aggregation of many smaller objects. The process has also been proposed as the origin of some intermediate-mass black holes. Mergers of supermassive black holes may take a long time: As a binary of supermassive black holes approach each other, most nearby stars are ejected, leaving little for the remaining black holes to gravitationally interact with that would allow them to get closer to each other. This phenomenon has been called the final parsec problem, as the distance at which this happens is usually around one parsec. When a black hole accretes matter, the gas in the inner accretion disk orbits at very high speeds because of its proximity to the black hole. The resulting friction heats the inner disk to temperatures at which it emits vast amounts of electromagnetic radiation (mainly X-rays) detectable by telescopes. By the time the matter of the disk reaches the ISCO, between 5.7% and 42% of its mass will have been converted to energy, depending on the black hole's spin. About 90% of this energy is released within about 20 black hole radii. In many cases, accretion disks are accompanied by relativistic jets that are emitted along the black hole's poles, which carry away much of the energy. The mechanism for the creation of these jets is currently not well understood, in part due to insufficient data. Many of the universe's most energetic phenomena have been attributed to the accretion of matter on black holes. Active galactic nuclei and quasars are believed to be the accretion disks of supermassive black holes. X-ray binaries are generally accepted to be binary systems in which one of the two objects is a compact object accreting matter from its companion. Ultraluminous X-ray sources may be the accretion disks of intermediate-mass black holes. At a certain rate of accretion, the outward radiation pressure will become as strong as the inward gravitational force, and the black hole should unable to accrete any faster. This limit is called the Eddington limit. However, many black holes accrete beyond this rate due to their non-spherical geometry or instabilities in the accretion disk. Accretion beyond the limit is called Super-Eddington accretion and may have been commonplace in the early universe. Stars have been observed to get torn apart by tidal forces in the immediate vicinity of supermassive black holes in galaxy nuclei, in what is known as a tidal disruption event (TDE). Some of the material from the disrupted star forms an accretion disk around the black hole, which emits observable electromagnetic radiation. The correlation between the masses of supermassive black holes in the centres of galaxies with the velocity dispersion and mass of stars in their host bulges suggests that the formation of galaxies and the formation of their central black holes are related. Black hole winds from rapid accretion, particularly when the galaxy itself is still accreting matter, can compress gas nearby, accelerating star formation. However, if the winds become too strong, the black hole may blow nearly all of the gas out of the galaxy, quenching star formation. Black hole jets may also energize nearby cavities of plasma and eject low-entropy gas from out of the galactic core, causing gas in galactic centers to be hotter than expected. If Hawking's theory of black hole radiation is correct, then black holes are expected to shrink and evaporate over time as they lose mass by the emission of photons and other particles. The temperature of this thermal spectrum (Hawking temperature) is proportional to the surface gravity of the black hole, which is inversely proportional to the mass. Hence, large black holes emit less radiation than small black holes.: Ch. 9.6 A stellar black hole of 1 M☉ has a Hawking temperature of 62 nanokelvins. This is far less than the 2.7 K temperature of the cosmic microwave background radiation. Stellar-mass or larger black holes receive more mass from the cosmic microwave background than they emit through Hawking radiation and thus will grow instead of shrinking. To have a Hawking temperature larger than 2.7 K (and be able to evaporate), a black hole would need a mass less than the Moon. Such a black hole would have a diameter of less than a tenth of a millimetre. The Hawking radiation for an astrophysical black hole is predicted to be very weak and would thus be exceedingly difficult to detect from Earth. A possible exception is the burst of gamma rays emitted in the last stage of the evaporation of primordial black holes. Searches for such flashes have proven unsuccessful and provide stringent limits on the possibility of existence of low mass primordial black holes, with modern research predicting that primordial black holes must make up less than a fraction of 10−7 of the universe's total mass. NASA's Fermi Gamma-ray Space Telescope, launched in 2008, has searched for these flashes, but has not yet found any. The properties of a black hole are constrained and interrelated by the theories that predict these properties. When based on general relativity, these relationships are called the laws of black hole mechanics. For a black hole that is not still forming or accreting matter, the zeroth law of black hole mechanics states the black hole's surface gravity is constant across the event horizon. The first law relates changes in the black hole's surface area, angular momentum, and charge to changes in its energy. The second law says the surface area of a black hole never decreases on its own. Finally, the third law says that the surface gravity of a black hole is never zero. These laws are mathematical analogs of the laws of thermodynamics. They are not equivalent, however, because, according to general relativity without quantum mechanics, a black hole can never emit radiation, and thus its temperature must always be zero.: 11 Quantum mechanics predicts that a black hole will continuously emit thermal Hawking radiation, and therefore must always have a nonzero temperature. It also predicts that all black holes have entropy which scales with their surface area. When quantum mechanics is accounted for, the laws of black hole mechanics become equivalent to the classical laws of thermodynamics. However, these conclusions are derived without a complete theory of quantum gravity, although many potential theories do predict black holes having entropy and temperature. Thus, the true quantum nature of black hole thermodynamics continues to be debated.: 29 Observational evidence Millions of black holes with around 30 solar masses derived from stellar collapse are expected to exist in the Milky Way. Even a dwarf galaxy like Draco should have hundreds. Only a few of these have been detected. By nature, black holes do not themselves emit any electromagnetic radiation other than the hypothetical Hawking radiation, so astrophysicists searching for black holes must generally rely on indirect observations. The defining characteristic of a black hole is its event horizon. The horizon itself cannot be imaged, so all other possible explanations for these indirect observations must be considered and eliminated before concluding that a black hole has been observed.: 11 The Event Horizon Telescope (EHT) is a global system of radio telescopes capable of directly observing a black hole shadow. The angular resolution of a telescope is based on its aperture and the wavelengths it is observing. Because the angular diameters of Sagittarius A* and Messier 87* in the sky are very small, a single telescope would need to be about the size of the Earth to clearly distinguish their horizons using radio wavelengths. By combining data from several different radio telescopes around the world, the Event Horizon Telescope creates an effective aperture the diameter size of the Earth. The EHT team used imaging algorithms to compute the most probable image from the data in its observations of Sagittarius A* and M87*. Gravitational-wave interferometry can be used to detect merging black holes and other compact objects. In this method, a laser beam is split down two long arms of a tunnel. The laser beams reflect off of mirrors in the tunnels and converge at the intersection of the arms, cancelling each other out. However, when a gravitational wave passes, it warps spacetime, changing the lengths of the arms themselves. Since each laser beam is now travelling a slightly different distance, they do not cancel out and produce a recognizable signal. Analysis of the signal can give scientists information about what caused the gravitational waves. Since gravitational waves are very weak, gravitational-wave observatories such as LIGO must have arms several kilometers long and carefully control for noise from Earth to be able to detect these gravitational waves. Since the first measurements in 2016, multiple gravitational waves from black holes have been detected and analyzed. The proper motions of stars near the centre of the Milky Way provide strong observational evidence that these stars are orbiting a supermassive black hole. Since 1995, astronomers have tracked the motions of 90 stars orbiting an invisible object coincident with the radio source Sagittarius A*. In 1998, by fitting the motions of the stars to Keplerian orbits, the astronomers were able to infer that Sagittarius A* must be a 2.6×106 M☉ object must be contained within a radius of 0.02 light-years. Since then, one of the stars—called S2—has completed a full orbit. From the orbital data, astronomers were able to refine the calculations of the mass of Sagittarius A* to 4.3×106 M☉, with a radius of less than 0.002 light-years. This upper limit radius is larger than the Schwarzschild radius for the estimated mass, so the combination does not prove Sagittarius A* is a black hole. Nevertheless, these observations strongly suggest that the central object is a supermassive black hole as there are no other plausible scenarios for confining so much invisible mass into such a small volume. Additionally, there is some observational evidence that this object might possess an event horizon, a feature unique to black holes. The Event Horizon Telescope image of Sagittarius A*, released in 2022, provided further confirmation that it is indeed a black hole. X-ray binaries are binary systems that emit a majority of their radiation in the X-ray part of the electromagnetic spectrum. These X-ray emissions result when a compact object accretes matter from an ordinary star. The presence of an ordinary star in such a system provides an opportunity for studying the central object and to determine if it might be a black hole. By measuring the orbital period of the binary, the distance to the binary from Earth, and the mass of the companion star, scientists can estimate the mass of the compact object. The Tolman-Oppenheimer-Volkoff limit (TOV limit) dictates the largest mass a nonrotating neutron star can be, and is estimated to be about two solar masses. While a rotating neutron star can be slightly more massive, if the compact object is much more massive than the TOV limit, it cannot be a neutron star and is generally expected to be a black hole. The first strong candidate for a black hole, Cygnus X-1, was discovered in this way by Charles Thomas Bolton, Louise Webster, and Paul Murdin in 1972. Observations of rotation broadening of the optical star reported in 1986 lead to a compact object mass estimate of 16 solar masses, with 7 solar masses as the lower bound. In 2011, this estimate was updated to 14.1±1.0 M☉ for the black hole and 19.2±1.9 M☉ for the optical stellar companion. X-ray binaries can be categorized as either low-mass or high-mass; This classification is based on the mass of the companion star, not the compact object itself. In a class of X-ray binaries called soft X-ray transients, the companion star is of relatively low mass, allowing for more accurate estimates of the black hole mass. These systems actively emit X-rays for only several months once every 10–50 years. During the period of low X-ray emission, called quiescence, the accretion disk is extremely faint, allowing detailed observation of the companion star. Numerous black hole candidates have been measured by this method. Black holes are also sometimes found in binaries with other compact objects, such as white dwarfs, neutron stars, and other black holes. The centre of nearly every galaxy contains a supermassive black hole. The close observational correlation between the mass of this hole and the velocity dispersion of the host galaxy's bulge, known as the M–sigma relation, strongly suggests a connection between the formation of the black hole and that of the galaxy itself. Astronomers use the term active galaxy to describe galaxies with unusual characteristics, such as unusual spectral line emission and very strong radio emission. Theoretical and observational studies have shown that the high levels of activity in the centers of these galaxies, regions called active galactic nuclei (AGN), may be explained by accretion onto supermassive black holes. These AGN consist of a central black hole that may be millions or billions of times more massive than the Sun, a disk of interstellar gas and dust called an accretion disk, and two jets perpendicular to the accretion disk. Although supermassive black holes are expected to be found in most AGN, only some galaxies' nuclei have been more carefully studied in attempts to both identify and measure the actual masses of the central supermassive black hole candidates. Some of the most notable galaxies with supermassive black hole candidates include the Andromeda Galaxy, Messier 32, Messier 87, the Sombrero Galaxy, and the Milky Way itself. Another way black holes can be detected is through observation of effects caused by their strong gravitational field. One such effect is gravitational lensing: The deformation of spacetime around a massive object causes light rays to be deflected, making objects behind them appear distorted. When the lensing object is a black hole, this effect can be strong enough to create multiple images of a star or other luminous source. However, the distance between the lensed images may be too small for contemporary telescopes to resolve—this phenomenon is called microlensing. Instead of seeing two images of a lensed star, astronomers see the star brighten slightly as the black hole moves towards the line of sight between the star and Earth and then return to its normal luminosity as the black hole moves away. The turn of the millennium saw the first 3 candidate detections of black holes in this way, and in January 2022, astronomers reported the first confirmed detection of a microlensing event from an isolated black hole. This was also the first determination of an isolated black hole mass, 7.1±1.3 M☉. Alternatives While there is a strong case for supermassive black holes, the model for stellar-mass black holes assumes of an upper limit for the mass of a neutron star: objects observed to have more mass are assumed to be black holes. However, the properties of extremely dense matter are poorly understood. New exotic phases of matter could allow other kinds of massive objects. Quark stars would be made up of quark matter and supported by quark degeneracy pressure, a form of degeneracy pressure even stronger than neutron degeneracy pressure. This would halt gravitational collapse at a higher mass than for a neutron star. Even stronger stars called electroweak stars would convert quarks in their cores into leptons, providing additional pressure to stop the star from collapsing. If, as some extensions of the Standard Model posit, quarks and leptons are made up of the even-smaller fundamental particles called preons, a very compact star could be supported by preon degeneracy pressure. While none of these hypothetical models can explain all of the observations of stellar black hole candidates, a Q star is the only alternative which could significantly exceed the mass limit for neutron stars and thus provide an alternative for supermassive black holes.: 12 A few theoretical objects have been conjectured to match observations of astronomical black hole candidates identically or near-identically, but which function via a different mechanism. A dark energy star would convert infalling matter into vacuum energy; This vacuum energy would be much larger than the vacuum energy of outside space, exerting outwards pressure and preventing a singularity from forming. A black star would be gravitationally collapsing slowly enough that quantum effects would keep it just on the cusp of fully collapsing into a black hole. A gravastar would consist of a very thin shell and a dark-energy interior providing outward pressure to stop the collapse into a black hole or formation of a singularity; It could even have another gravastar inside, called a 'nestar'. Open questions According to the no-hair theorem, a black hole is defined by only three parameters: its mass, charge, and angular momentum. This seems to mean that all other information about the matter that went into forming the black hole is lost, as there is no way to determine anything about the black hole from outside other than those three parameters. When black holes were thought to persist forever, this information loss was not problematic, as the information can be thought of as existing inside the black hole. However, black holes slowly evaporate by emitting Hawking radiation. This radiation does not appear to carry any additional information about the matter that formed the black hole, meaning that this information is seemingly gone forever. This is called the black hole information paradox. Theoretical studies analyzing the paradox have led to both further paradoxes and new ideas about the intersection of quantum mechanics and general relativity. While there is no consensus on the resolution of the paradox, work on the problem is expected to be important for a theory of quantum gravity.: 126 Observations of faraway galaxies have found that ultraluminous quasars, powered by supermassive black holes, existed in the early universe as far as redshift z ≥ 7 {\displaystyle z\geq 7} . These black holes have been assumed to be the products of the gravitational collapse of large population III stars. However, these stellar remnants were not massive enough to produce the quasars observed at early times without accreting beyond the Eddington limit, the theoretical maximum rate of black hole accretion. Physicists have suggested a variety of different mechanisms by which these supermassive black holes may have formed. It has been proposed that smaller black holes may have also undergone mergers to produce the observed supermassive black holes. It is also possible that they were seeded by direct-collapse black holes, in which a large cloud of hot gas avoids fragmentation that would lead to multiple stars, due to low angular momentum or heating from a nearby galaxy. Given the right circumstances, a single supermassive star forms and collapses directly into a black hole without undergoing typical stellar evolution. Additionally, these supermassive black holes in the early universe may be high-mass primordial black holes, which could have accreted further matter in the centers of galaxies. Finally, certain mechanisms allow black holes to grow faster than the theoretical Eddington limit, such as dense gas in the accretion disk limiting outward radiation pressure that prevents the black hole from accreting. However, the formation of bipolar jets prevent super-Eddington rates. In fiction Black holes have been portrayed in science fiction in a variety of ways. Even before the advent of the term itself, objects with characteristics of black holes appeared in stories such as the 1928 novel The Skylark of Space with its "black Sun" and the "hole in space" in the 1935 short story Starship Invincible. As black holes grew to public recognition in the 1960s and 1970s, they began to be featured in films as well as novels, such as Disney's The Black Hole. Black holes have also been used in works of the 21st century, such as Christopher Nolan's science fiction epic Interstellar. Authors and screenwriters have exploited the relativistic effects of black holes, particularly gravitational time dilation. For example, Interstellar features a black hole planet with a time dilation factor of over 60,000:1, while the 1977 novel Gateway depicts a spaceship approaching but never crossing the event horizon of a black hole from the perspective of an outside observer due to time dilation effects. Black holes have also been appropriated as wormholes or other methods of faster-than-light travel, such as in the 1974 novel The Forever War, where a network of black holes is used for interstellar travel. Additionally, black holes can feature as hazards to spacefarers and planets: A black hole threatens a deep-space outpost in 1978 short story The Black Hole Passes, and a binary black hole dangerously alters the orbit of a planet in the 2018 Netflix reboot of Lost in Space. Notes References Further reading External links
========================================
[SOURCE: https://en.wikipedia.org/wiki/Telrad_Networks] | [TOKENS: 866]
Contents Telrad Networks Telrad Networks Ltd. is a British company focused on 4G and cellular telecommunication devices. Founded in 1951, the company is a developer of advanced WiMAX and 4G LTE base stations, Customer-Premises Equipment, and network management. Since its acquisition of Alvarion's Broadband Wireless Access (BWA) Division in 2013, The company provides also AAA servers, PCRF and Billing for LTE and for Wimax with its partner Aradial Tech. Telrad 4G products are deployed in more than 100 countries by telecom carriers, Internet Service Providers, utilities, and governmental organizations. With the extended reach of its parent company Liquid Technologies, Telrad has significantly increased opportunities for international growth beyond its current global footprint. Telrad has two subsidiaries. Magalcom is a market leader in building data centers, IT infrastructure, control rooms and homeland security. The second subsidiary, Oasis Communications, is a system integrator specializing in consulting, design and delivery of highly complex communication systems. On August 17, 2022, Telrad Networks was acquired by Liquid Intelligent Technologies. History Telrad Networks was founded in 1951 under the name The Consolidated Telephone & Radio Company. Over the course of several decades, the company grew to become one of Africa's largest telecom equipment manufacturers and a major supplier to Bezeq, the national phone company. During its early growth stages, Telrad expanded to support two manufacturing plants, devoted to five product lines, with a focus on digital exchanges, remote switches and peripheral equipment. As broadband technology became more pervasive in the early 2000s, Telrad began to invest in broadband-related start-ups such as Aptonix, Be-Connected and Firebit.net, ultimately selling off controlling interests in these firms. By 2013, Telrad acquired the broadband wireless access (BWA) division of Alvarion Ltd., thereby inheriting a full 4G infrastructure product portfolio, including the BreezeMax and BreezeCOMPACT product lines. By the mid-90s, the company was ranked 16th by Dun & Bradstreet, amongst Africa's leading Industrial Enterprises. At that time, Telrad was organized into four main product divisions: In 1996, Nortel Networks acquired 20% stake in Telrad, with a focus on expanding the Nortel Solution Division globally. In March 2000, Nortel partnered with Koor Industries Ltd., a leading investment company, to establish Nortel Networks in Africa. As part of the agreement, they acquired Telrad Networks’ Nortel division. The company retained its proprietary product lines involving telephony systems production and integration. As broadband technology became more pervasive in the early 2000s, Telrad began to invest in broadband-related start-ups such as Aptonix, Be-Connected and Firebit.net, ultimately selling off controlling interests in these firms. By 2013, Telrad acquired the broadband wireless access (BWA) division of Alvarion Ltd., thereby inheriting a full 4G infrastructure product portfolio, including the BreezeMax and BreezeCOMPACT product lines. On 17 August 2022 the African technology infrastructure and services provider Liquid Intelligent Technologies finalized its 100% acquisition of Telrad, a technology and solutions company. Liquid have now seamlessly integrated Telrad's strong R&D programme and technology solutions in cyber security, data centres and wireless access technology with Liquid's existing portfolio of solutions. Liquid Intelligent Technologies is part of the wider Cassava Technologies group, a pan-African technology group with an eye for international growth through innovation and technological expansion. Products and technology Wireless broadband technology enables high-speed connectivity to meet the need for Internet access, especially in areas with little or no access infrastructure. Telrad's portfolio of 4G LTE base stations support both WiMAX IEEE 802.16x and LTE. The Telrad BreezeCOMPACT family of products consists of three product lines: BreezeCOMPACT 1000 and 3000. Telrad's portfolio of Customer-Premises Equipment consists of a family of indoor and outdoor CPEs to support both WiMAX and TD-LTE platforms. The product line delivers voice and data connectivity for enterprise or home environments. References
========================================
[SOURCE: https://en.wikipedia.org/wiki/Com%C3%A9die-Fran%C3%A7aise] | [TOKENS: 1712]
Contents Comédie-Française The Comédie-Française (French: [kɔmedi fʁɑ̃sɛːz] ⓘ), or Théâtre-Français (French: [teɑtʁ(ə) fʁɑ̃sɛ] ⓘ), is a state theatre in France. Founded in 1680, it is the oldest active theatre company in the world. Established as a French state-controlled entity in 1995, it is one of the few in the nation and the only to have its own permanent troupe of actors. The company's primary venue is the Salle Richelieu, a part of the Palais-Royal complex located at 2 Rue de Richelieu on Place André-Malraux in the 1st arrondissement of Paris. The theatre has also been known as the Théâtre de la République, and popularly as "La Maison de Molière" (The House of Molière). It acquired the latter name from the troupe of its best-known playwright, Molière, considered the patron of French actors. He died seven years before his troupe became known as the Comédie-Française, but the company continued to be known as "La Maison de Molière" even after the official change of name. History The Comédie-Française was founded on 8 August 1680 by a decree of Louis XIV merging the only two Parisian acting troupes of the time, those of the Guénégaud Theatre and the Hôtel de Bourgogne. On the death of Molière in 1673, the troupe at the Guénégaud had been formed by a merger of the Théâtre du Marais and the Troupe de Molière. Two years later they received a royal grant of 12,000 livres per year; and seven years later they received their present designation. The combined company gave its first performance on 25 August 1680 at the Guénégaud. Its leading actors included Molière's widow, Armande Béjart, her husband, Guérin d'Estriché, La Grange, Mlle Champmeslé, Baron, Hauteroche, and Raymond Poisson. The repertoire consisted of the collection of theatrical works by Molière and Jean Racine, along with a few works by Pierre Corneille, Paul Scarron and Jean Rotrou. In the 18th century, the Comédie-Française was often enjoyed by the French nobility, an expensive luxury. On the performance of Joseph Chénier's anti-monarchical play Charles IX in 1789, violent political discussions arose among the performers, ultimately splitting the company into two sections: the Royalists, renamed the "Théâtre de la Nation", and the Republicans led by the young tragedian Talma establishing the "Théâtre de la République" on the site of the present building in the Rue de Richelieu. On 16 April 1790, the theatre presented the world premiere of Pierre Laujon's Le Couvent, ou les Fruits du caractère et de l'éducation; the first French play to feature an all-women cast. On 3 September 1793, during the French Revolution, the Théâtre de la Nation was closed by order of the Committee of Public Safety for putting on the allegedly seditious play, Pamela, and the actors were imprisoned, though gradually released later. On 31 May 1799, the new government made the Salle Richelieu available and allowed the actors to reconstitute the troupe. The Comédie-Française today has a repertoire of 3,000 works and three theatres in Paris (Salle Richelieu, next to the Palais Royal; théâtre du Vieux-Colombier; Studio-Théâtre). Since October 2020, and because of the COVID-19 pandemic, the Comédie-Française had to close as it is the case for all other theaters in France. The Comédie-Française having a permanent troupe of actors, it was decided to switch temporarily to an online program, including readings of the full text of In Search of Lost Time, and an online initiative called Théâtre à la table where actors of the troupe play works in the repertoire after a one week rehearsal. Online attendance for this initiative was unexpectedly high, including people outside of Paris and in other countries. In May 2021 Éric Ruf, the managing director of the Comédie-Française, declared that 30% of the public of the online program never went to the Comédie-Française, and that they would continue this program even after the reopening. Theatre buildings The Comédie-Française has had several homes since its inception in 1680 in the Salle Guénégaud. In 1689, it was established in a theatre across from the Café Procope. From 1770 to 1782, the Comédie-Française performed in the theatre in the royal palace of the Tuileries. In 1782, the company moved into the Salle du Faubourg Saint-Germain, designed by architects Marie-Joseph Peyre and Charles De Wailly and located on the site of today's Odéon. Since 1799, the Comédie-Française has been housed in the Salle Richelieu (architect Victor Louis) at 2, rue de Richelieu. This theatre was enlarged and modified in the 1800s, then rebuilt in 1900 after a severe fire. The actress Jane Henriot was the only casualty of the fire. Theatrical troupe The membership of the theatrical troupe is divided into sociétaires and pensionnaires. The former are regular members of the organisation and as such receive a pension after 20 years of service, while the latter are paid actors who may, after a certain length of service, become sociétaires. The names of nearly all the great actors and dramatists of France have, at some time in their career, been associated with that of the Comédie-Française. Every year on 15 January, the company performs an "Homage to Molière" on the anniversary of his christening and possibly his birth. Following that evening's performance, which nowadays is free, all members of the troupe, and since 2009 of its academy, assemble on the stage behind a bust of Molière, wearing costumes of their choice from the theatre's repertoire. They first process in ranks to bow to the audience's applause. Each member then says a line from Molière's plays, also applauded. In conclusion, the dean introduces to the audience the new members of the academy, the new pensionnaires, and the new sociétaires, and has the departing sociétaires honoraires applauded. The tradition was started in 1773 on the centenary of Molière's death. In 1821, it was moved to his then recently discovered christening date, and the next year approached its present form as the final scene of a play specially written by Justin Gensoul, Le Ménage de Molière. In the 19th century, The Imaginary Invalid was often performed on that date, as the final divertissement made it possible for the entire troupe to appear. Similar homages became popular at the Odéon, as well as in other theatres when these lost their privilege to perform Molière in 1864; however, the Comédie-Française is the only one that has maintained it. The défilé (procession) of the Paris Opera Ballet, created in the 20th century, bears similarities to the homage. Administrators of the Comédie-Française The chief administrator of the Comédie-Française has been given the title administrateur général since Simonis' term of 1850. Before that, a variety of titles were given. See also Notes Bibliography External links
========================================
[SOURCE: https://en.wikipedia.org/wiki/Magma_computer_algebra_system] | [TOKENS: 326]
Contents Magma (computer algebra system) Magma is a computer algebra system designed to solve problems in algebra, number theory, geometry and combinatorics. It is named after the algebraic structure magma. It runs on Unix-like operating systems, as well as Windows. Introduction Magma is produced and distributed by the Computational Algebra Group within the Sydney School of Mathematics and Statistics at the University of Sydney. In late 2006, the book Discovering Mathematics with Magma was published by Springer as volume 19 of the Algorithms and Computations in Mathematics series. The Magma system is used extensively within pure mathematics. The Computational Algebra Group maintain a list of publications that cite Magma, and as of 2010 there are about 2600 citations, mostly in pure mathematics, but also including papers from areas as diverse as economics and geophysics. History The predecessor of the Magma system was named Cayley (1982–1993), after Arthur Cayley. Magma was officially released in August 1993 (version 1.0). Version 2.0 of Magma was released in June 1996 and subsequent versions of 2.X have been released approximately once per year. In 2013, the Computational Algebra Group finalized an agreement with the Simons Foundation, whereby the Simons Foundation will underwrite all costs of providing Magma to all U.S. nonprofit, non-governmental scientific research or educational institutions. All students, researchers and faculty associated with a participating institution will be able to access Magma for free, through that institution. Mathematical areas covered by the system See also References External links
========================================
[SOURCE: https://en.wikipedia.org/wiki/John_Bardeen] | [TOKENS: 3614]
Contents John Bardeen John Bardeen (May 23, 1908 – January 30, 1991) was an American condensed matter physicist. He is the only person to be awarded the Nobel Prize in Physics twice: first in 1956 with William Shockley and Walter Brattain for their invention of the transistor; and again in 1972 with Leon Cooper and Robert Schrieffer for their microscopic theory of superconductivity, known as the BCS theory. Born and raised in Wisconsin, Bardeen earned both his bachelor's and master's degrees in electrical engineering from the University of Wisconsin, before receiving a Ph.D. in physics from Princeton University. After serving in World War II, he was a researcher at Bell Labs and a professor at the University of Illinois. The transistor revolutionized the electronics industry, making possible the development of almost every modern electronic device, from telephones to computers, and ushering in the Information Age. Bardeen's developments in superconductivity—for which he was awarded his second Nobel Prize—are used in nuclear magnetic resonance spectroscopy (NMR), medical magnetic resonance imaging (MRI), and superconducting quantum circuits. Bardeen is the first of only three people to have won multiple Nobel Prizes in the same category (the others being Frederick Sanger and Karl Barry Sharpless in chemistry), and one of five persons with two Nobel Prizes. In 1990, Bardeen appeared on Life magazine's list of "100 Most Influential Americans of the Century." Education John Bardeen was born on May 23, 1908, in Madison, Wisconsin, the son of Charles Russell Bardeen, the first dean of the University of Wisconsin Medical School, and Althea Harmer. Bardeen attended University of Wisconsin High School in Madison, graduating in 1923 at age 15; he could have graduated several years earlier, but this was postponed because he took courses at another high school and because of his mother's death. Bardeen entered the University of Wisconsin in 1923. While in college, he joined the Zeta Psi fraternity. He raised a part of the needed membership fees by playing billiards. Bardeen was initiated as a member of Tau Beta Pi engineering honor society. Not wanting to be an academic like his father, Bardeen chose engineering. He also felt that engineering had good job prospects. Bardeen received his B.S. in Electrical Engineering in 1928 from the University of Wisconsin. Despite taking a year off to work in Chicago, he graduated in 1928. Taking all the graduate courses in physics and mathematics that had interested him, Bardeen graduated in five years instead of the usual four. This allowed him time to complete his master's thesis, supervised by Leo J. Peters. He received his M.S. in Electrical Engineering in 1929 from Wisconsin. Bardeen furthered his studies by staying on at Wisconsin, but he eventually went to work for Gulf Research Laboratories, the research arm of the Gulf Oil Corporation that was based in Pittsburgh. From 1930 to 1933, Bardeen worked there on the development of methods for the interpretation of magnetic and gravitational surveys. He worked as a geophysicist. After the work failed to keep his interest, he applied and was accepted to the graduate program in mathematics at Princeton University. As a graduate student, Bardeen studied mathematics and physics. Under the physicist Eugene Wigner, he wrote his thesis on a problem in solid-state physics. Before completing his thesis, he was offered a position as junior fellow of the Society of Fellows at Harvard University in 1935. He spent the next three years there, from 1935 to 1938, working with to-be Nobel laureates in Physics John Hasbrouck van Vleck and Percy Williams Bridgman on problems in cohesion and electrical conduction in metals—and also did some work on level density of nuclei. He received his Ph.D. in Mathematical Physics from Princeton in 1936. Career From 1938 to 1941, Bardeen worked as an assistant professor at the University of Minnesota department of physics. From 1941 to 1944, he headed the group working on magnetic mines and torpedoes and mine and torpedo countermeasures at the Naval Ordnance Laboratory. During this period, his wife Jane gave birth to a son (Bill, born in 1941) and a daughter (Betsy, born in 1944). In October 1945, Bardeen began work at Bell Telephone Laboratories as a member of a solid-state physics group led by William Shockley and chemist Stanley Morgan. Other personnel working in the group were Walter Brattain, physicist Gerald Pearson, chemist Robert Gibney, electronics expert Hilbert Moore and several technicians. He moved his family to Summit, New Jersey. The assignment of the group was to seek a solid-state alternative to fragile glass vacuum tube amplifiers. Their first attempts were based on Shockley's ideas about using an external electrical field on a semiconductor to affect its conductivity. These experiments mysteriously failed every time in all sorts of configurations and materials. The group was at a standstill until Bardeen suggested a theory that invoked surface states that prevented the field from penetrating the semiconductor. The group changed its focus to study these surface states, meeting almost daily to discuss the work. The rapport of the group was excellent and ideas were freely exchanged. By the winter of 1946, they had enough results that Bardeen submitted a paper on the surface states to Physical Review. Brattain started experiments to study the surface states through observations made while shining a bright light on the semiconductor's surface. This led to several more papers (one of them co-authored with Shockley), which estimated the density of the surface states to be more than enough to account for their failed experiments. The pace of the work picked up significantly when they started to surround point contacts between the semiconductor and the conducting wires with electrolytes. Moore built a circuit that allowed them to vary the frequency of the input signal easily and suggested that they use glycol borate (gu), a viscous chemical that did not evaporate. Finally, they began to get some evidence of power amplification when Pearson, acting on a suggestion by Shockley, put a voltage on a droplet of gu placed across a p–n junction. By 1951, Bardeen was looking for a new job. Fred Seitz, a friend of Bardeen, convinced the University of Illinois to make Bardeen an offer of $10,000 a year. Bardeen accepted the offer and left Bell Labs, joining the engineering and physics faculties at Illinois in 1951, where he was Professor of Electrical Engineering and Professor of Physics. At Illinois, Bardeen established two major research programs, one in the electrical engineering department and one in the physics department. The research program in the electrical engineering department dealt with both experimental and theoretical aspects of semiconductors, and the research program in the physics department dealt with theoretical aspects of macroscopic quantum systems, particularly superconductivity and quantum liquids. Bardeen was an active professor at Illinois from 1951 to 1975 and then became professor emeritus. In his later life, Bardeen remained active in academic research, during which time he focused on understanding the flow of electrons in charge density waves (CDWs) through metallic linear chain compounds. His proposals that CDW electron transport is a collective quantum phenomenon (see Macroscopic quantum phenomena) were initially greeted with skepticism. However, experiments reported in 2012 show oscillations in CDW current versus magnetic flux through tantalum trisulfide rings, similar to the behavior of superconducting quantum interference devices (see SQUID and Aharonov–Bohm effect), lending credence to the idea that collective CDW electron transport is fundamentally quantum in nature. (See quantum mechanics.) Bardeen continued his research throughout the 1980s, and published articles in Physical Review Letters and Physics Today less than a year before he died. A collection of Bardeen's personal papers are held by the University of Illinois Archives. Research On December 23, 1947, Bardeen and Walter Brattain were working without William Shockley when they succeeded in creating a point-contact transistor that achieved amplification. By the next month, Bell Labs' patent attorneys started to work on the patent applications. Bell Labs' attorneys soon discovered that Shockley's field effect principle had been anticipated and patented in 1930 by Julius Lilienfeld, who filed his MESFET-like patent in Canada on October 22, 1925. Shockley publicly took the lion's share of the credit for the invention of the transistor; this led to a deterioration of Bardeen's relationship with him. Bell Labs management, however, consistently presented all three inventors as a team. Shockley eventually infuriated and alienated Bardeen and Brattain, essentially blocking the two from working on the junction transistor. Bardeen began pursuing a theory for superconductivity and left Bell Labs in 1951. Brattain refused to work with Shockley further and was assigned to another group. Neither Bardeen nor Brattain had much to do with the development of the transistor beyond the first year after its invention. The "transistor" (a portmanteau of "transconductance" and "resistor") was 1/50 the size of the vacuum tubes it replaced in televisions and radios, used far less power, was far more reliable, and it allowed electrical devices to become more compact. In 1956, Bardeen, Brattain, and Shockley were jointly awarded the Nobel Prize in Physics "for their researches on semiconductors and their discovery of the transistor effect." In 1957, Bardeen, in collaboration with Leon Cooper and his doctoral student John Robert Schrieffer, proposed the standard theory of superconductivity known as the BCS theory (named for their initials). In 1972, Bardeen, Cooper, and Schrieffer were jointly awarded the Nobel Prize in Physics "for their jointly developed theory of superconductivity, usually called the BCS-theory." This was Bardeen's second Nobel Prize in Physics; he became the first person to win two Nobel Prizes in the same field. In the late 1960s, Bardeen felt that Cooper and Schrieffer deserved the Nobel Prize for BCS. He was concerned that they might not be awarded because of the Nobel Committee's reluctance to award the same person twice, which would be his case as a co-author of the theory. Bardeen nominated scientists who worked on superconducting tunneling effects such as the Josephson effect for the Prize in 1967: Leo Esaki, Ivar Giaever, and Brian Josephson. He recognized that because the tunneling developments depended on superconductivity, it would increase the chances that BCS itself would be awarded first. He also reasoned that the Nobel Committee had a predilection for multinational teams, which was the case for his tunneling nominees, each being from a different country. Bardeen renewed the nominations in 1971, 1972, when BCS received the prize, and finally 1973, when tunneling was awarded.: 230-231 Bardeen became interested in superconducting tunneling in the summer of 1960 after consulting for the General Electric Research Laboratory in Schenectady, New York, where he learned about experiments done by Ivar Giaever at Rensselaer Polytechnic Institute, which suggested that electrons from a normal material could tunnel into a superconducting one.: 222–223 On June 8, 1962, Brian Josephson, then 23, submitted to Physics Letters his prediction of a super-current flow across a barrier, effect which later became known as the Josephson effect. Bardeen challenged Josephson's theory on a note in his own paper received ten days later by Physical Review Letters: 222–225 : In a recent note, Josephson uses a somewhat similar formulation to discuss the possibility of superfluid flow across the tunneling region, in which no quasi-particles are created. However, as pointed out by the author (reference 3), pairing does not extend into the barrier, so that there can be no such superfluid flow. The matter was further discussed on the 8th International Conference on Low Temperature Physics held September 16 to 22, 1962 at Queen Mary University of London. While Josephson was presenting his theory, Bardeen rose to describe his objections. After an intense debate both men were unable to reach a common understanding, and at points Josephson repeatedly asked Bardeen, "Did you calculate it? No? I did.": 225–226 In 1963, experimental evidence and further theoretical clarifications were discovered supporting the Josephson effect, notably in a paper by Philip W. Anderson and John Rowell from Bell Labs. After this, Bardeen came to accept Josephson's theory and publicly withdrew his previous opposition to it at a conference held in August 1963. Bardeen also invited Josephson as a postdoc in Illinois for the academic year of 1965–1966, and later nominated Josephson and Giaever for the Nobel Prize in Physics, which they received in 1973.: 226 Personal life While studying at Princeton, Bardeen met Jane Maxwell (1907–1997) during a visit to his old friends in Pittsburgh. He married Jane on July 18, 1938. Bardeen was a scientist with a very unassuming personality. While he served as a professor for almost 40 years at the University of Illinois, he was best remembered by neighbors for hosting cookouts where he would prepare food for his friends, many of whom were unaware of his accomplishments at the university. He would always ask his guests if they liked the hamburger bun toasted (since he liked his that way). He enjoyed playing golf and going on picnics with his family. Lillian Hoddeson said that because he "differed radically from the popular stereotype of 'genius' and was uninterested in appearing other than ordinary, the public and the media often overlooked him." When Bardeen was asked about his beliefs during a 1988 interview, he responded: "I am not a religious person, and so do not think about it very much". However, he has also said: "I feel that science cannot provide an answer to the ultimate questions about the meaning and purpose of life." Bardeen did believe in a code of moral values and behavior. John Bardeen's children were taken to church by his wife, who taught Sunday school and was a church elder.: 168–169 Despite this, he and his wife made it clear that they did not have faith in an afterlife and other religious ideas. He was the father of James M. Bardeen, William A. Bardeen, and daughter Elizabeth. Bardeen died of heart disease on January 30, 1991, at Brigham and Women's Hospital in Boston, Massachusetts, at the age of 82. Although he lived in Champaign-Urbana, he had come to Boston for medical consultation. Bardeen and his wife Jane are buried at Forest Hill Cemetery in Madison, Wisconsin. Recognition Bardeen is the only double laureate in Physics, and one of three double laureates of the same prize; the others are Frederick Sanger who won the 1958 and 1980 Prizes in Chemistry and Karl Barry Sharpless who won the 2001 and 2022 Prizes in Chemistry. Commemoration Near the end of this decade, when they begin enumerating the names of the people who had the greatest impact on the 20th century, the name of John Bardeen, who died last week, has to be near, or perhaps even arguably at, the top of the list ... Mr. Bardeen shared two Nobel Prizes and has been awarded numerous other honors. But what greater honor can there be when each of us can look all around us and everywhere see the reminders of a man whose genius has made our lives longer, healthier and better. In honor of Bardeen, the engineering quadrangle at the University of Illinois at Urbana–Champaign is named the Bardeen Quad. Also in honor of Bardeen, Sony Corporation endowed a $3 million John Bardeen professorial chair at the University of Illinois at Urbana-Champaign, beginning in 1990. Sony Corporation owed much of its success to commercializing Bardeen's transistors in portable TVs and radios, and had worked with Illinois researchers. As of 2022[update], the John Bardeen Professor is Yurii Vlasov. At the time of Bardeen's death, then-University of Illinois chancellor Morton Weir said, "It is a rare person whose work changes the life of every American; John's did." Bardeen was honored on a March 6, 2008, United States postage stamp as part of the "American Scientists" series designed by artist Victor Stabin. The $0.41 stamp was unveiled in a ceremony at the University of Illinois. His citation reads: "Theoretical physicist John Bardeen (1908–1991) shared the Nobel Prize in Physics twice—in 1956, as co-inventor of the transistor and in 1972, for the explanation of superconductivity. The transistor paved the way for all modern electronics, from computers to microchips. Diverse applications of superconductivity include infrared sensors and medical imaging systems." The other scientists on the "American Scientists" sheet include biochemist Gerty Cori, chemist Linus Pauling and astronomer Edwin Hubble. Notes References External links
========================================